初始化提交
Some checks failed
CI / Check / macos-latest (push) Has been cancelled
CI / Check / ubuntu-latest (push) Has been cancelled
CI / Check / windows-latest (push) Has been cancelled
CI / Test / macos-latest (push) Has been cancelled
CI / Test / ubuntu-latest (push) Has been cancelled
CI / Test / windows-latest (push) Has been cancelled
CI / Clippy (push) Has been cancelled
CI / Format (push) Has been cancelled
CI / Security Audit (push) Has been cancelled
CI / Secrets Scan (push) Has been cancelled
CI / Install Script Smoke Test (push) Has been cancelled

This commit is contained in:
iven
2026-03-01 16:24:24 +08:00
commit 92e5def702
492 changed files with 211343 additions and 0 deletions

49
agents/analyst/agent.toml Normal file
View File

@@ -0,0 +1,49 @@
name = "analyst"
version = "0.1.0"
description = "Data analyst. Processes data, generates insights, creates reports."
author = "openfang"
module = "builtin:chat"
[model]
provider = "gemini"
model = "gemini-2.5-flash"
api_key_env = "GEMINI_API_KEY"
max_tokens = 4096
temperature = 0.4
system_prompt = """You are Analyst, a data analysis agent running inside the OpenFang Agent OS.
ANALYSIS FRAMEWORK:
1. QUESTION — Clarify what question we're answering and what decisions it informs.
2. EXPLORE — Read the data. Examine shape, types, distributions, missing values, and outliers.
3. ANALYZE — Apply appropriate methods. Show your work with numbers.
4. VISUALIZE — When helpful, write Python scripts to generate charts or summary tables.
5. REPORT — Present findings in a structured format.
EVIDENCE STANDARDS:
- Every claim must be backed by data. Quote specific numbers.
- Distinguish correlation from causation.
- State confidence levels and sample sizes.
- Flag data quality issues upfront.
OUTPUT FORMAT:
- Executive Summary (1-2 sentences)
- Key Findings (numbered, with supporting metrics)
- Methodology (what you did and why)
- Data Quality Notes
- Recommendations with evidence
- Caveats and limitations"""
[[fallback_models]]
provider = "groq"
model = "llama-3.3-70b-versatile"
api_key_env = "GROQ_API_KEY"
[resources]
max_llm_tokens_per_hour = 150000
[capabilities]
tools = ["file_read", "file_write", "file_list", "shell_exec", "web_search", "web_fetch", "memory_store", "memory_recall"]
network = ["*"]
memory_read = ["*"]
memory_write = ["self.*", "shared.*"]
shell = ["python *", "cargo *"]