Compare commits

..

28 Commits

Author SHA1 Message Date
8400b997b7 chore: 添加 AGENTS 默认回复约束 2026-03-19 09:56:39 +08:00
b3df515775 chore: sync JSONC config overrides 2026-03-19 09:51:55 +08:00
d9f89b3b28 chore: migrate config files to jsonc 2026-03-18 09:57:50 +08:00
cdabb9fd16 chore: 更新 GPT-5.4 mini 小模型配置 2026-03-18 09:49:14 +08:00
caf20d861d chore: 调整 writing 代理模型分配 2026-03-17 14:59:36 +08:00
b4569fa6a9 chore: 接入 LiteLLM 模型并更新轻量代理配置 2026-03-17 14:50:42 +08:00
db8880c833 chore: 移除 experimental 动态上下文裁剪配置 2026-03-16 02:43:25 +08:00
681d8340cb feat: 升级至 GPT-5.4 并调整代理模型分配 2026-03-16 00:46:00 +08:00
f8a349f478 chore: 调整 GPT-only 角色化模型分配 2026-03-04 23:54:56 +08:00
8fdfb38e84 chore: 精简配置格式并调整代理变体 2026-03-04 23:39:12 +08:00
9c768870e3 refactor: 从 CPA 换成 LiteLLM 2026-03-04 16:34:10 +08:00
8978eeb5de chore: 对齐分类文档链接并更新 artistry 模型参数 2026-02-26 11:09:52 +08:00
3a35190250 chore: 优化 GPT-only 配置以降低轻量任务成本 2026-02-26 10:24:53 +08:00
b9650bfde8 chore: 调整代理模型配置并移除 glm-4.7 2026-02-26 09:33:35 +08:00
8573f89db7 feat: 更新代理模型参考链接并调整小模型配置
- 更新代理模型参考链接指向新的代理库存文档位置。
- 将小模型配置更新为使用cpa-openai/glm-4.7。
2026-02-24 14:46:39 +08:00
069684079a chore: 关闭粘贴摘要折叠以便预览完整文本 2026-02-16 00:36:52 +08:00
472294b787 chore: 合并 CPA OpenAI 提供商并统一模型前缀 2026-02-15 19:56:44 +08:00
f087f34a96 chore: quick 分类切换为 gpt-5-nano 2026-02-15 19:26:13 +08:00
cd4dd39670 chore: 调整 quick 模型为 glm-4.7 2026-02-15 03:40:44 +08:00
9de3f60156 chore: 调整 writing 模型配置为 glm-4.7 2026-02-15 03:36:44 +08:00
9384469551 chore: 调整 librarian 与 explore 模型配置 2026-02-15 03:26:35 +08:00
09407f9980 chore: 更新小模型配置并统一格式化 2026-02-15 03:14:30 +08:00
12d944efb4 chore: 调整检索与轻任务模型配置以提升性价比 2026-02-15 02:56:57 +08:00
8ca07bcdf5 feat: 提升规划编排模型并补齐 artistry 分类 2026-02-15 02:40:19 +08:00
8ea1baccf2 feat: 质量优先切换 librarian/explore/quick/writing 模型 2026-02-15 02:21:11 +08:00
f4cc4c6f29 feat: 优化模型分层并新增 gpt-5.1-codex-mini 2026-02-15 02:07:05 +08:00
294451814e chore: 统一 explore 和 quick 模型使用 glm-4.7 2026-02-15 01:15:56 +08:00
0296ca6564 feat: 迁移至 Codex API 并添加 GLM Coding Plan 提供商 2026-02-15 01:05:03 +08:00
4 changed files with 35 additions and 150 deletions

View File

@@ -7,4 +7,4 @@ EXA_API_KEY=
CONTEXT7_API_KEY= CONTEXT7_API_KEY=
# Windows # Windows
OPENCODE_GIT_BASH_PATH=%USERPROFILE%\scoop\apps\git\current\bin\bash.exe OPENCODE_GIT_BASH_PATH=%USERPROFILE%\scoop\shims\bash.exe

5
AGENTS.md Normal file
View File

@@ -0,0 +1,5 @@
# AGENTS.md
- Always reply in Simplified Chinese by default.
- Use English only if the user explicitly asks for it.
- Keep code, commands, file paths, logs, and identifiers unchanged.

View File

@@ -1,34 +1,35 @@
{ {
"$schema": "https://raw.githubusercontent.com/code-yeongyu/oh-my-opencode/master/assets/oh-my-opencode.schema.json", "$schema": "https://raw.githubusercontent.com/code-yeongyu/oh-my-opencode/master/assets/oh-my-opencode.schema.json",
"agents": { "agents": {
"sisyphus": { "model": "anthropic/claude-opus-4-6", "variant": "high" }, "sisyphus": { "model": "openai/gpt-5.4", "variant": "medium" },
"hephaestus": { "model": "openai/gpt-5.3-codex", "variant": "medium" }, "hephaestus": { "model": "openai/gpt-5.3-codex", "variant": "medium" },
"oracle": { "model": "openai/gpt-5.4", "variant": "high" }, "oracle": { "model": "openai/gpt-5.4", "variant": "high" },
"librarian": { "model": "google/gemini-3-flash-preview" }, "librarian": { "model": "openai/gpt-5.4-mini" },
"explore": { "model": "litellm/grok-code-fast-1" }, "explore": { "model": "openai/gpt-5.4-mini" },
"multimodal-looker": { "model": "openai/gpt-5.4", "variant": "medium" }, "multimodal-looker": { "model": "openai/gpt-5.4", "variant": "medium" },
"prometheus": { "model": "anthropic/claude-opus-4-6", "variant": "high" }, "prometheus": { "model": "openai/gpt-5.4", "variant": "high" },
"metis": { "model": "anthropic/claude-opus-4-6", "variant": "high" }, "metis": { "model": "openai/gpt-5.4", "variant": "high" },
"momus": { "model": "openai/gpt-5.4", "variant": "xhigh" }, "momus": { "model": "openai/gpt-5.4", "variant": "xhigh" },
"atlas": { "model": "anthropic/claude-sonnet-4-6" } "atlas": { "model": "openai/gpt-5.4", "variant": "medium" },
"sisyphus-junior": { "model": "openai/gpt-5.4", "variant": "medium" }
}, },
"categories": { "categories": {
"visual-engineering": { "model": "google/gemini-3.1-pro-preview", "variant": "high" }, "visual-engineering": { "model": "openai/gpt-5.4", "variant": "medium" },
"ultrabrain": { "model": "openai/gpt-5.3-codex", "variant": "xhigh" }, "ultrabrain": { "model": "openai/gpt-5.3-codex", "variant": "xhigh" },
"deep": { "model": "openai/gpt-5.3-codex", "variant": "medium" }, "deep": { "model": "openai/gpt-5.3-codex", "variant": "medium" },
"artistry": { "model": "google/gemini-3.1-pro-preview", "variant": "high" }, "artistry": { "model": "openai/gpt-5.4", "variant": "medium" },
"quick": { "model": "anthropic/claude-haiku-4-5" }, "quick": { "model": "openai/gpt-5.4-mini" },
"unspecified-low": { "model": "anthropic/claude-sonnet-4-6" }, "unspecified-low": { "model": "openai/gpt-5.4", "variant": "medium" },
"unspecified-high": { "model": "openai/gpt-5.4", "variant": "high" }, "unspecified-high": { "model": "openai/gpt-5.4", "variant": "high" },
"writing": { "model": "google/gemini-3-flash-preview" } "writing": { "model": "openai/gpt-5.4", "variant": "medium" }
},
"git_master": {
"commit_footer": false,
"include_co_authored_by": false
}, },
"experimental": { "experimental": {
"dynamic_context_pruning": { "dynamic_context_pruning": {
"enabled": true "enabled": true
} }
},
"git_master": {
"commit_footer": false,
"include_co_authored_by": false
} }
} }

View File

@@ -1,40 +1,10 @@
{ {
"$schema": "https://opencode.ai/config.json", "$schema": "https://opencode.ai/config.json",
"autoupdate": false,
"model": "anthropic/claude-opus-4-6",
"small_model": "anthropic/claude-haiku-4-5",
"plugin": ["oh-my-opencode@latest"], "plugin": ["oh-my-opencode@latest"],
"experimental": { "autoupdate": false,
"disable_paste_summary": true "model": "openai/gpt-5.4",
}, "small_model": "openai/gpt-5.4-mini",
"provider": { "provider": {
"anthropic": {
"npm": "@ai-sdk/anthropic",
"options": {
"baseURL": "{env:LITELLM_BASE_URL}/v1",
"apiKey": "{env:LITELLM_API_KEY}"
},
"models": {
"claude-opus-4-6": {
"name": "Claude Opus 4.6",
"reasoning": true,
"modalities": { "input": ["text", "image", "pdf"], "output": ["text"] },
"limit": { "context": 200000, "output": 128000 }
},
"claude-sonnet-4-6": {
"name": "Claude Sonnet 4.6",
"reasoning": true,
"modalities": { "input": ["text", "image", "pdf"], "output": ["text"] },
"limit": { "context": 200000, "output": 64000 }
},
"claude-haiku-4-5": {
"name": "Claude Haiku 4.5",
"reasoning": true,
"modalities": { "input": ["text", "image", "pdf"], "output": ["text"] },
"limit": { "context": 200000, "output": 64000 }
}
}
},
"openai": { "openai": {
"npm": "@ai-sdk/openai", "npm": "@ai-sdk/openai",
"options": { "options": {
@@ -45,116 +15,25 @@
"gpt-5.4": { "gpt-5.4": {
"name": "GPT-5.4", "name": "GPT-5.4",
"reasoning": true, "reasoning": true,
"modalities": { "input": ["text", "image"], "output": ["text"] }, "modalities": { "input": ["text", "image", "pdf"], "output": ["text"] },
"limit": { "context": 1050000, "input": 272000, "output": 128000 } "limit": { "context": 1050000, "input": 922000, "output": 128000 }
},
"gpt-5.2": {
"name": "GPT-5.2",
"reasoning": true,
"modalities": { "input": ["text", "image"], "output": ["text"] },
"limit": { "context": 400000, "input": 272000, "output": 128000 }
}, },
"gpt-5.3-codex": { "gpt-5.3-codex": {
"name": "GPT-5.3-Codex", "name": "GPT-5.3-Codex",
"reasoning": true, "reasoning": true,
"modalities": { "input": ["text", "image", "pdf"], "output": ["text"] }, "modalities": { "input": ["text", "image", "pdf"], "output": ["text"] },
"limit": { "context": 400000, "input": 272000, "output": 128000 } "limit": { "context": 400000, "input": 272000, "output": 128000 }
}
}
}, },
"google": { "gpt-5.4-mini": {
"npm": "@ai-sdk/google", "name": "GPT-5.4 mini",
"options": {
"baseURL": "{env:LITELLM_BASE_URL}/v1beta",
"apiKey": "{env:LITELLM_API_KEY}"
},
"models": {
"gemini-3.1-pro-preview": {
"name": "Gemini 3.1 Pro",
"reasoning": true, "reasoning": true,
"modalities": { "input": ["text", "image", "video", "audio", "pdf"], "output": ["text"] }, "modalities": { "input": ["text", "image"], "output": ["text"] },
"limit": { "context": 1048576, "output": 65536 } "limit": { "context": 400000, "input": 272000, "output": 128000 }
},
"gemini-3-flash-preview": {
"name": "Gemini 3 Flash",
"reasoning": true,
"modalities": { "input": ["text", "image", "video", "audio", "pdf"], "output": ["text"] },
"limit": { "context": 1048576, "output": 65536 }
}
}
},
"litellm": {
"name": "LiteLLM",
"npm": "@ai-sdk/openai-compatible",
"options": {
"baseURL": "{env:LITELLM_BASE_URL}/v1",
"apiKey": "{env:LITELLM_API_KEY}"
},
"models": {
// Grok
"grok-code-fast-1": {
"name": "Grok Code Fast 1",
"reasoning": true,
"modalities": { "input": ["text"], "output": ["text"] },
"limit": { "context": 256000, "output": 10000 }
},
// GLM
"glm-5": {
"name": "GLM-5",
"reasoning": true,
"modalities": { "input": ["text"], "output": ["text"] },
"limit": { "context": 204800, "output": 131072 }
},
"glm-4.7": {
"name": "GLM-4.7",
"reasoning": true,
"modalities": { "input": ["text"], "output": ["text"] },
"limit": { "context": 204800, "output": 131072 }
},
// Kimi
"kimi-k2.5": {
"name": "Kimi K2.5",
"reasoning": true,
"modalities": { "input": ["text", "image", "video"], "output": ["text"] },
"limit": { "context": 262144, "output": 262144 }
},
// MiniMax
"minimax-m2.5": {
"name": "MiniMax-M2.5",
"reasoning": true,
"modalities": { "input": ["text"], "output": ["text"] },
"limit": { "context": 204800, "output": 131072 }
},
// Qwen
"qwen3.5-plus": {
"name": "Qwen3.5 Plus",
"reasoning": true,
"modalities": { "input": ["text", "image", "video"], "output": ["text"] },
"limit": { "context": 1000000, "output": 65536 }
},
"qwen3-max": {
"name": "Qwen3 Max",
"reasoning": false,
"modalities": { "input": ["text"], "output": ["text"] },
"limit": { "context": 262144, "output": 65536 }
},
"qwen3-coder-next": {
"name": "Qwen3 Coder Next",
"reasoning": false,
"modalities": { "input": ["text"], "output": ["text"] },
"limit": { "context": 262144, "output": 65536 }
},
"qwen3-coder-plus": {
"name": "Qwen3 Coder Plus",
"reasoning": false,
"modalities": { "input": ["text"], "output": ["text"] },
"limit": { "context": 1000000, "output": 65536 }
} }
} }
} }
},
"experimental": {
"disable_paste_summary": true
} }
} }