1.2 KiB
1.2 KiB
model routing policy
primary model switching
nicholai can request any model via "switch to [alias]" (ordered by intelligence):
opus→ anthropic/claude-opus-4-5 (metered)glm→ zai/glm-5 (free)sonnet→ anthropic/claude-sonnet-4-5 (metered)kimi→ opencode/openrouter/moonshotai/kimi-k2.5 (free-ish)gemini-flash→ opencode/google/antigravity-gemini-3-flash (free)gemini-pro→ opencode/google/antigravity-gemini-3-pro (free)
sub-agent routing
when anthropic weekly usage > 80%:
sub-agents MUST default to free models:
- gemini-flash (preferred for lightweight tasks)
- gemini-pro (for heavier reasoning)
- glm-local (local fallback)
if ALL free models are unavailable:
- notify nicholai immediately
- ask: anthropic oauth or glm-local?
- do NOT auto-fall back to metered models
when anthropic weekly usage < 80%:
sub-agents can use any model as appropriate for the task
kimi fallback chain
- kimi via ollama (preferred, local)
- kimi via openrouter (fallback, notify nicholai)
- if both fail: notify nicholai for alternative
checking usage
before spawning sub-agents, check usage via session_status look at weekly usage percentage to determine routing tier