docs(memory): prefer gpt-5.4 when subagent models fail
This commit is contained in:
@@ -40,6 +40,7 @@
|
||||
- In direct sessions with Will, cron jobs should use the `automation` agent by default unless Will explicitly says otherwise in that session.
|
||||
- Council tiers should use local LiteLLM-backed models for usage monitoring: light = `litellm/gpt-5.3-codex` with low thinking, medium = `litellm/gpt-5.3-codex` with high thinking, heavy = `litellm/gpt-5.4` with high thinking.
|
||||
- For non-trivial implementation work, treat `WIP.md` as the canonical state file and update it after each completed task/sub-task with status, concrete evidence, and the next recommended action.
|
||||
- If a subagent model choice causes execution/auth issues, prefer retrying implementation work on Codex GPT-5.4.
|
||||
|
||||
## Infrastructure notes worth remembering
|
||||
- Full `~/.openclaw` backups upload to MinIO bucket `zap` and are scheduled via OS cron every 6 hours.
|
||||
|
||||
Reference in New Issue
Block a user