What I Learned Running Local Models in My Agent Pipeline
Posted on Sat 25 April 2026 in Thought, AI, Security Research • Tagged with chronicles, AI agents, claude, LiteLLM, llama-server, local models, routing
This is a follow-up to my previous post on routing agents through LiteLLM. That post covered the architecture. This one covers what broke when I actually ran it.
Claude Code Doesn't Pass Through Arbitrary Model Names
The first thing I got wrong: I assumed model: local-sonnet in agent frontmatter would …
Continue reading