Discussion about this post

User's avatar
JP's avatar

The org structure framing is spot on. I keep running into people who think better prompts will fix context spillover, but it won't. You need isolation between concerns. What I find interesting is how skills fit into this picture too. The multi-agent setup works better when each agent has clearly scoped instructions it can load on demand rather than dumping everything into one massive system prompt. I dug into the fragmentation side of this recently; every tool from Cursor to Copilot has invented its own way of handling agent instructions, which makes the skills interoperability angle even more important: https://blog.devgenius.io/agent-skills-explained-teaching-your-ai-assistant-new-tricks-c6ee66a632c4

Pawel Jozefiak's avatar

The organizational structure angle is spot on. I've been managing a multi-agent setup for a couple months and the biggest surprise wasn't technical capability - it was how much it feels like managing actual junior engineers.

You write tasks, check their work, tell them when they're going down a rabbit hole. The failure modes are even similar: agents get stuck on details, sometimes ignore constraints, occasionally deliver exactly what you asked for but not what you meant.

Has anyone figured out good patterns for code review with agents? Right now I'm treating it like PR review, but that might not be the right mental model.

No posts

Ready for more?