You need to understand and manage the context that is being ingested while you’re interacting with an LLM. There’s a finite window, and when you start iterating and going back and forth with it, that context window is filling up, and eventually the stuff you started with starts to fall out of the context.What've noted is that LLMs are good if you have a full specification first - then run through once to produce code.
I suspect the attention mechanisms fail to maintain the specification once it becomes an iterative definition. This latter limitation results in poor architecture and systems integrations from what I've seen so far.
Do you use more persona based rule sets in the context? For example - adding the security analyst, the operational monitoring, the financial cost management (for both cost visibility and savings etc)?
Similarly with tools and MCP, there is a finite amount you can use and have it remain useful. 30 tools is about the upper limit that can be used.
There are various context management strategies that improve things. Using a scratch memory to track tasks etc. Claude does this when you ask it to plan something.
Also, smaller contexts are generally better. If you have 10 rules in a prompt, it’ll be pretty good at sticking to them. Have 100 rules, much less so.