Notable Quotes
"The context window is the main constraint that most AI coding agents face these days."
— [Speaker]
"The more information that you give a model, the worse it's going to perform."
— [Speaker]
"All models suffer from a problem of retrieving information from their own context. This is the classic needle in a haystack problem."
— [Speaker]
"The information in the middle of the chat is going to be less prioritized by the LLM."
— [Speaker]
"Models just do better with less, more focused information, just like humans do."
— [Speaker]
"Regularly clearing your coding agent chats will refresh the agents memory and clear its context window making for much better performance."
— [Speaker]
"Whenever you're working with a coding agent, you really do need full transparency, full understanding of what's happening in your context window at any time."
— [Speaker]
"When you're assessing an LLM, you shouldn't just look at how big the context window is. You should look at how well it retrieves information from its context window."
— [Speaker]
"MCP servers are super attractive because they allow you to plug and play with different pre-made tool sets out there in the ecosystem, but they can bloat your context incredibly rapidly."
— [Speaker]