I work on research at Chroma, and I just published our latest technical report on context rot.
TLDR: Model performance is non-uniform across context lengths, including state-of-the-art GPT-4.1, Claude 4, Gemini 2.5, and Qwen3 models.
This highlights the need for context engineering. Whether relevant information is present in a model’s context is not all that matters; what matters more is how that information is presented.
Here is the complete open-source codebase to replicate our results: https://github.com/chroma-core/context-rot
Comments URL: https://news.ycombinator.com/item?id=44564248
Points: 24
# Comments: 1
Jelentkezéshez jelentkezzen be
EGYÉB POSTS Ebben a csoportban

Article URL: https://underwriting-superintelligence.com/
Comments URL: http

Article URL: https://github.com/hazelgrove/hazel
Comments URL: https://news.ycombin

Article URL: https://aethermug.com/posts/human-stigmergy
Comments URL: http

Article URL: https://cartesia.ai/blog/hierarchical-modeling

Article URL: https://calv.info/openai-reflections
Comments URL: https://news.ycomb