this looks like a very good discussion on how context helps save token cost for AI models. I like the technical depth comparing the different techniques used in the blog.
This is a very clear comparison of file-based context vs a memory layer. I liked the way it derived the queries into different categories, it makes it easy to understand the metrics.