If you're using ChatGPT or other AI models, you've probably noticed they sometimes give incorrect information or hallucinate.
RAG helps solve this by searching through external documents, but this new approach takes a completely different approach - and it might just be what you need!
Good morning everyone!
As always, this is Louis-François, co-founder and CTO at Towards AI, and today, we'll dive deep into something really exciting: Cache-Augmented Generation, or CAG.
In the early days of LLMs, context windows, which is what we send them as text, were small, often capped at just 4,000 tokens (or 3,000 words), making it impossible to load all relevant context.
This limitation gave rise to approaches like Retrieval-Augmented Generation (RAG) in 2023, which dynamically fetches the necessary context.
As LLMs evolved to support much larger context windows—up to 100k or even millions of tokens—new approaches like caching, or CAG, began to emerge, offering a true alternative to RAG...
Learn more in the video (or written article here):
And that's it for this iteration! I'm incredibly grateful that the What's AI newsletter is now read by over 20,000 incredible human beings. Click here to share this iteration with a friend if you learned something new!
Looking for more cool AI stuff? 👇
Looking for AI news, code, learning resources, papers, memes, and more? Follow our weekly newsletter at Towards AI!
Looking to connect with other AI enthusiasts? Join the Discord community: Learn AI Together!
Want to share a product, event or course with my AI community? Reply directly to this email, or visit my Passionfroot profile to see my offers.
Thank you for reading, and I wish you a fantastic week! Be sure to have enough sleep and physical activities next week!
Louis-François Bouchard