Large Language Models (LLMs) are full of promise: instant access to a vast ocean of knowledge.
But what if you need a river, not an ocean? And what if you want the finest, freshest water? That’s when you need to go to the source.
Although LLMs capture the knowledge up to their training date, they are plagued by knowledge cut-offs, prone to "hallucinations", lack specialized domain knowledge, and they don’t like to cite their sources.
That is changing, as products like ChatGPT Plus gain the ability to dip into the web right now. But the web is a big place - despite seemingly infinite information, it often lacks the specialised data, those needles in the haystack, that many businesses require. Even when your LLM “browses the web”, you cannot be certain it is doing so meaningfully.
In other words, mass-market LLMs perform poorly on both recency and relevance. That’s why many businesses wanting a specialist, up-to-date knowledge engine are turning toward Retrieval Augmented Generation (RAG).