Intuition vs. Information: GPT's Know-How vs. Google's Database

Intuition vs. Information: GPT's Know-How vs. Google's Database

and why Retrieval Augmented Generation strategies are critical for LLMs


There is a big misconception surrounding GPT and other Large Language Models (LLMs): many believe that these models function as massive databases, tapping into the entirety of internet to craft responses to users questions.

It doesn’t work this way.

First, just considering the data size. While GPT's size is impressive, it's estimated to be under 1 million terabytes.

In contrast, the vast expanse of the internet size is more in the billions of terabytes. Comparatively, it's kind of contrasting a water bottle with a few Olympic-sized swimming pools.

Implicit (Tacit) knowledge vs Explicit knowledge

Going beyond data size, LLMs exhibit an amazing understanding of language, enabling them to abstract and infer based on context. This capability doesn’t come from accessing a vast database of facts but from what we call "parametric knowledge." With its parameters, LLMs generate the most probable text sequences in response to a given input.

However, it's crucial to distinguish: parametric knowledge isn't synonymous with facts. It is similar to what we use to call “tacit knowledge” in the earlier days of AI.

Think of it this way: an LLM's knowledge reservoir is like our human memory. We recall certain facts or phrases from all we've learned, connecting ideas and thoughts seamlessly. Interestingly, common sense facts (like: “fever is bad”) that we experience (and document) over and over is the type of knowledge, intuitive and non implicit, that makes its way into the LLM parameters.

Most of the time, our recollections are accurate, but occasionally, just like anyone, we might mix up details or even embellish a bit. After all, who hasn't done that once in a while? ?? I certainly have, only to be met with friends or family asking: "Show me the data! You are making up the numbers!".

Hallucinations is a big no no for B2B

This is called hallucinations in the context of LLMs is the #1 reason why, in B2B, LLMs and generative AI haven't (and cannot?) succeeded for tasks heavily dependent on content generation.

Using only parametric knowledge for answering questions is like asking a fresh MBA assistant to help only using its memory , not giving they access to a library nor to Google. It is like going back before writing and printing and relying only on “brain to brain” knowledge transfer.

The invention of writing and printing has indeed been a condition for business. Think about it: banking accounts, laws, contracts, records, product descriptions…. there is no business without factual recorded knowledge. Business and B2B demands explicit knowledge.

While it's OK and even desirable to have creative liberties in ad copy or artistic creations (hallucinations are not really a problem for image generation), the stakes are undeniably higher in B2B. B2B applications demand precision, accountability, and trustworthiness and answers must not only be accurate but also verifiable and dependable.

The RAG architecture

The result is that for B2B, an application cannot depend on parametric knowledge only and has to rely on external sources, using an architecture called Retrieval Augmented Generation.

This architecture works this way:

1- a body of knowledge, specific to the B2B task is built and encoded, sometimes in different formats (semantic vectors, plain text, knowledge graph). Let's call it the RAG Knowledge set

2- at inference time, the user query is compared to the RAG knowledge set and the most relevant articles/part of articles are added to the context of the query

3- the augmented prompt is sent to the LLM with instruction to formulate an answer relying on the information present in the context.

A more developed architecture is described in this document from Andreessen Horowitz. It shows (upper right) the need for a Vector database that encodes custom knowledge and is retrieved and passed as context at inference time.

Conclusion

The consequences for Corporations is that B2B generative applications are pivoting towards data and search challenges:

  • Data Collection Concerns:Availability and Context: Do we possess the necessary data, and can we pinpoint the relevant context? Should we enrich the data with public, social, private data that are commercially available?Accuracy and Organization: Is our data precise, and systematically arranged?Completeness: Are there gaps in our data repository?
  • Data Structuration Dilemmas:Storage Modalities: Should our data be housed within a file system, a document management platform, an SQL database, a semantic vector database through embeddings, or perhaps a knowledge graph?Granularity of Data: What should be the optimal size of our "data chunks"?Embedding Strategies: Which technique is most suitable for embedding our data?
  • Search Challenges:Volume: How many documents should our search retrieve? Keep in mind that although LLMs allow larger and larger contexts, most business models are “pay per token” and the larger the context the more we pay. Moreover, it's uncertain how effective very large contexts are. When the prompt is buried in a large context, precision diminishes.Matching: How can we best align the user's "prompt" with our knowledge repository?
  • Opportunities with NLP:At every juncture of this process, there are potential enhancements and optimizations that Natural Language Processing (NLP) can offer.

Overall this is not really new. The challenges and approaches resonate with AI & NLP projects that industry professionals have been grappling with since the 1980s.



Yann Gourvennec

Digital Strategist , Photographer, B2B marketer, Lecturer, Author and Entrepreneur

1 年

Thanks for shariing Dominique. However, I'm not sure a statistical parrot is entitled to any "know-how." In fact it doesn't "know" anything per se. Besides, Know-how implies that you know how to "do" things and GPT doesn't do anything apart from writing unless I am mistaken.

回复
Augie Ray

Expert in Customer Experience (CX) & Voice of the Customer (VoC) practices. Tracking COVID-19 and its continuing impact on health, the economy & business.

1 年

Very interesting. I appreciate you sharing some wisdom. I can't say I understand all the nuance with AI, and I appreciate people like you taking the time to help educate others.

Dominique Lahaix

CEO | Social Data | Social Intelligence | NLP | Artificial Intelligence | LLM

1 年

Here is a link to our Newsletter: wware.ai

回复

要查看或添加评论,请登录

社区洞察

其他会员也浏览了