Introducing The Context and Outputs Library (COOL) Framework For AI Agents:
Franck Boullier
Chief Digital Officer @ UNIQGIFT | Advisor | Entrepreneur | FinTech, DLT, AI | INSEAD MBA
Executive Summary:
Large Language Models (LLMs) like GPT (Open AI), Gemini (Google), and Claude (Anthropic) are becoming increasingly powerful and capable.
Research suggests these models are converging towards a shared understanding of the world.
Context and prompt engineering will be essential for AI Agents to deliver the best possible responses.
However, this creates new challenges:
Doing this well is critical to ensure accuracy, build user trust, enhance decision-making, enable continuous learning, and comply with regulatory and ethical standards.
The COOL Framework, inspired by the widely used IT Infrastructure Library (ITIL) Framework, is a comprehensive set of principles, processes, and practices designed to track and manage:
COOL's key objectives include:
Adopting COOL can significantly improve how organizations harness AI technology and AI Agents, ensuring AI developments align with business goals and contribute meaningfully to decision-making processes.
From Attention To Context:
The 11 pages Attention Is All You Need paper written in 2017 quick-started the Generative AI revolution that gave us Chat GPT, Gemini, Claude, Midjourney, and many more Gen AI-based tools.
The paper introduced the Transformer Architecture and the concept of Attention.
Attention allows AI models to focus on and weigh the importance of what the AI knows (the input) to generate the best possible response (the output) based on the description of the task to perform (the prompt).
In the research paper The Platonic Representation Hypothesis - 13 May 2024, the authors argue that AI models are converging:
Neural networks, trained with different objectives on different data and modalities, are converging to a shared statistical model of reality in their representation spaces.
If true, this finding has major implications for the future of LLMs: they will all have, eventually, more or less the same "model of reality".
With the continuing trend of models scaling up, (...) model alignment will increase over time – we might expect that the next generation of bigger, better models will be even more aligned with each other.
For AI Agents, that means that the quality of the information you provide (the input) and how well you describe the task to perform (the prompts) will be THE key differentiators between AI Agents tomorrow, not the LLMs they are based on anymore.
What Is Context - Why It Matters?
The concept of Context (or Context Window) in Large Language Models (LLMs) today derives from the Attention mechanism we mentioned earlier.
Context refers to the surrounding information that helps the model understand the meaning of words, images, and any other types of content that form the input they receive from the users.
Using the provided context and then combining that context with the user inputs and previous interactions an LLM like Open AI's GPT, Google's Gemini, or Anthropic's Claude can generate coherent, relevant responses even when there are complex interactions with the user.
In essence, AI Agents are like interns - they NEED context to give the best possible responses.
If you have ever built an Open AI GPT (a type of AI Agent), you know that the quality of the responses you'll get from your GPT is directly related to the quality of the prompt AND the quality of the context (the additional information) that you have provided to your AI Agent:
This information allows the AI agent to "narrow down" the scope of the conversation.
It also gives the AI Agent useful information and data it can use when it interacts with users.
Techniques such as Retrieval Augmented Generation (RAG) automatically add additional context to whatever the user is asking. This helps improve the quality of the AI Agent's response, reduces the risks of hallucinations, and is one of the most efficient techniques to ensure that the AI Agent stays on topic.
Context Windows Are Getting Larger:
Open AI Chat GPT was launched in November 2022 with a context window of 4,096 tokens, the equivalent of 3,000 to 4,000 words or about 5 to 6 pages of text.
The context had to stay small to avoid hitting the context window limit.
The size of the context that you can pass to the model became one of the key differentiators between LLMs: more context leads to better responses.
Researchers are working on the concept of "infinite context", a context window that would equip an AI Agent with the capability to process and utilize ALL the context that you could provide:
We may soon be able to create AI Agents with a perfect memory of everything they were exposed to.
How to efficiently format, update, manage, and access this context information?
We Need A Framework To Manage Context:
In the article Moving Past Gen AI Honeymoon Phase - from May 2024 McKinsey highlights:
Context WILL change over time as you get more data and the business environment evolves.
Getting this right requires significant human oversight from people with relevant expertise.
Managing And Monitoring AI Outputs:
Outputs are the content generated by AI Agents based on:
Outputs generated by AI Agents are the tangible results of the AI Agent's work.
If context is key to making sure that an AI Agent has all the information it needs to perform effectively, the AI Agent's outputs must also be tracked and managed efficiently.
In an article from August 2020, I highlighted the 5 reasons why you need explainable AI. The AI Verify Foundation, in collaboration with Singapore's IMDA (Infocomm Media Development Authority), has proposed a Model AI Governance Framework for Generative AI - June 2024 has proposed a comprehensive approach towards Generative AI governance.
Today more than ever, you need to be able to track, explain, and audit the outputs generated by AI and AI Agents.
Ensuring Accuracy and Relevance:
AI Agents can sometimes generate outputs that are inaccurate or irrelevant.
By tracking and managing these outputs, organizations can quickly identify and correct any discrepancies, ensuring the information provided is both accurate and pertinent to the user's needs.
User Trust and Brand Reputation:
The quality and reliability of AI Agent outputs directly impact user trust and the overall reputation of your brand.
By ensuring that outputs are consistently high-quality, relevant, and free from errors, organizations can build and maintain trust, which is crucial for long-term success.
领英推荐
Enhancing Decision-Making Processes:
In many cases, the outputs of AI Agents are used to inform decision-making processes.
Efficient tracking and management ensure that these decisions are based on the most accurate and up-to-date information, leading to better outcomes.
Continuous Learning and Improvement:
By monitoring the outputs of AI Agents, organizations can gather valuable data on their performance, which can be used to train the models further or improve the AI Agent itself.
This continuous learning loop will improve your AI Agents over time, making sure that they are adapting to new data and becoming more effective in their tasks.
Identify And Prevent "AI Hallucination":
Large Language Models can sometimes generate plausible but entirely fabricated information: the infamous "hallucinations."
Monitoring outputs is crucial for identifying and mitigating these occurrences and preventing the creation of false or misleading information.
Intellectual Property and Data Security:
Outputs generated by AI Agents may contain sensitive or proprietary information.
Managing AI Agent's outputs helps protect intellectual property and adhere to data security protocols, preventing unauthorized access or misuse of sensitive data.
Regulatory Compliance and Ethical Considerations:
In many industries, regulatory compliance mandates strict oversight of automated decision-making processes.
Most AI regulations that are being drafted around the world are focusing on the need for AI Agents to be auditable and explainable.
Tracking outputs is essential for demonstrating compliance with regulations. Moreover, it'll help your organization follow ethical standards, ensuring that AI Agents do not generate biased, discriminatory, or harmful content.
The Egg Theory of AI Agents:
In The Egg Theory of AI Agents - 30 May 2024 Rex Woodbury reminds us of the importance of making sure that the user can contribute to and work with AI Agents instead of being passive consumers of the outputs generated.
The egg theory is a consumer psychology concept that explains why people are more likely to use a product if they feel like they have contributed to it in some way:
When instant cake mixes came out, they sold poorly. Making a cake was too quick and simple. People felt guilty about not contributing to the baking. So companies started requiring you to add an egg, which made people feel like they contributed. Sales soared. It turns out that there’s such a thing as too easy
Rex uses the following example to illustrate his point:
I might say to an AI agent, “Book me a flight to Paris on July 3rd.” It would be uncomfortable to remove me entirely from the workflow. I might feel lazy, guilty, even nervous about it. Did the bot book the right flight? Does it know I prefer morning flights, hate red eyes, and am a loyal Delta SkyMiles member?
The Context and Outputs Library (COOL) Framework:
Introducing the Context and Outputs Library (COOL) Framework.
The COOL framework is designed to help organizations manage:
The COOL framework is a set of principles, processes, and practices to manage the documents, data, and information (the artifacts) that are:
The COOL framework covers the full lifecycle of these artifacts.
If this sounds familiar, it's because the COOL framework is directly inspired by the IT Infrastructure Library (ITIL) Framework. ITIL is a widely adopted framework for IT service management.
Let's explore the parallels.
COOL and ITIL Share Common Principles:
ITIL emphasizes viewing IT as a service provider focused on customer needs, managing expectations, and continuously improving. It aims to align IT with the business, optimize costs, and enhance quality across strategy, design, transition, operations, and improvement.
Similarly, COOL positions AI Agents as a service to the organization.
It focuses on:
?The objectives of the COOL framework are to:
COOL Principles:
ITIL best practices are guided by 7 key principles:
These principles map remarkably well to the world of AI Agents:
Implementing The COOL Framework:
Putting the COOL framework into practice involves steps that mirror an ITIL implementation:
Wrapping It Up:
By adopting the COOL framework and its ITIL-inspired principles, organizations can:
Borrowing from Rex Woodbury's article again:
The best companies will be savvy in how they embed human decision-making into workflows, rather than removing the need for human input altogether. (...) Winning products (...) will be those that offer a bridge from the world of human work to the world of software work, making us feel comfortable and in control along the ride.
The COOL framework can help build that bridge.
Your Turn!
This article is just a starting point.
Share your insights and experiences in the comments below.
I'm eager to hear your thoughts!
Sources:
?
Senior Manager at Neoris || INSEAD MBA || Co-founder BolzanoSlushD
4 个月Cool summary Franck Boullier! I wonder how close is the "infinite context" you mention with "common sense" and "tacit knowledge", and how they overlap. From what I have seen, it is somewhat easy (kinda of) to convert explicit knowledge into something actionable via GenAI. Depending on the final user's maturity, this can be well documented. However, to capture everything else you really have to build that layer on top to monitor and audit that you mention, but it is challenging to capture every possible scenario. This until we don't have a general model (very close to ideas of AGI) that is able to do this without explicit instruction, but i have no clue on how close it could be.
SID Accredited Director - Director Strategy & Innovation | Business Development | Entrepreneur | Solution Finder > 20yr+ Exp
4 个月Very insightful. Interesting approach to structure the Data and AI governance in any structure.
Building Generative AI , Single and Multiple Agents for Enterprises | Mentor | Agentic AI expert | SAP BTP &AI| Advisor | Gen AI Lead/Architect | SAP Business AI | SAP Joule
4 个月It would be prudent to explore how generative AI can enhance decision-making processes in sectors with stringent compliance and governance requirements.
Walk the Talk - Generative AI - Thinkers50 Radar - VP, Co-lead of The Management Lab by Capgemini Invent
4 个月Very comprehensive and insightful. Very interesting the context windows size exponential growth. For AI Agents, unfortunately I have no longer access to ChatGPT Plus, so I can't assess latest developments on the GPT Builder. But the direction is clearly that you described.