OpenAI DevDay: Unleashes GPT-4 Turbo. Kills traditional GenAI RAG Solutions
Siva Appavoo
Senior Technology Leader - Unleashing AI/ML, Data Analytics, and RPA to Propel Innovation and Transform Organizations | Forbes Interviewee | FT Intelligent Business Award Winner
OpenAI hosted its inaugural developer event on Monday, and it made quite an impact. The company unveiled upgraded models and new APIs. In this article, I'll provide a summary of the noteworthy features, how they relate to existing offerings that business and tech leaders should be aware of in this rapidly evolving GenAI landscape.
Highlights
New GPT-4 Turbo released with impressive features:
New GPT-3.5 Turbo released with enhanced features:
Enhanced Function calls.
Reproducible outputs.
New “Assistant API” (a game changer).
Custom Models - Selected organizations with extremely large proprietary data (billions of tokens) can work with OpenAI researchers to train custom GPT-4 models to their specific domain and keep it private.
Copyright Shield to defend and cover legal costs for customers facing copyright infringement claims.
Some of these features deserve a closer look to understand their game-changing potential.
Let's explore a few of them in more detail:
1.?Function Calls
What? Connects OpenAI models to your internal or external APIs and data sources.
Why? ?You can leverage the power of LLMs and your deep domain knowledge to:
The Old:
New:
2.?Reproducible Outputs
What? Allows higher degree of control over the model behavior ensuring consistent completions by the LLM (mostly).
Why? Consistency and reliability are crucial for customer-facing use cases that rely on your internal knowledge. You wouldn't want your chatbot to provide different variations of the same response with each interaction.
The Old:
Model output is mostly random (non-deterministic). Dependable "Repeatable output" (deterministic) was a distant dream. The popular ways to maximize deterministic outputs are to adjust the Temperature and Top-K/Top-P settings yet falls short for many use cases that need consistent output.
The New:
To control the output, set a consistent integer value as the seed parameter across requests (e.g., 12345), and maintain uniform values for all other parameters (prompt, temperature, top_p, etc.) across requests, and monitor the system_fingerprint field in the response to track model and configuration changes.
?
3. Assistants API
What? An assistant is a purpose-built AI that follows specific instructions, utilizes additional knowledge (your proprietary data), and can invoke models and tools to perform tasks (send emails, post in social media etc.).
Why? This feature enables businesses to create custom agents with distinct personalities and enhanced capabilities to meet their specific needs.
The Old:
Build your own agents to assemble required tasks to complete a unit of work.
New:
Retrieval feature eliminates the need for complex computations and storage of document embeddings, as well as the implementation of chunking and search algorithms. The Assistants API optimizes the choice of retrieval technique based on OpenAI's experience in building knowledge retrieval in ChatGPT. It requires gpt-3.5-turbo-1106?and?gpt-4-1106-preview?models.
Overall, this is a formidable feature that has the potential to render RAG architecture (Extraction, Chunking, Embedding, Indexing, Relevant retrieval, Context size management, Conversation State management) obsolete.
I am excited to experiment with GPT 4 Turbo.