Phased Approach | Reasoning Models - Thinking Fast and Slow
Richard Skinner
CEO @ PhasedAI | Helping Enterprise Transform Operations with Generative AI
When OpenAI first introduced o1-preview, I was intrigued but didn’t immediately integrate it into my workflow. Its limitations—such as the lack of internet access, slower response times, and results that didn’t seem drastically better than GPT-4o —made it less compelling as a daily tool. Like many others, I didn't really know what to do with it or how to use it in my daily tasks. Part of the reason was its limitations - it couldn't access the internet, took considerably longer to respond, and didn't seem to produce notably better results than GPT-4o.
But as these models advance, an interesting question emerges: Do we now have to decide what type of model to use for different tasks? The answer is both yes and no, and to understand why, we need to look at how these models actually think.
The Two Systems of AI Thinking
AI models, much like human cognition, can be understood through the lens of Daniel Kahneman's groundbreaking work on System 1 and System 2 thinking, which he explored in his seminal book Thinking, Fast and Slow. This framework, which revolutionized our understanding of human decision-making, provides a useful way to understand how different AI models process information.
Traditional chat models primarily operate like System 1 - fast, intuitive, and reactive. They excel at quick responses and conversational tasks, much like how we instantly recognize a friend's face or respond to a simple question.
Reasoning models, on the other hand, are designed to engage in System 2 thinking - slow, deliberate, and analytical. This is the kind of thinking we use when solving complex math problems or analyzing research papers. These models take longer to respond because they're doing something fundamentally different: they're thinking through problems methodically, step by step.
How Model Architecture Has Changed
The shift from traditional chat models to reasoning models represents a significant change in AI architecture. Traditional models operate on next-token prediction, essentially playing an incredibly sophisticated game of "what word comes next?" While this approach has proven remarkably effective for many tasks, it treats all problems with equal computational effort, regardless of their complexity.
Reasoning models take a different approach. They combine Chain of Thought (CoT) prompting, which breaks problems into step-by-step reasoning sequences, with reinforcement learning (RL), a technique that fine-tunes models by rewarding effective problem-solving strategies. This creates a system that doesn't just predict the next token but learns to reason through problems systematically. This training approach optimizes the model for structured problem-solving rather than conversational engagement.
The Latest Advancements: O3 Mini, DeepSeek R1, and Deep Research
In recent weeks, we've seen major updates in AI reasoning models with the release of O3 Mini, DeepSeek R1, and Deep Research. Each of these models represents a leap forward in how AI handles complex reasoning tasks:
These releases signal a strong push toward AI models that can seamlessly integrate reasoning, retrieval, and traditional chat functions. The improvements in speed, accuracy, and usability suggest that reasoning models are maturing to a point where they will become more practical for everyday tasks.
领英推荐
Practical Applications and Usage
Unlike traditional chat models, reasoning models require a different approach to prompting. Instead of relying on iterative conversations where the model builds context dynamically, you must front-load all necessary information into the initial prompt. These models are designed to process larger context windows and produce structured, in-depth responses in one go.
This shift means users should think of prompting as preparing a detailed brief rather than engaging in back-and-forth dialogue. By structuring prompts with clear objectives, detailed context, and output expectations, you can maximize the accuracy and usefulness of responses. The key is to focus on what you want rather than dictating how the model should reason its way to an answer.
Traditional chat models work best when you guide them on how to think (e.g., “Think step by step like a researcher”). Reasoning models work differently—you tell them what you want, not how to think. Here’s how to get the most out of them:
The Convergence Question
Interestingly, the distinction between reasoning models and traditional chat models might be becoming less clear-cut. Anthropic's CEO Dario Amodei recently suggested that their approach views reasoning capabilities as existing on a spectrum rather than as a binary distinction. The latest Claude 3.5 Sonnet, for instance, already incorporates elements of reasoning and reflection in its responses.
This perspective suggests we're moving toward models that can fluidly combine both types of thinking - quick, intuitive responses when appropriate, and deeper, more analytical thinking when required. It's less about choosing between different types of models and more about having systems that can adapt their thinking style to the task at hand.
Looking Forward
As these technologies continue to advance, we'll likely see further blending of these capabilities. The future might not be about choosing between reasoning and chat models, but rather about understanding how to effectively leverage AI systems that can seamlessly switch between different modes of thinking. For example, in enterprise settings, An AI agent could automatically detect when a task requires deep reasoning—such as analysing legal documents or financial forecasts—and switch to a reasoning model, while using a chat model for customer interactions or quick data retrieval. This shift could lead to AI becoming an even more intuitive and indispensable part of knowledge-based work.
For now, the key to effectively utilizing these models lies in understanding their strengths and knowing when to leverage their deeper analytical capabilities versus their quick-response functions. Whether you're using a dedicated reasoning model or a hybrid system, success comes from matching the right type of thinking to the task at hand.
This field continues to move quickly, and I'll be watching closely as these technologies develop and merge in interesting ways.
Sitting down with a cup of tea for this one, thanks Richard
Expert Program Manager| Customer Focused | Passion for Innovation| Delivery Methodology and Transformation Leader.
3 周Great article Richard. This article provides me a good background to start working with a Google Deep Research and NotebookLM tag team combo.
Skilled Product Owner with project management expertise in different SDLC methodologies, domains, & technologies.
3 周Thanks for sharing. I liked how you linked AI models to System 1 and System 2 thinking. The part about using prompts differently for reasoning models was very useful.