AI can't learn without asking

AI can't learn without asking

I ran across this great video from IBM Technology on "Can AI Think? " It points out that LLMs (and neural networks in general) perform Probability Pattern Matching. It mentions the development of Inference-Time Compute (aka Test-Time Compute), in which Agentic LLMs become more iterative and have them check and refine results by spending longer "thinking." The results of this approach are quite impressive. YouTuber Matthew Berman demos DeepSeek R1, showing R1 planning and chaining 'reasoning' together to create the game Tetris from a 1-sentence prompt.

?This advancement increases the scope of what GPT AI can tackle. OpenAI's paper from this week describes how their Inference-Time Compute improved the quality and robustness of their o1-preview model. To oversimplify, agents run a prompt through a sequence of LLMs, enabling them to check and refine the output multiple times for a better result.? Another outcome of this approach is better protection against adversarial attacks, such as intentionally malformed prompts (i.e., soft-token.)

?However, one thing that continues to be glaringly missing is the ability of these agentic models to ask questions. ?Even in the Inference-time compute, the model is essentially 'asking' other models (which are always limited) for further input and analysis. Using the Socratic method to learn is completely missing from AI Models today. At this time, any Socratic dialog between humans and AI is one-directional only. The human can learn more by asking questions of the GPT chatbot. But it doesn't go the other way, and it should.

To more fully simulate human intelligence, AI models need to learn by asking questions. They need to know when to ask someone else a question, and more importantly, they need to know how to find the right person to ask. Then, they need to incorporate what they've learned (and probably verify it) into their model.

Yes, some AI models ask questions, but this is typically only to help clarify gaps identified in the prompt. Sometimes, the LLM model may generate the questions only because of its probability pattern matching.? The one method I'm aware of that comes closer to truly 'learning' AI is Reinforcement Learning/Training.? But some are looking in the right direction. A just-published paper titled "Collaborative Framework for Dynamic Knowledge Updating and Transparent Reasoning with Large Language Models" looks at how LLMs can be combined with Knowledge Graphs that can update dynamically. Google DeepMind released a paper on how a 'Boundless Socratic Learning" approach could enable AI models to be self-improving.

Until an AI agent can dynamically ask me questions like? "Why do you want to know?" or "Who is the doctor you're mentioning?" AI remains a helpful information savant but is still far from being able to reason or think. Anyone telling the press we're close to AI actually thinking or even the singularity is only trying to cash in on the hype or delusional. Or both. There's still a lot of (human) work to move AI forward, but that's what keeps it interesting. I’m looking forward to it.

?

?#ArtificialIntelligence

#LLM #LargeLanguageModels

#LargeConceptModels

#AIResearch

#MachineLearning

#NaturalLanguageProcessing

#TechInnovation

#AITrends

#SemanticAI

#FutureOfAI

?

[Rick Munoz started working in AI at Symbolics, Inc. in the 1980s. He went on to incorporate AI components like Expert Systems, Natural Language Processing, and Fuzzy Logic into multiple systems. He currently designs and implements large cloud-based applications that include AI capabilities.]

Rick Munoz

Chief Architect and CTO at T4S Partners

1 个月

Here's a good example of why this is an issue for #AI: https://futurism.com/openai-asks-permission-important

回复

要查看或添加评论,请登录

Rick Munoz的更多文章

  • Why We Need LLM Output Gatekeeping - Policing AI Agents

    Why We Need LLM Output Gatekeeping - Policing AI Agents

    In Computerphile's recent video on "Generative AI's Greatest Flaw," Mike Pound shows how indirect prompt injection…

  • Dividing Event-Sourcing into nanoservices

    Dividing Event-Sourcing into nanoservices

    A few years ago, I led the design of a cloud-native financial management system meant to run globally, with users in…

  • Which Low-code development platform is best?

    Which Low-code development platform is best?

    Are you grappling with deciding which low-code application development platform to use? The landscape of low-code (or…

    2 条评论
  • DeepSeek's AI improves when rewarded

    DeepSeek's AI improves when rewarded

    Reinforcement Learning (RL) has been a popular approach to training and improving AI models. Here's an…

    2 条评论
  • Meta AI's LCM: a good evolutionary step

    Meta AI's LCM: a good evolutionary step

    Meta's recent paper on Large Concept Models (LCMs) is being heralded as the next generation of AI technology. This is…

  • Three Keys to Successful App Transformation

    Three Keys to Successful App Transformation

    How old are your applications? Your database systems? Is any of your technology nearing (or way past) its end-of-life…

社区洞察

其他会员也浏览了