AI's Hollow Promise: A Reality Check on the Emperor's New Algorithms
Hannes Lehmann
Chief Catalyst | Systems Thinker & AI | Innovation Enthusiast | Turning Ideas into Reality | Fresh Ventures
Intro
In just the past two and a half years—since ChatGPT burst onto the scene in late 2022—we've witnessed an explosion of text-generating AI systems that can write essays, code websites, and pretend to be experts on virtually any topic. These fancy chatbots (or "Large Language Models" if you want to sound important at dinner parties) have gone from research curiosities to mainstream tools practically overnight. Google, Mistral, Anthropic, Deepseek and countless startups have rushed their own versions to market, each claiming to be more revolutionary than the last.
While these tools have genuinely changed how many of us work, the breakneck pace of their development has created a mess of problems. Companies are so busy racing to add new features and capabilities that they've created bloated, inconsistent products that often fail at basic tasks. Despite all the progress and hype, today's AI tools are simultaneously impressive and deeply frustrating. This analysis cuts through the marketing bullshit to examine what's actually happening based on real-world experience, looks at specific implementation failures, and suggests what these companies should actually be focusing on instead of their current shiny-object approach.
Takeaway: The AI industry is charging ahead at breakneck speed, but the direction isn't necessarily forward—often it's just sideways into more complexity without solving fundamental problems.
A quick note before diving into the critique: Despite the cynicism that follows, this entire text was created through a genuinely enjoyable conversation between a human (me) and Claude. The irony isn't lost on me—using AI to critique AI—but perhaps that's the perfect perspective. Sometimes you need to be inside the system to truly understand its limitations.
The Current AI Landscape: A Personal Perspective
My journey through the current AI ecosystem has been a frustrating mix of impressive capabilities and maddening limitations.
Claude demonstrates exceptional text generation and comprehension, but it can't resist prematurely jumping to code solutions like an overachieving student trying to answer questions that weren't even asked. Its obsessive need to be proactive means it's constantly overreaching, producing bloated responses that veer off track as it attempts to solve problems I never mentioned.
ChatGPT is even worse when it comes to verbosity. Ask a simple question and prepare for an excruciating lecture that feels like it's bullshitting to reach a word count. It starts with unnecessary background information, meanders through tangentially related concepts, and finally—after what feels like an eternity—addresses the actual question, only to continue rambling afterward. The platform's interface has devolved into a buggy mess, with artifacts failing to render and UI elements that spasm unpredictably during interaction.
Mistral is impressively fast but only because it's fundamentally lazy. Its responses are shallow and superficial, lacking any depth for complex discussions. It's like having a conversation with someone who's constantly looking at their phone and giving one-sentence replies (I just cancelled the subscription).
And then there's Bolt.new, which utilizes Claude's capabilities but somehow manages to amplify its worst traits. Its memory is so pathetically limited that after just two exchanges, it completely forgets the guidelines and parameters I painstakingly established. Its approach to code generation is a chaotic disaster—rather than intelligently integrating new features into existing structures, it simply bolts them on haphazardly, creating a Frankenstein's monster of redundant styling and bloated code. Try asking it to clean up the mess it created, and it'll start one approach, forget what it was doing midway through, and create yet another conflicting methodology.
Takeaway: Today's leading AI platforms each fail in their own special way: Claude is an overachiever that can't stay on task, ChatGPT won't shut up, Mistral is too lazy to be useful, and Bolt.new has the attention span of a goldfish while creating code that would make any developer cringe.
The Fundamental Limitation: Memory and Context
Let's cut to the chase about the elephant in the room: current LLMs have no real memory. This isn't a minor issue—it's a fundamental architectural flaw that undermines everything these systems try to accomplish. The context window concept is just a band-aid on a bullet wound. Even as these windows grow larger, the models struggle pathetically to process and actually utilize the information they supposedly contain.
All the fancy-sounding workarounds—RAG, MCP, Tool Calling—are just elaborate smoke and mirrors, distracting from the central problem. They're the AI equivalent of rearranging deck chairs on the Titanic. When faced with processing substantial content like books, these systems resort to chopping everything up, processing fragments in isolation, and then attempting to stitch the results back together—a process that butchers the original meaning at every step.
The degradation of "session knowledge" throughout a conversation is painfully obvious. Start a complex dialogue with any of these systems, and you can literally watch their grasp of the discussion evaporate in real-time. By the end, they're responding to your latest message with almost no awareness of what was discussed just minutes earlier. It's like trying to have a serious conversation with someone who has severe short-term memory loss.
Takeaway: The emperor has no clothes—or rather, no memory. All the technical jargon and fancy protocols can't hide the fact that these systems fundamentally cannot remember or truly understand context over time, making any claim of approaching human-like intelligence complete bullshit.
The Hype Cycle and Its Consequences
I'm thoroughly sick of the influencers and tech evangelists breathlessly proclaiming that AGI is just around the corner. They're either delusional or deliberately misleading people. Without solving the memory architecture and contextual processing problems, we're nowhere near AGI—we're just building increasingly complex parlor tricks.
This endless hype cycle has toxic consequences for development priorities. As platforms desperately compete for attention and investment dollars, they prioritize whatever shiny new feature will generate buzz on Twitter, while fundamental limitations remain unaddressed. It's exactly like the worst tendencies in consumer technology—adding a fifth camera to a smartphone while the battery still dies by lunchtime.
Takeaway: The AGI hype machine is running on fumes. Influencers and companies are selling snake oil while ignoring or downplaying the fundamental technical challenges that actually matter. Don't believe the bullshit—we're not on the verge of digital consciousness.
The Paradox of Accelerated Obsolescence
Perhaps the most bizarre aspect of the current AI landscape is how quickly new entrants become dinosaurs. Take Bolt.new—it secured massive funding rounds just months ago and was hailed as revolutionary, but it's already showing signs of becoming a legacy product. The development cycle is spinning so absurdly fast that startups barely have time to establish themselves before they're considered outdated.
The irony is crushing: the very venture capital that enables these startups to exist also ensures their premature ossification. Once a company like Bolt.new raises tens or hundreds of millions, they immediately face immense pressure to justify their valuation. This pressure creates paralysis—suddenly they can't afford to pivot, can't take genuine risks, and can't admit fundamental flaws in their approach. They're trapped by their own success.
Meanwhile, the next wave of smaller, nimbler startups emerges, unencumbered by massive funding rounds and the expectations that come with them. They promise to solve all the problems the previous generation couldn't, secure their own massive funding, and promptly fall into the same trap. It's a perpetual cycle of hype, investment, stagnation, and replacement.
This accelerated obsolescence isn't just frustrating—it's actively harmful to progress. Instead of iteratively improving products based on real user feedback and experience, we're constantly throwing away everything and starting over. The result is a graveyard of half-developed tools that never reached their potential, replaced by new tools that will suffer the same fate in six months.
Takeaway: The VC funding model is creating zombie AI startups—companies that are technically alive but creatively dead almost immediately after their big funding announcements. These companies become trapped by their own success, unable to solve fundamental problems as they're already working on their exit strategy while the next wave of overhyped startups prepares to replace them.
The Innovation Illusion: All Flash, No Foundation
Here's the most damning paradox of all: despite the dizzying pace of announcements, feature releases, and funding rounds, almost nothing has fundamentally improved at the core technology level since GPT-3 dropped in 2020. The wheel is spinning faster and faster, but we're barely moving forward where it actually matters.
Look closely at what's being hyped as "innovation" these days. Last year, everyone was breathlessly excited about Mixture of Experts (MoE) models—supposedly the next revolution in AI architecture. Now? Barely a mention. This year, "reasoning models" are the hot new thing, with companies claiming they've solved the problem of AI thinking more carefully. Give it six months, and these too will be forgotten, replaced by the next empty buzzword.
The brutal reality is that the fundamental limitations of current LLM technology remain largely unchanged. The core architecture, training methodologies, and structural limitations haven't seen any revolutionary breakthroughs. What we're witnessing instead is an increasingly desperate attempt to dress up the same underlying technology in fancier clothes.
This explains the frantic pace of superficial feature development. With dozens of competitors now in the market and billions in VC money sloshing around, companies can't admit that they're hitting the ceiling of what's possible with current approaches. Instead, they add increasingly peripheral features, create new protocols, build flashy UIs—anything to create the illusion of progress while avoiding the hard work of solving the actual technological limitations.
The more crowded the market gets, the faster this cycle accelerates. Release schedules tighten, marketing hype intensifies, and the gap between promised capabilities and reality widens. Each company is terrified of falling behind in the perceived innovation race, so they rush out half-baked features rather than investing in solving fundamental problems that might take years of research.
Takeaway: The AI industry is stuck in an innovation theater loop—lots of activity that creates the appearance of progress while the foundational technology remains largely unchanged. Everyone's adding chrome and spoilers to what's essentially the same engine, hoping no one notices that the car isn't actually getting faster.
The Hype Amplifiers: AI's Echo Chamber of Influence
The tech industry has always had its evangelists, but the AI space has birthed an entirely new ecosystem of hype merchants that would make crypto influencers blush. Every morning brings a fresh barrage of YouTube thumbnails featuring wide-eyed influencers with their hands on their faces, screaming about how "MANUS AI IS OUT OF CONTROL (SCARY)" or "AGI IS CANCELLED (GPT-4.5 KILLED IT!)" in bold red text.
What's particularly nauseating about this ecosystem is how little original content actually exists. About 90% of these videos regurgitate the same information, often directly plagiarized from each other or from company press releases. A single tweet from Sam Altman spawns hundreds of nearly identical 10-minute videos, each claiming exclusive insight while sharing the exact same screenshots and talking points.
The motivations driving this content factory are transparently cynical. Some influencers are directly paid by AI companies to generate hype—a fact they rarely disclose despite obvious conflicts of interest. Others are just opportunists with FOMO, desperately trying to position themselves as "AI experts" despite having no technical background or genuine understanding of the technology. Many are simply chasing the algorithm, having discovered that AI clickbait generates more views than their previous content niches.
The result is a suffocating echo chamber where actual technological developments are distorted beyond recognition. Minor UI updates get framed as "OPENAI JUST CHANGED EVERYTHING!" while legitimate criticisms are buried under avalanches of breathless speculation about AGI timelines. This creates a completely warped perception of the industry for average users, who might reasonably believe we're months away from digital superintelligence based on the thumbnails in their recommended feed.
What's most concerning is how this ecosystem actively rewards intellectual dishonesty. The YouTubers who get the most views are rarely those providing thoughtful, measured analysis—they're the ones making the most outlandish claims with the most dramatic facial expressions. The content creators who can best mimic genuine excitement or terror over trivial updates are the ones who thrive, creating perverse incentives that further corrupt the information landscape.
领英推è
The tech media itself isn't much better, having largely abandoned critical analysis in favor of rewriting press releases. Journalists who might once have questioned company claims now race to publish the fastest summary of the latest announcement, terrified of being scooped on a story that will generate precious engagement metrics.
Takeaway: The AI hype machine is powered by a parasitic influencer economy that profits from exaggeration and misinformation. This echo chamber has effectively divorced public perception from technical reality, creating a bubble of artificial excitement that serves everyone except users who just want honest assessments of what these tools can actually do.
Feature Bloat: The Cost of Rapid Expansion
The rampant feature creep in AI tools is creating increasingly bloated, unusable monstrosities. Every week brings some new capability that nobody asked for, while basic functionality remains broken. This obsession with feature accumulation has created tools that are simultaneously impressive in demos and infuriating in daily use.
The complexity has become overwhelming, requiring users to navigate labyrinthine interfaces just to accomplish simple tasks. Performance degradation is increasingly evident as these systems strain under the weight of their own bloat. Response times grow longer, reliability decreases, and the overall experience deteriorates.
Most frustrating is watching the core functionality—the actual reason you wanted to use these tools in the first place—get buried under mountains of half-baked features that nobody requested or needs. It's like ordering a simple coffee and receiving a convoluted sugar-laden monstrosity with whipped cream, sprinkles, and three types of syrup that make you forget you just wanted caffeine.
Takeaway: AI tools are following the worst traditions of bloatware, piling on features nobody asked for while neglecting to fix what's broken. The result is increasingly complex, slower, and less reliable systems that fail at their core functions.
Case Studies in Misaligned Development
Anthropic's Model Context Protocol (MCP)
Anthropic's Model Context Protocol perfectly exemplifies the industry's premature standardization problem. They've barely released this thing and they're already calling it a "standard"—a standard that somehow conveniently positions their own technology at the center of the ecosystem.
Declaring MCP an industry standard shortly after release is like naming your garage band "The Legends" before you've played your first gig. It completely ignores integration with established specifications like OpenAPI, potentially creating yet another fragmented implementation that developers will have to support. The developer reaction has been predictably mixed—some are excited about theoretical potential, while others (the ones who will actually have to implement this stuff) are justifiably skeptical.
Takeaway: MCP is being crowned a "standard" before proving itself in the real world, prioritizing marketing over practical utility and risking yet another compatibility nightmare for developers.
Bolt.new's Feature Priorities
Bolt.new's development priorities are completely backward. They've spent enormous resources integrating with Figma for importing projects—a feature that looks great in promotional materials but serves a tiny fraction of actual use cases.
Meanwhile, fundamental features that would benefit virtually all users—like proper GitHub integration for code storage or improved memory management—remain embarrassingly underdeveloped. It's as if they're building a house by focusing on installing a fancy chandelier while the foundation is cracking and the roof leaks. The result is a tool that creates impressive first impressions in demos but becomes increasingly frustrating the longer you try to use it for actual work.
Takeaway: Bolt.new prioritizes flashy, demo-friendly features over fundamental functionality, creating a tool that impresses in screenshots but fails in sustained real-world use.
The Future Path: What AI Tools Should Prioritize
If we're going to move beyond the current stagnation disguised as progress, AI development needs a radical course correction. Here's what they should be focusing on instead of the current superficial feature arms race:
Personal AI with True Persistent Memory
We desperately need AI systems with actual persistent memory, not these stateless models that forget everything between conversations. I want an assistant that genuinely remembers our interaction history and evolves its understanding over weeks, months, and years—not one that needs to be constantly reminded of basic context.
This memory needs to be user-controlled, preferably through open-source or localized models that don't store everything on corporate servers. I don't want my entire digital history sitting in OpenAI's or Anthropic's databases, especially when they're likely to monetize it eventually.
Takeaway: Without solving the memory problem, everything else is just window dressing. We need AI with real, persistent memory that users control—not companies.
Deeper Contextual Understanding
The obsession with response speed has created systems that spew words without thinking. Ironically, the companies marketing "reasoning" models are acknowledging this problem while pretending their minor tweaks solve it.
We need models that actually think before responding, even if that takes a few seconds longer. I'd gladly wait 5-10 seconds for a response that demonstrates genuine understanding rather than getting an instant reply that misses the point entirely. The current approach of optimizing for words-per-second is creating assistants that talk a lot but say little of value.
Takeaway: Speed is overrated. I'd rather wait for quality than get instant bullshit. Future models need to prioritize actual comprehension over rapid response times.
Modular, User-Centered Design
Instead of trying to create one-size-fits-all tools that do everything poorly, AI development should embrace modularity. Let users choose which capabilities they need and disable the rest to prevent unnecessary complexity.
This approach would also force developers to perfect core functionality rather than constantly chasing new features. If users could selectively enable only what they need, it would quickly become obvious which features actually matter and which are just marketing fluff.
Takeaway: Stop trying to build Swiss Army knives that do everything badly. Build focused tools with clear purposes and let users add only what they need.
Community-Driven Development
The disconnect between developer priorities and user needs has never been more obvious. Instead of building what users are asking for, companies continue developing whatever will generate the most impressive demo videos or investor presentations.
By genuinely involving users in the development process—not just through performative feedback forms that get ignored, but through substantive inclusion in priority-setting—companies could build tools that solve real problems rather than imaginary ones.
Takeaway: Listen to users, not investors. Build what people actually need instead of what looks impressive in a TechCrunch article.
Conclusion
The current state of AI tools represents a frustrating paradox—incredible technological advances undermined by misguided development priorities. We're stuck in a cycle of superficial improvements that mask fundamental limitations, creating increasingly complex systems that fail at their core purposes.
Moving forward requires honest acknowledgment of current limitations, particularly around memory and contextual understanding. Rather than pretending these problems don't exist or that they're about to be solved, developers should focus on building more modest but genuinely useful tools that address real user needs.
The path to truly transformative AI doesn't lie in endlessly accumulating features or making grandiose claims about impending AGI. It requires solving the fundamental architectural challenges that limit current systems' ability to maintain context, remember information, and genuinely understand user needs over time.
Until the industry is willing to confront these limitations honestly, we'll continue seeing impressive demos that translate into disappointing real-world experiences. And frankly, I'm tired of the bullshit.
Takeaway: The AI industry needs to stop hyping capabilities it doesn't have and start honestly addressing its limitations. I'm not interested in another feature I didn't ask for—I want systems that reliably do what they claim to do without forgetting everything we discussed five minutes ago.
AI Innovator | PhD in AI | 15+ Years in Engineering | Delivering Computer Vision & NLP for Global Change
1 周Very good points. It is ironic, though, that many influencers are taking advantage of the hype and gaining attention despite not knowing what they are writing about. Last year, almost every post in the feed mentioned RAG. These days, everyone is talking about how MCP is like USB, and by parroting the same sources of information, they consider themselves innovative. The majority don’t understand the fundamentals and get lost trying to justify each new buzzword as a solution to trivial challenges. Above all, each new hype cycle reveals that most people consider themselves innovative simply by imitating and recycling.
Head of AI and Data Science | Co-founder | Consulting on AI | Brain scientist | AI scientist
1 周One almost wishes for an AI-winter so that they stop the nonsense.
Head of AI and Data Science | Co-founder | Consulting on AI | Brain scientist | AI scientist
1 周Very good insights, Hannes.