The AI Trust Fall

The AI Trust Fall

If I asked you, “When someone turns in a work assignment, how accurate is it? 80%, 90%, 95% or perhaps 100%?”

We don’t think this way about coworkers’ spreadsheets. But we will probably think this way about AI & this will very likely change the way product managers on-board users.

When was the last time you signed up for a SaaS & wondered : Would the data be accurate? Would the database corrupt my data? Would the report be correct?

But today, with every AI software now tucking a disclaimer at the bottom of the page, we will be wondering. “Gemini may display inaccurate info, including about people, so double-check its responses” & “ChatGPT/Claude can make mistakes. Check important info” are two examples.

In the early days of this epoch, mistakes will be common. Over time, less so, as accuracies improve.

The more important the work, the greater peoples’ need to be confident the AI is correct. We will demand much better than human error rates. Self-driving cars provide an extreme example of this trust fall. Waymo & Cruise have published data arguing self-driving cars are 65-94% safer.

Yet, 2/3 of Americans surveyed by the AAA fear them.

We suffer from a cognitive bias : work performed by a human is likely more trustworthy because we understand the biases & the limitations. AIs are a Schrodinger’s cat stuffed in a black box. We don’t comprehend how the box works (yet), nor can we believe our eyes if the feline is dead or alive when we see it.

New product on-boarding will need to mitigate this bias.

One path may be starting with low-value tasks where the software-maker has tested exhaustively the potential inputs & outputs. Another tactic may be to provide a human-in-the-loop to check the AI’s work. Citations, references, & other forms of fact-checking will be a core part of the product experience. Independent testing might be another path.

As with any new colleague, the first impressions & a series of small wins will determine the person’s trust. Severe errors in the future will erode confidence, that must be rebuilt - likely with the help of human support teams who will explain, develop tests for the future, & assure users.

I recently asked a financial LLM to analyze NVIDIA’s annual report. A question about the company’s increase in dividend amount vaporized its credibility, raising the question : is it less work to do the analysis myself than to check the AI’s work?

That will be the trust fall for AI. Will the software catch us if we trust it?

Jason M. Lemkin

SaaStr Annual 2025 is May 13-15 in SF Bay!! See You There!!

9 个月

It in essence passes the test by default for jobs no one is willing to do anymore

Kirill Solovjov

AI LQA safely & reliably deployed to any TMS · LLM evaluation & translator feedback for Localization teams · Linguistic Quality Management expert · B2B SaaS founder · 20yrs in tech & int'l

9 个月

Trust but verify, this is what we always say at ContentQuo! We’ve built the company to help buyers of AI-powered services develop trust towards their vendors based on easy to understand & openly shared quality metrics. In our first vertical, this really helped alleviate trust issue over time and move to constructive data-based vendor performance conversations and procurement processes. Expecting to see more of it as LLM space grows!

Konstantin Ristl

CEO @ Solace | Success Leaves Clues – AI, SaaS, Mindset & More | Post-Exit Investor

9 个月

?love your post! I’m a big advocate for pointing out that not only AI but also humans make mistakes. When trusting someone with a task for the first time, I’m more skeptical than after seeing them perform it multiple times. As for people: Not every task fits every person, just like not every task fits AI.

Jeff Dodge

Chief Revenue Officer at Fog Solutions - Building AI Foundations and Solutions on Azure

9 个月

Great piece. Your point about humans trusting self-driving cars is particularly poignant as the data is pretty overwhelming that lives would be saved if self-driving cars were broadly adopted in suitable circumstances. Ultimately humans do want to know "who is accountible" and "who will be punished if something goes wrong" even if there are less occurances of something going wrong. When it's a human at fault, society punishes that human and it 'feels right' ie justice is served. When AI is wrong, the only 'person' to hold accountable in most circumstances is a corporation, and generally a trillion dollar corporation and they have a lot to lose finincially so the risk/reward for those corporations is tricky and even more so the legal ramification.

Luke Shalom

CEO @ Atticus | Helping CEOs turn LinkedIn into predictable pipeline with inbound, outbound & AI tech.

9 个月

Trusting AI without fully knowing it can be worrying.

回复

要查看或添加评论,请登录

Tomasz Tunguz的更多文章

  • The Implications of the Wiz/Google Deal

    The Implications of the Wiz/Google Deal

    Is tech M&A back? Google announced its intention to buy Wiz for $32b today. If approved by regulators, it would be the…

    6 条评论
  • Halving R&D with AI & the Impact to Valuation

    Halving R&D with AI & the Impact to Valuation

    Engineering teams within AI application startups are much smaller than a classic software company - maybe half the size…

    8 条评论
  • The Mirage in the Software Clouds

    The Mirage in the Software Clouds

    Public SaaS companies’ growth rates have halved since 2023, as David Spitz pointed, from 36% to 17%. Why? There are…

    12 条评论
  • This Analysis Cost 27 Cents

    This Analysis Cost 27 Cents

    Monday’s analysis cost about 27 cents to produce. This little screenshot is of Claude Code, the product I use now to…

    9 条评论
  • Positioning Startups in the Age of AI

    Positioning Startups in the Age of AI

    How do you position and scale an AI company in a rapidly evolving market? Join us for an in-person Office Hours session…

    6 条评论
  • How Much Is A Venture Firm Worth?

    How Much Is A Venture Firm Worth?

    A small spin-out from a publicly traded behemoth launched with the ambitious vision of transforming their entire…

    5 条评论
  • Why War & Peace Is Killing Your Data Budget

    Why War & Peace Is Killing Your Data Budget

    Imagine if every time you edited a document, the word processor forced you to retype everything that had been written…

    3 条评论
  • A Founder's Guide: Essential Management Advice for Startups

    A Founder's Guide: Essential Management Advice for Startups

    As startups scale, effective management becomes the difference between chaotic growth and sustainable success. After…

    10 条评论
  • Lopsided AI Revenues

    Lopsided AI Revenues

    Which is the best business in AI at the moment? I analyzed Q4 revenue data from publicly traded companies across…

    8 条评论
  • Four Marketing Principles That Redefine Markets from Klaviyo's Former CMO

    Four Marketing Principles That Redefine Markets from Klaviyo's Former CMO

    During a recent Theory Office Hours with Kady Srinivasan, former CMO at Lightspeed Commerce, Dropbox, and Klaviyo, we…

    5 条评论

社区洞察

其他会员也浏览了