The Brief on AI #18
The meteoric rise of infrastructural AI and the market's struggle to adapt
Recent developments
Quote of the week
The question of whether a computer can think is no more interesting than the question of whether a submarine can swim.
Edsger W. Dijkstra
What to Make of This
Two years ago, Nvidia was worth $400bn. Now it stands at $3.3tn. While AI-driven companies surge in value, traditional sectors are struggling. What's happening in the market?
Investors are pouring their money into AI infrastructure companies, while the rest of the companies are being penalized for not yet capitalizing on AI's promised benefits. The market appears to be overestimating how quickly companies can use AI to reduce costs and boost revenues. Investors want to see results now, but adopting this new technology is more tricky than most think. It’s going to be a gradual process and it’s going to take much longer than everyone is betting on. Three main factors are compounding the issue.
1. Building a layer on top of foundation models is necessary but risky
Investments in early stage generative AI have nosedived in Q1 2024. Many of the startups that got funded in Q3 and Q4 of 2023 shouldn’t have as they became obsolete after new frontier model releases. Companies like these have been dubbed ‘thin wrappers’. Early stage investors are holding back on investing in them to see how the foundational models evolve. This is an unfortunate development because a user friendly layer on top of foundation models is exactly what companies need to be able to benefit from AI.
领英推荐
2. Limitations of LLMs are still unclear
In the context of limitations there is a lot of talk about hallucinations. Let us stop using this word and be more specific about the unwanted results we are getting from machines. The recently published study by Standford on the Reliability of Leading AI Legal Research Tools did a great job at this. The paper talks about ‘groundedness’ when the researchers refer to the question whether the key factual propositions make valid references to relevant legal documents. They talk of ‘correctness’ when determining whether a response is factually correct and relevant. This is a great first step, but if we truly want to uncover what’s holding us back in AI we need to:
Erik J. Larson’s The Myth of Artificial Intelligence: Why Computers Can't Think the Way We Do and The Atomic Human by Neil D. Lawrence both do a terrific job at the first three points I mentioned above. The last one, devising strategies to bridge the gaps, is where to opportunity lies for new AI companies to add value. Which leads me to the third and maybe biggest issue of them all.
3. AI companies tell lies
To raise funding and land new clients as an AI company at the application level you need a clear cut answer to a problem that existing AI companies can’t solve at the foundation level. And here’s the thing: most of these companies are coming up with solutions that only partially solve a problem. They might be aware of this, but they can’t get away with saying: “here is a solution that helps a little bit.”
To understand what I mean, let’s dive into Stanford’s paper on legal research some more. We discussed how groundedness and correctness are important. But to get good results from a research tool it needs to be able to synthesize facts while keeping the appropriate legal context in mind, select relevant documents based on other factors than text (e.g. legal hierarchy) and give more than one clear cut answer to a legal prompt. Even if legal research tools are able to eliminate ungrounded and incorrect answers, which they clearly are not, we are still left with these other crucial elements in the research process that currently can only be done by humans. That’s a really hard sell to investors and potential new clients. So what happens? AI companies tell an oversimplified story. Here are three examples from the Stanford paper:
And here is an overview of what is actually happening:
Benchmarks and patience
The Stanford paper obviously stirred up significant discussion and debate within the academic and technological communities. The Artificial Lawyer recently shared Thomson Reuters and LexisNexis are going to create “…a consortium of stakeholders to work together to develop and maintain industry standard benchmarks across a range of legal use cases.” This is exactly the direction we should be going as an industry. We need to understand the limitations of the status quo, create solutions at an application level and measure these solutions scientifically and objectively. This will take years, decades even. And as we progress there will be gradual change, because from what I’ve learned from Larson’s and Lawrence’s books, we are far from replicating the magical machine between our ears.