The rising cost of LLM-based search
Exponential View via Dalle-3

The rising cost of LLM-based search

Large Language Models like ChatGPT are shifting the paradigm of information retrieval and search. These models go beyond merely providing a list of possible answers; they analyse, interpret, and contextualise queries, delivering nuanced and detailed responses. For instance, I regularly use Perplexity to help make sense of questions as diverse as tank survivability rates, what kind of lights I should get for my bookshelves, or understanding the regulatory burden across industrial sectors. LLMs have the potential to move beyond the search engine, towards synthesising information, a process I described three years ago in my essay “A Short History of Knowledge Technologies”.

However, this computational wizardry comes at a significant energetic cost. Each interaction with ChatGPT may consume up to 2.9 watt-hours (Wh) of energy, enough to boil two tablespoons of water and nearly ten times the 0.3 Wh energy cost of a standard Google search. And if Google replaced its current search algorithms with LLMs, SemiAnalysis estimates that each search request would cost up to 8.9 Wh of energy. Given Google’s 9 billion search requests per day this would lead to an annual energy use of 29.2 terawatt-hours (TWh) on their servers, equivalent to the entire annual energy consumption of Ireland.

Of course, several variables could offset this theoretical surge in energy use, including bottlenecks in server manufacturing, improvements in energy efficiency, and Google simply not wanting to pay 30 times the energy costs for a search request. Nevertheless, it illustrates how emerging technologies often come with heightened energetic demands.

As we look ahead, our digital world’s computational appetite shows no signs of abating - even without AI developments. Efficiency gains in data centres, which have historically helped mitigate increased energy consumption, are plateauing. MIT predicts that by 2030, data centres could devour up to 21% of the world’s total electricity supply, up from 1-1.5% currently. Other more conservative forecasts suggest energy demand will still increase by 67% by 2028. Either way, new technologies and the demand for computation will pressure the energy transition.

Yet this pressure can often stimulate the quest for greater efficiency. We’ve seen this dynamic play out in cryptocurrency. Initially, blockchain technologies like Ethereum relied on energy-intensive proof-of-work mechanisms. As energy concerns mounted, the industry pivoted towards more sustainable proof-of-stake models. We are already seeing promising avenues within AI, such as gains in the performance of smaller models. However, this cycle of new technologies raising energy consumption before efficiencies eventually kick in will likely continue into the future.

Of course, the long-term goal is to make these concerns irrelevant in an age of abundant, clean energy — we’re looking at you, nuclear fusion.?

Raghav Sehgal

Helping Enterprises deploy #GenerativeAI in their workflows !

1 å¹´

Yep, now I get the data center in space idea.

赞
回复
Nagesh Setty K S

Business Automation Consultant | Go High Level Automations | Make Automations | Social Media Automations | Lead Generation Expert | No-Code AI Automation Coach | AI & Automations Training

1 å¹´

Thanks for sharing this content. I always tried to see the positive side of Using AI or Adopting AI but never thought about the negative impact it has had. So we need to use AI wisely and try to avoid using it to a limit of Usage for Productivity. When we were kids, when a new toy was given to us, most of the time, we always tried to play with the new toy keeping away the old toys. Now the AI Tools also have become the same, unnecessarily, everyone is jumping to see how to use them and how to learn them on their own instead of learning from Experts. I appeal to the Experts of AI, to save Energy, to come forward and Train and Educate People, on how to use AI Tools effectively and effortlessly, so that we can save Energy.

赞
回复
Christian Graham

Experiments team Co-Lead at Friends of the Earth

1 å¹´

Think this will be a short-medium term problem as cost reduction is a strong incentive to produce more efficient chips, models etc.?Also there will be some displacement - I almost never use traditional search now. That’s not only saving search costs, but also the linked websites (when did you last find exactly the answer you wanted on the first search/website you did/browsed?) However, it may mean some of the existing players don’t make it across this chasm, and could see some desperate scrabbling in the short term around business model innovation for these companies (particularly on freemium offers).? But wouldn’t be surprised if we are looking back on this in 5-10 years as a problem that gently faded away.?New stuff is often expensive, but gets cheaper via wrights law.? There will be other, some foreseen/others not, problems we’ll be dealing with due to poor incentive alignment - but this doesn’t feel like one of them.

Miguel R.

CTO| MBA | Invited Professor| Innovation | Agile | AI | Eager to create a more equitable and sustainable future through technology

1 å¹´

Where is the raw data to support the statement cost for each query Azeem Azhar ?

赞
回复

要查看或添加评论,请登录

Azeem Azhar的更多文章

  • ?? What's the deal with Manus AI?

    ?? What's the deal with Manus AI?

    Six things you need to know to understand the hype The online discourse around Manus AI typically falls into three…

    8 条评论
  • AI’s productivity paradox

    AI’s productivity paradox

    I want to play a game of counterfactuals..

    11 条评论
  • Why the AI surge isn't like 1999

    Why the AI surge isn't like 1999

    Economist Paul Krugman sees parallels between the late-90s tech bubble and today’s AI frenzy. In my conversation with…

    4 条评论
  • What OpenAI’s Deep research means for search

    What OpenAI’s Deep research means for search

    Originally published in Exponential View on 4 February OpenAI released yet another add-on on to its growing suite of AI…

    4 条评论
  • ??DeepSeek: everything you need to know right now.

    ??DeepSeek: everything you need to know right now.

    My WhatsApp exploded over the weekend as we received an early Chinese New Year surprise from DeepSeek. The Chinese AI…

    38 条评论
  • ?? Stargate & DeepSeek R-1 – What matters

    ?? Stargate & DeepSeek R-1 – What matters

    In the past week, a lot was written about the US government’s “Stargate” partnership with OpenAI AND DeepSeek R-1…

    12 条评论
  • Davos Daily, Day 1

    Davos Daily, Day 1

    The energy here is different this year, so I’ll share my daily takes from the Forum to help you understand what it’s…

    6 条评论
  • ?? Join me live on AI, deep tech & geopolitics

    ?? Join me live on AI, deep tech & geopolitics

    Hi all, I am going live in two hours from DLD — one of Europe’s most important annual events focused on the…

    5 条评论
  • Five contrarian ideas about genAI in the workplace

    Five contrarian ideas about genAI in the workplace

    ChatGPT alone sees over 300 million weekly users—roughly 7% of all mobile phone owners worldwide. Nearly a third of…

    13 条评论
  • ?? AGI in 2025?

    ?? AGI in 2025?

    We can't ignore Sam's bet that..

    11 条评论

社区洞察

其他会员也浏览了