Chess as a barometer for AI
Chess Board

Chess as a barometer for AI

For over a century, the game of chess was considered the ultimate benchmark for measuring intelligence. The belief was that if a machine could excel at chess, it would mimic human thought processes. Chess was widely regarded as a complex game with countless combinations, and the creation of a machine capable of mastering it seemed like an insurmountable task. However, the way chess was eventually conquered by machines offers valuable insights into the development and understanding of current large language models (LLMs).

The Chess Conundrum

As a kid, I have spent countless hours playing and mastering chess. And at one point, I ranked in the top 1% on MSN Chess. My approach was rather simple: play lots of games with higher ranked players than me. During one year I spent about 8 hours per day. In terms of MSN rankings, the brute force approach to learning worked out. But, to use the math analogy, I achieved only a local maximum. I became a top player within a “local” community of players. By no means, did I become a top player for the “global” community of players. I, maybe, would have made it to the top 10%. It was not an optimal way to be a good chess player, and frankly the career in chess was not for me.

There are, of course, others approaches to getting good at chess. One includes reading various combinations from top players. Basically memorizing as many games and combinations as possible. But lots of people memorize those combinations and still do not come close to the level of a world champion such as Kasparov. Perhaps there is indeed an element of raw intelligence for the game of chess that you need to get to that level.

On the subject of raw intelligence I have one striking memory - that of re-watching Kasparov’s game against a super computer. At the time, the chess community believed that only artificial intelligence could truly master the game, given the vast number of possible combinations (a number with 48 “0s”). When IBM's Deep Blue defeated the world champion Garry Kasparov in 1997, it seemed like artificial intelligence had arrived. And AI finally possessed levels of raw intelligence greater than that of the world champion.

The Misconception About Chess and AI

As it turned out, the approach used to defeat Kasparov in chess was not based on replicating human thought processes, or acquiring that “raw intelligence”. Instead of creating a machine that was good at winning, the task was to create a machine that was good at not losing.

Deep Blue leveraged a form min-max method, which relies on brute force computing power. To win, this method does NOT require calculating all possible combinations but only a few more than what humans can calculate independently.

To use my MSN analogy, Deep Blue found a local maximum in which it was able to beat Kasparov. It did not need to find a global winning strategy - just a winning one against “humans”. The victory of Deep Blue over Kasparov was not a testament to true raw (artificial) intelligence but rather an example of exceptional computing power.

Lessons from Chess for Large Language Models

The success and capabilities of large language models (LLMs) have led many to believe that language is a better proxy for human thought process than chess.

The success and capabilities of large language models (LLMs) have led many to believe that language is a better proxy for human thought process than chess.

Consequently, some may argue that mastering language through LLMs signifies the achievement of true artificial intelligence. However, the lesson from chess serves as a cautionary tale. Just like the brute force approach in chess, LLMs may be avoiding losing rather than truly winning. The question then arises: Are LLMs genuinely replicating human thought processes, or are they just efficiently leveraging computational power (and storage)?

Why would this matter. For one, if we assume AI is going to be integrated in our everyday’s life, then we should actually care:

  • whether we are given the shortest estimated driving directions or the average one,
  • whether our health diagnosis is given with the best optimal assessment or whether it is normalized based on averages of similar groups,

One of the ways a bunch of folks recently started testing LLMs is by using SAT and GMAT tests. The results are phenomenal, but perhaps a bit overblown. It has been awhile, but as I recall, a major winning strategy for both tests—especially in the essay writing section—is a min-max strategy of avoiding a mistake (avoiding a “loss”)…

要查看或添加评论,请登录

SG Mir的更多文章

  • Developing AI Applications

    Developing AI Applications

    Taking a closer look at AI feature development lifecycle Airtable Formula Suggest Those of us who grew up learning…

  • Omni Analytics

    Omni Analytics

    Since before it became the 4th largest acquisition in Google’s history, Looker had a target on its back as every other…

    1 条评论
  • Applying LLMs to Data Analytics

    Applying LLMs to Data Analytics

    One of the areas in Data I am most passionate about is ANALYTICS AS CODE. I am bullish about this because I believe…

    5 条评论
  • New Year's Resolutions

    New Year's Resolutions

    As we embark on the New Year, I find there is too much happening globally to talk about my private resolutions…

    3 条评论
  • How Databricks won the battle, for now - Conference Recap - Part I

    How Databricks won the battle, for now - Conference Recap - Part I

    The battle between Product led Growth and Sales (SLG) “Databricks is a $38 billion dollar mistake” wrote Benn Stancil…

    2 条评论
  • Plug & Play: Speaking on Generative AI

    Plug & Play: Speaking on Generative AI

    Talking Privacy, Data Ownership, Biases, and Worker Employment A common theme in many conversations at most conferences…

  • Paul Graham has 49.3 times more "AI tokens", but does that actually mean what we think it does?

    Paul Graham has 49.3 times more "AI tokens", but does that actually mean what we think it does?

    1. The challenge of AI understanding Understanding language remains one of the most complex challenges in artificial…

  • Generative AI

    Generative AI

    Hello there, Today, we'll explore several pivotal aspects of generative AI. Let’s dive right in… AI Infrastructure:…

    1 条评论
  • ChatGPT, Emotional AI, and the Future of Communication

    ChatGPT, Emotional AI, and the Future of Communication

    Emotional AI is a broad range of technologies aimed at automating objective measurement of opinions, feelings, and…

  • Why you can only eat that sandwich once…

    Why you can only eat that sandwich once…

    When I met my wife, she wanted to open a restaurant. Great, I thought — I love eating… I mean, everyone does.

社区洞察

其他会员也浏览了