Charting the Path to AGI
Michael Spencer
A.I. Writer, researcher and curator - full-time Newsletter publication manager.
Hi Everyone,
This a guest post by Abhinav Upadhyay, who has an incredible substack called Confessions of a Code Addict. His summaries of papers also shines a light of clarity on the difficult topics he chooses.
So I asked him to tackle the most difficult topic of all of our times: AGI. As fate would have it, Google decided to publish a rather lucid paper on this recently.
?? From our sponsor: ??
Sell Smarter with AI in 2024! New Data from HubSpot and G2 dives into insights from 600+ sales pros and leaders across B2B and B2C teams.
Now more than ever, sales teams to turn to AI-based tools to optimize the sales process, automate manual tasks, helping them actually spend their time connecting with prospects and closing more deals.
By Abhinav Upadhyay, Hyderabad India, December, 2023.
Following are some of his articles which have made an impact on the readers:
Charting the Path to AGI: DeepMind's Levels and Risks Framework
A year has passed since the release of ChatGPT, and AI has progressed rapidly in this short amount of time. This progress has also triggered discussions about artificial general intelligence (AGI). Some people believe that ChatGPT has shown sparks of AGI, while some believe that state-of-the-art large language models (LLM) are already AGI. And, some scientists believe that we are not yet at the level of AGI but accelerating towards it at a rapid pace.
When we talk about AGI, it also brings up the debate around the risks it poses to society. A part of the industry is even asking for regulations to control the research and development of AI to mitigate the potential risks. However, everyone in this debate has their own definition of AGI, which makes the discussion very subjective.
Before we talk about regulations, it’s important that the scientists come up with an objective definition of AGI, design a benchmark to test for AGI, and create a framework to assess the associated risks based on the capabilities of these models.
Till now, very little progress has happened on this front. However, recently a team at DeepMind released a paper which may build the foundation for defining AGI and the associated risks. The paper outlines a gradual pathway toward AGI going from “no AI” to “artificial superintelligence”, along with a framework for assessing the risks associated with each level of AGI.
This paper might be the start of an important discussion for formulating the risks of AGI. This article will unpack the paper for you, and highlight the important insights. So, without wasting anymore words, let’s dive in.
For those looking to stay on top of the latest developments in AI policy discussions, consider subscribing to “AI Policy Perspectives”, a Substack curated by a team of researchers at DeepMind. They publish a comprehensive monthly report highlighting the most significant news and updates around AI policy. ( Harry Law and friends).
Six Key Principles for Defining Levels of AGI
Scientists and philosophers have long thought about AGI, and it is important to consider how they have thought about it in the past. The authors analyze nine definitions of AGI proposed between the period of 1950 and 2023 [Turing 1950, Searle 1980, Legg 2008, Shanahan 2015, OpenAI 2018, Marcus 2022, Suleyman and Bhaskar 2023, Norvig et al. 2023] and come up with six key principles that should form the basis of a framework for defining AGI and its risks. These six principles are as follows:
领英推荐
Six Levels of AGI
Using these six key principles, the authors delineate six gradual levels of AGI. These levels are defined using ‘performance’ and ‘generality’ as the two dimensions to measure them.?
Here, performance refers to how well an AI system can perform a task in comparison to a skilled human. While, generality is concerned with the breadth of the tasks that the AI can perform. For instance, is it skilled in a narrow domain such as protein folding, or is it capable of performing well on tasks from a wide array of domains.
The following table shows these six levels of AGI as defined in the paper. The two columns, Narrow and General, specify the generality of the AI, while the six rows define the six levels of AGI with gradually increasing levels of capabilities.?
Table 1 from the paper showing Levels of AGI with examples (above)
Although, these leveled definitions are straightforward, few important points are worth highlighting:
Defining a Benchmark for AGI
The 6 levels of AGI as defined in the paper give us a pathway to classify the progress of AI systems. However, it does not specify any benchmark that should be used to test these systems to certify them as belonging to one of these levels.
These tasks need to be diverse, challenging, and relevant to real-world use cases. This is a serious undertaking which needs to include multiple perspectives. The benchmarks need to measure both cognitive and metacognitive abilities, and include tasks from diverse areas such as mathematical and logical reasoning, linguistics, coding, spatial reasoning, social intelligence, the ability to learn new skills, and creativity.?
Exhaustively enumerating tasks for such a benchmark is a monumental task and impossible to get right in the first attempt. Additionally, this benchmark needs to be a living and breathing piece of work which can get updated with new tasks as we understand more about the AI systems and their capabilities. For these reasons, the authors leave out the definition of a representative benchmark from the paper. However, they note that this is an important goal for the AI community to strive for.?
A Framework for Assessing Risks of AGI?
Just defining levels of AGI is a job half done, this also needs to be accompanied by a framework for assessing risks associated with each of these levels. For defining this framework, the authors introduce the concept of autonomy.
The autonomy of an AI system depends on its capabilities, and the environment in which it is operating. The environment here means the interface which enables human-AI interaction. The authors introduce six levels of autonomy which are directly correlated with the levels of AGI. Progressing through the levels of AGI unlocks higher levels of autonomy in the model. Because of this, the interface design of these systems is going to play a crucial role in the safe deployment of AGI in the real-world.
The following table from the paper shows these six levels of AGI autonomy, and examples of some of the associated risks:
Table 2 from the paper showing autonomy levels of AI and associated risks
I would highlight couple of points about this framework:
Conclusion
The rate at which AI technologies are advancing demands a cohesive framework for recognizing AGI and assessing its potential impact. The DeepMind paper sets a foundation by charting a structured trajectory with clear AGI levels. What remains critical is the development of detailed benchmarks that will reliably measure AI capabilities across these levels. These benchmarks require input from a broad spectrum of fields to ensure they encompass the necessary complexities of real-world tasks.
Moreover, a nuanced risk assessment framework is essential for the development of advanced AI systems. Each incremental step of AI advancement brings with it distinctive challenges and risks that must be anticipated and managed. A well-constructed framework will dispel misconceptions and mitigate the rush to implement potentially stifling regulations that could hinder progress in AI research.
The AI community needs to extend the dialog started by the paper, filling in gaps with in-depth analyses and tools for testing AGI systems. Only through such concerted efforts will we be able to align the march toward AGI with prudent oversight and ethical considerations, ensuring that the transition to higher levels of intelligence is both beneficial and secure.
Many thanks to the author of this paper Google DeepMind , Meredith Ringel Morris, Jascha Sohl-dickstein, Noah Fiedel, Tris Warkentin, Allan Dafoe, Aleksandra Faust, Clement Farabet, Shane Legg, - published November 4th, 2023.
About Me
Abhinav is a seasoned software engineer with over a decade of experience in the industry in various roles ranging from dev ops, backend engineering to ML. He is an explorer who likes to break things open to understand how they work from the inside. This passion of learning has led him on the path of writing to share the insider’s perspective with his audience.
On his Substack, “Confessions of a Code Addict”, he talks about a myriad of topics including AI, programming languages, compilers, databases and many more. His in-depth exploration and insights offer readers a unique understanding of these subjects from a practitioner's viewpoint, making complex concepts accessible and engaging.?
Following are some of his articles which have made an impact on the readers:
This article is a guest post on A.I. Supremacy this week, my main Newsletter related to A.I. and its impact.
Hospitality Driver/Social Media/Writer/Old Alaskan Scout.
1 年Try playing a play by email online game with an AI as your game master?
President and CEO, Cyber Intelligent Partners LLC | CISO | NYU Adjunct Instructor
1 年?? updating ‘21 research on AI competition! “With the era of the Fourth Industrial revolution spreading across the world, AI has become the commanding heights of the new millennial.” ? Lack of AI Ethics and safety. Chinese leadership has unaccountable history of practices that ignore ethical standards. According to a British Medical Journal,the WHO confirmed breaches of safety procedures at one of Beijing's top virology laboratories were the probable cause of SARS virus outbreaks. ? AI advances in military capabilities (with no red lines). Beijing has proven its desire to develop AI-enabled warfare such as swarm intelligence, integration with command decision making, and innovations for Artificial General Intelligence (AGI), which former Google China CEO called, “the holy grail of AI.” ? Digital socialism and government use of data. In China, social credit uses big-data collection and analysis, to monitor, shape, and rate individual’s behavior. ???? “Smart-City” services in over 40 countries proves China’s mass surveillance program as the most prolific and accessible in the world. https://www.dhirubhai.net/pulse/2021-research-united-kingdom-jeff-hoffmann?utm_source=share&utm_medium=member_ios&utm_campaign=share_via
Manuf. Engineer at Ford Motor Company
1 年But who is defining the choreography & trajectyof ethical codices being breathed into this new life form, as it takes shape and begins its ascent toward autonomy?