Global AI Adoption is Outpacing Risk Understanding Warns MIT CSAIL
I have been writing about the risks of AI for some time and the imperative to uplift our AI digital literacy knowledge and governance practices to design, develop, and implement AI solutions with more robust controls. MIT CSAIL and MIT FutureTech researchers have just released the world's first foundational AI risk framework to help leaders understand and build risk mitigation plans applicable to the different AI AI methods.
I am very excited to see this development and the leadership that the academic community has taken to advance the knowledge of AI risks and expose the gaps we have in understanding AI risk. This world-class repository will not only help business leaders to manage AI risk more effectively, but it will also support policy and legislation development, and standard bodies to govern AI risk with more rigor. This further creates an opportunity to increase knowledge amongst educators in developing AI educational risk programs and certifications.
Below is a summary of the major highlights released by MIT CSAIL and MIT FutureTech as of August 14, 2024. This is a very positive development in our AI industry and in a recent conversation with Tony Gaffney President of the Vector Institute , a leading AI Canadian research organization, Tony advised that they will also be releasing AI governance toolkits to help business leaders design, develop, and sustain more effective AI practices to derisk AI practices. The understanding of the imperative to adopt AI and sustain it accelerating
The adoption of AI is rapidly increasing; census data shows a significant (47%) rise in AI usage within US industries, jumping from 3.7% to 5.45% between September 2023 and February 2024. The comprehensive review from MIT CSAIL and MIT FutureTech has uncovered critical gaps in existing AI risk frameworks. Their analysis reveals that even the most thorough individual framework overlooks approximately 30% of the risks identified across all reviewed frameworks.
To help address this, other academic institutions were involved from the University of Queensland, Future of Life Institute, KU Leuven, and Harmony Intelligence, to release the first-ever AI Risk Repository: a comprehensive and accessible living database of 700+ risks posed by AI that will be expanded and updated to ensure that it remains current and relevant.
“Since the AI risk literature is scattered across peer-reviewed journals, preprints, and industry reports, and quite varied, I worry that decision-makers may unwittingly consult incomplete overviews, miss important concerns, and develop collective blind spots,” says Dr. Peter Slattery, an incoming postdoc at the MIT FutureTech Lab and current project lead.
After searching several academic databases, engaging experts, and retrieving more than 17,000 records, the researchers identified 43 existing AI risk classification frameworks. From these, they extracted more than 700 risks. They then used approaches that they developed from two existing frameworks to categorize each risk by cause (e.g., when or why it occurs), risk domain (e.g., “Misinformation”), and risk subdomain (e.g., “False or misleading information”).?
Examples of risks identified include “Unfair discrimination and misrepresentation”, “Fraud, scams, and targeted manipulation”, and “Overreliance and unsafe use.” More of the risks analyzed were attributed to AI systems (51%) than humans (34%) and presented as emerging after AI was deployed (65%) rather than during its development (10%).? The most frequently addressed risk domains included “AI system safety, failures, and limitations” (76% of documents); “Socioeconomic and environmental harms” (73%); “Discrimination and toxicity” (71%); “Privacy and security” (68%); and “Malicious actors and misuse” (68%). In contrast, “Human-computer interaction” (41%) and “Misinformation” (44%) received comparatively less attention.?
Some risk subdomains were discussed more frequently than others. For example, “Unfair discrimination and misrepresentation” (63%), “Compromise of privacy” (61%), and “Lack of capability or robustness” (59%), were mentioned in more than 50% of documents. Others, like “AI welfare and rights” (2%), “Pollution of information ecosystem and loss of consensus reality” (12%), and “Competitive dynamics” (12%), were mentioned by less than 15% of documents.?
On average, frameworks mentioned just 34% of the 23 risk subdomains identified, with nearly a quarter covering less than 20%. No document or overview mentioned all 23 risk subdomains, and the most comprehensive (Gabriel et al., 2024) covered only 70%.?
The work addresses the urgent need to help decision-makers in government, research, and industry understand and prioritize the risks associated with AI and work together to address them. “Many AI governance initiatives are emerging across the world focused on addressing key risks from AI,” says collaborator Risto Uuk, EU Research Lead at the Future of Life Institute. “These institutions need a more comprehensive and complete understanding of the risk landscape.”?
Researchers and risk evaluation professionals are also impeded by the fragmentation of current literature. "It is hard to find specific studies of risk in some niche domains where AI is used, such as weapons and military decision support systems,” explains Taniel Yusef, a Research Affiliate, at The Centre for the Study of Existential Risk, at the University of Cambridge who was not involved in the research. “Without referring to these studies, it can be difficult to speak about technical aspects of AI risk to non-technical experts. This repository helps us do that.”
"There's a significant need for a comprehensive database of risks from advanced AI which safety evaluators like Harmony Intelligence can use to identify and catch risks systematically,” argues collaborator Soroush Pour, CEO & Co-founder of AI safety evaluations and red teaming company Harmony Intelligence. “Otherwise, it’s unclear what risks we should be looking for, or what tests need to be done. It becomes much more likely that we miss something by simply not being aware of it”.
领英推荐
AI’s Risky Business?
The researchers built on two frameworks (Yampolskiy 2016 & Weidinger et al., 2022) in categorizing the risks they extracted. Based on these approaches, they group the risks in two ways.
First by causal factors:?
Second, by seven AI risk domains:?
These are further divided into 23 subdomains (full descriptions here):
"The AI Risk Repository is, to our knowledge, the first attempt to rigorously curate, analyze, and extract AI risk frameworks into a publicly accessible, comprehensive, extensible, and categorized risk database. It is part of a larger effort to understand how we are responding to AI risks and to identify if there are gaps in our current approaches," says Dr. Neil Thompson, head of the MIT FutureTech Lab and one of the lead researchers on the project. "We are starting with a comprehensive checklist, to help us understand the breadth of potential risks. We plan to use this to identify shortcomings in organizational responses. For instance, if everyone focuses on one type of risk while overlooking others of similar importance, that's something we should notice and address."
The next phase will involve experts evaluating and prioritizing the risks within the repository, then using it to analyze public documents from influential AI developers and large companies. The analysis will examine if organizations respond to risks from AI -- and do so in proportion to experts’ concerns -- and compare risk management approaches across different industries and sectors.
The repository is freely available online to download, copy, and use. Feedback and suggestions can be submitted here.
Website: https://airisk.mit.edu/
Serious about sales acceleration? Welcome ?? Devoted to raise your win rates with proven systems. Results: +25% pricing power, +25% win rate, +25% sales velocity. Book a call: strategysprints.com
3 个月This is a groundbreaking development in AI governance!
Lead Future Tech with Human Impact| CEO & Founder, Top 100 Women of the Future | Award winning Fintech and Future Tech Influencer| Educator| Keynote Speaker | Advisor| Responsible AI, VR, Metaverse Web3
3 个月Thanks for sharing- very helpful Risk farmework from MIT!
Virtual CMO and Go-to-Market Builder for Video Tech Companies
3 个月This new AI risk framework from MIT CSAIL and MIT FutureTech is essential for startup founders. It offers a structured approach to identifying and mitigating over 700 potential AI risks, which can help you build safer, more reliable products. As Clayton Christensen often emphasized, understanding risks early on can be crucial for sustainable innovation.
Lead at the AI Risk Repository | MIT FutureTech
3 个月Thank you for sharing, Cindy! We welcome feedback, if you have any, here: https://docs.google.com/forms/u/2/d/1tDd-0Olru5dYHY9bjs3oHj9cg3-QRJqf6lMHn4lEVRc/edit