AI companies promised to self-regulate one year ago. What’s changed?
Sarah Rogers/MITTR | Getty

AI companies promised to self-regulate one year ago. What’s changed?

One year ago, seven leading AI companies—Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI—committed with the White House to a set of eight voluntary commitments on how to develop AI in a safe and trustworthy way. These included promises to do things like improve the testing and transparency around AI systems, and share information on potential harms and risks. In this edition of What’s Next in Tech, learn what progress the tech sector has made since then and what work still needs to be done.

Our weekly artificial intelligence newsletter, The Algorithm, demystifies the latest AI breakthroughs and cuts through the hype. Sign up for free today.

The White House’s voluntary AI commitments have brought better red-teaming practices and watermarks, but no meaningful transparency or accountability.

On the first anniversary of these voluntary commitments, MIT Technology Review asked the AI companies that signed the commitments for details on their work so far. Their replies show that the tech sector has made some welcome progress, with big caveats.

Commitment 1: The companies commit to internal and external security testing of their AI systems before their release. This testing, which will be carried out in part by independent experts, guards against some of the most significant sources of AI risks, such as biosecurity and cybersecurity, as well as its broader societal effects.

All the companies (excluding Inflection, which chose not to comment) say they conduct red-teaming exercises that get both internal and external testers to probe their models for flaws and risks. OpenAI says it has a separate preparedness team that tests models for cybersecurity, chemical, biological, radiological, and nuclear threats and for situations where a sophisticated AI model can do or persuade a person to do things that might lead to harm. Anthropic and OpenAI also say they conduct these tests with external experts before launching their new models.?

For example, for the launch of Anthropic’s latest model, Claude 3.5, the company conducted predeployment testing with experts at the UK’s AI Safety Institute. Anthropic has also allowed METR, a research nonprofit, to do an “initial exploration” of Claude 3.5’s capabilities for autonomy.?

Google says it also conducts internal red-teaming to test the boundaries of its model, Gemini, around election-related content, societal risks, and national security concerns. Microsoft says it has worked with third-party evaluators at NewsGuard, an organization advancing journalistic integrity, to evaluate risks and mitigate the risk of abusive deepfakes in Microsoft’s text-to-image tool. In addition to red-teaming, Meta says, it evaluated its latest model, Llama 3, to understand its performance in a series of risk areas like weapons, cyberattacks, and child exploitation.?

But when it comes to testing, it’s not enough to just report that a company is taking actions, says Rishi Bommasani, the society lead at the Stanford Center for Research on Foundation Models, who also reviewed the responses for MIT Technology Review. For example, Amazon and Anthropic said they had worked with the nonprofit Thorn to combat risks to child safety posed by AI. Bommasani would have wanted to see more specifics about how the interventions that companies are implementing actually reduce those risks.?

“It should become clear to us that it’s not just that companies are doing things but those things are having the desired effect,” Bommasani says.??

Result: Good. The push for red-teaming and testing for a wide range of risks is a good and important one. However, Merve Hickok, the president and research director of the Center for AI and Digital Policy, who also reviewed the companies’ replies, would have liked to see independent researchers get broader access to companies’ models.?

Read the story for the rest of our assessment of the progress AI companies have made in the past year.

Get ahead with these related stories:

  1. What is AI? Everyone thinks they know but no one can agree. And that’s a problem.
  2. Meta says AI-generated election content is not happening at a “systemic level” Speaking at MIT Technology Review's EmTech Digital conference, Meta's global affairs boss Nick Clegg discussed the company's efforts at combating misinformation, labeling AI-generated content, and more.?
  3. Opinion: Tech workers should shine a light on the industry’s secretive work with the military Despite what happens in executive suites, workers themselves have tools to force change. Here's how.

Image: Sarah Rogers/MITTR | Getty


Everyone should listen to Musk’s take on this when he was interviewed by Joe Rogan. Sort of chilling especially his take on Altman and Open AI.

回复

Apocalypse... epilogue... If I start working, then everyone (not Me) will stop (cannot) work... Who am I? ...

回复
Ian Yang

?? "Always learning, Always thinking, Always moving" | Red Team Tenth Man | AI, Energy, National Defense, Aerospace | Senior at Auburn University

4 个月

The progress made by AI companies in the past year under voluntary commitments is noteworthy. While we’ve seen improvements in areas like red-teaming and watermarking, the lack of meaningful transparency and accountability is concerning. More robust measures and comprehensive federal legislation are needed to ensure that AI development truly serves the public good. The strides made are a step in the right direction, but there’s still a long way to go. I wonder what others think about the balance between innovation and regulation in the AI industry.

Europe has led the way with serious AI regulation that is much needed. In the US, some individual states have taken serious steps but the federal government is lagging. This year and next promise to be an important time for responsible AI.

回复
Leonard Lee

Tech Industry Advisor & Realist dedicated to making you constructively uncomfortable. Ring the bell ?? and subscribe to next-curve.com for the tech and industry insights the matter!

4 个月

Self-regulation is quite humorous. Let's be honest. At best, it is an absence of governance and outcome. I think it only gets worse from there. Correction, comical.

要查看或添加评论,请登录

社区洞察

其他会员也浏览了