Social Media

Light
Dark

Top AI companies visit the White House to make ‘voluntary’ safety commitments

While substantive AI legislation may still be years away, the industry is moving at light speed and many — including the White House — are worried that it may get carried away. So the Biden administration has collected “voluntary commitments” from seven of the biggest AI developers to pursue shared safety and transparency goals ahead of a planned executive order.

OpenAI, Anthropic, Google, Inflection, Microsoft, Meta and Amazon are the companies taking part in this non-binding agreement, and will send representatives to the White House to meet with President Biden today.

To be clear, there is no rule or enforcement being proposed here — the practices agreed to are purely voluntary. But although no government agency will hold a company accountable if it shirks a few, it will also likely be a matter of public record.

Here’s the list of attendees at the White House gig:

Brad Smith, President, Microsoft
Kent Walker, President, Google
Dario Amodei, CEO, Anthropic
Mustafa Suleyman, CEO, Inflection AI
Nick Clegg, President, Meta
Greg Brockman, President, OpenAI
Adam Selipsky, CEO, Amazon Web Services
No underlings, but no billionaires, either. (And no women.)

The seven companies (and likely others that didn’t get the red carpet treatment but will want to ride along) have committed to the following:

Internal and external security tests of AI systems before release, including adversarial “red teaming” by experts outside the company.
Share information across government, academia and “civil society” on AI risks and mitigation techniques (such as preventing “jailbreaking”).
Invest in cybersecurity and “insider threat safeguards” to protect private model data like weights. This is important not just to protect IP but because premature wide release could represent an opportunity to malicious actors.
Facilitate third-party discovery and reporting of vulnerabilities, e.g. a bug bounty program or domain expert analysis.
Develop robust watermarking or some other way of marking AI-generated content.
Report AI systems’ “capabilities, limitations, and areas of appropriate and inappropriate use.” Good luck getting a straight answer on this one.
Prioritize research on societal risks like systematic bias or privacy issues.
Develop and deploy AI “to help address society’s greatest challenges” like cancer prevention and climate change. (Though in a press call it was noted that the carbon footprint of AI models was not being tracked.)
Though the above are voluntary, one can easily imagine that the threat of an executive order — they are “currently developing” one — is there to encourage compliance. For instance, if some companies fail to allow external security testing of their models before release, the EO may develop a paragraph directing the FTC to look closely at AI products claiming robust security. (One EO is already in force asking agencies to watch out for bias in development and use of AI.)

The White House is plainly eager to get out ahead of this next big wave of tech, having been caught somewhat flat-footed by the disruptive capabilities of social media. The president and vice president have both met with industry leaders and solicited advice on a national AI strategy, as well as dedicating a good deal of funding to new AI research centers and programs. Of course, the national science and research apparatus is well ahead of them, as this highly comprehensive (though necessarily slightly out of date) research challenges and opportunities report from the DOE and National Labs shows.

Leave a Reply

Your email address will not be published. Required fields are marked *