AI companies, including OpenAI, Alphabet, and Meta Platforms, have made voluntary commitments to the White House to enhance the safety of artificial intelligence technology. President Joe Biden announced this positive development on Friday, emphasizing that more collaborative efforts are needed to address potential threats from emerging technologies to U.S. democracy.
During a White House event, Biden acknowledged the growing concerns surrounding AI’s potential for disruptive applications and stressed the importance of being vigilant in safeguarding against such risks.
The companies involved which also include Anthropic, Inflection, Amazon.com, and Microsoft (as an OpenAI partner), pledged to implement measures such as thorough testing of AI systems before release and sharing information to reduce risks and invest in cybersecurity.
This move marks a significant milestone for the Biden administration’s efforts to regulate AI, given the technology’s rapid growth in investment and popularity among consumers.
In response to the announcement, Microsoft expressed its support, commending the president’s leadership in bringing the tech industry together to create concrete steps that will enhance the safety, security, and public benefits of AI.
Generative AI, exemplified by ChatGPT’s human-like prose creation, has seen a surge in popularity this year. Consequently, lawmakers worldwide have started examining ways to mitigate potential dangers posed by this emerging technology to national security and the economy.
While the U.S. has been slower than the EU in addressing AI regulation, Congress is currently considering legislation that would require political ads to disclose whether AI was used to create imagery or other content.
Biden also shared that he is working on an executive order and bipartisan legislation focused on AI technology.
As part of the commitment, the seven companies pledged to develop a watermark system for all AI-generated content, including text, images, audios, and videos. This watermark will enable users to identify when AI technology has been utilized, potentially aiding in spotting deep-fake images or audios that could perpetuate misinformation or harm.
However, details on how this watermark will be evident in content sharing remain unclear.
Furthermore, the companies have promised to prioritize user privacy as AI evolves, ensuring the technology remains free of bias and isn’t used to discriminate against vulnerable groups. In addition, they committed to using AI solutions to address scientific challenges, such as medical research and climate change mitigation.
While these voluntary commitments represent a positive step forward, the efforts to regulate AI technology and promote its responsible use are likely to continue as advancements in the field rapidly reshape the technological landscape.