Former OpenAI, Google and Meta employees testified before Congress Tuesday about the threat posed by AI reaching human-level intelligence, calling on members of the Senate Subcommittee on Privacy, Technology and the Law to advance U.S. AI policy that protects against harms caused by AI.
Artificial general intelligence (AGI), is an AI system that reaches almost human-level cognition. William Saunders, a former member of technical staff at OpenAI who resigned from the company in February 2024, testified during the hearing that AGI could cause “catastrophic harm” through systems autonomously conducting cyberattacks or assisting in the creation of novel biological weapons.
Saunders said that while there are significant gaps to close in AGI development, it’s plausible an AGI system can be built in as little as three years.
“AI companies are making rapid progress towards building AGI,” Saunders said, pointing to OpenAI’s recent announcement of GPT-o1. “AGI would cause significant changes to society, including radical changes to the economy and employment.”
He added that no one knows how to ensure AGI systems will be safe and controlled, meaning they could be deceptive and hide misbehaviors. OpenAI has “repeatedly prioritized speed of deployment over rigor,” which also leaves vulnerabilities and heightens threats such as theft from foreign adversaries of the U.S., Saunders said. During his employment at OpenAI, he noted that OpenAI did not prioritize internal security. He said there would be long periods where vulnerabilities would’ve allowed employees to bypass access controls and steal the company’s most advanced AI systems, including GPT-4.
“OpenAI will say they are improving,” he said. “I and other employees who resigned doubt they will be ready in time. This is true not just with OpenAI. The incentives to prioritize rapid deployment apply to the entire industry. This is why a policy response is needed.”
AGI, lack of AI policy top insiders’ concerns
Saunders called on policymakers to advance policies prioritizing testing of AI systems before and after deployment, requiring testing results to be shared, and implementing protections for whistleblowers.
“I resigned from OpenAI because I lost faith that by themselves they would make responsible decisions about AGI,” he said during the hearing.
Helen Toner, who served on OpenAI’s nonprofit board from 2021 until November 2023, testified that AGI is a goal many AI companies think they could reach soon, making federal AI policy necessary. Toner serves as director of strategy and foundational research grants at Georgetown University’s Center for Security and Emerging Technology.
“Many top AI companies including OpenAI, Google, Anthropic, are treating building AGI as an entirely serious goal,” Toner said. “A goal that many people inside those companies believe that, if they succeed in building computers that are as smart as humans or perhaps far smarter than humans, that technology will be, at a minimum, extraordinarily disruptive. At a maximum, could lead to literal human extinction.”
Margaret Mitchell, a former Microsoft and Google research scientist who now works as chief ethics scientist at AI startup Hugging Face, said policymakers must address the many gaps in companies’ AI practices that could lead to harm. Indeed, voluntary self-regulation on safe and secure AI, something multiple AI companies committed to last year, does not work, David Harris, senior policy advisor at UC Berkeley’s California Initiative for Technology and Democracy, said during the hearing.
Harris, who worked at Meta on the civic integrity and responsible AI teams from 2018 to 2023, said those two safety teams no longer exist, pointing out how trust and safety teams at tech firms generally have “shrunk dramatically” over the last two years.
Harris said many Congressional AI bills provide solid frameworks for AI safety and fairness. Indeed, multiple AI bills are awaiting floor votes in both the House and Senate. But Congress has yet to pass AI legislation.
“My fear is we’ll make the same mistake we did with social media, which is too little, too late,” Sen. Richard Blumenthal, D-Conn., chair of the subcommittee, said during the hearing. “What we should learn from social media is don’t trust big tech. We can’t rely on them to do the job.”
Makenzie Holland is a senior news writer covering big tech and federal regulation. Prior to joining TechTarget Editorial, she was a general assignment reporter for the Wilmington StarNews and a crime and education reporter at the Wabash Plain Dealer.
Source link
#OpenAI #associates #fear #AGI #lack #U.S #policy
Unlock the potential of cutting-edge AI solutions with our comprehensive offerings. As a leading provider in the AI landscape, we harness the power of artificial intelligence to revolutionize industries. From machine learning and data analytics to natural language processing and computer vision, our AI solutions are designed to enhance efficiency and drive innovation. Explore the limitless possibilities of AI-driven insights and automation that propel your business forward. With a commitment to staying at the forefront of the rapidly evolving AI market, we deliver tailored solutions that meet your specific needs. Join us on the forefront of technological advancement, and let AI redefine the way you operate and succeed in a competitive landscape. Embrace the future with AI excellence, where possibilities are limitless, and competition is surpassed.