Complete Story
 

12/18/2023

OpenAI’s Sutskever Possesses a Plan for Keeping AI in Check

The super-alignment team has devised a way to guide AI models' behavior

OpenAI was founded on a promise to build artificial intelligence that benefits all of humanity—even when that AI becomes considerably smarter than its creators. Since the debut of ChatGPT last year and during the company's recent governance crisis, its commercial ambitions have been more prominent. Now, the company says a new research group working on wrangling the supersmart AIs of the future is starting to bear fruit.

"AGI is very fast approaching," said Leopold Aschenbrenner, a researcher at OpenAI involved with the Superalignment research team established in July. "We're gonna see superhuman models, they're gonna have vast capabilities and they could be very, very dangerous... we don't yet have the methods to control them." OpenAI has said it will dedicate a fifth of its available computing power to the Superalignment project.

A research paper released by OpenAI today touts results from experiments designed to test a way to let an inferior AI model guide the behavior of a much smarter one without making it less smart. Although the technology involved is far from surpassing the flexibility of humans, the scenario was designed to stand in for a future time when humans must work with AI systems more intelligent than themselves.

Please select this link to read the complete article from WIRED.

Printer-Friendly Version