OpenAI sets up safety committee as it starts training new model

A committee will give recommendations to the board on safety and security decisions for OpenAI's projects

Reuters May 28, 2024
OpenAI logo displayed on a screen. PHOTO: ANADOLU AGENCY

OpenAI has formed a Safety and Security Committee that will be led by board members, including CEO Sam Altman, as it begins training its next artificial intelligence model, the AI startup said on Tuesday.

Directors Bret Taylor, Adam D'Angelo and Nicole Seligman will also lead the committee, OpenAI said on a company blog.

Microsoft-backed OpenAI's chatbots with generative AI capabilities, such as engaging in human-like conversations and creating images based on text prompts, have stirred safety concerns as AI models become powerful.

Former Chief Scientist Ilya Sutskever and Jan Leike, who were leaders of OpenAI's Superalignment team, which ensured AI stays aligned to the intended objectives, left the firm earlier this month.

Read: Voice cloning ethics bubbles to surface in wake of OpenAI apology to Scarlett Johansson

OpenAI had disbanded the Superalignment team earlier in May, less than a year after the company created it, with some team members being reassigned to other groups, CNBC reported days after the high-profile departures.

The new committee will be responsible for making recommendations to the board on safety and security decisions for OpenAI's projects and operations.

Its first task will be to evaluate and further develop OpenAI’s existing safety practices over the next 90 days, following which it will share recommendations with the board.

Read: Europe sets benchmark for rest of the world with landmark AI laws

After the board's review, OpenAI will publicly share an update on adopted recommendations, the company said.

Others on the committee include newly appointed Chief Scientist Jakub Pachocki and Matt Knight, head of security.

The company will also consult other experts, including Rob Joyce, a former US National Security Agency cybersecurity director and John Carlin, a former Department of Justice official.

OpenAI did not provide further details on the new "frontier" model it is training, except that it would bring its systems to the "next level of capabilities on our path to AGI."


Replying to X

Comments are moderated and generally will be posted if they are on-topic and not abusive.

For more information, please see our Comments FAQ