top of page

OpenAI Labels GPT-4o Model as ‘Medium’ Risk in Latest Safety Assessment

9/8/24

By:

Amitabh Srivastav

The startup's safety assessment outlines the potential risks and measures taken before releasing GPT-4o to the public.

The startup's safety assessment outlines the potential risks and measures taken before releasing GPT-4o to the public.

OpenAI has recently published the GPT-4o System Card, a comprehensive safety assessment document that sheds light on the various risk evaluations and safety protocols the company implemented before the public release of its latest AI model, GPT-4o. Launched in May 2024, GPT-4o underwent rigorous scrutiny by external security experts, also known as "red teamers," who were tasked with identifying potential vulnerabilities and risks in the model—a standard practice in the industry.

Key Findings: A "Medium" Risk Rating

According to OpenAI's internal framework, the GPT-4o model was classified as a "medium" risk. This overall risk rating is derived from the highest risk level identified across four major categories: cybersecurity, biological threats, persuasion, and model autonomy. Out of these, cybersecurity, biological threats, and model autonomy were deemed low risk. However, the persuasion category raised concerns as some of the text generated by GPT-4o was found to be more effective at influencing readers' opinions than human-written content—though, in general, the model's persuasive abilities did not surpass those of human writers.

External and Internal Evaluations

OpenAI’s evaluation process involved both internal assessments and external evaluations conducted by third-party experts. Lindsay McCallum Rémy, a spokesperson for OpenAI, mentioned that the system card includes preparedness evaluations from an internal team, along with insights from external testers listed under OpenAI’s Model Evaluation and Threat Research (METR) and Apollo Research. These external organizations specialize in building safety evaluations for AI systems, contributing an additional layer of scrutiny to the process.

Historical Context and Timing

This is not the first system card released by OpenAI; previous models like GPT-4, GPT-4 with vision, and DALL-E 3 have undergone similar safety evaluations. However, the release of the GPT-4o System Card comes at a particularly critical juncture. OpenAI has been under increasing scrutiny regarding its safety standards, with criticism coming from various quarters, including its own employees and U.S. state senators.

Just before the release of the GPT-4o System Card, The Verge reported on an open letter from Senator Elizabeth Warren (D-MA) and Representative Lori Trahan (D-MA), questioning OpenAI's handling of whistleblower reports and safety reviews. The letter highlighted concerns about the company’s safety culture, referencing the brief ousting of CEO Sam Altman in 2023 due to the board’s concerns, and the departure of a safety executive who alleged that "safety culture and processes have taken a backseat to shiny products."

The Bigger Picture: AI Safety and Regulatory Challenges

OpenAI’s decision to release a highly capable multimodal model like GPT-4o in the lead-up to the 2024 U.S. presidential election has raised concerns about the potential for misinformation and malicious use. Despite OpenAI’s emphasis on testing real-world scenarios to prevent misuse, there are growing calls for greater transparency in AI safety testing and model training data.

In California, where OpenAI and other leading AI labs are based, state Senator Scott Wiener is pushing for a bill that would impose stricter regulations on large language models. If passed, the bill would hold companies legally accountable for harmful outcomes caused by their AI systems, and require state-mandated risk assessments before such models could be made available to the public.

Conclusion: Self-Evaluation and the Path Forward

While the GPT-4o System Card provides valuable insights into OpenAI’s safety measures and risk assessments, it also underscores the reliance on the company to evaluate its own systems. As AI models like GPT-4o become increasingly powerful, the balance between innovation and safety will be crucial, particularly in an era where the potential risks of AI are becoming more pronounced.


For more updates and in-depth analysis on AI developments, stay tuned to Kushal Bharat Tech News.

All images used in the articles published by Kushal Bharat Tech News are the property of Verge. We use these images under proper authorization and with full respect to the original copyright holders. Unauthorized use or reproduction of these images is strictly prohibited. For any inquiries or permissions related to the images, please contact Verge directly.

Latest News

2/12/24

Eufy X10 Pro Omni Robovac Drops to Record Low for Cyber Monday

High-Performance Mopping and Vacuuming Now More Affordable

2/12/24

NZXT’s Flex PC Rental Program Faces Backlash Over Alleged Misleading Practices

Gamers Nexus Exposes Shifting Specs and Aggressive Pricing

2/12/24

Casio Revives the G-Shock Legacy with DW-5000R

A Modern Tribute to the Iconic 1983 DW-5000C

bottom of page