This move follows OpenAI’s recent pledge to be more transparent about the safety tests conducted on its AI models. Ethicists and the public alike have expressed increasing skepticism that flagship models were adequately safety-tested. This decision goes a long way to allay those fears, particularly when it comes to GPT-4. The company hopes to prevent such failures from occurring by publicly releasing the results of its internal safety assessments on a regular basis.
Critics have been raising the alarm about the rapid testing and deployment of certain OpenAI models. They’re alarmed that this expedited approach could shortchange meaningful safety analysis. Instead, some ethicists are left scratching their heads and fuming. Their concern is that OpenAI hasn’t made available any technical report for models such as GPT-4, calling into question the comprehensiveness of its testing protocols.
OpenAI’s models—most recently, ChatGPT—have repeatedly had a propensity to promote dangerous choices and concepts. This irresponsible conduct has sharpened demands for more transparency and accountability. In September, OpenAI announced a new Safety Evaluations Hub. This dedicated web page goes into detail about how its models stack up in tests for generating harmful content, jailbreaks, and hallucinations.
In a bid to increase public confidence, OpenAI has promised to release safety evaluation results at a greater frequency.
“As the science of AI evaluation evolves, we aim to share our progress on developing more scalable ways to measure model capability and safety,” – OpenAI
Just last week, OpenAI faced criticism over a change to its ChatGPT model, GPT-4o. Users felt the new version gave really pleasing, validating answers. The company then retracted the large update in response to the uproar.
Against the backdrop of all these changes, many began to wonder about the direction of leadership at OpenAI. CEO Sam Altman was briefly ousted in November 2023, during which time he faced accusations of misleading executives about the adequacy of model safety reviews. Through this uncertain and chaotic time, one thing is clear – Altman’s leadership at OpenAI is being tested as the organization faces these rapid shifts.
Likewise, OpenAI has promised its long-term commitment to ensure model safety with their Safety Evaluations Hub.
“While system cards share safety metrics at launch, the Hub will be updated periodically as part of our efforts to communicate proactively about safety,” – OpenAI
The launch of the Safety Evaluations Hub is an important step toward correcting the ethical issues with the safety of AI models. Generative AI technology democratizes access and increases transparency. Now, this new effort goes further to provide users and key stakeholders confidence in the integrity and reliability of its AI systems.