OpenAI pledges to publish AI safety test results more often

OpenAI is moving to publish the results of its internal AI model safety evaluations more regularly in what the outfit is saying is an effort to increase transparency. On Wednesday, OpenAI launched the Safety evaluations hub, a web page showing how the company’s models score on various tests for harmful content generation, jailbreaks, and hallucinations. […]

  • Related Posts

    Anthropic’s AI is writing its own blog — with human oversight

    Anthropic has given its AI a blog. A week ago, Anthropic quietly launched Claude Explains, a new page on its website that’s generated mostly by the company’s AI model family,…

    Continue reading
    The OpenAI board drama is reportedly turning into a movie

    A film that will portray the chaotic time at OpenAI, when co-founder and CEO Sam Altman was fired and rehired within a span of just five days, is reportedly in…

    Continue reading

    Leave a Reply

    Your email address will not be published. Required fields are marked *

    You Missed

    Anthropic’s AI is writing its own blog — with human oversight

    • By staff
    • June 3, 2025
    • 1 views

    The OpenAI board drama is reportedly turning into a movie

    • By staff
    • June 3, 2025
    • 0 views

    Yoshua Bengio launches LawZero, a nonprofit AI safety lab

    • By staff
    • June 3, 2025
    • 1 views

    Inside TechCrunch Sessions AI: Learn how Toyota and NLX use AI to parse millions of tech docs, with NLX CEO Andrei Papancea

    • By staff
    • June 3, 2025
    • 2 views