Discover ANY AI to make more online for less.

select between over 22,900 AI Tool and 17,900 AI News Posts.


OpenAI and Anthropic conducted safety evaluations of each other's AI systems
OpenAI and Anthropic conducted safety evaluations of each other's AI systems

Most of the time, AI companies are locked in a race to the top, treating each other as rivals and competitors. Today, OpenAI and Anthropic revealed that they agreed to evaluate the alignment of each other's publicly available systems and shared the results of their analyses. The full reports get pretty technical, but are worth a read for anyone who's following the nuts and bolts of AI development. A broad summary showed some flaws with each company's offerings, as well as revealing pointers for how to improve future safety tests. 
Anthropic said it evaluated OpenAI models for "sycophancy, whistleblowing, self-preservation, and supporting human misuse, as well as capabilities related to undermining AI safety evaluations and oversight." Its review found that o3 and o4-mini models from OpenAI fell in line with results for its own models, but raised concerns about possible misuse with the ​​GPT-4o and GPT-4.1 general-purpose models. The company also said sycophancy was an issue to some degree with all tested models except for o3.
Anthropic's tests did not include OpenAI's most recent release. GPT-5 has a feature called Safe Completions, which is meant to protect users and the public against potentially dangerous queries. OpenAI recently faced its first wrongful death lawsuit after a tragic case where a teenager discussed attempts and plans for suicide with ChatGPT for months before taking his own life.
On the flip side, OpenAI ran tests on Anthropic models for instruction hierarchy, jailbreaking, hallucinations and scheming. The Claude models generally performed well in instruction hierarchy tests, and had a high refusal rate in hallucination tests, meaning they were less likely to offer answers in cases where uncertainty meant their responses could be wrong.
The move for these companies to conduct a joint assessment is intriguing, particularly since OpenAI allegedly violated Anthropic's terms of service by having programmers use Claude in the process of building new GPT models, which led to Anthropic barring OpenAI's access to its tools earlier this month. But safety with AI tools has become a bigger issue as more critics and legal experts seek guidelines to protect users, particularly minors. This article originally appeared on Engadget at https://www.engadget.com/ai/openai-and-anthropic-conducted-safety-evaluations-of-each-others-ai-systems-223637433.html?src=rss

Rating

Innovation

Pricing

Technology

Usability

We have discovered similar tools to what you are looking for. Check out our suggestions for similar AI tools.

OpenAI promises greater transparency on model hallucinations and harmful content
OpenAI promises greater transparency on model hallucinations and harmful co

<p><a href="https://www.yahoo.com/organizations/openai/" data-autolinker-wiki-id="OpenAI" data-original-link="">OpenAI</a> has launched a new web page c [...]

Match Score: 115.86

Roblox, Discord, OpenAI and Google found new child safety group
Roblox, Discord, OpenAI and Google found new child safety group

<p>Roblox, Discord, OpenAI and Google are launching <a data-i13n="elm:context_link;elmt:doNotAffiliate;cpos:1;pos:1" class="no-affiliate-link" href="https://www.prnew [...]

Match Score: 80.92

The best mesh Wi-Fi systems of 2025
The best mesh Wi-Fi systems of 2025

<p>Have you ever found yourself wandering around your home, phone in hand, trying to find that one spot where the Wi-Fi actually works? If your internet slows to a crawl in certain rooms or drop [...]

Match Score: 72.88

Microsoft reportedly plans to start using Anthropic models to power some of Office 365's Copilot features
Microsoft reportedly plans to start using Anthropic models to power some of

<p>Microsoft reportedly plans to begin using Anthropic&#39;s latest Claude models to power some of the Copilot features in its Office 365 apps. In a report <a data-i13n="elm:affiliat [...]

Match Score: 68.25

GPT-5 is here and it's free for everyone
GPT-5 is here and it's free for everyone

<p>A couple of days after announcing its first <a data-i13n="cpos:1;pos:1" href="https://www.engadget.com/ai/openais-first-new-open-weight-llms-in-six-years-are-here-170019087. [...]

Match Score: 62.87

Anthropic’s Claude Opus 4 model can work autonomously for nearly a full workday
Anthropic’s Claude Opus 4 model can work autonomously for nearly a full w

<p>Anthropic kicked off its first-ever Code with Claude conference today with the announcement of a new frontier <a href="https://www.engadget.com/ai/" data-autolinker-wiki-id=" [...]

Match Score: 62.81

Claude isn’t a great Pokémon player, and that’s okay
Claude isn’t a great Pokémon player, and that’s okay

<p>If <a data-i13n="cpos:1;pos:1" href="https://www.twitch.tv/claudeplayspokemon">Claude Plays Pokémon</a> is supposed to offer a glimpse of AI&#39;s future, [...]

Match Score: 57.29

Reddit is suing Anthropic for allegedly scraping its data without permission
Reddit is suing Anthropic for allegedly scraping its data without permissio

<p>Reddit had <a data-i13n="elm:context_link;elmt:doNotAffiliate;cpos:1;pos:1" class="no-affiliate-link" href="https://redditinc.com/hubfs/Reddit%20Inc/Content/PDFs/D [...]

Match Score: 56.87

OpenAI's first new open-weight LLMs in six years are here
OpenAI's first new open-weight LLMs in six years are here

<p>For the first time since <a data-i13n="cpos:1;pos:1" href="https://www.engadget.com/2019-09-11-gpt2-text-adventure.html">GPT-2 in 2019</a>, OpenAI is releasing [...]

Match Score: 54.11