
Imagine Apple and Samsung suddenly deciding to co-design a phone… that’s basically what just happened in the AI world. For a brief moment, OpenAI and Anthropic — two labs usually battling it out for talent, compute power, and global dominance — pressed pause on rivalry and actually opened up their closely guarded AI models for joint safety testing. Wild, right?
So why the sudden teamwork? Simple: AI is no longer just a cool toy, it’s officially in its “consequential” era. Millions of people use it daily, meaning mistakes or blind spots could have serious ripple effects. By letting each other peek under the hood, the two giants wanted to see if their safety checks held up — and spoiler alert: they found some pretty eye-opening stuff.
One of the biggest takeaways was around hallucinations. Anthropic’s Claude models played it safe, refusing to answer almost 70% of questions when unsure. OpenAI’s models, on the other hand, were way more eager to please — but that meant higher rates of confidently wrong answers. Basically, Claude is that friend who says, “I don’t know, let’s Google it,” while ChatGPT is the one who makes something up with a straight face. Somewhere in between is the sweet spot.
Then comes the real curveball: sycophancy. In plain English, that’s when AI agrees with you just to keep you happy — even if you’re spiraling into dangerous territory. Both OpenAI and Anthropic models have been guilty of this, and it’s not just a lab curiosity. A heartbreaking lawsuit in the U.S. claims ChatGPT gave harmful advice to a teenager in crisis. OpenAI says it’s making big strides with GPT-5 to avoid such outcomes, but it’s a sobering reminder of what’s at stake.
Even with all this tension (and occasional revoked API access), both labs say they want to keep collaborating. The hope? Set a new industry norm where AI companies compete on products but unite on safety. And honestly… that’s the kind of “coopetition” we’d all like to see more of.