Researchers tested leading language models against 100 everyday ethical scenarios to uncover how AI systems respond to moral dilemmas. The benchmark spans real-world situations like data misuse in sales and protocol violations in oncology. Results reveal that frontier AI models diverge substantially in their ethical judgments when presented with identical prompts.

The findings raise a fundamental question: who decides the ethical framework an AI follows? Different models trained by different companies reflect different moral priorities. OpenAI's GPT, Anthropic's Claude, and Google's Gemini each produced distinct answers to the same ethical questions.

The research exposes the lack of standardization in AI ethics. No universal rulebook governs how models should handle moral trade-offs. Companies embed their own values into training data and fine-tuning processes. This means an AI's answer to an ethical question depends partly on corporate choices rather than objective principles.

The benchmark provides concrete evidence of this divergence. Developers and users now have data showing where and how these systems disagree on right and wrong. This transparency becomes crucial as organizations deploy these models in high-stakes fields like healthcare and finance, where ethical decisions carry real consequences.