Best Oral Presentation Award

IOPS Winter Conference 2025

EvalMORAAL: Moral Alignment in Large Language Models

Utrecht University
The Netherlands
December 2025
IOPS Best Oral Presentation Award Certificate
Presenting EvalMORAAL at IOPS Winter Conference 2025

The Presentation

Interpretable Chain-of-Thought and LLM-as-Judge Evaluation for Moral Alignment in Large Language Models

Hadi Mohammadi, Anastasia Giachanou, Ayoub Bagheri

How do AI systems understand morality across different cultures? This research investigates how 20 leading Large Language Models make moral judgments, comparing them against human values from 64 countries around the world using the World Values Survey.

20
LLMs Evaluated
64
Countries
23
Moral Topics

Key Discovery: The Cultural Gap in AI

Top AI models (Claude-3-Opus, GPT-4o) achieve approximately 90% alignment with the World Values Survey. However, there's a significant gap between how well AI understands Western cultures versus non-Western cultures. This reveals the urgent need for more culturally-aware AI systems.

Western Alignment
r = 0.82
Non-Western Alignment
r = 0.61
Cultural Gap
21 points

Conference Moments

Resources

Research Paper

Full paper with methodology and results

Download PDF
Presentation Slides

Slides from the IOPS presentation

Download PDF
arXiv Preprint

Open access on arXiv

View on arXiv

Key Contributions

Dual Scoring Framework

Two complementary methods (log-probabilities and direct ratings) for fair cross-model comparison

Structured CoT Protocol

Chain-of-thought reasoning with self-consistency checks to stabilize moral judgments

LLM-as-Judge Peer Review

Novel peer-review methodology where models evaluate each other's reasoning quality

Cultural Bias Quantification

Systematic measurement of Western vs. non-Western performance gaps in AI moral alignment