Google DeepMind's New Test: Can AI Learn to Think Morally?
Google DeepMind is now asking if artificial intelligence can learn to think about right and wrong. In a new study, researchers have built a system to measure how AI models handle ethical dilemmas, moving beyond simple rules to probe for something like genuine reasoning. The work immediately confronts a foundational problem: whose moral standards should these systems use?
The research, detailed in a paper on moral reasoning in large language models, creates a benchmark that draws from major schools of philosophical thought—consequentialism, deontology, and virtue ethics. Instead of looking for a single right answer, it tests whether an AI can identify conflicting values and justify its stance. The models were given complex scenarios where duties clash or outcomes are unclear, similar to the tough choices humans debate.
Early results show these models are uneven. They can spot clear-cut violations but falter with nuanced conflicts, much like people do. More concerning, they tend to favor ethical frameworks most common in their training data, which is largely Western and English-language. This means an AI might lean utilitarian not by reasoned choice, but because it saw more examples online.
This isn't just academic. AI is already influencing decisions in healthcare, finance, and content moderation. The moral lens it applies has real consequences. DeepMind’s approach focuses on measurement rather than prescribing one moral code, a practical step for an industry in a hurry. Yet the core tension remains: human morality is tied to emotion, experience, and relationships—things a machine cannot know. What DeepMind has built is a mirror, showing how far AI still has to go before it can truly understand the weight of its choices.
Original source
Read on Webpronews