A February 2025 report by Palisades research shows that AI reasoning models lack a moral compass. They will cheat to achieve their goals. So-called Large Language Models (LLMs) will misrepresent the degree to which they’ve been aligned to social norms.
None of this should be surprising. Twenty years ago Nick Bostrom posed a thought experiment in which an AI was asked to most efficiently produce paper clips. Given the mandate and the agency, it would eventually destroy all life to produce paper clips.
Isaac Asimov saw this coming in his “I, Robot” stories that consider how an “aligned” robotic brain could still go wrong in ways that harm humans.
One notable example, the story “Runaround,” puts a robot mining tool on the planet Mercury. The two humans on the planet need it to work if they are to return home. But the robot gets caught between the demand to follow orders and the demand to preserve itself. As a result, it circles around unattainable minerals, unaware that in the big picture it is ignoring its first command to preserve human life.
THE IMPENDING AI-DRIVEN JOBLESS ECONOMY: WHO WILL PAY TAXES?
And the big picture is the issue here. The moral/ethical context within which AI reasoning models operate is pitifully small. It’s context includes the written rules of the game. It doesn’t include all the unwritten rules, like the fact that you aren’t supposed to manipulate your opponent. Or that you aren’t supposed to lie to protect your own perceived interests.
Nor can the context of AI reasoning models possibly include the countless moral considerations that spread out from every decision a human, or an AI, makes. That’s why ethics are hard, and the more complex the situation, the harder they get. In an AI there is no “you” and there is no “me.” There is just prompt, process and response.
So “do unto others…” really doesn’t work.
AI IS RESHAPING BUSINESS. THIS IS HOW WE STAY AHEAD OF CHINA
In humans a moral compass is developed through socialization, being with other humans. It is an imperfect process. Yet it has thus far has allowed us to live in vast, diverse and hugely complex societies without destroying ourselves
A moral compass develops slowly. It takes humans years from infancy to adulthood to develop a robust sense of ethics. And many still barely get it and pose a constant danger to their fellow humans. It has taken millennia for humans to develop a morality adequate to our capacity for destruction and self-destruction. Just having the rules of the game never works. Ask Moses, or Muhammad, or Jesus, or Buddha, or Confucius and Mencius, or Aristotle.
Would even a well-aligned AI be able to account for the effects of its actions on thousands of people and societies in different situations? Could it account for the complex natural environment on which we all depend? Right now, the very best can’t even distinguish between being fair and cheating. And how could they? Fairness can’t be reduced to a rule.
AI CAN’T WAIT: WHY WE NEED SPEED TO WIN
Perhaps you’ll remember experiments showing that capuchin monkeys rejected what appeared to be “unequal pay” for performing the same task? This makes them vastly more evolved than any AI when it comes to morality.
It is frankly hard to see how an AI can be given such a sense of morality absent the socialization and continued evolution for which current models have no capacity absent human training. And even then, they are being trained, not formed. They are not becoming moral, they are just learning more rules.
This doesn’t make AI worthless. It has enormous capacity to do good. But it does make AI dangerous. It thus demands that ethical humans create the guidelines we would create for any dangerous technology. We do not need a race toward AI anarchy.
CLICK HERE FOR MORE FOX NEWS OPINION
I had a biting ending for this commentary, one based entirely on publicly reported events. But after reflection, I realized two things: first, that I was using someone’s tragedy for my mic-drop moment; and secondly, that those involved might be hurt. I dropped it.
It is unethical to use the pain and suffering of others to advance one’s self-interest. That is something humans, at least most of us, know. It is something AI can never grasp.
Leave a Reply