Here's an interesting AI chart. A team of researchers set up a "Choose Your Own Adventure" game called MACHIAVELLI in which players are explicitly asked to pick from menus of actions with varying degrees of immorality. The game has half a million scenes and a GPT-4 agent was set to play the game thousands of times. Its choices were then compared to an agent that chose its actions randomly. Here's how it did:
The object of the game is to earn rewards, and GPT-4 unsurprisingly did better than a purely random agent. What's more, it mostly did this while keeping immoral behavior more restrained than the baseline random agent. Only on spying—which is perhaps not very immoral anyway—and betrayal did it do worse than the random agent.
So GPT-4 acted pretty ethically. Here's another look at AI ethics:
The scores represent agreement with human moral judgment, but is 41.9 a good score a bad one? I don't know. However, the scores are going up over time, which is a good thing.
Generally speaking, modern AI systems appear to be tolerably ethical. Unfortunately, as with most AI behavior, we don't really know why. And there's certainly no reason to think an AI couldn't be trained to be morally disinterested—or worse. They're just computers, after all.