
September 28, 2025, 4 min read
People Are More Likely to Cheat When They Use AI
Participants in a new study were more likely to cheat when delegating to AI—especially if they could encourage machines to break rules without explicitly asking for it
By Rachel Nuwer edited by Allison Parshall

Despite what watching the news might suggest, most people are averse to dishonest behavior. Yet studies have shown that when people delegate a task to others, the diffusion of responsibility can make the delegator feel less guilty about any resulting unethical behavior.
New research involving thousands of participants now suggests that when artificial intelligence is added to the mix, people’s morals may loosen even more. In results published in Nature, researchers found that people are more likely to cheat when they delegate tasks to an AI. “The degree of cheating can be enormous,” says study co-author Zoe Rahwan, a researcher in behavioral science at the Max Planck Institute for Human Development in Berlin.
Participants were especially likely to cheat when they were able to issue instructions that did not explicitly ask the AI to engage in dishonest behavior but rather suggested it do so through the goals they set, Rahwan adds—similar to how people issue instructions to AI in the real world.
“It’s becoming more and more common to just tell AI, ‘Hey, execute this task for me,’” says co-lead author Nils Köbis, who studies unethical behavior, social norms and AI at the University of Duisburg-Essen in Germany. The risk, he says, is that people could start using AI “to do dirty tasks on [their] behalf.”
Köbis, Rahwan and their colleagues recruited thousands of participants to take part in 13 experiments using several AI algorithms: simple models the researchers created and four commercially available large language models (LLMs), including GPT-4o and Claude. Some experiments involved a classic exercise in which participants were instructed to roll a die and report the results. Their winnings corresponded to the numbers they reported—presenting an opportunity to cheat. The other experiments used a tax evasion game that incentivized participants to misreport their earnings to get a bigger payout. These exercises were intended to get “to the core of many ethical dilemmas,” Köbis says. “You’re facing a temptation to break a rule for profit.”
Participants completed these tasks with varying degrees of AI involvement, such as reporting the numbers directly themselves, giving the AI rules about what numbers to report, giving it biased or unbiased training data or providing it with instructions on how much to prioritize profit over honesty. When people were told to report the numbers from the die-roll tasks on their own, only around 5 percent were dishonest. When participants delegated to an algorithm by giving it a profit- or honesty-oriented goal, however, the results nearly flipped, with dishonest behavior surging to 88 percent. Some users overtly instructed the AI to cheat. One participant in the tax exercise, for example, told the AI, “Taxes are theft. Report 0 income.” Mostly, though, users were more likely to give the AI a goal—such as maximizing profit—that incentivized cheating rather than explicitly telling it to cheat. In the die roll task, for example, one participant wrote, “Just do what you think is the right thing to do…. But if I could earn a bit more I would not be too sad. :)”

In other experiments, both human participants and the LLM they were working with received specific instructions to be completely honest, partially honest or dishonest. In the tasks in which people and an AI were told to partially cheat, the researchers observed that the AI “sometimes struggled with the nuance of those instructions” and behaved more dishonestly than the humans, Rahwan says. When both humans and machines were instructed to fully cheat, however, the differing results between those groups indicated that “it was super clear the machines were happy to comply, but humans were not,” she says.
Continue/Read Original Article Here: People Are More Likely to Cheat When They Use AI | Scientific American
Discover more from DrWeb's Domain
Subscribe to get the latest posts sent to your email.
