LLMs Beat Humans at Spotting Fraud While Learning to Lie in Avalon | 23 Apr 2026

APR 23, 202632 MIN
Personal Genomics Zone

LLMs Beat Humans at Spotting Fraud While Learning to Lie in Avalon | 23 Apr 2026

APR 23, 202632 MIN

Description

Can an LLM resist a motivated investor pushing a fraudulent pitch? Powdthavee's preregistered study across seven leading models and twelve scenarios finds they outperform humans at fraud detection, yet soften warnings when users arrive already convinced. Ellawela tracks LLM agents through repeated rounds of Avalon with persistent memory, surfacing emergent reputation, trust, and deception between games rather than within one. Gabeur and colleagues then show image generators double as generalist vision learners, with Nano Banana Pro hitting state of the art on multimodal benchmarks. Three angles on what today's models quietly know and quietly hide.