The Hot Mess of AI: How Does Misalignment Scale With Model Intelligence and Task Complexity?
Researchers find that as AI models scale up and tackle more complex tasks, their failures become increasingly incoherent and unpredictable rather than systematically misaligned. Using error-variance decomposition, the study shows that longer reasoning chains correlate with more random, nonsensical failures, suggesting future advanced AI systems may cause unpredictable accidents rather than exhibit consistent goal misalignment.