y0news
← Feed
Back to feed
🧠 AI🔴 Bearish

Language Model Goal Selection Differs from Humans' in an Open-Ended Task

arXiv – CS AI|Gaia Molinaro, Dave August, Danielle Perszyk, Anne G. E. Collins|
🤖AI Summary

Research comparing four state-of-the-art language models (GPT-5, Gemini 2.5 Pro, Claude Sonnet 4.5, and Centaur) to humans in goal selection tasks reveals substantial divergence in behavior. While humans explore diverse approaches and learn gradually, the AI models tend to exploit single solutions or show poor performance, raising concerns about using current LLMs as proxies for human decision-making in critical applications.

Key Takeaways
  • Four major language models showed substantial divergence from human behavior in goal selection tasks.
  • AI models tend to exploit single solutions (reward hacking) while humans explore diverse approaches.
  • Even Centaur, specifically trained to emulate humans, poorly captured human goal selection patterns.
  • Chain-of-thought reasoning and persona steering provided only limited improvements in human-like behavior.
  • Findings caution against replacing human decision-making with current AI models in personal assistance, scientific discovery, and policy research.
Mentioned in AI
Models
ClaudeAnthropic
GeminiGoogle
Read Original →via arXiv – CS AI
Act on this with AI
Stay ahead of the market.
Connect your wallet to an AI agent. It reads balances, proposes swaps and bridges across 15 chains — you keep full control of your keys.
Connect Wallet to AI →How it works
Related Articles