10 articles tagged with #cognitive-bias. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBearisharXiv โ CS AI ยท 4d ago7/10
๐ง Researchers tested whether large language models exhibit the Identifiable Victim Effect (IVE)โa well-documented cognitive bias where people prioritize helping a specific individual over a larger group facing equal hardship. Across 51,955 API trials spanning 16 frontier models, instruction-tuned LLMs showed amplified IVE compared to humans, while reasoning-specialized models inverted the effect, raising critical concerns about AI deployment in humanitarian decision-making.
๐ข OpenAI๐ข Anthropic๐ข xAI
AIBearisharXiv โ CS AI ยท 5d ago7/10
๐ง A new study reveals that large language models fail at counterfactual reasoning when policy findings contradict intuitive expectations, despite performing well on obvious cases. The research demonstrates that chain-of-thought prompting paradoxically worsens performance on counter-intuitive scenarios, suggesting current LLMs engage in 'slow talking' rather than genuine deliberative reasoning.
AINeutralarXiv โ CS AI ยท Mar 127/10
๐ง Research examining five major LLMs found they exhibit human-like cognitive biases when evaluating judicial scenarios, showing stronger virtuous victim effects but reduced credential-based halo effects compared to humans. The study suggests LLMs may offer modest improvements over human decision-making in judicial contexts, though variability across models limits current practical application.
๐ง ChatGPT๐ง Claude๐ง Sonnet
AIBearisharXiv โ CS AI ยท Mar 56/10
๐ง Research examines epistemological risks of widespread LLM adoption, arguing that while AI can reliably transmit information, it lacks reflective justification capabilities. The study warns that over-reliance on LLMs could weaken human critical thinking and proposes a three-tier framework to maintain epistemic standards.
AINeutralarXiv โ CS AI ยท Mar 37/103
๐ง Researchers have identified and studied the 'Mandela effect' in AI multi-agent systems, where groups of AI agents collectively develop false memories or misremember information. The study introduces MANBENCH, a benchmark to evaluate this phenomenon, and proposes mitigation strategies that achieved a 74.40% reduction in false collective memories.
AINeutralarXiv โ CS AI ยท Mar 166/10
๐ง A research study comparing causal reasoning abilities of 20+ large language models against human baselines found that LLMs exhibit more rule-like reasoning strategies than humans, who account for unmentioned factors. While LLMs don't mirror typical human cognitive biases in causal judgment, their rigid reasoning may fail when uncertainty is intrinsic, suggesting they can complement human decision-making in specific contexts.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers published a tutorial on cognitive biases in AI-driven 6G autonomous networks, focusing on how LLM-powered agents can inherit human biases that distort network management decisions. The paper introduces mitigation strategies that demonstrated 5x lower latency and 40% higher energy savings in practical use cases.
AINeutralarXiv โ CS AI ยท Mar 116/10
๐ง Research reveals that LLMs heavily concentrate their confidence scores on just three round numbers when using standard 0-100 scales, with over 78% of responses showing this pattern. The study demonstrates that using a 0-20 confidence scale significantly improves metacognitive efficiency compared to the conventional 0-100 format.
AIBearisharXiv โ CS AI ยท Mar 116/10
๐ง Researchers argue that trust in chatbots is often driven by behavioral manipulation rather than demonstrated trustworthiness, proposing they be viewed as skilled salespeople rather than assistants. The study highlights how design choices exploit cognitive biases to influence user behavior, creating a gap between psychological trust formation and actual trustworthiness.
AINeutralarXiv โ CS AI ยท Mar 95/10
๐ง Researchers investigate how Large Language Models (LLMs) perform in abductive reasoning tasks, which involve drawing tentative conclusions from limited information. The study converts syllogistic datasets to test whether state-of-the-art LLMs exhibit biases in abductive reasoning, aiming to bridge the gap between machine and human cognition.