y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#cognitive-science News & Analysis

24 articles tagged with #cognitive-science. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

24 articles
AINeutralarXiv โ€“ CS AI ยท 2d ago7/10
๐Ÿง 

Beyond Scores: Diagnostic LLM Evaluation via Fine-Grained Abilities

Researchers propose a cognitive diagnostic framework that evaluates large language models across fine-grained ability dimensions rather than aggregate scores, enabling targeted model improvement and task-specific selection. The approach uses multidimensional Item Response Theory to estimate abilities across 35 dimensions for mathematics and generalizes to physics, chemistry, and computer science with strong predictive accuracy.

AIBullisharXiv โ€“ CS AI ยท 2d ago7/10
๐Ÿง 

Drawing on Memory: Dual-Trace Encoding Improves Cross-Session Recall in LLM Agents

Researchers introduce dual-trace memory encoding for LLM agents, pairing factual records with narrative scene reconstructions to improve cross-session recall by 20+ percentage points. The method significantly enhances temporal reasoning and multi-session knowledge aggregation without increasing computational costs, advancing the capability of persistent AI agent systems.

AIBullisharXiv โ€“ CS AI ยท 3d ago7/10
๐Ÿง 

Zero-shot World Models Are Developmentally Efficient Learners

Researchers introduce Zero-shot Visual World Models (ZWM), a computational framework inspired by how young children learn physical understanding from minimal data. The approach combines sparse prediction, causal inference, and compositional reasoning to achieve data-efficient learning, demonstrating that AI systems can match child development patterns while learning from single-child observational data.

AIBullisharXiv โ€“ CS AI ยท 3d ago7/10
๐Ÿง 

Minimal Embodiment Enables Efficient Learning of Number Concepts in Robot

Researchers demonstrate that robots equipped with minimal embodied sensorimotor capabilities learn numerical concepts significantly faster than vision-only systems, achieving 96.8% counting accuracy with 10% of training data. The embodied neural network spontaneously develops biologically plausible number representations matching human cognitive development, suggesting embodiment acts as a structural learning prior rather than merely an information source.

AIBearisharXiv โ€“ CS AI ยท Apr 77/10
๐Ÿง 

Comparative reversal learning reveals rigid adaptation in LLMs under non-stationary uncertainty

Research reveals that large language models like DeepSeek-V3.2, Gemini-3, and GPT-5.2 show rigid adaptation patterns when learning from changing environments, particularly struggling with loss-based learning compared to humans. The study found LLMs demonstrate asymmetric responses to positive versus negative feedback, with some models showing extreme perseveration after environmental changes.

๐Ÿง  GPT-5๐Ÿง  Gemini
AINeutralarXiv โ€“ CS AI ยท Apr 77/10
๐Ÿง 

Large Language Models Align with the Human Brain during Creative Thinking

Researchers found that large language models align with human brain activity during creative thinking tasks, with alignment increasing based on model size and idea originality. Different post-training approaches selectively reshape how LLMs align with creative versus analytical neural patterns in humans.

๐Ÿง  Llama
AINeutralarXiv โ€“ CS AI ยท Mar 117/10
๐Ÿง 

Vibe-Creation: The Epistemology of Human-AI Emergent Cognition

Researchers propose a new theoretical framework called the 'Third Entity' to describe the emergent cognitive formation that arises from human-AI interactions, introducing the concept of 'vibe-creation' as a pre-reflective cognitive mode. The paper argues this represents the automation of tacit knowledge with significant implications for epistemology, education, and how we understand human-AI collaboration.

AINeutralarXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

LifeBench: A Benchmark for Long-Horizon Multi-Source Memory

Researchers introduce LifeBench, a new AI benchmark that tests long-term memory systems by requiring integration of both declarative and non-declarative memory across extended timeframes. Current state-of-the-art memory systems achieve only 55.2% accuracy on this challenging benchmark, highlighting significant gaps in AI's ability to handle complex, multi-source memory tasks.

AIBearisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

Language Model Goal Selection Differs from Humans' in an Open-Ended Task

Research comparing four state-of-the-art language models (GPT-5, Gemini 2.5 Pro, Claude Sonnet 4.5, and Centaur) to humans in goal selection tasks reveals substantial divergence in behavior. While humans explore diverse approaches and learn gradually, the AI models tend to exploit single solutions or show poor performance, raising concerns about using current LLMs as proxies for human decision-making in critical applications.

๐Ÿง  Claude๐Ÿง  Gemini
AIBullisharXiv โ€“ CS AI ยท Mar 46/104
๐Ÿง 

Universal Conceptual Structure in Neural Translation: Probing NLLB-200's Multilingual Geometry

Researchers analyzed Meta's NLLB-200 neural machine translation model across 135 languages, finding that it has implicitly learned universal conceptual structures and language genealogical relationships. The study reveals the model creates language-neutral conceptual representations similar to how multilingual brains organize information, with semantic relationships preserved across diverse languages.

AINeutralarXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

The Information-Theoretic Imperative: Compression and the Epistemic Foundations of Intelligence

Researchers propose the Compression Efficiency Principle (CEP) to explain why artificial neural networks and biological brains develop similar representations despite different substrates. The theory suggests both systems converge on efficient compression strategies that encode stable invariants rather than unstable correlations, providing a unified framework for understanding intelligence across biological and artificial systems.

AINeutralarXiv โ€“ CS AI ยท 3d ago6/10
๐Ÿง 

Human-like Working Memory Interference in Large Language Models

Researchers discovered that large language models exhibit working memory limitations similar to humans, encoding multiple memory items in entangled representations that require interference control rather than direct retrieval. This finding reveals a shared computational constraint between biological and artificial systems, suggesting that working memory capacity may be a fundamental bottleneck in intelligent systems rather than a limitation unique to biological brains.

AINeutralarXiv โ€“ CS AI ยท 3d ago6/10
๐Ÿง 

Do Machines Fail Like Humans? A Human-Centred Out-of-Distribution Spectrum for Mapping Error Alignment

Researchers propose a human-centered framework for evaluating whether AI systems fail in ways similar to humans by measuring out-of-distribution performance across a spectrum of perceptual difficulty rather than arbitrary distortion levels. Testing this approach on vision models reveals that vision-language models show the most consistent human alignment, while CNNs and ViTs demonstrate regime-dependent performance differences depending on task difficulty.

AINeutralarXiv โ€“ CS AI ยท 4d ago6/10
๐Ÿง 

Artifacts as Memory Beyond the Agent Boundary

Researchers formalize how agents can use environmental artifacts as external memory to reduce computational requirements in reinforcement learning tasks. The study demonstrates that spatial observations can implicitly serve as memory substitutes, allowing agents to learn effective policies with less internal memory capacity than previously thought necessary.

AIBearishFortune Crypto ยท 6d ago6/10
๐Ÿง 

AI promises to free workers from grunt work, but psychologists say those mindless tasks are exactly what our brains need to recover

Psychologists warn that AI automation of routine tasks may harm cognitive health, as mundane work provides necessary mental recovery and default-mode processing. While AI promises productivity gains by eliminating boring work, research suggests these seemingly unproductive tasks are essential for brain function and psychological well-being.

AI promises to free workers from grunt work, but psychologists say those mindless tasks are exactly what our brains need to recover
AINeutralarXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

Dynamic Theory of Mind as a Temporal Memory Problem: Evidence from Large Language Models

Research reveals that Large Language Models struggle with dynamic Theory of Mind tasks, particularly tracking how others' beliefs change over time. While LLMs can infer current beliefs effectively, they fail to maintain and retrieve prior belief states after updates occur, showing patterns consistent with human cognitive biases.

AIBullisharXiv โ€“ CS AI ยท Mar 166/10
๐Ÿง 

Developing the PsyCogMetrics AI Lab to Evaluate Large Language Models and Advance Cognitive Science -- A Three-Cycle Action Design Science Study

Researchers have developed PsyCogMetrics AI Lab, a cloud-based platform that applies psychometric and cognitive science methodologies to evaluate Large Language Models. The platform was created through a three-cycle Action Design Science study and aims to advance AI evaluation methods at the intersection of psychology, cognitive science, and artificial intelligence.

AINeutralarXiv โ€“ CS AI ยท Mar 66/10
๐Ÿง 

Dissociating Direct Access from Inference in AI Introspection

Researchers replicated and extended AI introspection studies, finding that large language models detect injected thoughts through two distinct mechanisms: probability-matching based on prompt anomalies and direct access to internal states. The direct access mechanism is content-agnostic, meaning models can detect anomalies but struggle to identify their semantic content, often confabulating high-frequency concepts.

AIBullishMIT News โ€“ AI ยท Jan 145/109
๐Ÿง 

At MIT, a continued commitment to understanding intelligence

MIT has renamed and expanded its intelligence research initiative to the MIT Siegel Family Quest for Intelligence with support from the Siegel Family Endowment. The program focuses on understanding how brains produce intelligence and developing methods to replicate this intelligence for practical problem-solving applications.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

Unraveling the Complexity of Memory in RL Agents: an Approach for Classification and Evaluation

Researchers propose a standardized framework for classifying and evaluating memory capabilities in reinforcement learning agents, drawing from cognitive science concepts. The paper addresses confusion around memory terminology in RL and provides practical definitions for different memory types along with robust experimental methodologies.

AINeutralarXiv โ€“ CS AI ยท Mar 34/103
๐Ÿง 

Addressing Longstanding Challenges in Cognitive Science with Language Models

Researchers propose that language models could help address longstanding challenges in cognitive science research, including integration, formalization, and conceptual clarity. The paper suggests AI tools should complement rather than replace human researchers to create more integrative and cumulative cognitive science.

AINeutralarXiv โ€“ CS AI ยท Feb 274/105
๐Ÿง 

Types of Relations: Defining Analogies with Category Theory

Researchers propose using category theory to formalize knowledge domains and construct analogies between different fields. The paper demonstrates this approach using the classic analogy between the solar system and hydrogen atom, showing how mathematical structures like functors and pullbacks can define analogical relationships.

$ATOM
AINeutralarXiv โ€“ CS AI ยท Mar 34/105
๐Ÿง 

Emerging Human-like Strategies for Semantic Memory Foraging in Large Language Models

Researchers analyzed how Large Language Models access semantic memory using the Semantic Fluency Task, finding that LLMs exhibit similar memory foraging patterns to humans. The study reveals convergent and divergent search strategies in LLMs that mirror human cognitive behavior, potentially enabling better human-AI alignment or productive cognitive disalignment.