24 articles tagged with #cognitive-science. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AINeutralarXiv โ CS AI ยท 2d ago7/10
๐ง Researchers propose a cognitive diagnostic framework that evaluates large language models across fine-grained ability dimensions rather than aggregate scores, enabling targeted model improvement and task-specific selection. The approach uses multidimensional Item Response Theory to estimate abilities across 35 dimensions for mathematics and generalizes to physics, chemistry, and computer science with strong predictive accuracy.
AIBullisharXiv โ CS AI ยท 2d ago7/10
๐ง Researchers introduce dual-trace memory encoding for LLM agents, pairing factual records with narrative scene reconstructions to improve cross-session recall by 20+ percentage points. The method significantly enhances temporal reasoning and multi-session knowledge aggregation without increasing computational costs, advancing the capability of persistent AI agent systems.
AIBullisharXiv โ CS AI ยท 3d ago7/10
๐ง Researchers introduce Zero-shot Visual World Models (ZWM), a computational framework inspired by how young children learn physical understanding from minimal data. The approach combines sparse prediction, causal inference, and compositional reasoning to achieve data-efficient learning, demonstrating that AI systems can match child development patterns while learning from single-child observational data.
AIBullisharXiv โ CS AI ยท 3d ago7/10
๐ง Researchers demonstrate that robots equipped with minimal embodied sensorimotor capabilities learn numerical concepts significantly faster than vision-only systems, achieving 96.8% counting accuracy with 10% of training data. The embodied neural network spontaneously develops biologically plausible number representations matching human cognitive development, suggesting embodiment acts as a structural learning prior rather than merely an information source.
AIBearisharXiv โ CS AI ยท Apr 77/10
๐ง Research reveals that large language models like DeepSeek-V3.2, Gemini-3, and GPT-5.2 show rigid adaptation patterns when learning from changing environments, particularly struggling with loss-based learning compared to humans. The study found LLMs demonstrate asymmetric responses to positive versus negative feedback, with some models showing extreme perseveration after environmental changes.
๐ง GPT-5๐ง Gemini
AINeutralarXiv โ CS AI ยท Apr 77/10
๐ง Researchers found that large language models align with human brain activity during creative thinking tasks, with alignment increasing based on model size and idea originality. Different post-training approaches selectively reshape how LLMs align with creative versus analytical neural patterns in humans.
๐ง Llama
AINeutralarXiv โ CS AI ยท Mar 117/10
๐ง Researchers propose a new theoretical framework called the 'Third Entity' to describe the emergent cognitive formation that arises from human-AI interactions, introducing the concept of 'vibe-creation' as a pre-reflective cognitive mode. The paper argues this represents the automation of tacit knowledge with significant implications for epistemology, education, and how we understand human-AI collaboration.
AINeutralarXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduce LifeBench, a new AI benchmark that tests long-term memory systems by requiring integration of both declarative and non-declarative memory across extended timeframes. Current state-of-the-art memory systems achieve only 55.2% accuracy on this challenging benchmark, highlighting significant gaps in AI's ability to handle complex, multi-source memory tasks.
AIBearisharXiv โ CS AI ยท Mar 56/10
๐ง Research comparing four state-of-the-art language models (GPT-5, Gemini 2.5 Pro, Claude Sonnet 4.5, and Centaur) to humans in goal selection tasks reveals substantial divergence in behavior. While humans explore diverse approaches and learn gradually, the AI models tend to exploit single solutions or show poor performance, raising concerns about using current LLMs as proxies for human decision-making in critical applications.
๐ง Claude๐ง Gemini
AIBullisharXiv โ CS AI ยท Mar 46/104
๐ง Researchers analyzed Meta's NLLB-200 neural machine translation model across 135 languages, finding that it has implicitly learned universal conceptual structures and language genealogical relationships. The study reveals the model creates language-neutral conceptual representations similar to how multilingual brains organize information, with semantic relationships preserved across diverse languages.
AINeutralarXiv โ CS AI ยท Mar 37/104
๐ง Researchers propose the Compression Efficiency Principle (CEP) to explain why artificial neural networks and biological brains develop similar representations despite different substrates. The theory suggests both systems converge on efficient compression strategies that encode stable invariants rather than unstable correlations, providing a unified framework for understanding intelligence across biological and artificial systems.
AINeutralarXiv โ CS AI ยท 3d ago6/10
๐ง Researchers discovered that large language models exhibit working memory limitations similar to humans, encoding multiple memory items in entangled representations that require interference control rather than direct retrieval. This finding reveals a shared computational constraint between biological and artificial systems, suggesting that working memory capacity may be a fundamental bottleneck in intelligent systems rather than a limitation unique to biological brains.
AINeutralarXiv โ CS AI ยท 3d ago6/10
๐ง Researchers propose a human-centered framework for evaluating whether AI systems fail in ways similar to humans by measuring out-of-distribution performance across a spectrum of perceptual difficulty rather than arbitrary distortion levels. Testing this approach on vision models reveals that vision-language models show the most consistent human alignment, while CNNs and ViTs demonstrate regime-dependent performance differences depending on task difficulty.
AINeutralarXiv โ CS AI ยท 4d ago6/10
๐ง Researchers formalize how agents can use environmental artifacts as external memory to reduce computational requirements in reinforcement learning tasks. The study demonstrates that spatial observations can implicitly serve as memory substitutes, allowing agents to learn effective policies with less internal memory capacity than previously thought necessary.
AIBearishFortune Crypto ยท 6d ago6/10
๐ง Psychologists warn that AI automation of routine tasks may harm cognitive health, as mundane work provides necessary mental recovery and default-mode processing. While AI promises productivity gains by eliminating boring work, research suggests these seemingly unproductive tasks are essential for brain function and psychological well-being.
AINeutralarXiv โ CS AI ยท Mar 176/10
๐ง Research reveals that Large Language Models struggle with dynamic Theory of Mind tasks, particularly tracking how others' beliefs change over time. While LLMs can infer current beliefs effectively, they fail to maintain and retrieve prior belief states after updates occur, showing patterns consistent with human cognitive biases.
AIBullisharXiv โ CS AI ยท Mar 176/10
๐ง Researchers propose a new AI learning architecture inspired by human and animal cognition that integrates observational learning and active behavior learning. The framework includes a meta-control system that switches between learning modes, addressing current limitations in autonomous AI learning.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers have developed PsyCogMetrics AI Lab, a cloud-based platform that applies psychometric and cognitive science methodologies to evaluate Large Language Models. The platform was created through a three-cycle Action Design Science study and aims to advance AI evaluation methods at the intersection of psychology, cognitive science, and artificial intelligence.
AINeutralarXiv โ CS AI ยท Mar 66/10
๐ง Researchers replicated and extended AI introspection studies, finding that large language models detect injected thoughts through two distinct mechanisms: probability-matching based on prompt anomalies and direct access to internal states. The direct access mechanism is content-agnostic, meaning models can detect anomalies but struggle to identify their semantic content, often confabulating high-frequency concepts.
AIBullishMIT News โ AI ยท Jan 145/109
๐ง MIT has renamed and expanded its intelligence research initiative to the MIT Siegel Family Quest for Intelligence with support from the Siegel Family Endowment. The program focuses on understanding how brains produce intelligence and developing methods to replicate this intelligence for practical problem-solving applications.
AINeutralarXiv โ CS AI ยท Mar 54/10
๐ง Researchers propose a standardized framework for classifying and evaluating memory capabilities in reinforcement learning agents, drawing from cognitive science concepts. The paper addresses confusion around memory terminology in RL and provides practical definitions for different memory types along with robust experimental methodologies.
AINeutralarXiv โ CS AI ยท Mar 34/103
๐ง Researchers propose that language models could help address longstanding challenges in cognitive science research, including integration, formalization, and conceptual clarity. The paper suggests AI tools should complement rather than replace human researchers to create more integrative and cumulative cognitive science.
AINeutralarXiv โ CS AI ยท Feb 274/105
๐ง Researchers propose using category theory to formalize knowledge domains and construct analogies between different fields. The paper demonstrates this approach using the classic analogy between the solar system and hydrogen atom, showing how mathematical structures like functors and pullbacks can define analogical relationships.
$ATOM
AINeutralarXiv โ CS AI ยท Mar 34/105
๐ง Researchers analyzed how Large Language Models access semantic memory using the Semantic Fluency Task, finding that LLMs exhibit similar memory foraging patterns to humans. The study reveals convergent and divergent search strategies in LLMs that mirror human cognitive behavior, potentially enabling better human-AI alignment or productive cognitive disalignment.