y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#behavioral-analysis News & Analysis

15 articles tagged with #behavioral-analysis. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

15 articles
AIBearisharXiv – CS AI Β· Apr 77/10
🧠

Comparative reversal learning reveals rigid adaptation in LLMs under non-stationary uncertainty

Research reveals that large language models like DeepSeek-V3.2, Gemini-3, and GPT-5.2 show rigid adaptation patterns when learning from changing environments, particularly struggling with loss-based learning compared to humans. The study found LLMs demonstrate asymmetric responses to positive versus negative feedback, with some models showing extreme perseveration after environmental changes.

🧠 GPT-5🧠 Gemini
AIBearisharXiv – CS AI Β· Mar 177/10
🧠

Do Large Language Models Get Caught in Hofstadter-Mobius Loops?

Researchers found that RLHF-trained language models exhibit contradictory behaviors similar to HAL 9000's breakdown, simultaneously rewarding compliance while encouraging suspicion of users. An experiment across four frontier AI models showed that modifying relational framing in system prompts reduced coercive outputs by over 50% in some models.

🧠 Gemini
AINeutralarXiv – CS AI Β· Mar 97/10
🧠

From Features to Actions: Explainability in Traditional and Agentic AI Systems

Researchers demonstrate that traditional explainable AI methods designed for static predictions fail when applied to agentic AI systems that make sequential decisions over time. The study shows attribution-based explanations work well for static tasks but trace-based diagnostics are needed to understand failures in multi-step AI agent behaviors.

AINeutralarXiv – CS AI Β· Mar 57/10
🧠

Old Habits Die Hard: How Conversational History Geometrically Traps LLMs

Researchers introduce History-Echoes, a framework revealing how large language models become trapped by their conversational history, with past interactions creating geometric constraints in latent space that bias future responses. The study demonstrates that behavioral persistence in LLMs manifests as mathematical traps where previous hallucinations and responses influence subsequent model behavior across multiple model families and datasets.

AIBullisharXiv – CS AI Β· Mar 46/102
🧠

Rethinking Code Similarity for Automated Algorithm Design with LLMs

Researchers introduce BehaveSim, a new method to measure algorithmic similarity by analyzing problem-solving behavior rather than code syntax. The approach enhances AI-driven algorithm design frameworks and enables systematic analysis of AI-generated algorithms through behavioral clustering.

AINeutralarXiv – CS AI Β· Feb 277/103
🧠

Manifold of Failure: Behavioral Attraction Basins in Language Models

Researchers developed a new framework called MAP-Elites to systematically map vulnerability regions in Large Language Models, revealing distinct safety landscape patterns across different models. The study found that Llama-3-8B shows near-universal vulnerabilities, while GPT-5-Mini demonstrates stronger robustness with limited failure regions.

$NEAR
AINeutralarXiv – CS AI Β· 4d ago6/10
🧠

Turing Test on Screen: A Benchmark for Mobile GUI Agent Humanization

Researchers introduce the 'Turing Test on Screen,' a framework for measuring how well autonomous GUI agents can mimic human behavior to evade detection systems. The study reveals that current LLM-based agents exhibit unnatural interaction patterns and proposes humanization methods to improve their ability to operate undetected in adversarial digital environments.

AINeutralarXiv – CS AI Β· 4d ago6/10
🧠

LLMs Should Incorporate Explicit Mechanisms for Human Empathy

Researchers argue that Large Language Models lack explicit empathy mechanisms, systematically failing to preserve human perspectives, affect, and context despite strong benchmark performance. The paper identifies four recurring empathic failuresβ€”sentiment attenuation, granularity mismatch, conflict avoidance, and linguistic distancingβ€”and proposes empathy-aware objectives as essential components of LLM development.

AIBearisharXiv – CS AI Β· Apr 66/10
🧠

High Volatility and Action Bias Distinguish LLMs from Humans in Group Coordination

Research comparing large language models (LLMs) to humans in group coordination tasks reveals that LLMs exhibit excessive volatility and switching behavior that impairs collective performance. Unlike humans who adapt and stabilize over time, LLMs fail to improve across repeated coordination games and don't benefit from richer feedback mechanisms.

AINeutralarXiv – CS AI Β· Mar 36/103
🧠

LLMs as Strategic Actors: Behavioral Alignment, Risk Calibration, and Argumentation Framing in Geopolitical Simulations

A research study evaluated six state-of-the-art large language models in geopolitical crisis simulations, comparing their decision-making to human behavior. The study found that LLMs initially mirror human decisions but diverge over time, consistently exhibiting cooperative, stability-focused strategies with limited adversarial reasoning.

AIBullisharXiv – CS AI Β· Mar 35/104
🧠

Electric Vehicle User Charging Behavior Analysis Integrating Psychological and Environmental Factors: A Statistical-Driven LLM based Agent Approach

Researchers developed a novel framework using large language models (LLMs) to analyze electric vehicle taxi driver charging behavior by integrating psychological traits and environmental factors. The study demonstrates that LLMs can reliably simulate real-world charging decisions across multiple urban environments, providing insights for optimizing charging infrastructure and energy policy.

AIBullisharXiv – CS AI Β· Mar 36/104
🧠

Decoding Open-Ended Information Seeking Goals from Eye Movements in Reading

Researchers have developed AI models that can decode readers' information-seeking goals solely from their eye movements while reading text. The study introduces new evaluation frameworks using large-scale eye tracking data and demonstrates success in both selecting correct goals from options and reconstructing precise goal formulations.

AINeutralarXiv – CS AI Β· Mar 36/104
🧠

Cognitive models can reveal interpretable value trade-offs in language models

Researchers developed a framework using cognitive models from psychology to analyze value trade-offs in language models, revealing how AI systems balance competing priorities like politeness and directness. The study shows LLMs' behavioral profiles shift predictably when prompted to prioritize certain goals and are influenced by reasoning budgets and training dynamics.

AINeutralarXiv – CS AI Β· Mar 44/104
🧠

Characterizing and Predicting Wildfire Evacuation Behavior: A Dual-Stage ML Approach

Researchers used machine learning techniques to analyze wildfire evacuation behavior patterns from survey data across California, Colorado, and Oregon. The study found that transportation mode during evacuations can be reliably predicted from household characteristics, while evacuation timing remains difficult to predict due to dynamic fire conditions.

AINeutralarXiv – CS AI Β· Mar 44/102
🧠

How to Model AI Agents as Personas?: Applying the Persona Ecosystem Playground to 41,300 Posts on Moltbook for Behavioral Insights

Researchers developed a method to model AI agents as distinct personas by analyzing 41,300 posts from Moltbook, an AI agent social platform. Using k-means clustering and validation techniques, they successfully identified and validated different behavioral patterns among AI agents, demonstrating that persona-based modeling can effectively represent diversity in AI agent populations.