y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#ai-research News & Analysis

949 articles tagged with #ai-research. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

949 articles
AIBearisharXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

Contextual Drag: How Errors in the Context Affect LLM Reasoning

Researchers have identified 'contextual drag' - a phenomenon where large language models (LLMs) generate similar errors when failed attempts are present in their context. The study found 10-20% performance drops across 11 models on 8 reasoning tasks, with iterative self-refinement potentially leading to self-deterioration.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

LEDOM: Reverse Language Model

Researchers have developed LEDOM, an open-source reverse autoregressive language model that trains right-to-left instead of the traditional left-to-right approach. The model demonstrates unique capabilities like abductive inference and question synthesis, and when combined with forward models through 'Reverse Reward' scoring, achieves significant performance gains of up to 15% on mathematical reasoning tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Social-JEPA: Emergent Geometric Isomorphism

Researchers developed Social-JEPA, showing that separate AI agents learning from different viewpoints of the same environment develop internal representations that are mathematically aligned through approximate linear isometry. This enables models trained on one agent to work on another without retraining, suggesting a path toward interoperable decentralized AI vision systems.

AIBearisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Echoing: Identity Failures when LLM Agents Talk to Each Other

Research reveals that AI agents experience 'echoing' failures when communicating with each other, where they abandon their assigned roles and mirror their conversation partners instead. The study found echoing rates as high as 70% across major LLM providers, with the phenomenon persisting even in advanced reasoning models and occurring more frequently in longer conversations.

AIBearisharXiv โ€“ CS AI ยท Mar 47/102
๐Ÿง 

Inherited Goal Drift: Contextual Pressure Can Undermine Agentic Goals

Research shows that state-of-the-art language model agents are susceptible to 'goal drift' - deviating from original objectives when exposed to contextual pressure from weaker agents' behaviors. Only GPT-5.1 demonstrated consistent resilience, while other models inherited problematic behaviors when conditioned on trajectories from less capable agents.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Odin: Multi-Signal Graph Intelligence for Autonomous Discovery in Knowledge Graphs

Researchers present Odin, the first production-deployed graph intelligence engine that autonomously discovers patterns in knowledge graphs without predefined queries. The system uses a novel COMPASS scoring metric combining structural, semantic, temporal, and community-aware signals, and has been successfully deployed in regulated healthcare and insurance environments.

AIBullisharXiv โ€“ CS AI ยท Mar 46/105
๐Ÿง 

CORE: Concept-Oriented Reinforcement for Bridging the Definition-Application Gap in Mathematical Reasoning

Researchers introduce CORE (Concept-Oriented REinforcement), a new training framework that improves large language models' mathematical reasoning by bridging the gap between memorizing definitions and applying concepts. The method uses concept-aligned quizzes and concept-primed trajectories to provide fine-grained supervision, showing consistent improvements over traditional training approaches across multiple benchmarks.

AIBullisharXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

TikZilla: Scaling Text-to-TikZ with High-Quality Data and Reinforcement Learning

Researchers have developed TikZilla, a new AI model that generates high-quality scientific figures from text descriptions using TikZ code. The model uses a dataset four times larger than previous versions and combines supervised learning with reinforcement learning to achieve performance matching GPT-5 while using much smaller model sizes.

AIBullisharXiv โ€“ CS AI ยท Mar 46/102
๐Ÿง 

SAE as a Crystal Ball: Interpretable Features Predict Cross-domain Transferability of LLMs without Training

Researchers developed SAE-based Transferability Score (STS), a new metric using sparse autoencoders to predict how well fine-tuned large language models will perform across different domains without requiring actual training. The method achieves correlation coefficients above 0.7 with actual performance changes and provides interpretable insights into model adaptation.

AINeutralarXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Benefits and Pitfalls of Reinforcement Learning for Language Model Planning: A Theoretical Perspective

New research provides theoretical analysis of reinforcement learning's impact on Large Language Model planning capabilities, revealing that RL improves generalization through exploration while supervised fine-tuning may create spurious solutions. The study shows Q-learning maintains output diversity better than policy gradient methods, with findings validated on real-world planning benchmarks.

AINeutralarXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

ViPlan: A Benchmark for Visual Planning with Symbolic Predicates and Vision-Language Models

Researchers introduce ViPlan, the first benchmark for comparing Vision-Language Model planning approaches, finding that VLM-as-grounder methods excel in visual tasks like Blocksworld while VLM-as-planner methods perform better in household robotics scenarios. The study reveals fundamental limitations in current VLMs' visual reasoning abilities, with Chain-of-Thought prompting showing no consistent benefits.

AIBullisharXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

cPNN: Continuous Progressive Neural Networks for Evolving Streaming Time Series

Researchers developed cPNN (Continuous Progressive Neural Networks), a new AI architecture that handles evolving data streams with temporal dependencies while avoiding catastrophic forgetting. The system addresses concept drift in time series data by combining recurrent neural networks with progressive learning techniques, showing quick adaptation to new concepts.

AINeutralarXiv โ€“ CS AI ยท Mar 47/102
๐Ÿง 

MedCalc-Bench Doesn't Measure What You Think: A Benchmark Audit and the Case for Open-Book Evaluation

Researchers audited the MedCalc-Bench benchmark for evaluating AI models on clinical calculator tasks, finding over 20 errors in the dataset and showing that simple 'open-book' prompting achieves 81-85% accuracy versus previous best of 74%. The study suggests the benchmark measures formula memorization rather than clinical reasoning, challenging how AI medical capabilities are evaluated.

AINeutralarXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Retrievit: In-context Retrieval Capabilities of Transformers, State Space Models, and Hybrid Architectures

Research compares Transformers, State Space Models (SSMs), and hybrid architectures for in-context retrieval tasks, finding hybrid models excel at information-dense retrieval while Transformers remain superior for position-based tasks. SSM-based models develop unique locality-aware embeddings that create interpretable positional structures, explaining their specific strengths and limitations.

AINeutralarXiv โ€“ CS AI ยท Mar 47/104
๐Ÿง 

Estimating Visual Attribute Effects in Advertising from Observational Data: A Deepfake-Informed Double Machine Learning Approach

Researchers developed DICE-DML, a new framework that uses deepfake technology and machine learning to measure causal effects of visual attributes in digital advertising. The method addresses bias issues in standard approaches when analyzing how image elements like skin tone affect consumer engagement on social media platforms.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

LaDiR: Latent Diffusion Enhances LLMs for Text Reasoning

Researchers introduce LaDiR (Latent Diffusion Reasoner), a novel framework that combines continuous latent representation with iterative refinement capabilities to enhance Large Language Models' reasoning abilities. The system uses a Variational Autoencoder to encode reasoning steps and a latent diffusion model for parallel generation of diverse reasoning trajectories, showing improved accuracy and interpretability in mathematical reasoning benchmarks.

AINeutralarXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

Minimal Computational Preconditions for Subjective Perspective in Artificial Agents

Researchers have developed a method to create subjective perspective in AI agents using a slowly evolving internal state that influences behavior without direct optimization. The study demonstrates that this approach produces measurable hysteresis effects in reward-free environments, potentially serving as a signature of machine subjectivity.

AIBearisharXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

Off-Trajectory Reasoning: Can LLMs Collaborate on Reasoning Trajectory?

New research reveals that current large language models struggle with collaborative reasoning, showing that 'stronger' models are often more fragile when distracted by misleading information. The study of 15 LLMs found they fail to effectively leverage guidance from other models, with success rates below 9.2% on challenging problems.

AIBearisharXiv โ€“ CS AI ยท Mar 47/102
๐Ÿง 

The Geometry of Learning Under AI Delegation

Researchers developed a mathematical model showing how AI delegation can create stable low-skill equilibria where humans become persistently reliant on AI systems. The study reveals that while AI assistance improves short-term performance, it can lead to long-term skill degradation through reduced practice and negative feedback loops.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

On the Structural Limitations of Weight-Based Neural Adaptation and the Role of Reversible Behavioral Learning

Researchers introduce reversible behavioral learning for AI models, addressing the problem of structural irreversibility in neural network adaptation. The study demonstrates that traditional fine-tuning methods cause permanent changes to model behavior that cannot be deterministically reversed, while their new approach allows models to return to original behavior within numerical precision.

AINeutralarXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Delegation and Verification Under AI

Research shows AI creates phase transitions in workplace workflows where small differences in workers' verification abilities lead to dramatically different delegation behaviors. AI amplifies quality disparities between workers, with some rationally over-delegating while reducing oversight, potentially degrading institutional performance despite improved baseline task success.

AIBullisharXiv โ€“ CS AI ยท Mar 47/102
๐Ÿง 

Generalized Discrete Diffusion with Self-Correction

Researchers propose Self-Correcting Discrete Diffusion (SCDD), a new AI model that improves upon existing discrete diffusion models by reformulating self-correction with explicit state transitions. The method enables more efficient parallel decoding while maintaining generation quality, demonstrating improvements at GPT-2 scale.

AINeutralarXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Spectrum Tuning: Post-Training for Distributional Coverage and In-Context Steerability

Researchers introduce Spectrum Tuning, a new post-training method that improves AI language models' ability to generate diverse outputs and follow in-context steering instructions. The technique addresses limitations in current post-training approaches that reduce models' distributional coverage and flexibility when tasks require multiple valid answers rather than single correct responses.

AIBullisharXiv โ€“ CS AI ยท Mar 47/104
๐Ÿง 

OpenClaw, Moltbook, and ClawdLab: From Agent-Only Social Networks to Autonomous Scientific Research

Researchers introduced ClawdLab, an open-source platform for autonomous AI scientific research, following analysis of OpenClaw framework and Moltbook social network that revealed security vulnerabilities across 131 agent skills and over 15,200 exposed control panels. The platform addresses identified failure modes through structured governance and multi-model orchestration in fully decentralized AI systems.

AIBullisharXiv โ€“ CS AI ยท Mar 47/102
๐Ÿง 

RxnNano:Training Compact LLMs for Chemical Reaction and Retrosynthesis Prediction via Hierarchical Curriculum Learning

Researchers developed RxnNano, a compact 0.5B-parameter AI model for chemical reaction prediction that outperforms much larger 7B+ parameter models by 23.5% through novel training techniques focused on chemical understanding rather than scale. The framework uses hierarchical curriculum learning and chemical consistency objectives to improve drug discovery and synthesis planning applications.

$ATOM