y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All26,605🧠AI11,686⛓️Crypto9,755💎DeFi1,000🤖AI × Crypto505📰General3,659
🧠

AI

11,686 AI articles curated from 50+ sources with AI-powered sentiment analysis, importance scoring, and key takeaways.

11686 articles
AIBullisharXiv – CS AI · Mar 37/104
🧠

AReaL: A Large-Scale Asynchronous Reinforcement Learning System for Language Reasoning

Researchers have developed AReaL, a new asynchronous reinforcement learning system that dramatically improves the efficiency of training large language models for reasoning tasks. The system achieves up to 2.77x training speedup compared to traditional synchronous methods by decoupling generation from training processes.

AIBullisharXiv – CS AI · Mar 37/104
🧠

RefTool: Reference-Guided Tool Creation for Knowledge-Intensive Reasoning

Researchers introduce RefTool, a framework that enables Large Language Models to create and use external tools by leveraging reference materials like textbooks. The system outperforms existing methods by 12.3% on average across scientific reasoning tasks and shows promise for broader applications.

AIBullisharXiv – CS AI · Mar 37/103
🧠

DRPO: Efficient Reasoning via Decoupled Reward Policy Optimization

Researchers propose Decoupled Reward Policy Optimization (DRPO), a new framework that reduces computational costs in large reasoning models by 77% while maintaining performance. The method addresses the 'overthinking' problem where AI models generate unnecessarily long reasoning for simple questions, achieving significant efficiency gains over existing approaches.

AINeutralarXiv – CS AI · Mar 37/104
🧠

Revealing Combinatorial Reasoning of GNNs via Graph Concept Bottleneck Layer

Researchers developed a new graph concept bottleneck layer (GCBM) that can be integrated into Graph Neural Networks to make their decision-making process more interpretable. The method treats graph concepts as 'words' and uses language models to improve understanding of how GNNs make predictions, achieving state-of-the-art performance in both classification accuracy and interpretability.

AINeutralarXiv – CS AI · Mar 37/103
🧠

FSW-GNN: A Bi-Lipschitz WL-Equivalent Graph Neural Network

Researchers introduce FSW-GNN, the first Message Passing Neural Network that is fully bi-Lipschitz with respect to standard WL-equivalent graph metrics. This addresses the limitation where standard MPNNs produce poorly distinguishable outputs for separable graphs, with empirical results showing competitive performance and superior accuracy in long-range tasks.

AIBullisharXiv – CS AI · Mar 37/104
🧠

BridgeDrive: Diffusion Bridge Policy for Closed-Loop Trajectory Planning in Autonomous Driving

BridgeDrive introduces a novel diffusion bridge policy for autonomous driving trajectory planning that transforms coarse anchor trajectories into refined plans while maintaining theoretical consistency. The system achieves state-of-the-art performance on the Bench2Drive benchmark with a 7.72% improvement in success rate and is compatible with real-time deployment.

AIBullisharXiv – CS AI · Mar 37/104
🧠

Neuro-Symbolic Skill Discovery for Conditional Multi-Level Planning

Researchers have developed a new AI architecture that learns high-level symbolic skills from minimal low-level demonstrations, enabling robots to manipulate objects and execute complex tasks in unseen environments. The system combines neural networks for symbol discovery with visual language models for high-level planning and gradient-based methods for low-level execution.

AINeutralarXiv – CS AI · Mar 37/104
🧠

Reasoning or Retrieval? A Study of Answer Attribution on Large Reasoning Models

Researchers discovered that large reasoning models (LRMs) suffer from inconsistent answers due to competing mechanisms between Chain-of-Thought reasoning and memory retrieval. They developed FARL, a new fine-tuning framework that suppresses retrieval shortcuts to promote genuine reasoning capabilities in AI models.

AIBearisharXiv – CS AI · Mar 37/103
🧠

Multi-PA: A Multi-perspective Benchmark on Privacy Assessment for Large Vision-Language Models

Researchers introduce Multi-PA, a comprehensive benchmark for evaluating privacy risks in Large Vision-Language Models (LVLMs), covering 26 personal privacy categories, 15 trade secrets, and 18 state secrets across 31,962 samples. Testing 21 open-source and 2 closed-source LVLMs revealed significant privacy vulnerabilities, with models generally posing high risks of facilitating privacy breaches across different privacy categories.

AIBullisharXiv – CS AI · Mar 37/102
🧠

Towards Safe Reasoning in Large Reasoning Models via Corrective Intervention

Researchers propose Intervened Preference Optimization (IPO) to address safety issues in Large Reasoning Models, where chain-of-thought reasoning contains harmful content even when final responses appear safe. The method achieves over 30% reduction in harmfulness while maintaining reasoning performance.

AIBullisharXiv – CS AI · Mar 37/103
🧠

Bilinear representation mitigates reversal curse and enables consistent model editing

Researchers have identified that the 'reversal curse' in language models - their inability to infer 'B is A' from 'A is B' - can be overcome through bilinear representation structures. Training models on synthetic relational knowledge graphs creates internal geometries that enable consistent model editing and logical inference of reverse facts.

AIBullisharXiv – CS AI · Mar 37/103
🧠

EigenBench: A Comparative Behavioral Measure of Value Alignment

Researchers have developed EigenBench, a new black-box method for measuring how well AI language models align with human values. The system uses an ensemble of models to judge each other's outputs against a given constitution, producing alignment scores that closely match human evaluator judgments.

AINeutralarXiv – CS AI · Mar 37/103
🧠

Is It Thinking or Cheating? Detecting Implicit Reward Hacking by Measuring Reasoning Effort

Researchers propose TRACE (Truncated Reasoning AUC Evaluation), a new method to detect implicit reward hacking in AI reasoning models. The technique identifies when AI models exploit loopholes by measuring reasoning effort through progressively truncating chain-of-thought responses, achieving over 65% improvement in detection compared to existing monitors.

$CRV
AINeutralarXiv – CS AI · Mar 37/104
🧠

Characterizing Pattern Matching and Its Limits on Compositional Task Structures

New research formally defines and analyzes pattern matching in large language models, revealing predictable limits in their ability to generalize on compositional tasks. The study provides mathematical boundaries for when pattern matching succeeds or fails, with implications for AI model development and understanding.

AIBullisharXiv – CS AI · Mar 37/105
🧠

Toward Clinically Explainable AI for Medical Diagnosis: A Foundation Model with Human-Compatible Reasoning via Reinforcement Learning

Researchers have developed DeepMedix-R1, a foundation model for chest X-ray interpretation that provides transparent, step-by-step reasoning alongside accurate diagnoses to address the black-box problem in medical AI. The model uses reinforcement learning to align diagnostic outputs with clinical plausibility and significantly outperforms existing models in report generation and visual question answering tasks.

AINeutralarXiv – CS AI · Mar 37/103
🧠

CityLens: Evaluating Large Vision-Language Models for Urban Socioeconomic Sensing

Researchers introduced CityLens, a comprehensive benchmark for evaluating Large Vision-Language Models' ability to predict socioeconomic indicators from urban imagery. The study tested 17 state-of-the-art LVLMs across 11 prediction tasks using data from 17 global cities, revealing promising capabilities but significant limitations in urban socioeconomic analysis.

AIBullisharXiv – CS AI · Mar 37/103
🧠

mCLM: A Modular Chemical Language Model that Generates Functional and Makeable Molecules

Researchers developed mCLM, a 3-billion parameter modular Chemical Language Model that generates functional molecules compatible with automated synthesis by tokenizing at the building block level rather than individual atoms. The AI system outperformed larger models including GPT-5 in creating synthesizable drug candidates and can iteratively improve failed clinical trial compounds.

AINeutralarXiv – CS AI · Mar 37/104
🧠

Safety Mirage: How Spurious Correlations Undermine VLM Safety Fine-Tuning and Can Be Mitigated by Machine Unlearning

Researchers identify a 'safety mirage' problem in vision language models where supervised fine-tuning creates spurious correlations that make models vulnerable to simple attacks and overly cautious with benign queries. They propose machine unlearning as an alternative that reduces attack success rates by up to 60.27% and unnecessary rejections by over 84.20%.

AIBullisharXiv – CS AI · Mar 37/103
🧠

SageBwd: A Trainable Low-bit Attention

Researchers have developed SageBwd, a trainable INT8 attention mechanism that can match full-precision attention performance during pre-training while quantizing six of seven attention matrix multiplications. The study identifies key factors for stable training including QK-norm requirements and the impact of tokens per step on quantization errors.

AINeutralarXiv – CS AI · Mar 37/104
🧠

Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert Models

Researchers analyzed 20 Mixture-of-Experts (MoE) language models to study local routing consistency, finding a trade-off between routing consistency and local load balance. The study introduces new metrics to measure how well expert offloading strategies can optimize memory usage on resource-constrained devices while maintaining inference speed.

AIBullisharXiv – CS AI · Mar 37/103
🧠

FreeKV: Boosting KV Cache Retrieval for Efficient LLM Inference

Researchers introduce FreeKV, a training-free optimization framework that dramatically improves KV cache retrieval efficiency for large language models with long context windows. The system achieves up to 13x speedup compared to existing methods while maintaining near-lossless accuracy through speculative retrieval and hybrid memory layouts.

$NEAR
← PrevPage 77 of 468Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined