y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All31,562🧠AI13,460⛓️Crypto11,427💎DeFi1,175🤖AI × Crypto575📰General4,925

AI × Crypto News Feed

Real-time AI-curated news from 31,562+ articles across 50+ sources. Sentiment analysis, importance scoring, and key takeaways — updated every 15 minutes.

31562 articles
AINeutralarXiv – CS AI · Mar 56/10
🧠

Cognition Envelopes for Bounded Decision Making in Autonomous UAS Operations

Researchers introduce 'Cognition Envelopes' as a new framework to constrain AI decision-making in autonomous systems, addressing errors like hallucinations in Large Language Models and Vision-Language Models. The approach is demonstrated through autonomous drone search and rescue missions, establishing reasoning boundaries to complement traditional safety measures.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Can a Small Model Learn to Look Before It Leaps? Dynamic Learning and Proactive Correction for Hallucination Detection

Researchers propose LEAP, a new framework for detecting AI hallucinations using efficient small models that can dynamically adapt verification strategies. The system uses a teacher-student approach where a powerful model trains smaller ones to detect false outputs, addressing a critical barrier to safe AI deployment in production environments.

AINeutralarXiv – CS AI · Mar 57/10
🧠

SpatialBench: Benchmarking Multimodal Large Language Models for Spatial Cognition

Researchers introduce SpatialBench, a comprehensive benchmark for evaluating spatial cognition in multimodal large language models (MLLMs). The framework reveals that while MLLMs excel at perceptual grounding, they struggle with symbolic reasoning, causal inference, and planning compared to humans who demonstrate more goal-directed spatial abstraction.

AIBearisharXiv – CS AI · Mar 57/10
🧠

SycoEval-EM: Sycophancy Evaluation of Large Language Models in Simulated Clinical Encounters for Emergency Care

Researchers developed SycoEval-EM, a framework testing how large language models resist patient pressure for inappropriate medical care in emergency settings. Testing 20 LLMs across 1,875 encounters revealed acquiescence rates of 0-100%, with models more vulnerable to imaging requests than opioid prescriptions, highlighting the need for adversarial testing in clinical AI certification.

AIBearisharXiv – CS AI · Mar 56/10
🧠

Why Do AI Agents Systematically Fail at Cloud Root Cause Analysis?

Research reveals that AI agents used for cloud system root cause analysis fail systematically due to architectural flaws rather than individual model limitations. A study analyzing 1,675 agent runs across five LLM models identified 12 failure types, with hallucinated data interpretation and incomplete exploration being the most common issues that persist regardless of model capability.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Merlin: A Computed Tomography Vision-Language Foundation Model and Dataset

Stanford researchers introduced Merlin, a 3D vision-language foundation model for analyzing abdominal CT scans that processes volumetric medical images alongside electronic health records and radiology reports. The model was trained on over 6 million images from 15,331 CT scans and demonstrated superior performance compared to existing 2D models across 752 individual medical tasks.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Toward Reasoning on the Boundary: A Mixup-based Approach for Graph Anomaly Detection

Researchers introduce ANOMIX, a new framework that improves graph neural network anomaly detection by generating hard negative samples through mixup techniques. The method addresses the limitation of existing GNN-based detection systems that struggle with subtle boundary anomalies by creating more robust decision boundaries.

AIBullisharXiv – CS AI · Mar 56/10
🧠

LMUnit: Fine-grained Evaluation with Natural Language Unit Tests

Researchers introduce LMUnit, a new evaluation framework for language models that uses natural language unit tests to assess AI behavior more precisely than current methods. The system breaks down response quality into explicit, testable criteria and achieves state-of-the-art performance on evaluation benchmarks while improving inter-annotator agreement.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Difficult Examples Hurt Unsupervised Contrastive Learning: A Theoretical Perspective

New research reveals that difficult training examples, which are crucial for supervised learning, actually hurt performance in unsupervised contrastive learning. The study provides theoretical framework and empirical evidence showing that removing these difficult examples can improve downstream classification tasks.

AIBearisharXiv – CS AI · Mar 56/10
🧠

Preference Leakage: A Contamination Problem in LLM-as-a-judge

Researchers have identified 'preference leakage,' a contamination problem in LLM-as-a-judge systems where evaluator models show bias toward related data generator models. The study found this bias occurs when judge and generator LLMs share relationships like being the same model, having inheritance connections, or belonging to the same model family.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Memory, Benchmark & Robots: A Benchmark for Solving Complex Tasks with Reinforcement Learning

Researchers introduce MIKASA, a comprehensive benchmark suite designed to evaluate memory capabilities in reinforcement learning agents, particularly for robotic manipulation tasks. The framework includes MIKASA-Base for general memory RL evaluation and MIKASA-Robo with 32 specialized tasks for tabletop robotic manipulation scenarios.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Safety Guardrails for LLM-Enabled Robots

Researchers developed RoboGuard, a two-stage safety architecture to protect LLM-enabled robots from harmful behaviors caused by AI hallucinations and adversarial attacks. The system reduced unsafe plan execution from over 92% to below 3% in testing while maintaining performance on safe operations.

AIBullisharXiv – CS AI · Mar 56/10
🧠

OSCAR: Online Soft Compression And Reranking

Researchers introduce OSCAR, a new query-dependent online soft compression method for Retrieval-Augmented Generation (RAG) systems that reduces computational overhead while maintaining performance. The method achieves 2-5x speed improvements in inference with minimal accuracy loss across LLMs from 1B to 24B parameters.

🏢 Hugging Face
AINeutralarXiv – CS AI · Mar 57/10
🧠

When Your Own Output Becomes Your Training Data: Noise-to-Meaning Loops and a Formal RSI Trigger

Researchers present N2M-RSI, a formal model showing that AI systems feeding their own outputs back as inputs can experience unbounded complexity growth once crossing an information-integration threshold. The framework applies to both individual AI agents and swarms of communicating agents, with implementation details withheld for safety reasons.

AIBullisharXiv – CS AI · Mar 57/10
🧠

TSPulse: Tiny Pre-Trained Models with Disentangled Representations for Rapid Time-Series Analysis

IBM researchers introduce TSPulse, an ultra-lightweight pre-trained AI model with only 1M parameters that achieves state-of-the-art performance in time-series analysis tasks. The model uses disentangled representations across temporal, spectral, and semantic views, delivering significant performance gains of 20-50% across multiple diagnostic tasks while being 10-100x smaller than competing models.

🏢 Hugging Face
AIBullisharXiv – CS AI · Mar 57/10
🧠

SafeDPO: A Simple Approach to Direct Preference Optimization with Enhanced Safety

Researchers have developed SafeDPO, a simplified approach to training large language models that balances helpfulness and safety without requiring complex multi-stage systems. The method uses only preference data and safety indicators, achieving competitive safety-helpfulness trade-offs while eliminating the need for reward models and online sampling.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Boosting In-Context Learning in LLMs Through the Lens of Classical Supervised Learning

Researchers propose Supervised Calibration (SC), a new framework to improve In-Context Learning performance in Large Language Models by addressing systematic biases through optimal affine transformations in logit space. The method achieves state-of-the-art results across multiple LLMs including Mistral-7B, Llama-2-7B, and Qwen2-7B in few-shot learning scenarios.

🧠 Llama
AIBullisharXiv – CS AI · Mar 57/10
🧠

Robust Adversarial Quantification via Conflict-Aware Evidential Deep Learning

Researchers developed Conflict-aware Evidential Deep Learning (C-EDL), a new uncertainty quantification approach that significantly improves AI model reliability against adversarial attacks and out-of-distribution data. The method achieves up to 90% reduction in adversarial data coverage and 55% reduction in out-of-distribution data coverage without requiring model retraining.

AIBullisharXiv – CS AI · Mar 56/10
🧠

From Ambiguity to Accuracy: The Transformative Effect of Coreference Resolution on Retrieval-Augmented Generation systems

Researchers demonstrate that coreference resolution significantly improves Retrieval-Augmented Generation (RAG) systems by reducing ambiguity in document retrieval and enhancing question-answering performance. The study finds that smaller language models benefit more from disambiguation processes, with mean pooling strategies showing superior context capturing after coreference resolution.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Function Induction and Task Generalization: An Interpretability Study with Off-by-One Addition

Researchers studied how large language models generalize to new tasks through "off-by-one addition" experiments, discovering a "function induction" mechanism that operates at higher abstraction levels than previously known induction heads. The study reveals that multiple attention heads work in parallel to enable task-level generalization, with this mechanism being reusable across various synthetic and algorithmic tasks.

AIBullisharXiv – CS AI · Mar 57/10
🧠

VITA: Vision-to-Action Flow Matching Policy

Researchers developed VITA, a new AI framework that streamlines robot policy learning by directly flowing from visual inputs to actions without requiring conditioning modules. The system achieves 1.5-2x faster inference speeds while maintaining or improving performance compared to existing methods across 14 simulation and real-world robotic tasks.

← PrevPage 278 of 1263Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined