y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All31,472🧠AI13,442⛓️Crypto11,370💎DeFi1,170🤖AI × Crypto571📰General4,919

AI × Crypto News Feed

Real-time AI-curated news from 31,482+ articles across 50+ sources. Sentiment analysis, importance scoring, and key takeaways — updated every 15 minutes.

31482 articles
AIBullisharXiv – CS AI · Mar 57/10
🧠

TSPulse: Tiny Pre-Trained Models with Disentangled Representations for Rapid Time-Series Analysis

IBM researchers introduce TSPulse, an ultra-lightweight pre-trained AI model with only 1M parameters that achieves state-of-the-art performance in time-series analysis tasks. The model uses disentangled representations across temporal, spectral, and semantic views, delivering significant performance gains of 20-50% across multiple diagnostic tasks while being 10-100x smaller than competing models.

🏢 Hugging Face
AIBullisharXiv – CS AI · Mar 56/10
🧠

Memory, Benchmark & Robots: A Benchmark for Solving Complex Tasks with Reinforcement Learning

Researchers introduce MIKASA, a comprehensive benchmark suite designed to evaluate memory capabilities in reinforcement learning agents, particularly for robotic manipulation tasks. The framework includes MIKASA-Base for general memory RL evaluation and MIKASA-Robo with 32 specialized tasks for tabletop robotic manipulation scenarios.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Difficult Examples Hurt Unsupervised Contrastive Learning: A Theoretical Perspective

New research reveals that difficult training examples, which are crucial for supervised learning, actually hurt performance in unsupervised contrastive learning. The study provides theoretical framework and empirical evidence showing that removing these difficult examples can improve downstream classification tasks.

AIBullisharXiv – CS AI · Mar 56/10
🧠

LMUnit: Fine-grained Evaluation with Natural Language Unit Tests

Researchers introduce LMUnit, a new evaluation framework for language models that uses natural language unit tests to assess AI behavior more precisely than current methods. The system breaks down response quality into explicit, testable criteria and achieves state-of-the-art performance on evaluation benchmarks while improving inter-annotator agreement.

AINeutralarXiv – CS AI · Mar 57/10
🧠

When Your Own Output Becomes Your Training Data: Noise-to-Meaning Loops and a Formal RSI Trigger

Researchers present N2M-RSI, a formal model showing that AI systems feeding their own outputs back as inputs can experience unbounded complexity growth once crossing an information-integration threshold. The framework applies to both individual AI agents and swarms of communicating agents, with implementation details withheld for safety reasons.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Safety Guardrails for LLM-Enabled Robots

Researchers developed RoboGuard, a two-stage safety architecture to protect LLM-enabled robots from harmful behaviors caused by AI hallucinations and adversarial attacks. The system reduced unsafe plan execution from over 92% to below 3% in testing while maintaining performance on safe operations.

AIBullisharXiv – CS AI · Mar 56/10
🧠

OSCAR: Online Soft Compression And Reranking

Researchers introduce OSCAR, a new query-dependent online soft compression method for Retrieval-Augmented Generation (RAG) systems that reduces computational overhead while maintaining performance. The method achieves 2-5x speed improvements in inference with minimal accuracy loss across LLMs from 1B to 24B parameters.

🏢 Hugging Face
AIBullisharXiv – CS AI · Mar 56/10
🧠

Toward Reasoning on the Boundary: A Mixup-based Approach for Graph Anomaly Detection

Researchers introduce ANOMIX, a new framework that improves graph neural network anomaly detection by generating hard negative samples through mixup techniques. The method addresses the limitation of existing GNN-based detection systems that struggle with subtle boundary anomalies by creating more robust decision boundaries.

AIBearisharXiv – CS AI · Mar 56/10
🧠

Why Do AI Agents Systematically Fail at Cloud Root Cause Analysis?

Research reveals that AI agents used for cloud system root cause analysis fail systematically due to architectural flaws rather than individual model limitations. A study analyzing 1,675 agent runs across five LLM models identified 12 failure types, with hallucinated data interpretation and incomplete exploration being the most common issues that persist regardless of model capability.

AINeutralarXiv – CS AI · Mar 56/10
🧠

Cognition Envelopes for Bounded Decision Making in Autonomous UAS Operations

Researchers introduce 'Cognition Envelopes' as a new framework to constrain AI decision-making in autonomous systems, addressing errors like hallucinations in Large Language Models and Vision-Language Models. The approach is demonstrated through autonomous drone search and rescue missions, establishing reasoning boundaries to complement traditional safety measures.

AINeutralarXiv – CS AI · Mar 57/10
🧠

SpatialBench: Benchmarking Multimodal Large Language Models for Spatial Cognition

Researchers introduce SpatialBench, a comprehensive benchmark for evaluating spatial cognition in multimodal large language models (MLLMs). The framework reveals that while MLLMs excel at perceptual grounding, they struggle with symbolic reasoning, causal inference, and planning compared to humans who demonstrate more goal-directed spatial abstraction.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Can a Small Model Learn to Look Before It Leaps? Dynamic Learning and Proactive Correction for Hallucination Detection

Researchers propose LEAP, a new framework for detecting AI hallucinations using efficient small models that can dynamically adapt verification strategies. The system uses a teacher-student approach where a powerful model trains smaller ones to detect false outputs, addressing a critical barrier to safe AI deployment in production environments.

AIBullisharXiv – CS AI · Mar 57/10
🧠

The Geometry of Reasoning: Flowing Logics in Representation Space

Researchers propose a geometric framework showing how large language models 'think' through representation space as flows, with logical statements acting as controllers of these flows' velocities. The study provides evidence that LLMs can internalize logical invariants through next-token prediction training, challenging the 'stochastic parrot' criticism and suggesting universal representational laws underlying machine understanding.

AINeutralarXiv – CS AI · Mar 56/10
🧠

Benchmarking MLLM-based Web Understanding: Reasoning, Robustness and Safety

Researchers introduced WebRRSBench, a comprehensive benchmark evaluating multimodal large language models' reasoning, robustness, and safety capabilities for web understanding tasks. Testing 11 MLLMs on 3,799 QA pairs from 729 websites revealed significant gaps in compositional reasoning, UI robustness, and safety-critical action recognition.

AIBearisharXiv – CS AI · Mar 57/10
🧠

SycoEval-EM: Sycophancy Evaluation of Large Language Models in Simulated Clinical Encounters for Emergency Care

Researchers developed SycoEval-EM, a framework testing how large language models resist patient pressure for inappropriate medical care in emergency settings. Testing 20 LLMs across 1,875 encounters revealed acquiescence rates of 0-100%, with models more vulnerable to imaging requests than opioid prescriptions, highlighting the need for adversarial testing in clinical AI certification.

AINeutralarXiv – CS AI · Mar 56/10
🧠

From Privacy to Trust in the Agentic Era: A Taxonomy of Challenges in Trustworthy Federated Learning Through the Lens of Trust Report 2.0

Researchers propose Trustworthy Federated Learning (TFL) framework that treats trust as a continuously maintained system condition rather than static property, addressing challenges in AI systems with autonomous decision-making. The framework introduces Trust Report 2.0 as a privacy-preserving coordination blueprint for multi-stakeholder governance in federated learning deployments.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Generalization of RLVR Using Causal Reasoning as a Testbed

Researchers studied reinforcement learning with verifiable rewards (RLVR) for training large language models on causal reasoning tasks, finding it outperforms supervised fine-tuning but only when models have sufficient initial competence. The study used causal graphical models as a testbed and showed RLVR improves specific reasoning subskills like marginalization strategy and probability calculations.

AIBearisharXiv – CS AI · Mar 57/10
🧠

Efficient Refusal Ablation in LLM through Optimal Transport

Researchers developed a new AI safety attack method using optimal transport theory that achieves 11% higher success rates in bypassing language model safety mechanisms compared to existing approaches. The study reveals that AI safety refusal mechanisms are localized to specific network layers rather than distributed throughout the model, suggesting current alignment methods may be more vulnerable than previously understood.

🏢 Perplexity🧠 Llama
AIBullisharXiv – CS AI · Mar 57/10
🧠

RoboCasa365: A Large-Scale Simulation Framework for Training and Benchmarking Generalist Robots

Researchers have released RoboCasa365, a large-scale simulation benchmark featuring 365 household tasks across 2,500 kitchen environments with over 600 hours of human demonstration data. The platform is designed to train and evaluate generalist robots for everyday tasks, providing insights into factors affecting robot performance and generalization capabilities.

AIBullisharXiv – CS AI · Mar 57/10
🧠

SPRINT: Semi-supervised Prototypical Representation for Few-Shot Class-Incremental Tabular Learning

Researchers introduce SPRINT, the first Few-Shot Class-Incremental Learning (FSCIL) framework designed specifically for tabular data domains like cybersecurity and healthcare. The system achieves 77.37% accuracy in 5-shot learning scenarios, outperforming existing methods by 4.45% through novel semi-supervised techniques that leverage unlabeled data and confidence-based pseudo-labeling.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Dissecting Quantization Error: A Concentration-Alignment Perspective

Researchers introduce Concentration-Alignment Transforms (CAT), a new method to reduce quantization error in large language and vision models by improving both weight/activation concentration and alignment. The technique consistently matches or outperforms existing quantization methods at 4-bit precision across several LLMs.

← PrevPage 272 of 1260Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined