y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All31,634🧠AI13,484⛓️Crypto11,461💎DeFi1,179🤖AI × Crypto583📰General4,927

AI × Crypto News Feed

Real-time AI-curated news from 31,634+ articles across 50+ sources. Sentiment analysis, importance scoring, and key takeaways — updated every 15 minutes.

31634 articles
AIBearisharXiv – CS AI · Mar 47/104
🧠

Quantifying Frontier LLM Capabilities for Container Sandbox Escape

Researchers introduced SANDBOXESCAPEBENCH, a new benchmark that measures large language models' ability to break out of Docker container sandboxes commonly used for AI safety. The study found that LLMs can successfully identify and exploit vulnerabilities in sandbox environments, highlighting significant security risks as AI agents become more autonomous.

AINeutralarXiv – CS AI · Mar 46/102
🧠

The Malignant Tail: Spectral Segregation of Label Noise in Over-Parameterized Networks

Researchers identify the 'Malignant Tail' phenomenon where over-parameterized neural networks segregate signal from noise during training, leading to harmful overfitting. They demonstrate that Stochastic Gradient Descent pushes label noise into high-frequency orthogonal subspaces while preserving semantic features in low-rank subspaces, and propose Explicit Spectral Truncation as a post-hoc solution to recover optimal generalization.

AIBearisharXiv – CS AI · Mar 47/103
🧠

ZeroDayBench: Evaluating LLM Agents on Unseen Zero-Day Vulnerabilities for Cyberdefense

Researchers introduced ZeroDayBench, a new benchmark testing LLM agents' ability to find and patch 22 critical vulnerabilities in open-source code. Testing on frontier models GPT-5.2, Claude Sonnet 4.5, and Grok 4.1 revealed that current LLMs cannot yet autonomously solve cybersecurity tasks, highlighting limitations in AI-powered code security.

AIBullisharXiv – CS AI · Mar 46/103
🧠

Preconditioned Score and Flow Matching

Researchers propose a new preconditioning method for flow matching and score-based diffusion models that improves training optimization by reshaping the geometry of intermediate distributions. The technique addresses optimization bias caused by ill-conditioned covariance matrices, preventing training from stagnating at suboptimal weights and enabling better model performance.

AIBullisharXiv – CS AI · Mar 46/102
🧠

RIVA: Leveraging LLM Agents for Reliable Configuration Drift Detection

Researchers introduce RIVA, a multi-agent AI system that uses specialized verification agents and cross-validation to detect infrastructure configuration drift more reliably. The system improves accuracy from 27.3% to 50% when dealing with erroneous tool responses, addressing a critical reliability issue in cloud infrastructure management.

AIBullisharXiv – CS AI · Mar 46/104
🧠

Large Electron Model: A Universal Ground State Predictor

Researchers introduce Large Electron Model, a neural network that uses Fermi Sets architecture to predict ground state wavefunctions of interacting electrons across different Hamiltonian parameters. The model demonstrates accurate predictions for up to 50 particles and generalizes across unseen coupling strengths, potentially advancing material discovery beyond density functional theory limitations.

AIBullisharXiv – CS AI · Mar 46/102
🧠

PlayWrite: A Multimodal System for AI Supported Narrative Co-Authoring Through Play in XR

PlayWrite is a new mixed-reality AI system that allows users to create stories by directly manipulating virtual characters and props in XR, rather than through traditional text prompts. The system uses multi-agent AI to interpret user actions into structured narrative elements and generates final stories via large language models, demonstrating a novel approach to AI-human creative collaboration.

AIBullisharXiv – CS AI · Mar 46/102
🧠

Rigidity-Aware Geometric Pretraining for Protein Design and Conformational Ensembles

Researchers introduce RigidSSL, a new geometric pretraining framework for protein design that improves designability by up to 43% and enhances success rates in protein generation tasks. The two-phase approach combines geometric learning from 432K protein structures with molecular dynamics refinement to better capture protein conformational dynamics.

AIBullisharXiv – CS AI · Mar 47/103
🧠

MIRAGE: Knowledge Graph-Guided Cross-Cohort MRI Synthesis for Alzheimer's Disease Prediction

Researchers introduce MIRAGE, a novel AI framework that uses knowledge graphs and electronic health records to predict Alzheimer's disease when MRI scans are unavailable. The system improves AD classification rates by 13% compared to single-modality approaches by creating synthetic representations without expensive 3D brain scan reconstruction.

AIBullisharXiv – CS AI · Mar 47/103
🧠

Can Computational Reducibility Lead to Transferable Models for Graph Combinatorial Optimization?

Researchers developed a new neural solver model using GCON modules and energy-based loss functions that achieves state-of-the-art performance across multiple graph combinatorial optimization tasks. The study demonstrates effective transfer learning between related optimization problems through computational reducibility-informed pretraining strategies, representing progress toward foundational AI models for combinatorial optimization.

AIBullisharXiv – CS AI · Mar 47/103
🧠

Learning Object-Centric Spatial Reasoning for Sequential Manipulation in Cluttered Environments

Researchers developed Unveiler, a robotic manipulation framework that uses object-centric spatial reasoning to retrieve items from cluttered environments. The system achieves up to 97.6% success in simulation by separating high-level spatial reasoning from low-level action execution, and demonstrates zero-shot transfer to real-world scenarios.

AINeutralarXiv – CS AI · Mar 46/105
🧠

Human-Certified Module Repositories for the AI Age

Researchers propose Human-Certified Module Repositories (HCMRs) as a new framework to ensure trustworthy software development in the AI era. The system combines human oversight with automated analysis to certify and curate reusable code modules, addressing growing security concerns as AI increasingly generates and assembles software components.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Bridging Diffusion Guidance and Anderson Acceleration via Hopfield Dynamics

Researchers have developed Geometry Aware Attention Guidance (GAG), a new method that improves diffusion model generation quality by optimizing attention-space extrapolation. The approach models attention dynamics as fixed-point iterations within Modern Hopfield Networks and applies Anderson Acceleration to stabilize the process while reducing computational costs.

AIBullisharXiv – CS AI · Mar 47/104
🧠

CoDAR: Continuous Diffusion Language Models are More Powerful Than You Think

Researchers propose CoDAR, a new continuous diffusion language model framework that addresses key bottlenecks in token rounding through a two-stage approach combining continuous diffusion with an autoregressive decoder. The model demonstrates substantial improvements in generation quality over existing latent diffusion methods and becomes competitive with discrete diffusion language models.

AIBullisharXiv – CS AI · Mar 46/103
🧠

Through the Lens of Contrast: Self-Improving Visual Reasoning in VLMs

Researchers introduce VC-STaR, a new framework that improves visual reasoning in vision-language models by using contrastive image pairs to reduce hallucinations. The approach creates VisCoR-55K, a new dataset that outperforms existing visual reasoning methods when used for model fine-tuning.

AIBullisharXiv – CS AI · Mar 47/103
🧠

CAPT: Confusion-Aware Prompt Tuning for Reducing Vision-Language Misalignment

Researchers propose CAPT, a Confusion-Aware Prompt Tuning framework that addresses systematic misclassifications in vision-language models like CLIP by learning from the model's own confusion patterns. The method uses a Confusion Bank to model persistent category misalignments and introduces specialized modules to capture both semantic and sample-level confusion cues.

AINeutralarXiv – CS AI · Mar 46/102
🧠

How Controllable Are Large Language Models? A Unified Evaluation across Behavioral Granularities

Researchers introduce SteerEval, a new benchmark for evaluating how controllable Large Language Models are across language features, sentiment, and personality domains. The study reveals that current steering methods often fail at finer-grained control levels, highlighting significant risks when deploying LLMs in socially sensitive applications.

AIBullisharXiv – CS AI · Mar 46/102
🧠

GPUTOK: GPU Accelerated Byte Level BPE Tokenization

Researchers developed GPUTOK, a GPU-accelerated tokenizer for large language models that processes text significantly faster than existing CPU-based solutions. The optimized version shows 1.7x speed improvement over tiktoken and 7.6x over HuggingFace's GPT-2 tokenizer while maintaining output quality.

AIBullisharXiv – CS AI · Mar 46/103
🧠

Detecting Structural Heart Disease from Electrocardiograms via a Generalized Additive Model of Interpretable Foundation-Model Predictors

Researchers developed an interpretable AI framework for detecting structural heart disease from electrocardiograms, achieving better performance than existing deep-learning methods while providing clinical transparency. The model demonstrated improvements of nearly 1% across key metrics using the EchoNext benchmark of over 80,000 ECG-ECHO pairs.

AIBullisharXiv – CS AI · Mar 46/103
🧠

Robust Heterogeneous Analog-Digital Computing for Mixture-of-Experts Models with Theoretical Generalization Guarantees

Researchers propose a heterogeneous computing framework for Mixture-of-Experts AI models that combines analog in-memory computing with digital processing to improve energy efficiency. The approach identifies noise-sensitive experts for digital computation while running the majority on analog hardware, eliminating the need for costly retraining of large models.

AINeutralarXiv – CS AI · Mar 47/102
🧠

Credibility Governance: A Social Mechanism for Collective Self-Correction under Weak Truth Signals

Researchers propose Credibility Governance (CG), a new mechanism that improves collective decision-making on online platforms by dynamically scoring agent and opinion credibility based on alignment with emerging evidence. Testing in simulated environments shows CG outperforms traditional voting and stake-weighted systems, offering better resistance to misinformation and manipulation.

← PrevPage 291 of 1266Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined