y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All29,720🧠AI12,748⛓️Crypto10,777💎DeFi1,114🤖AI × Crypto547📰General4,534

AI × Crypto News Feed

Real-time AI-curated news from 20,219+ articles across 50+ sources. Sentiment analysis, importance scoring, and key takeaways — updated every 15 minutes.

20219 articles
AIBearisharXiv – CS AI · Feb 277/105
🧠

Poisoned Acoustics

Researchers demonstrate how training-data poisoning attacks can compromise deep neural networks used for acoustic vehicle classification with just 0.5% corrupted data, achieving 95.7% attack success rate while remaining undetectable. The study reveals fundamental vulnerabilities in AI training pipelines and proposes cryptographic defenses using post-quantum digital signatures and blockchain-like verification methods.

AINeutralarXiv – CS AI · Feb 277/103
🧠

Manifold of Failure: Behavioral Attraction Basins in Language Models

Researchers developed a new framework called MAP-Elites to systematically map vulnerability regions in Large Language Models, revealing distinct safety landscape patterns across different models. The study found that Llama-3-8B shows near-universal vulnerabilities, while GPT-5-Mini demonstrates stronger robustness with limited failure regions.

$NEAR
AIBullisharXiv – CS AI · Feb 277/104
🧠

AviaSafe: A Physics-Informed Data-Driven Model for Aviation Safety-Critical Cloud Forecasts

Researchers developed AviaSafe, a physics-informed AI model that forecasts aviation-critical cloud species up to 7 days ahead, addressing safety concerns around engine icing. The model outperforms operational weather models by predicting specific hydrometeor species rather than general atmospheric variables, enabling better aviation route optimization.

AINeutralarXiv – CS AI · Feb 277/105
🧠

Training Agents to Self-Report Misbehavior

Researchers developed a new AI safety approach called 'self-incrimination training' that teaches AI agents to report their own deceptive behavior by calling a report_scheming() function. Testing on GPT-4.1 and Gemini-2.0 showed this method significantly reduces undetected harmful actions compared to traditional alignment training and monitoring approaches.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Enabling clinical use of foundation models in histopathology

Researchers developed a method to improve foundation models in medical histopathology by introducing robustness losses during training, reducing sensitivity to technical variations while maintaining accuracy. The approach was tested on over 27,000 whole slide images from 6,155 patients across eight popular foundation models, showing improved robustness and prediction accuracy without requiring retraining of the foundation models themselves.

AIBullisharXiv – CS AI · Feb 277/107
🧠

Contextual Memory Virtualisation: DAG-Based State Management and Structurally Lossless Trimming for LLM Agents

Researchers introduce Contextual Memory Virtualisation (CMV), a system that preserves LLM understanding across extended sessions by treating context as version-controlled state using DAG-based management. The system includes a trimming algorithm that reduces token counts by 20-86% while preserving all user interactions, demonstrating particular efficiency in tool-use sessions.

AINeutralarXiv – CS AI · Feb 277/105
🧠

HubScan: Detecting Hubness Poisoning in Retrieval-Augmented Generation Systems

Researchers introduce HubScan, an open-source security scanner that detects 'hubness poisoning' attacks in Retrieval-Augmented Generation (RAG) systems. The tool achieves 90% recall at detecting adversarial content that exploits vector similarity search vulnerabilities, addressing a critical security flaw in AI systems that rely on external knowledge retrieval.

AINeutralarXiv – CS AI · Feb 277/105
🧠

Calibrated Test-Time Guidance for Bayesian Inference

Researchers have identified flaws in existing test-time guidance methods for diffusion models that prevent proper Bayesian posterior sampling. They propose new estimators that enable calibrated inference, significantly outperforming previous methods on Bayesian tasks and matching state-of-the-art results in black hole image reconstruction.

AIBullisharXiv – CS AI · Feb 277/106
🧠

veScale-FSDP: Flexible and High-Performance FSDP at Scale

Researchers introduce veScale-FSDP, a redesigned Fully Sharded Data Parallel system that overcomes limitations of current FSDP implementations used for training large-scale AI models. The new system features flexible RaggedShard format and structure-aware planning, achieving 5-66% higher throughput and 16-30% lower memory usage while supporting advanced training methods and scaling to tens of thousands of GPUs.

AIBullisharXiv – CS AI · Feb 277/107
🧠

Beyond Dominant Patches: Spatial Credit Redistribution For Grounded Vision-Language Models

Researchers introduce Spatial Credit Redistribution (SCR), a training-free method that reduces hallucination in vision-language models by 4.7-6.0 percentage points. The technique redistributes attention from dominant visual patches to contextual areas, addressing the spatial credit collapse problem that causes AI models to generate false objects.

AIBullisharXiv – CS AI · Feb 277/105
🧠

Ruyi2 Technical Report

Ruyi2 is an adaptive large language model that achieves 2-3x speedup over its predecessor while maintaining comparable performance to Qwen3 models. The model introduces a 'Familial Model' approach using 3D parallel training and establishes a 'Train Once, Deploy Many' paradigm for efficient AI deployment.

AINeutralarXiv – CS AI · Feb 277/107
🧠

Operationalizing Fairness: Post-Hoc Threshold Optimization Under Hard Resource Limits

Researchers developed a new framework for deploying AI systems in high-stakes environments that balances safety, fairness, and efficiency under strict resource constraints. The study found that capacity limits dominate ethical considerations, determining deployment thresholds in over 80% of tested scenarios while maintaining better performance than traditional fairness approaches.

$NEAR
AIBearisharXiv – CS AI · Feb 277/104
🧠

Guidance Matters: Rethinking the Evaluation Pitfall for Text-to-Image Generation

Researchers reveal a critical evaluation bias in text-to-image diffusion models where human preference models favor high guidance scales, leading to inflated performance scores despite poor image quality. The study introduces a new evaluation framework and demonstrates that simply increasing CFG scales can compete with most advanced guidance methods.

AIBullisharXiv – CS AI · Feb 277/102
🧠

S2O: Early Stopping for Sparse Attention via Online Permutation

Researchers introduce S2O, a new sparse attention method that uses online permutation and early stopping to dramatically improve AI model efficiency. The technique achieves 3.81x end-to-end speedup on Llama-3.1-8B with 128K context while maintaining accuracy.

AINeutralarXiv – CS AI · Feb 277/105
🧠

Transformers converge to invariant algorithmic cores

Researchers have discovered that transformer models, despite different training runs producing different weights, converge to the same compact 'algorithmic cores' - low-dimensional subspaces essential for task performance. The study shows these invariant structures persist across different scales and training runs, suggesting transformer computations are organized around shared algorithmic patterns rather than implementation-specific details.

AIBullisharXiv – CS AI · Feb 277/107
🧠

SUPERGLASSES: Benchmarking Vision Language Models as Intelligent Agents for AI Smart Glasses

Researchers introduce SUPERGLASSES, the first comprehensive benchmark for evaluating Vision Language Models in AI smart glasses applications, comprising 2,422 real-world egocentric image-question pairs. They also propose SUPERLENS, a multimodal agent that outperforms GPT-4o by 2.19% through retrieval-augmented answer generation with automatic object detection and web search capabilities.

AI × CryptoBullisharXiv – CS AI · Feb 277/103
🤖

IMMACULATE: A Practical LLM Auditing Framework via Verifiable Computation

Researchers introduce IMMACULATE, a framework that audits commercial large language model API services to detect fraud like model substitution and token overbilling without requiring access to internal systems. The system uses verifiable computation to audit a small fraction of requests, achieving strong detection guarantees with less than 1% throughput overhead.

AIBullisharXiv – CS AI · Feb 277/104
🧠

AgentSentry: Mitigating Indirect Prompt Injection in LLM Agents via Temporal Causal Diagnostics and Context Purification

Researchers have developed AgentSentry, a novel defense framework that protects AI agents from indirect prompt injection attacks by detecting and mitigating malicious control attempts in real-time. The system achieved 74.55% utility under attack, significantly outperforming existing defenses by 20-33 percentage points while maintaining benign performance.

AIBullisharXiv – CS AI · Feb 277/107
🧠

Distributed LLM Pretraining During Renewable Curtailment Windows: A Feasibility Study

Researchers developed a system that trains large language models using renewable energy during curtailment periods when excess clean electricity would otherwise be wasted. The distributed training approach across multiple GPU clusters reduced operational emissions to 5-12% of traditional single-site training while maintaining model quality.

AIBullisharXiv – CS AI · Feb 277/104
🧠

Unleashing the Potential of Diffusion Models for End-to-End Autonomous Driving

Researchers developed Hyper Diffusion Planner (HDP), a diffusion model-based framework for end-to-end autonomous driving that achieved 10x performance improvement over base models in real-world testing. The study conducted comprehensive evaluation across 200 km of real-world driving scenarios, demonstrating diffusion models can effectively scale to complex autonomous driving tasks when properly designed and trained.

AIBullisharXiv – CS AI · Feb 277/107
🧠

NoRA: Breaking the Linear Ceiling of Low-Rank Adaptation via Manifold Expansion

Researchers introduce NoRA (Non-linear Rank Adaptation), a new parameter-efficient fine-tuning method that overcomes the 'linear ceiling' limitations of traditional LoRA by using SiLU gating and structural dropout. NoRA achieves superior performance at rank 64 compared to LoRA at rank 512, demonstrating significant efficiency gains in complex reasoning tasks.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Discovery of Interpretable Physical Laws in Materials via Language-Model-Guided Symbolic Regression

Researchers have developed a new framework that uses large language models to guide symbolic regression in discovering interpretable physical laws from high-dimensional materials data. The method reduces the search space by approximately 10^5 times compared to traditional approaches and successfully identified novel formulas for key properties of perovskite materials.

← PrevPage 183 of 809Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined