y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All29,744🧠AI12,749⛓️Crypto10,788💎DeFi1,114🤖AI × Crypto547📰General4,546

AI × Crypto News Feed

Real-time AI-curated news from 20,278+ articles across 50+ sources. Sentiment analysis, importance scoring, and key takeaways — updated every 15 minutes.

20278 articles
AIBullisharXiv – CS AI · Feb 277/106
🧠

Enabling clinical use of foundation models in histopathology

Researchers developed a method to improve foundation models in medical histopathology by introducing robustness losses during training, reducing sensitivity to technical variations while maintaining accuracy. The approach was tested on over 27,000 whole slide images from 6,155 patients across eight popular foundation models, showing improved robustness and prediction accuracy without requiring retraining of the foundation models themselves.

AIBullisharXiv – CS AI · Feb 277/107
🧠

Contextual Memory Virtualisation: DAG-Based State Management and Structurally Lossless Trimming for LLM Agents

Researchers introduce Contextual Memory Virtualisation (CMV), a system that preserves LLM understanding across extended sessions by treating context as version-controlled state using DAG-based management. The system includes a trimming algorithm that reduces token counts by 20-86% while preserving all user interactions, demonstrating particular efficiency in tool-use sessions.

AINeutralarXiv – CS AI · Feb 277/105
🧠

HubScan: Detecting Hubness Poisoning in Retrieval-Augmented Generation Systems

Researchers introduce HubScan, an open-source security scanner that detects 'hubness poisoning' attacks in Retrieval-Augmented Generation (RAG) systems. The tool achieves 90% recall at detecting adversarial content that exploits vector similarity search vulnerabilities, addressing a critical security flaw in AI systems that rely on external knowledge retrieval.

AINeutralarXiv – CS AI · Feb 277/105
🧠

Calibrated Test-Time Guidance for Bayesian Inference

Researchers have identified flaws in existing test-time guidance methods for diffusion models that prevent proper Bayesian posterior sampling. They propose new estimators that enable calibrated inference, significantly outperforming previous methods on Bayesian tasks and matching state-of-the-art results in black hole image reconstruction.

AIBullisharXiv – CS AI · Feb 277/106
🧠

veScale-FSDP: Flexible and High-Performance FSDP at Scale

Researchers introduce veScale-FSDP, a redesigned Fully Sharded Data Parallel system that overcomes limitations of current FSDP implementations used for training large-scale AI models. The new system features flexible RaggedShard format and structure-aware planning, achieving 5-66% higher throughput and 16-30% lower memory usage while supporting advanced training methods and scaling to tens of thousands of GPUs.

AIBullisharXiv – CS AI · Feb 277/107
🧠

Beyond Dominant Patches: Spatial Credit Redistribution For Grounded Vision-Language Models

Researchers introduce Spatial Credit Redistribution (SCR), a training-free method that reduces hallucination in vision-language models by 4.7-6.0 percentage points. The technique redistributes attention from dominant visual patches to contextual areas, addressing the spatial credit collapse problem that causes AI models to generate false objects.

AIBullisharXiv – CS AI · Feb 277/105
🧠

Ruyi2 Technical Report

Ruyi2 is an adaptive large language model that achieves 2-3x speedup over its predecessor while maintaining comparable performance to Qwen3 models. The model introduces a 'Familial Model' approach using 3D parallel training and establishes a 'Train Once, Deploy Many' paradigm for efficient AI deployment.

AINeutralarXiv – CS AI · Feb 277/107
🧠

Operationalizing Fairness: Post-Hoc Threshold Optimization Under Hard Resource Limits

Researchers developed a new framework for deploying AI systems in high-stakes environments that balances safety, fairness, and efficiency under strict resource constraints. The study found that capacity limits dominate ethical considerations, determining deployment thresholds in over 80% of tested scenarios while maintaining better performance than traditional fairness approaches.

$NEAR
AIBearisharXiv – CS AI · Feb 277/104
🧠

Guidance Matters: Rethinking the Evaluation Pitfall for Text-to-Image Generation

Researchers reveal a critical evaluation bias in text-to-image diffusion models where human preference models favor high guidance scales, leading to inflated performance scores despite poor image quality. The study introduces a new evaluation framework and demonstrates that simply increasing CFG scales can compete with most advanced guidance methods.

AIBullisharXiv – CS AI · Feb 277/102
🧠

S2O: Early Stopping for Sparse Attention via Online Permutation

Researchers introduce S2O, a new sparse attention method that uses online permutation and early stopping to dramatically improve AI model efficiency. The technique achieves 3.81x end-to-end speedup on Llama-3.1-8B with 128K context while maintaining accuracy.

AINeutralarXiv – CS AI · Feb 277/105
🧠

Transformers converge to invariant algorithmic cores

Researchers have discovered that transformer models, despite different training runs producing different weights, converge to the same compact 'algorithmic cores' - low-dimensional subspaces essential for task performance. The study shows these invariant structures persist across different scales and training runs, suggesting transformer computations are organized around shared algorithmic patterns rather than implementation-specific details.

AIBullisharXiv – CS AI · Feb 277/107
🧠

SUPERGLASSES: Benchmarking Vision Language Models as Intelligent Agents for AI Smart Glasses

Researchers introduce SUPERGLASSES, the first comprehensive benchmark for evaluating Vision Language Models in AI smart glasses applications, comprising 2,422 real-world egocentric image-question pairs. They also propose SUPERLENS, a multimodal agent that outperforms GPT-4o by 2.19% through retrieval-augmented answer generation with automatic object detection and web search capabilities.

AI × CryptoBullisharXiv – CS AI · Feb 277/103
🤖

IMMACULATE: A Practical LLM Auditing Framework via Verifiable Computation

Researchers introduce IMMACULATE, a framework that audits commercial large language model API services to detect fraud like model substitution and token overbilling without requiring access to internal systems. The system uses verifiable computation to audit a small fraction of requests, achieving strong detection guarantees with less than 1% throughput overhead.

AIBullisharXiv – CS AI · Feb 277/104
🧠

AgentSentry: Mitigating Indirect Prompt Injection in LLM Agents via Temporal Causal Diagnostics and Context Purification

Researchers have developed AgentSentry, a novel defense framework that protects AI agents from indirect prompt injection attacks by detecting and mitigating malicious control attempts in real-time. The system achieved 74.55% utility under attack, significantly outperforming existing defenses by 20-33 percentage points while maintaining benign performance.

AIBullisharXiv – CS AI · Feb 277/107
🧠

Distributed LLM Pretraining During Renewable Curtailment Windows: A Feasibility Study

Researchers developed a system that trains large language models using renewable energy during curtailment periods when excess clean electricity would otherwise be wasted. The distributed training approach across multiple GPU clusters reduced operational emissions to 5-12% of traditional single-site training while maintaining model quality.

AIBullisharXiv – CS AI · Feb 277/104
🧠

Unleashing the Potential of Diffusion Models for End-to-End Autonomous Driving

Researchers developed Hyper Diffusion Planner (HDP), a diffusion model-based framework for end-to-end autonomous driving that achieved 10x performance improvement over base models in real-world testing. The study conducted comprehensive evaluation across 200 km of real-world driving scenarios, demonstrating diffusion models can effectively scale to complex autonomous driving tasks when properly designed and trained.

AIBullisharXiv – CS AI · Feb 277/107
🧠

NoRA: Breaking the Linear Ceiling of Low-Rank Adaptation via Manifold Expansion

Researchers introduce NoRA (Non-linear Rank Adaptation), a new parameter-efficient fine-tuning method that overcomes the 'linear ceiling' limitations of traditional LoRA by using SiLU gating and structural dropout. NoRA achieves superior performance at rank 64 compared to LoRA at rank 512, demonstrating significant efficiency gains in complex reasoning tasks.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Discovery of Interpretable Physical Laws in Materials via Language-Model-Guided Symbolic Regression

Researchers have developed a new framework that uses large language models to guide symbolic regression in discovering interpretable physical laws from high-dimensional materials data. The method reduces the search space by approximately 10^5 times compared to traditional approaches and successfully identified novel formulas for key properties of perovskite materials.

AIBullisharXiv – CS AI · Feb 277/107
🧠

Residual Koopman Spectral Profiling for Predicting and Preventing Transformer Training Instability

Researchers developed Residual Koopman Spectral Profiling (RKSP), a method that predicts transformer training instability from a single forward pass at initialization with 99.5% accuracy. The technique includes Koopman Spectral Shaping (KSS) which can prevent training divergence and enable 50-150% higher learning rates across various AI models including GPT-2 and LLaMA-2.

$NEAR
AIBullisharXiv – CS AI · Feb 277/107
🧠

LLMServingSim 2.0: A Unified Simulator for Heterogeneous and Disaggregated LLM Serving Infrastructure

Researchers have released LLMServingSim 2.0, a unified simulator that models the complex interactions between heterogeneous hardware and disaggregated software in large language model serving infrastructures. The simulator achieves 0.97% average error compared to real deployments while maintaining 10-minute simulation times for complex configurations.

$NEAR
AIBullisharXiv – CS AI · Feb 277/106
🧠

Affine-Scaled Attention: Towards Flexible and Stable Transformer Attention

Researchers propose Affine-Scaled Attention, a new mechanism that improves Transformer model training stability by introducing flexible scaling and bias terms to attention weights. The approach shows consistent improvements in optimization behavior and downstream task performance compared to standard softmax attention across multiple language model sizes.

AINeutralarXiv – CS AI · Feb 277/106
🧠

Accelerated Online Risk-Averse Policy Evaluation in POMDPs with Theoretical Guarantees and Novel CVaR Bounds

Researchers developed a new theoretical framework for accelerated risk-averse policy evaluation in partially observable Markov decision processes (POMDPs) using Conditional Value-at-Risk (CVaR) bounds. The method enables safe elimination of suboptimal actions while maintaining computational guarantees, achieving substantial speedups in autonomous agent decision-making under uncertainty.

← PrevPage 184 of 812Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined