y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All31,632🧠AI13,482⛓️Crypto11,461💎DeFi1,179🤖AI × Crypto583📰General4,927

AI × Crypto News Feed

Real-time AI-curated news from 31,634+ articles across 50+ sources. Sentiment analysis, importance scoring, and key takeaways — updated every 15 minutes.

31634 articles
AIBullisharXiv – CS AI · Mar 46/102
🧠

Predicting Tuberculosis from Real-World Cough Audio Recordings and Metadata

Researchers developed an AI system that can detect tuberculosis from cough recordings with 70% accuracy using audio alone, improving to 81% when combined with clinical metadata. The study used real-world data from a phone-based app across Africa and Asia, suggesting mobile applications could enhance TB diagnosis in community health settings.

$CRV
AIBullisharXiv – CS AI · Mar 47/103
🧠

Param$\Delta$ for Direct Weight Mixing: Post-Train Large Language Model at Zero Cost

Researchers introduce Param∆, a novel method for transferring post-training capabilities to updated language models without additional training costs. The technique achieves 95% performance of traditional post-training by computing weight differences between base and post-trained models, offering significant cost savings for AI model development.

AIBullisharXiv – CS AI · Mar 47/102
🧠

RxnNano:Training Compact LLMs for Chemical Reaction and Retrosynthesis Prediction via Hierarchical Curriculum Learning

Researchers developed RxnNano, a compact 0.5B-parameter AI model for chemical reaction prediction that outperforms much larger 7B+ parameter models by 23.5% through novel training techniques focused on chemical understanding rather than scale. The framework uses hierarchical curriculum learning and chemical consistency objectives to improve drug discovery and synthesis planning applications.

$ATOM
AIBullisharXiv – CS AI · Mar 47/103
🧠

ATPO: Adaptive Tree Policy Optimization for Multi-Turn Medical Dialogue

Researchers developed ATPO (Adaptive Tree Policy Optimization), a new AI algorithm for multi-turn medical dialogues that outperforms existing methods by better handling uncertainty in patient-doctor interactions. The algorithm enabled a smaller Qwen3-8B model to surpass GPT-4o's accuracy by 0.92% on medical dialogue benchmarks through improved value estimation and exploration strategies.

AIBullisharXiv – CS AI · Mar 46/102
🧠

Is Retraining-Free Enough? The Necessity of Router Calibration for Efficient MoE Compression

Researchers propose Router Knowledge Distillation (Router KD) to improve retraining-free compression of Mixture-of-Experts (MoE) models by calibrating routers while keeping expert parameters unchanged. The method addresses router-expert mismatch issues that cause performance degradation in compressed MoE models, showing particularly strong results in fine-grained MoE architectures.

AIBullisharXiv – CS AI · Mar 47/103
🧠

Self-Play Only Evolves When Self-Synthetic Pipeline Ensures Learnable Information Gain

Researchers propose a framework for sustainable AI self-evolution through triadic roles (Proposer, Solver, Verifier) that ensures learnable information gain across iterations. The study identifies three key system designs to prevent the common plateau effect in self-play AI systems: asymmetric co-evolution, capacity growth, and proactive information seeking.

AIBullisharXiv – CS AI · Mar 47/102
🧠

NExT-Guard: Training-Free Streaming Safeguard without Token-Level Labels

Researchers introduce NExT-Guard, a training-free framework for real-time AI safety monitoring that uses Sparse Autoencoders to detect unsafe content in streaming language models. The system outperforms traditional supervised training methods while requiring no token-level annotations, making it more cost-effective and scalable for deployment.

AINeutralarXiv – CS AI · Mar 47/103
🧠

Forecasting as Rendering: A 2D Gaussian Splatting Framework for Time Series Forecasting

Researchers introduce TimeGS, a novel time series forecasting framework that reimagines prediction as 2D generative rendering using Gaussian splatting techniques. The approach addresses key limitations in existing methods by treating future sequences as continuous latent surfaces and enforcing temporal continuity across periodic boundaries.

AIBullisharXiv – CS AI · Mar 46/103
🧠

MedFeat: Model-Aware and Explainability-Driven Feature Engineering with LLMs for Clinical Tabular Prediction

Researchers introduce MedFeat, a new AI framework that uses Large Language Models for healthcare feature engineering in clinical tabular predictions. The system incorporates model awareness and domain knowledge to discover clinically meaningful features that outperform traditional approaches and demonstrate robustness across different hospital settings.

AINeutralarXiv – CS AI · Mar 47/102
🧠

MedCalc-Bench Doesn't Measure What You Think: A Benchmark Audit and the Case for Open-Book Evaluation

Researchers audited the MedCalc-Bench benchmark for evaluating AI models on clinical calculator tasks, finding over 20 errors in the dataset and showing that simple 'open-book' prompting achieves 81-85% accuracy versus previous best of 74%. The study suggests the benchmark measures formula memorization rather than clinical reasoning, challenging how AI medical capabilities are evaluated.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Neural Paging: Learning Context Management Policies for Turing-Complete Agents

Researchers introduce Neural Paging, a new architecture that addresses the computational bottleneck of finite context windows in Large Language Models by implementing a hierarchical system that decouples reasoning from memory management. The approach reduces computational complexity from O(N²) to O(N·K²) for long-horizon reasoning tasks, potentially enabling more efficient AI agents.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Generalized Discrete Diffusion with Self-Correction

Researchers propose Self-Correcting Discrete Diffusion (SCDD), a new AI model that improves upon existing discrete diffusion models by reformulating self-correction with explicit state transitions. The method enables more efficient parallel decoding while maintaining generation quality, demonstrating improvements at GPT-2 scale.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Physics-Informed Neural Networks with Architectural Physics Embedding for Large-Scale Wave Field Reconstruction

Researchers developed Physics-Embedded PINNs (PE-PINN) that achieve 10x faster convergence than standard physics-informed neural networks and orders of magnitude memory reduction compared to traditional methods for large-scale wave field reconstruction. The breakthrough enables high-fidelity electromagnetic wave modeling for wireless communications, sensing, and room acoustics applications.

AINeutralarXiv – CS AI · Mar 47/103
🧠

Structured vs. Unstructured Pruning: An Exponential Gap

Research reveals an exponential gap between structured and unstructured neural network pruning methods. While unstructured weight pruning can approximate target functions with O(d log(1/ε)) neurons, structured neuron pruning requires Ω(d/ε) neurons, demonstrating fundamental limitations of structured approaches.

AIBullisharXiv – CS AI · Mar 46/104
🧠

Talking with Verifiers: Automatic Specification Generation for Neural Network Verification

Researchers have developed a framework that allows neural network verification tools to accept natural language specifications instead of low-level technical constraints. The system automatically translates human-readable requirements into formal verification queries, significantly expanding the practical applicability of neural network verification across diverse domains.

AINeutralarXiv – CS AI · Mar 46/104
🧠

CUDABench: Benchmarking LLMs for Text-to-CUDA Generation

Researchers introduce CUDABench, a comprehensive benchmark for evaluating Large Language Models' ability to generate CUDA code from text descriptions. The benchmark reveals significant challenges including high compilation success rates but low functional correctness, lack of domain-specific knowledge, and poor GPU hardware utilization.

AIBullisharXiv – CS AI · Mar 46/103
🧠

Concept Heterogeneity-aware Representation Steering

Researchers introduce CHaRS (Concept Heterogeneity-aware Representation Steering), a new method for controlling large language model behavior that uses optimal transport theory to create context-dependent steering rather than global directions. The approach models representations as Gaussian mixture models and derives input-dependent steering maps, showing improved behavioral control over existing methods.

AIBullisharXiv – CS AI · Mar 46/103
🧠

MEBM-Speech: Multi-scale Enhanced BrainMagic for Robust MEG Speech Detection

Researchers propose MEBM-Speech, a neural decoder that detects speech activity from brain signals using magnetoencephalography (MEG). The system achieved 89.3% F1 score on benchmark tests and could advance brain-computer interfaces for cognitive neuroscience and clinical applications.

AIBullisharXiv – CS AI · Mar 46/104
🧠

Universal Conceptual Structure in Neural Translation: Probing NLLB-200's Multilingual Geometry

Researchers analyzed Meta's NLLB-200 neural machine translation model across 135 languages, finding that it has implicitly learned universal conceptual structures and language genealogical relationships. The study reveals the model creates language-neutral conceptual representations similar to how multilingual brains organize information, with semantic relationships preserved across diverse languages.

AIBearisharXiv – CS AI · Mar 47/102
🧠

Silent Sabotage During Fine-Tuning: Few-Shot Rationale Poisoning of Compact Medical LLMs

Researchers discovered a new stealth poisoning attack method targeting medical AI language models during fine-tuning that degrades performance on specific medical topics without detection. The attack injects poisoned rationales into training data, proving more effective than traditional backdoor attacks or catastrophic forgetting methods.

AIBullisharXiv – CS AI · Mar 47/103
🧠

Social-JEPA: Emergent Geometric Isomorphism

Researchers developed Social-JEPA, showing that separate AI agents learning from different viewpoints of the same environment develop internal representations that are mathematically aligned through approximate linear isometry. This enables models trained on one agent to work on another without retraining, suggesting a path toward interoperable decentralized AI vision systems.

AIBullisharXiv – CS AI · Mar 46/103
🧠

PRISM: Exploring Heterogeneous Pretrained EEG Foundation Model Transfer to Clinical Differential Diagnosis

Researchers introduce PRISM, an EEG foundation model that demonstrates how diverse pretraining data leads to better clinical performance than narrow-source datasets. The study shows that geographically diverse EEG data outperforms larger but homogeneous datasets in medical diagnosis tasks, particularly achieving 12.3% better accuracy in distinguishing epilepsy from similar conditions.

$COMP
← PrevPage 290 of 1266Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined