y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🧠All17,571🧠AI12,705🤖AI × Crypto536📰General4,330
Home/AI Pulse

AI Pulse News

Models, papers, tools. 17,585 articles with AI-powered sentiment analysis and key takeaways.

17585 articles
AIBullisharXiv – CS AI · Mar 46/102
🧠

AI-for-Science Low-code Platform with Bayesian Adversarial Multi-Agent Framework

Researchers have developed a Bayesian adversarial multi-agent framework for AI-driven scientific code generation, featuring three coordinated LLM agents that work together to improve reliability and reduce errors. The Low-code Platform (LCP) enables non-expert users to generate scientific code through natural language prompts, demonstrating superior performance in benchmark tests and Earth Science applications.

AIBullisharXiv – CS AI · Mar 47/103
🧠

Self-Play Only Evolves When Self-Synthetic Pipeline Ensures Learnable Information Gain

Researchers propose a framework for sustainable AI self-evolution through triadic roles (Proposer, Solver, Verifier) that ensures learnable information gain across iterations. The study identifies three key system designs to prevent the common plateau effect in self-play AI systems: asymmetric co-evolution, capacity growth, and proactive information seeking.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Generalized Discrete Diffusion with Self-Correction

Researchers propose Self-Correcting Discrete Diffusion (SCDD), a new AI model that improves upon existing discrete diffusion models by reformulating self-correction with explicit state transitions. The method enables more efficient parallel decoding while maintaining generation quality, demonstrating improvements at GPT-2 scale.

AIBullisharXiv – CS AI · Mar 47/103
🧠

Density-Guided Response Optimization: Community-Grounded Alignment via Implicit Acceptance Signals

Researchers introduce Density-Guided Response Optimization (DGRO), a new AI alignment method that learns community preferences from implicit acceptance signals rather than explicit feedback. The technique uses geometric patterns in how communities naturally engage with content to train language models without requiring costly annotation or preference labeling.

AIBullisharXiv – CS AI · Mar 47/102
🧠

NExT-Guard: Training-Free Streaming Safeguard without Token-Level Labels

Researchers introduce NExT-Guard, a training-free framework for real-time AI safety monitoring that uses Sparse Autoencoders to detect unsafe content in streaming language models. The system outperforms traditional supervised training methods while requiring no token-level annotations, making it more cost-effective and scalable for deployment.

AINeutralarXiv – CS AI · Mar 47/103
🧠

Forecasting as Rendering: A 2D Gaussian Splatting Framework for Time Series Forecasting

Researchers introduce TimeGS, a novel time series forecasting framework that reimagines prediction as 2D generative rendering using Gaussian splatting techniques. The approach addresses key limitations in existing methods by treating future sequences as continuous latent surfaces and enforcing temporal continuity across periodic boundaries.

AIBullisharXiv – CS AI · Mar 46/103
🧠

MedFeat: Model-Aware and Explainability-Driven Feature Engineering with LLMs for Clinical Tabular Prediction

Researchers introduce MedFeat, a new AI framework that uses Large Language Models for healthcare feature engineering in clinical tabular predictions. The system incorporates model awareness and domain knowledge to discover clinically meaningful features that outperform traditional approaches and demonstrate robustness across different hospital settings.

AINeutralarXiv – CS AI · Mar 47/102
🧠

MedCalc-Bench Doesn't Measure What You Think: A Benchmark Audit and the Case for Open-Book Evaluation

Researchers audited the MedCalc-Bench benchmark for evaluating AI models on clinical calculator tasks, finding over 20 errors in the dataset and showing that simple 'open-book' prompting achieves 81-85% accuracy versus previous best of 74%. The study suggests the benchmark measures formula memorization rather than clinical reasoning, challenging how AI medical capabilities are evaluated.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Neural Paging: Learning Context Management Policies for Turing-Complete Agents

Researchers introduce Neural Paging, a new architecture that addresses the computational bottleneck of finite context windows in Large Language Models by implementing a hierarchical system that decouples reasoning from memory management. The approach reduces computational complexity from O(N²) to O(N·K²) for long-horizon reasoning tasks, potentially enabling more efficient AI agents.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Physics-Informed Neural Networks with Architectural Physics Embedding for Large-Scale Wave Field Reconstruction

Researchers developed Physics-Embedded PINNs (PE-PINN) that achieve 10x faster convergence than standard physics-informed neural networks and orders of magnitude memory reduction compared to traditional methods for large-scale wave field reconstruction. The breakthrough enables high-fidelity electromagnetic wave modeling for wireless communications, sensing, and room acoustics applications.

AINeutralarXiv – CS AI · Mar 47/103
🧠

Structured vs. Unstructured Pruning: An Exponential Gap

Research reveals an exponential gap between structured and unstructured neural network pruning methods. While unstructured weight pruning can approximate target functions with O(d log(1/ε)) neurons, structured neuron pruning requires Ω(d/ε) neurons, demonstrating fundamental limitations of structured approaches.

AIBullisharXiv – CS AI · Mar 47/104
🧠

Beyond Binary Preferences: A Principled Framework for Reward Modeling with Ordinal Feedback

Researchers present a new mathematical framework for training AI reward models using Likert scale preferences instead of simple binary comparisons. The approach uses ordinal regression to better capture nuanced human feedback, outperforming existing methods across chat, reasoning, and safety benchmarks.

AIBullisharXiv – CS AI · Mar 46/104
🧠

Talking with Verifiers: Automatic Specification Generation for Neural Network Verification

Researchers have developed a framework that allows neural network verification tools to accept natural language specifications instead of low-level technical constraints. The system automatically translates human-readable requirements into formal verification queries, significantly expanding the practical applicability of neural network verification across diverse domains.

AINeutralarXiv – CS AI · Mar 46/104
🧠

CUDABench: Benchmarking LLMs for Text-to-CUDA Generation

Researchers introduce CUDABench, a comprehensive benchmark for evaluating Large Language Models' ability to generate CUDA code from text descriptions. The benchmark reveals significant challenges including high compilation success rates but low functional correctness, lack of domain-specific knowledge, and poor GPU hardware utilization.

AIBullisharXiv – CS AI · Mar 46/103
🧠

Concept Heterogeneity-aware Representation Steering

Researchers introduce CHaRS (Concept Heterogeneity-aware Representation Steering), a new method for controlling large language model behavior that uses optimal transport theory to create context-dependent steering rather than global directions. The approach models representations as Gaussian mixture models and derives input-dependent steering maps, showing improved behavioral control over existing methods.

AIBullisharXiv – CS AI · Mar 46/104
🧠

Universal Conceptual Structure in Neural Translation: Probing NLLB-200's Multilingual Geometry

Researchers analyzed Meta's NLLB-200 neural machine translation model across 135 languages, finding that it has implicitly learned universal conceptual structures and language genealogical relationships. The study reveals the model creates language-neutral conceptual representations similar to how multilingual brains organize information, with semantic relationships preserved across diverse languages.

AIBullisharXiv – CS AI · Mar 46/103
🧠

MEBM-Speech: Multi-scale Enhanced BrainMagic for Robust MEG Speech Detection

Researchers propose MEBM-Speech, a neural decoder that detects speech activity from brain signals using magnetoencephalography (MEG). The system achieved 89.3% F1 score on benchmark tests and could advance brain-computer interfaces for cognitive neuroscience and clinical applications.

AIBearisharXiv – CS AI · Mar 47/102
🧠

Silent Sabotage During Fine-Tuning: Few-Shot Rationale Poisoning of Compact Medical LLMs

Researchers discovered a new stealth poisoning attack method targeting medical AI language models during fine-tuning that degrades performance on specific medical topics without detection. The attack injects poisoned rationales into training data, proving more effective than traditional backdoor attacks or catastrophic forgetting methods.

AIBullisharXiv – CS AI · Mar 47/103
🧠

Social-JEPA: Emergent Geometric Isomorphism

Researchers developed Social-JEPA, showing that separate AI agents learning from different viewpoints of the same environment develop internal representations that are mathematically aligned through approximate linear isometry. This enables models trained on one agent to work on another without retraining, suggesting a path toward interoperable decentralized AI vision systems.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Bridging Diffusion Guidance and Anderson Acceleration via Hopfield Dynamics

Researchers have developed Geometry Aware Attention Guidance (GAG), a new method that improves diffusion model generation quality by optimizing attention-space extrapolation. The approach models attention dynamics as fixed-point iterations within Modern Hopfield Networks and applies Anderson Acceleration to stabilize the process while reducing computational costs.

AIBullisharXiv – CS AI · Mar 47/104
🧠

CoDAR: Continuous Diffusion Language Models are More Powerful Than You Think

Researchers propose CoDAR, a new continuous diffusion language model framework that addresses key bottlenecks in token rounding through a two-stage approach combining continuous diffusion with an autoregressive decoder. The model demonstrates substantial improvements in generation quality over existing latent diffusion methods and becomes competitive with discrete diffusion language models.

AINeutralarXiv – CS AI · Mar 46/105
🧠

Human-Certified Module Repositories for the AI Age

Researchers propose Human-Certified Module Repositories (HCMRs) as a new framework to ensure trustworthy software development in the AI era. The system combines human oversight with automated analysis to certify and curate reusable code modules, addressing growing security concerns as AI increasingly generates and assembles software components.

AIBullisharXiv – CS AI · Mar 46/103
🧠

Through the Lens of Contrast: Self-Improving Visual Reasoning in VLMs

Researchers introduce VC-STaR, a new framework that improves visual reasoning in vision-language models by using contrastive image pairs to reduce hallucinations. The approach creates VisCoR-55K, a new dataset that outperforms existing visual reasoning methods when used for model fine-tuning.

AIBullisharXiv – CS AI · Mar 46/102
🧠

When Scaling Fails: Mitigating Audio Perception Decay of LALMs via Multi-Step Perception-Aware Reasoning

Researchers identified a critical problem in Large Audio-Language Models (LALMs) where audio perception deteriorates during extended reasoning processes. They developed MPAR² framework using reinforcement learning, which improved perception performance from 31.74% to 63.51% and achieved 74.59% accuracy on MMAU benchmark.

AIBullisharXiv – CS AI · Mar 46/103
🧠

PRISM: Exploring Heterogeneous Pretrained EEG Foundation Model Transfer to Clinical Differential Diagnosis

Researchers introduce PRISM, an EEG foundation model that demonstrates how diverse pretraining data leads to better clinical performance than narrow-source datasets. The study shows that geographically diverse EEG data outperforms larger but homogeneous datasets in medical diagnosis tasks, particularly achieving 12.3% better accuracy in distinguishing epilepsy from similar conditions.

$COMP
← PrevPage 147 of 704Next →
◆ AI Mentions
🏢OpenAI
101×
🏢Nvidia
58×
🧠GPT-5
41×
🧠Gemini
38×
🧠Claude
37×
🏢Anthropic
36×
🧠Llama
19×
🧠ChatGPT
19×
🧠GPT-4
18×
🏢Meta
11×
🏢xAI
9×
🏢Perplexity
9×
🧠Sonnet
9×
🧠Opus
8×
🏢Microsoft
7×
🏢Google
7×
🏢Hugging Face
6×
🧠Grok
6×
🧠o1
2×
🧠DALL E
1×
▲ Trending Tags
1#ai5132#iran4863#market3474#geopolitical3185#trump1146#openai977#security948#geopolitics699#inflation6810#artificial-intelligence6711#nvidia5712#machine-learning5613#geopolitical-risk5414#fed4815#google45
Tag Sentiment
#ai513 articles
#iran486 articles
#market347 articles
#geopolitical318 articles
#trump114 articles
#openai97 articles
#security94 articles
#geopolitics69 articles
#inflation68 articles
#artificial-intelligence67 articles
BullishNeutralBearish
Stay Updated
Models, papers, tools
Tag Connections
#geopolitical↔#iran
223
#iran↔#market
145
#geopolitical↔#market
119
#iran↔#trump
83
#ai↔#artificial-intelligence
55
#ai↔#market
52
#geopolitical↔#trump
43
#market↔#trump
41
#ai↔#openai
39
#ai↔#google
36
Filters
Sentiment
Importance
Sort
📡 See all 70+ sources
y0.exchange
Your AI agent for DeFi
Connect Claude or GPT to your wallet. AI reads balances, proposes swaps and bridges — you approve. Your keys never leave your device.
8 MCP tools · 15 chains · $0 fees
Connect Wallet to AI →How it works →
Viewing: AI Pulse feed
Filters
Sentiment
Importance
Sort
Stay Updated
Models, papers, tools
y0news
y0.exchangeLaunch AppDigestsSourcesAboutRSSAI NewsCrypto News
© 2026 y0.exchange