y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🧠All19,058🧠AI13,505🤖AI × Crypto594📰General4,959
Home/AI Pulse

AI Pulse News

Models, papers, tools. 19,058 articles with AI-powered sentiment analysis and key takeaways.

19058 articles
AINeutralarXiv – CS AI · Mar 176/10
🧠

Citation-Enforced RAG for Fiscal Document Intelligence: Cited, Explainable Knowledge Retrieval in Tax Compliance

Researchers have developed a new AI framework that uses citation-enforced retrieval-augmented generation (RAG) specifically for analyzing tax and fiscal documents. The system prioritizes transparency and explainability for tax authorities, showing improved citation accuracy and reduced AI hallucinations when tested on real IRS documents.

AINeutralarXiv – CS AI · Mar 176/10
🧠

Deeper Thought, Weaker Aim: Understanding and Mitigating Perceptual Impairment during Reasoning in Multimodal Large Language Models

Researchers have identified that multimodal large language models (MLLMs) lose visual focus during complex reasoning tasks, with attention becoming scattered across images rather than staying on relevant regions. They propose a training-free Visual Region-Guided Attention (VRGA) framework that improves visual grounding and reasoning accuracy by reweighting attention to question-relevant areas.

AIBullisharXiv – CS AI · Mar 176/10
🧠

Self-Indexing KVCache: Predicting Sparse Attention from Compressed Keys

Researchers propose a novel self-indexing KV cache system that unifies compression and retrieval for efficient sparse attention in large language models. The method uses 1-bit vector quantization and integrates with FlashAttention to reduce memory bottlenecks in long-context LLM inference.

AIBearisharXiv – CS AI · Mar 176/10
🧠

I'm Not Reading All of That: Understanding Software Engineers' Level of Cognitive Engagement with Agentic Coding Assistants

A research study reveals that software engineers' cognitive engagement consistently declines when working with agentic AI coding assistants, raising concerns about over-reliance and reduced critical thinking. The study found that current AI assistants provide limited support for reflection and verification, identifying design opportunities to promote deeper thinking in AI-assisted programming.

AIBullisharXiv – CS AI · Mar 176/10
🧠

Mitigating Overthinking in Large Reasoning Language Models via Reasoning Path Deviation Monitoring

Researchers propose a new early-exit method for Large Reasoning Language Models that detects and prevents overthinking by monitoring high-entropy transition tokens that indicate deviation from correct reasoning paths. The method improves performance and efficiency compared to existing approaches without requiring additional training overhead or limiting inference throughput.

AINeutralarXiv – CS AI · Mar 176/10
🧠

AEX: Non-Intrusive Multi-Hop Attestation and Provenance for LLM APIs

Researchers propose AEX, a new attestation protocol for LLM APIs that provides cryptographic proof that API responses actually correspond to client requests. The system addresses trust issues with hosted AI models by adding signed attestation objects to existing JSON-based APIs without disrupting current functionality.

🏢 OpenAI
AIBullisharXiv – CS AI · Mar 176/10
🧠

Deconfounded Lifelong Learning for Autonomous Driving via Dynamic Knowledge Spaces

Researchers propose DeLL, a new framework for autonomous driving systems that addresses lifelong learning challenges through dynamic knowledge spaces and causal inference mechanisms. The system uses Dirichlet process mixture models to prevent catastrophic forgetting and improve adaptability to new driving scenarios while maintaining previously learned knowledge.

AIBullisharXiv – CS AI · Mar 176/10
🧠

AerialVLA: A Vision-Language-Action Model for UAV Navigation via Minimalist End-to-End Control

Researchers propose AerialVLA, a minimalist end-to-end Vision-Language-Action framework for UAV navigation that directly maps visual observations and linguistic instructions to continuous control signals. The system eliminates reliance on external object detectors and dense oracle guidance, achieving nearly three times the success rate of existing baselines in unseen environments.

AIBullisharXiv – CS AI · Mar 176/10
🧠

OxyGen: Unified KV Cache Management for Vision-Language-Action Models under Multi-Task Parallelism

Researchers propose OxyGen, a unified KV cache management system for Vision-Language-Action Models that enables efficient multi-task parallelism in embodied AI agents. The system achieves up to 3.7x speedup by sharing computational resources across tasks and eliminating redundant processing of shared observations.

AIBullisharXiv – CS AI · Mar 176/10
🧠

From $\boldsymbol{\log\pi}$ to $\boldsymbol{\pi}$: Taming Divergence in Soft Clipping via Bilateral Decoupled Decay of Probability Gradient Weight

Researchers introduce Decoupled Gradient Policy Optimization (DGPO), a new reinforcement learning method that improves large language model training by using probability gradients instead of log-probability gradients. The technique addresses instability issues in current methods while maintaining exploration capabilities, showing superior performance across mathematical benchmarks.

AIBullisharXiv – CS AI · Mar 176/10
🧠

ES-Merging: Biological MLLM Merging via Embedding Space Signals

Researchers propose ES-Merging, a new framework for combining specialized biological multimodal large language models (MLLMs) by using embedding space signals rather than traditional parameter-based methods. The approach estimates merging coefficients at both layer-wise and element-wise granularities, outperforming existing merging techniques and even task-specific fine-tuned models on cross-modal scientific problems.

AIBullisharXiv – CS AI · Mar 176/10
🧠

Distilling Reasoning Without Knowledge: A Framework for Reliable LLMs

Researchers propose a new framework for large language models that separates planning from factual retrieval to improve reliability in fact-seeking question answering. The modular approach uses a lightweight student planner trained via teacher-student learning to generate structured reasoning steps, showing improved accuracy and speed on challenging benchmarks.

AINeutralarXiv – CS AI · Mar 176/10
🧠

Infinite Problem Generator: Verifiably Scaling Physics Reasoning Data with Agentic Workflows

Researchers introduce the Infinite Problem Generator (IPG), an AI framework that creates verifiable physics problems using executable Python code instead of probabilistic text generation. The system released ClassicalMechanicsV1, a dataset of 1,335 physics problems that demonstrates how code complexity can precisely measure problem difficulty for training large language models.

AIBullisharXiv – CS AI · Mar 176/10
🧠

VLA-Thinker: Boosting Vision-Language-Action Models through Thinking-with-Image Reasoning

Researchers introduce VLA-Thinker, a new AI framework that enhances Vision-Language-Action models by enabling dynamic visual reasoning during robotic tasks. The system achieved a 97.5% success rate on LIBERO benchmarks through a two-stage training pipeline combining supervised fine-tuning and reinforcement learning.

AINeutralarXiv – CS AI · Mar 176/10
🧠

MALicious INTent Dataset and Inoculating LLMs for Enhanced Disinformation Detection

Researchers released MALINT, the first human-annotated English dataset for detecting disinformation and its malicious intent, developed with expert fact-checkers. The study benchmarked 12 language models and introduced intent-based inoculation techniques that improved zero-shot disinformation detection across six datasets, five LLMs, and seven languages.

🧠 Llama
AIBearisharXiv – CS AI · Mar 176/10
🧠

The Scenic Route to Deception: Dark Patterns and Explainability Pitfalls in Conversational Navigation

Researchers warn that AI-powered conversational navigation systems using Large Language Models could transform route guidance from verifiable geometric tasks into manipulative dialogues. The study proposes a framework categorizing risks as dark patterns or explainability pitfalls, suggesting neuro-symbolic architectures to maintain trustworthiness.

AIBullisharXiv – CS AI · Mar 176/10
🧠

$PA^3$: $\textbf{P}$olicy-$\textbf{A}$ware $\textbf{A}$gent $\textbf{A}$lignment through Chain-of-Thought

Researchers developed PA³, a new method to improve AI assistant alignment with business policies by teaching models to recall and apply relevant rules during reasoning without including full policies in prompts. The approach reduces computational overhead by 40% while achieving 16-point performance improvements over baselines.

$PA
AINeutralarXiv – CS AI · Mar 176/10
🧠

Compute Allocation for Reasoning-Intensive Retrieval Agents

Researchers studied computational resource allocation in AI retrieval systems for long-horizon agents, finding that re-ranking stages benefit more from powerful models and deeper candidate pools than query expansion stages. The study suggests concentrating compute power on re-ranking rather than distributing it uniformly across pipeline stages for better performance.

🧠 Gemini
AIBullisharXiv – CS AI · Mar 176/10
🧠

Nudging Hidden States: Training-Free Model Steering for Chain-of-Thought Reasoning in Large Audio-Language Models

Researchers developed training-free model steering techniques to improve reasoning in large audio-language models (LALMs) through chain-of-thought prompting. The approach achieved up to 4.4% accuracy gains and demonstrated cross-modal transfer where text-derived steering vectors can effectively guide speech-based reasoning.

AIBullisharXiv – CS AI · Mar 176/10
🧠

MVHOI: Bridge Multi-view Condition to Complex Human-Object Interaction Video Reenactment via 3D Foundation Model

Researchers introduce MVHOI, a new AI framework that significantly improves human-object interaction video generation by handling complex 3D manipulations through a two-stage process using 3D foundation models. The system can create realistic long-duration videos showing intricate object manipulations from multiple viewpoints, addressing limitations of existing approaches that struggle with non-planar movements.

AIBullisharXiv – CS AI · Mar 176/10
🧠

Beyond Local Code Optimization: Multi-Agent Reasoning for Software System Optimization

Researchers introduced a multi-agent AI framework for whole-system software optimization that goes beyond local code improvements to analyze entire microservice architectures. The system uses coordinated agents for summarization, analysis, optimization, and verification, achieving 36.58% throughput improvement and 27.81% response time reduction in proof-of-concept testing.

AIBullisharXiv – CS AI · Mar 176/10
🧠

AdapterTune: Zero-Initialized Low-Rank Adapters for Frozen Vision Transformers

AdapterTune introduces a new method for efficiently fine-tuning Vision Transformers by using zero-initialized low-rank adapters that start at the pretrained function to prevent optimization instability. The technique achieves +14.9 point accuracy improvement over head-only transfer while using only 0.92% of parameters needed for full fine-tuning.

AIBullisharXiv – CS AI · Mar 176/10
🧠

Universe Routing: Why Self-Evolving Agents Need Epistemic Control

Researchers propose a 'universe routing' solution for AI agents that struggle to choose appropriate reasoning frameworks when faced with different types of questions. The study shows that hard routing to specialized solvers is 7x faster than soft mixing approaches, with a 465M-parameter router achieving superior generalization and zero forgetting in continual learning scenarios.

🏢 Meta
AIBullisharXiv – CS AI · Mar 176/10
🧠

SimCert: Probabilistic Certification for Behavioral Similarity in Deep Neural Network Compression

Researchers developed SimCert, a probabilistic certification framework that verifies behavioral similarity between compressed neural networks and their original versions. The framework addresses critical safety challenges in deploying compressed DNNs on resource-constrained systems by providing quantitative safety guarantees with adjustable confidence levels.

AIBullisharXiv – CS AI · Mar 176/10
🧠

RAZOR: Ratio-Aware Layer Editing for Targeted Unlearning in Vision Transformers and Diffusion Models

Researchers introduce RAZOR, a new framework for efficiently removing sensitive information from AI models like CLIP and Stable Diffusion without requiring full retraining. The method selectively edits specific layers and attention heads in transformer models to achieve targeted 'unlearning' while preserving overall performance.

🧠 Stable Diffusion
← PrevPage 360 of 763Next →
◆ AI Mentions
🏢OpenAI
68×
🏢Anthropic
40×
🧠Claude
35×
🏢Nvidia
31×
🧠Gemini
19×
🧠ChatGPT
19×
🧠GPT-4
18×
🧠GPT-5
15×
🧠Llama
15×
🏢Perplexity
12×
🧠Opus
9×
🏢xAI
8×
🏢Google
5×
🏢Hugging Face
5×
🧠Sonnet
5×
🏢Meta
4×
🧠Grok
3×
🧠Haiku
2×
🏢Microsoft
2×
🧠Sora
1×
▲ Trending Tags
1#ai2402#geopolitical-risk2283#geopolitics1874#iran1815#market-volatility1206#middle-east1077#oil-markets838#sanctions799#geopolitical7410#energy-markets7411#inflation6912#ai-infrastructure6213#openai6014#strait-of-hormuz5615#market53
Tag Sentiment
#ai240 articles
#geopolitical-risk228 articles
#geopolitics187 articles
#iran181 articles
#market-volatility120 articles
#middle-east107 articles
#oil-markets83 articles
#sanctions79 articles
#energy-markets74 articles
#geopolitical74 articles
BullishNeutralBearish
Stay Updated
Models, papers, tools
Tag Connections
#geopolitics↔#iran
53
#geopolitical↔#iran
42
#geopolitical-risk↔#market-volatility
41
#geopolitical-risk↔#oil-markets
40
#geopolitics↔#oil-markets
39
#geopolitical-risk↔#middle-east
36
#geopolitics↔#middle-east
33
#iran↔#trump
29
#geopolitical-risk↔#strait-of-hormuz
29
#energy-markets↔#geopolitical-risk
28
Filters
Sentiment
Importance
Sort
📡 See all 70+ sources
y0.exchange
Your AI agent for DeFi
Connect Claude or GPT to your wallet. AI reads balances, proposes swaps and bridges — you approve. Your keys never leave your device.
8 MCP tools · 15 chains · $0 fees
Connect Wallet to AI →How it works →
Viewing: AI Pulse feed
Filters
Sentiment
Importance
Sort
Stay Updated
Models, papers, tools
y0news
y0.exchangeLaunch AppDigestsSourcesAboutRSSAI NewsCrypto News
© 2026 y0.exchange