y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🧠All17,711🧠AI12,727🤖AI × Crypto545📰General4,439
Home/AI Pulse

AI Pulse News

Models, papers, tools. 17,716 articles with AI-powered sentiment analysis and key takeaways.

17716 articles
AINeutralarXiv – CS AI · Feb 277/106
🧠

On the Complexity of Neural Computation in Superposition

Researchers establish theoretical foundations for neural network superposition, proving lower bounds that require at least Ω(√m' log m') neurons and Ω(m' log m') parameters to compute m' features. The work demonstrates exponential complexity gaps between computing versus merely representing features and provides first subexponential bounds on network capacity.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Toward Automated Validation of Language Model Synthesized Test Cases using Semantic Entropy

Researchers introduce VALTEST, a framework that uses semantic entropy to automatically validate test cases generated by Large Language Models, addressing the problem of invalid or hallucinated tests that mislead AI programming agents. The system improves test validity by up to 29% and enhances code generation performance through better filtering of LLM-generated test cases.

AIBullisharXiv – CS AI · Feb 277/106
🧠

ViT-Linearizer: Distilling Quadratic Knowledge into Linear-Time Vision Models

Researchers developed ViT-Linearizer, a distillation framework that transfers Vision Transformer knowledge into linear-time models, addressing quadratic complexity issues for high-resolution inputs. The method achieves 84.3% ImageNet accuracy while providing significant speedups, bridging the gap between efficient RNN-based architectures and transformer performance.

AIBullisharXiv – CS AI · Feb 277/109
🧠

Towards a Sharp Analysis of Offline Policy Learning for $f$-Divergence-Regularized Contextual Bandits

Researchers achieved breakthrough sample complexity improvements for offline reinforcement learning algorithms using f-divergence regularization, particularly for contextual bandits. The study demonstrates optimal O(ε⁻¹) sample complexity under single-policy concentrability conditions, significantly improving upon existing bounds.

$NEAR
AINeutralarXiv – CS AI · Feb 277/105
🧠

Using the Path of Least Resistance to Explain Deep Networks

Researchers propose Geodesic Integrated Gradients (GIG), a new method for explaining AI model decisions that uses curved paths instead of straight lines to compute feature importance. The method addresses flawed attributions in existing approaches by integrating gradients along geodesic paths under a model-induced Riemannian metric.

AINeutralarXiv – CS AI · Feb 277/106
🧠

RaPA: Enhancing Transferable Targeted Attacks via Random Parameter Pruning

Researchers propose Random Parameter Pruning Attack (RaPA), a new method that improves targeted adversarial attacks by randomly pruning model parameters during optimization. The technique achieves up to 11.7% higher attack success rates when transferring from CNN to Transformer models compared to existing methods.

AIBullisharXiv – CS AI · Feb 277/104
🧠

Beyond the Monitor: Mixed Reality Visualization and Multimodal AI for Enhanced Digital Pathology Workflow

Researchers developed PathVis, a mixed-reality platform for Apple Vision Pro that revolutionizes digital pathology by allowing pathologists to examine gigapixel cancer diagnostic images through immersive visualization and multimodal AI assistance. The system replaces traditional 2D monitor limitations with natural interactions using eye gaze, hand gestures, and voice commands, integrated with AI agents for computer-aided diagnosis.

AIBullisharXiv – CS AI · Feb 277/105
🧠

Towards Autonomous Memory Agents

Researchers introduce U-Mem, an autonomous memory agent system that actively acquires and validates knowledge for large language models. The system uses cost-aware knowledge extraction and semantic Thompson sampling to improve performance, showing significant gains on benchmarks like HotpotQA and AIME25.

AINeutralarXiv – CS AI · Feb 277/106
🧠

Epistemic Filtering and Collective Hallucination: A Jury Theorem for Confidence-Calibrated Agents

Researchers propose a new framework for collective decision-making where AI agents can abstain from voting when uncertain, extending the Condorcet Jury Theorem to confidence-gated settings. The study shows this selective participation approach can improve group accuracy and potentially reduce hallucinations in large language model systems.

AIBullisharXiv – CS AI · Feb 277/105
🧠

Mirroring the Mind: Distilling Human-Like Metacognitive Strategies into Large Language Models

Researchers propose Metacognitive Behavioral Tuning (MBT), a new framework that addresses structural fragility in Large Reasoning Models by injecting human-like self-regulatory control into AI thought processes. The approach reduces reasoning collapse and improves accuracy while consuming fewer computational tokens across multi-hop question-answering benchmarks.

AINeutralarXiv – CS AI · Feb 277/106
🧠

VeRO: An Evaluation Harness for Agents to Optimize Agents

Researchers introduced VeRO (Versioning, Rewards, and Observations), a new evaluation framework for testing AI coding agents that can optimize other AI agents through iterative improvement cycles. The system provides reproducible benchmarks and structured execution traces to systematically measure how well coding agents can improve target agents' performance.

AIBullisharXiv – CS AI · Feb 277/109
🧠

ArchAgent: Agentic AI-driven Computer Architecture Discovery

ArchAgent, an AI-driven system built on AlphaEvolve, has achieved breakthrough results in automated computer architecture discovery by designing state-of-the-art cache replacement policies. The system achieved 5.3% performance improvements in just 2 days and 0.9% improvements in 18 days, working 3-5x faster than human-developed solutions.

AIBullisharXiv – CS AI · Feb 277/105
🧠

CourtGuard: A Model-Agnostic Framework for Zero-Shot Policy Adaptation in LLM Safety

Researchers introduce CourtGuard, a new framework for AI safety that uses retrieval-augmented multi-agent debate to evaluate LLM outputs without requiring expensive retraining. The system achieves state-of-the-art performance across 7 safety benchmarks and demonstrates zero-shot adaptability to new policy requirements, offering a more flexible approach to AI governance.

AINeutralarXiv – CS AI · Feb 277/108
🧠

A Mathematical Theory of Agency and Intelligence

Researchers propose a mathematical framework distinguishing agency from intelligence in AI systems, introducing 'bipredictability' as a measure of effective information sharing between observations, actions, and outcomes. Current AI systems achieve agency but lack true intelligence, which requires adaptive learning and self-monitoring capabilities.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Toward Personalized LLM-Powered Agents: Foundations, Evaluation, and Future Directions

Researchers published a comprehensive survey on personalized LLM-powered agents that can adapt to individual users over extended interactions. The study organizes these agents into four key components: profile modeling, memory, planning, and action execution, providing a framework for developing more user-aligned AI assistants.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Sparse Imagination for Efficient Visual World Model Planning

Researchers propose a new sparse imagination technique for visual world model planning that significantly reduces computational burden while maintaining task performance. The method uses transformers with randomized grouped attention to enable efficient planning in resource-constrained environments like robotics.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Know What You Know: Metacognitive Entropy Calibration for Verifiable RL Reasoning

Researchers propose EGPO, a new framework that improves large reasoning models by incorporating uncertainty awareness into reinforcement learning training. The approach addresses the "uncertainty-reward mismatch" where current training methods treat high and low-confidence solutions equally, preventing models from developing better reasoning capabilities.

AIBearisharXiv – CS AI · Feb 277/107
🧠

Bob's Confetti: Phonetic Memorization Attacks in Music and Video Generation

Researchers discovered a vulnerability in AI music and video generation systems where phonetic prompts can bypass copyright filters. The 'Adversarial PhoneTic Prompting' attack achieves 91% similarity to copyrighted content by using sound-alike phrases that preserve acoustic patterns while evading text-based detection.

$NEAR$APT
AIBullisharXiv – CS AI · Feb 277/106
🧠

LayerT2V: A Unified Multi-Layer Video Generation Framework

LayerT2V introduces a breakthrough multi-layer video generation framework that produces editable layered video components (background, foreground layers with alpha mattes) in a single inference pass. The system addresses professional workflow limitations of current text-to-video models by enabling semantic consistency across layers and introduces VidLayer, the first large-scale dataset for multi-layer video generation.

AINeutralarXiv – CS AI · Feb 277/106
🧠

Echoes of AI: Investigating the Downstream Effects of AI Assistants on Software Maintainability

A controlled study of 151 professional developers found that AI coding assistants like GitHub Copilot provide significant productivity gains (30.7% faster completion) but don't impact code maintainability when other developers later modify the code. The research suggests AI-assisted code is neither easier nor harder for subsequent developers to work with.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Supervised Reinforcement Learning: From Expert Trajectories to Step-wise Reasoning

Researchers propose Supervised Reinforcement Learning (SRL), a new training framework that helps small-scale language models solve complex multi-step reasoning problems by generating internal reasoning monologues and providing step-wise rewards. SRL outperforms traditional Supervised Fine-Tuning and Reinforcement Learning approaches, enabling smaller models to tackle previously unlearnable problems.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Intelligence per Watt: Measuring Intelligence Efficiency of Local AI

Researchers propose 'Intelligence per Watt' (IPW) as a metric to measure AI efficiency, finding that local AI models can handle 71.3% of queries while being 1.4x more energy efficient than cloud alternatives. The study demonstrates that smaller local language models (≤20B parameters) can redistribute computational demand from centralized cloud infrastructure.

AIBullisharXiv – CS AI · Feb 277/108
🧠

UniQL: Unified Quantization and Low-rank Compression for Adaptive Edge LLMs

Researchers introduce UniQL, a unified framework for quantizing and compressing large language models to run efficiently on mobile devices. The system achieves 4x-5.7x memory reduction and 2.7x-3.4x speed improvements while maintaining accuracy within 5% of original models.

AINeutralarXiv – CS AI · Feb 277/103
🧠

The Tool Decathlon: Benchmarking Language Agents for Diverse, Realistic, and Long-Horizon Task Execution

Researchers introduce Tool Decathlon (Toolathlon), a comprehensive benchmark for evaluating AI language agents across 32 software applications and 604 tools in realistic, multi-step scenarios. The benchmark reveals significant limitations in current AI models, with the best performer (Claude-4.5-Sonnet) achieving only 38.6% success rate on complex, real-world tasks.

AIBullisharXiv – CS AI · Feb 277/105
🧠

Compute-Optimal Quantization-Aware Training

Researchers developed a new approach to quantization-aware training (QAT) that optimizes compute allocation between full-precision and quantized training phases. They discovered that contrary to previous findings, the optimal ratio of QAT to full-precision training increases with total compute budget, and derived scaling laws to predict optimal configurations across different model sizes and bit widths.

← PrevPage 170 of 709Next →
◆ AI Mentions
🏢OpenAI
101×
🏢Nvidia
55×
🧠GPT-5
36×
🏢Anthropic
36×
🧠Claude
35×
🧠Gemini
33×
🧠ChatGPT
19×
🧠GPT-4
14×
🧠Llama
14×
🏢Meta
9×
🏢xAI
9×
🧠Sonnet
8×
🏢Perplexity
8×
🏢Microsoft
7×
🧠Opus
7×
🏢Google
7×
🧠Grok
5×
🏢Hugging Face
4×
🧠o1
2×
🧠Copilot
1×
▲ Trending Tags
1#ai4432#iran4363#market2834#geopolitical2575#geopolitics1206#geopolitical-risk1087#trump988#openai969#security8310#inflation7411#middle-east6212#artificial-intelligence6213#market-volatility5414#sanctions5415#nvidia53
Tag Sentiment
#ai443 articles
#iran436 articles
#market283 articles
#geopolitical257 articles
#geopolitics120 articles
#geopolitical-risk108 articles
#trump98 articles
#openai96 articles
#security83 articles
#inflation74 articles
BullishNeutralBearish
Stay Updated
Models, papers, tools
Tag Connections
#geopolitical↔#iran
189
#iran↔#market
124
#geopolitical↔#market
96
#iran↔#trump
72
#ai↔#artificial-intelligence
48
#ai↔#market
42
#geopolitical↔#trump
38
#ai↔#openai
38
#market↔#trump
36
#geopolitics↔#iran
36
Filters
Sentiment
Importance
Sort
📡 See all 70+ sources
y0.exchange
Your AI agent for DeFi
Connect Claude or GPT to your wallet. AI reads balances, proposes swaps and bridges — you approve. Your keys never leave your device.
8 MCP tools · 15 chains · $0 fees
Connect Wallet to AI →How it works →
Viewing: AI Pulse feed
Filters
Sentiment
Importance
Sort
Stay Updated
Models, papers, tools
y0news
y0.exchangeLaunch AppDigestsSourcesAboutRSSAI NewsCrypto News
© 2026 y0.exchange