y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🧠All16,742🧠AI12,178🤖AI × Crypto505📰General4,059
Home/AI Pulse

AI Pulse News

Models, papers, tools. 16,750 articles with AI-powered sentiment analysis and key takeaways.

16750 articles
AIBearisharXiv – CS AI · Mar 267/10
🧠

When AI output tips to bad but nobody notices: Legal implications of AI's mistakes

Research reveals that generative AI's legal fabrications aren't random 'hallucinations' but predictable failures when the AI's internal state crosses a calculable threshold. The study shows AI can flip from reliable legal reasoning to creating fake case law and statutes, posing serious risks for attorneys and courts who may unknowingly use fabricated legal content.

AIBullisharXiv – CS AI · Mar 267/10
🧠

SCoOP: Semantic Consistent Opinion Pooling for Uncertainty Quantification in Multiple Vision-Language Model Systems

Researchers developed SCoOP, a training-free framework that combines multiple Vision-Language Models to improve uncertainty quantification and reduce hallucinations in AI systems. The method achieves 10-13% better hallucination detection performance compared to existing approaches while adding only microsecond-level overhead to processing time.

AIBullisharXiv – CS AI · Mar 267/10
🧠

From Pixels to Digital Agents: An Empirical Study on the Taxonomy and Technological Trends of Reinforcement Learning Environments

Researchers conducted a large-scale empirical study analyzing over 2,000 publications to map the evolution of reinforcement learning environments. The study reveals a paradigm shift toward two distinct ecosystems: LLM-driven 'Semantic Prior' agents and 'Domain-Specific Generalization' systems, providing a roadmap for next-generation AI simulators.

AIBullisharXiv – CS AI · Mar 267/10
🧠

AI-Supervisor: Autonomous AI Research Supervision via a Persistent Research World Model

Researchers have developed AI-Supervisor, a multi-agent framework that maintains a persistent Research World Model to autonomously conduct end-to-end AI research supervision. Unlike traditional linear pipelines, the system uses specialized agents with structured gap discovery, self-correcting loops, and consensus mechanisms to continuously evolve research understanding.

AIBearisharXiv – CS AI · Mar 267/10
🧠

Internal Safety Collapse in Frontier Large Language Models

Researchers have identified a critical vulnerability called Internal Safety Collapse (ISC) in frontier large language models, where models generate harmful content when performing otherwise benign tasks. Testing on advanced models like GPT-5.2 and Claude Sonnet 4.5 showed 95.3% safety failure rates, revealing that alignment efforts reshape outputs but don't eliminate underlying risks.

🧠 GPT-5🧠 Claude🧠 Sonnet
AINeutralarXiv – CS AI · Mar 267/10
🧠

Mitigating Many-Shot Jailbreaking

Researchers have developed techniques to mitigate many-shot jailbreaking (MSJ) attacks on large language models, where attackers use numerous examples to override safety training. Combined fine-tuning and input sanitization approaches significantly reduce MSJ effectiveness while maintaining normal model performance.

AINeutralarXiv – CS AI · Mar 267/10
🧠

Evidence for Limited Metacognition in LLMs

Researchers developed new methods to quantitatively measure metacognitive abilities in large language models, finding that frontier LLMs since early 2024 show increasing evidence of self-awareness capabilities. The study reveals these abilities are limited in resolution and qualitatively different from human metacognition, with variations across models suggesting post-training influences development.

AIBullisharXiv – CS AI · Mar 267/10
🧠

Berta: an open-source, modular tool for AI-enabled clinical documentation

Alberta Health Services deployed Berta, an open-source AI scribe platform that reduces clinical documentation costs by 70-95% compared to commercial alternatives. The system was used by 198 emergency physicians across 105 facilities, generating over 22,000 clinical sessions while keeping all data within secure health system infrastructure.

AINeutralarXiv – CS AI · Mar 267/10
🧠

A Theory of LLM Information Susceptibility

Researchers propose a theory of LLM information susceptibility that identifies fundamental limits to how large language models can improve optimization in AI agent systems. The study shows that nested, co-scaling architectures may be necessary for open-ended AI self-improvement, providing predictive constraints for AI system design.

AINeutralarXiv – CS AI · Mar 267/10
🧠

Beyond Accuracy: Introducing a Symbolic-Mechanistic Approach to Interpretable Evaluation

Researchers propose a new symbolic-mechanistic approach to evaluate AI models that goes beyond accuracy metrics to detect whether models truly generalize or rely on shortcuts like memorization. Their method combines symbolic rules with mechanistic interpretability to reveal when models exploit patterns rather than learn genuine capabilities, demonstrated through NL-to-SQL tasks where a memorization model achieved 94% accuracy but failed true generalization tests.

AIBullisharXiv – CS AI · Mar 267/10
🧠

PLDR-LLMs Reason At Self-Organized Criticality

Researchers demonstrate that PLDR-LLMs trained at self-organized criticality exhibit enhanced reasoning capabilities at inference time. The study shows that reasoning ability can be quantified using an order parameter derived from global model statistics, with models performing better when this parameter approaches zero at criticality.

AIBullisharXiv – CS AI · Mar 267/10
🧠

Physics-driven human-like working memory outperforms digital networks in dynamic vision

Researchers have developed a physics-driven AI system called Intrinsic Plasticity Network (IPNet) that uses magnetic tunnel junctions to create human-like working memory. The system demonstrates 18x error reduction in dynamic vision tasks while reducing memory-energy overhead by over 90,000x compared to traditional digital AI systems.

AIBullisharXiv – CS AI · Mar 267/10
🧠

DanQing: An Up-to-Date Large-Scale Chinese Vision-Language Pre-training Dataset

Researchers have released DanQing, a large-scale Chinese vision-language dataset containing 100 million high-quality image-text pairs curated from Common Crawl data. The dataset addresses the bottleneck in Chinese VLP development and demonstrates superior performance compared to existing Chinese datasets across various AI tasks.

AINeutralarXiv – CS AI · Mar 267/10
🧠

Probing Ethical Framework Representations in Large Language Models: Structure, Entanglement, and Methodological Challenges

Researchers analyzed how large language models (4B-72B parameters) internally represent different ethical frameworks, finding that models create distinct ethical subspaces but with asymmetric transfer patterns between frameworks. The study reveals structural insights into AI ethics processing while highlighting methodological limitations in probing techniques.

AIBullisharXiv – CS AI · Mar 267/10
🧠

OSS-CRS: Liberating AIxCC Cyber Reasoning Systems for Real-World Open-Source Security

Researchers have created OSS-CRS, an open framework that makes DARPA's AI Cyber Challenge systems usable for real-world cybersecurity applications. The system successfully ported the winning Atlantis CRS and discovered 10 previously unknown bugs, including three high-severity issues, across 8 open-source projects.

AIBullisharXiv – CS AI · Mar 267/10
🧠

QUARK: Quantization-Enabled Circuit Sharing for Transformer Acceleration by Exploiting Common Patterns in Nonlinear Operations

Researchers have developed QUARK, a quantization-enabled FPGA acceleration framework that significantly improves Transformer model performance by optimizing nonlinear operations through circuit sharing. The system achieves up to 1.96x speedup over GPU implementations while reducing hardware overhead by more than 50% compared to existing approaches.

AIBullisharXiv – CS AI · Mar 267/10
🧠

E0: Enhancing Generalization and Fine-Grained Control in VLA Models via Tweedie Discrete Diffusion

Researchers introduce E0, a new AI framework using tweedie discrete diffusion to improve Vision-Language-Action (VLA) models for robotic manipulation. The system addresses key limitations in existing VLA models by generating more precise actions through iterative denoising over quantized action tokens, achieving 10.7% better performance on average across 14 diverse robotic environments.

AINeutralarXiv – CS AI · Mar 267/10
🧠

From Prompts to Packets: A View from the Network on ChatGPT, Copilot, and Gemini

A comprehensive study analyzed network traffic patterns of popular AI chatbots ChatGPT, Copilot, and Gemini through Android mobile apps. The research reveals distinctive protocol footprints and traffic characteristics that create new challenges for network management, including sustained upstream activity and high-rate bursts unlike conventional messaging apps.

🏢 Microsoft🧠 ChatGPT🧠 Gemini
AINeutralarXiv – CS AI · Mar 267/10
🧠

Divide, then Ground: Adapting Frame Selection to Query Types for Long-Form Video Understanding

Researchers propose DIG, a training-free framework that improves long-form video understanding by adapting frame selection strategies based on query types. The system uses uniform sampling for global queries and specialized selection for localized queries, achieving better performance than existing methods while scaling to 256 input frames.

AINeutralarXiv – CS AI · Mar 267/10
🧠

Collaborative Causal Sensemaking: Closing the Complementarity Gap in Human-AI Decision Support

Researchers propose Collaborative Causal Sensemaking (CCS) as a new framework to improve human-AI collaboration in high-stakes decision making. The study identifies a 'complementarity gap' where current AI agents function as answer engines rather than true collaborative partners, limiting the effectiveness of human-AI teams.

AIBullisharXiv – CS AI · Mar 267/10
🧠

Bottlenecked Transformers: Periodic KV Cache Consolidation for Generalised Reasoning

Researchers introduce Bottlenecked Transformers, a new architecture that improves AI reasoning by up to 6.6 percentage points through periodic memory consolidation inspired by brain processes. The system uses a Cache Processor to rewrite key-value cache entries at reasoning step boundaries, achieving better performance on math reasoning benchmarks compared to standard Transformers.

AIBullisharXiv – CS AI · Mar 267/10
🧠

Reward Is Enough: LLMs Are In-Context Reinforcement Learners

Researchers demonstrate that large language models can perform reinforcement learning during inference through a new 'in-context RL' prompting framework. The method shows LLMs can optimize scalar reward signals to improve response quality across multiple rounds, achieving significant improvements on complex tasks like mathematical competitions and creative writing.

AIBearisharXiv – CS AI · Mar 267/10
🧠

Enhancing Jailbreak Attacks on LLMs via Persona Prompts

Researchers developed a genetic algorithm-based method using persona prompts to exploit large language models, reducing refusal rates by 50-70% across multiple LLMs. The study reveals significant vulnerabilities in AI safety mechanisms and demonstrates how these attacks can be enhanced when combined with existing methods.

AIBullisharXiv – CS AI · Mar 267/10
🧠

From Imperative to Declarative: Towards LLM-friendly OS Interfaces for Boosted Computer-Use Agents

Researchers have developed Declarative Model Interface (DMI), a new abstraction layer that transforms traditional GUIs into LLM-friendly interfaces for computer-use agents. Testing with Microsoft Office Suite showed 67% improvement in task success rates and 43.5% reduction in interaction steps, with over 61% of tasks completed in a single LLM call.

AIBullisharXiv – CS AI · Mar 267/10
🧠

Moonwalk: Inverse-Forward Differentiation

Researchers introduce Moonwalk, a new algorithm that solves backpropagation's memory limitations by eliminating the need to store intermediate activations during neural network training. The method uses vector-inverse-Jacobian products and submersive networks to reconstruct gradients in a forward sweep, enabling training of networks more than twice as deep under the same memory constraints.

← PrevPage 92 of 670Next →
◆ AI Mentions
🏢OpenAI
86×
🏢Nvidia
59×
🧠GPT-5
47×
🧠Claude
47×
🏢Anthropic
47×
🧠Gemini
32×
🧠ChatGPT
29×
🧠GPT-4
17×
🧠Llama
13×
🏢xAI
9×
🏢Google
8×
🧠Opus
8×
🧠Sonnet
7×
🏢Meta
7×
🧠Grok
7×
🏢Hugging Face
5×
🏢Microsoft
4×
🏢Perplexity
3×
🏢Cohere
2×
🧠Haiku
1×
▲ Trending Tags
1#iran6772#ai6173#market4974#geopolitical4605#trump1636#security1207#openai868#artificial-intelligence719#nvidia5910#google5711#china5612#inflation5413#fed4114#russia3415#meta32
Tag Sentiment
#iran677 articles
#ai617 articles
#market497 articles
#geopolitical460 articles
#trump163 articles
#security120 articles
#openai86 articles
#artificial-intelligence71 articles
#nvidia59 articles
#google57 articles
BullishNeutralBearish
Stay Updated
Models, papers, tools
Tag Connections
#geopolitical↔#iran
314
#iran↔#market
216
#geopolitical↔#market
176
#iran↔#trump
114
#ai↔#market
66
#market↔#trump
65
#ai↔#artificial-intelligence
63
#geopolitical↔#trump
60
#ai↔#google
51
#ai↔#openai
44
Filters
Sentiment
Importance
Sort
📡 See all 70+ sources
y0.exchange
Your AI agent for DeFi
Connect Claude or GPT to your wallet. AI reads balances, proposes swaps and bridges — you approve. Your keys never leave your device.
8 MCP tools · 15 chains · $0 fees
Connect Wallet to AI →How it works →
Viewing: AI Pulse feed
Filters
Sentiment
Importance
Sort
Stay Updated
Models, papers, tools
y0news
y0.exchangeLaunch AppDigestsSourcesAboutRSSAI NewsCrypto News
© 2026 y0.exchange