y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🧠All17,509🧠AI12,683🤖AI × Crypto529📰General4,297
Home/AI Pulse

AI Pulse News

Models, papers, tools. 17,520 articles with AI-powered sentiment analysis and key takeaways.

17520 articles
AINeutralarXiv – CS AI · Mar 57/10
🧠

Can Large Language Models Derive New Knowledge? A Dynamic Benchmark for Biological Knowledge Discovery

Researchers have developed DBench-Bio, a dynamic benchmark system that automatically evaluates AI's ability to discover new biological knowledge using a three-stage pipeline of data acquisition, question-answer extraction, and quality filtering. The benchmark addresses the critical problem of data contamination in static datasets and provides monthly updates across 12 biomedical domains, revealing current limitations in state-of-the-art AI models' knowledge discovery capabilities.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Towards Self-Robust LLMs: Intrinsic Prompt Noise Resistance via CoIPO

Researchers propose CoIPO (Contrastive Learning-based Inverse Direct Preference Optimization), a new method to improve Large Language Model robustness against noisy or imperfect user prompts. The approach enhances LLMs' intrinsic ability to handle prompt variations without relying on external preprocessing tools, showing significant accuracy improvements on benchmark tests.

AIBullisharXiv – CS AI · Mar 56/10
🧠

DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following

Researchers introduce DIALEVAL, a new automated framework that uses dual LLM agents to evaluate how well AI models follow instructions. The system achieves 90.38% accuracy by breaking down instructions into verifiable components and applying type-specific evaluation criteria, showing 26.45% error reduction over existing methods.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Quantum-Inspired Self-Attention in a Large Language Model

Researchers developed a quantum-inspired self-attention (QISA) mechanism and integrated it into GPT-1's language modeling pipeline, marking the first such integration in autoregressive language models. The QISA mechanism demonstrated significant performance improvements over standard self-attention, achieving 15.5x better character error rate and 13x better cross-entropy loss with only 2.6x longer inference time.

AINeutralarXiv – CS AI · Mar 56/10
🧠

Automated Concept Discovery for LLM-as-a-Judge Preference Analysis

Researchers developed automated methods to discover biases in Large Language Models when used as judges, analyzing over 27,000 paired responses. The study found LLMs exhibit systematic biases including preference for refusing sensitive requests more than humans, favoring concrete and empathetic responses, and showing bias against certain legal guidance.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Discern Truth from Falsehood: Reducing Over-Refusal via Contrastive Refinement

Researchers introduce DCR (Discernment via Contrastive Refinement), a new method to reduce over-refusal in safety-aligned large language models. The approach helps LLMs better distinguish between genuinely toxic and seemingly toxic prompts, maintaining safety while improving helpfulness without degrading general capabilities.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Controlling Chat Style in Language Models via Single-Direction Editing

Researchers developed a training-free method to control stylistic attributes in large language models by identifying that different styles are encoded as linear directions in the model's activation space. The approach enables precise style control while preserving core capabilities and supports linear style composition across over a dozen tested models.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Controllable and explainable personality sliders for LLMs at inference time

Researchers propose Sequential Adaptive Steering (SAS), a new framework for controlling Large Language Model personalities at inference time without retraining. The method uses orthogonalized steering vectors to enable precise, multi-dimensional personality control by adjusting coefficients, validated on Big Five personality traits.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Test-Time Meta-Adaptation with Self-Synthesis

Researchers introduce MASS, a meta-learning framework that enables large language models to self-adapt at test time by generating synthetic training data and performing targeted self-updates. The system uses bilevel optimization to meta-learn data-attribution signals and optimize synthetic data through scalable meta-gradients, showing effectiveness in mathematical reasoning tasks.

AIBullisharXiv – CS AI · Mar 57/10
🧠

AutoHarness: improving LLM agents by automatically synthesizing a code harness

Researchers developed AutoHarness, a technique where smaller LLMs like Gemini-2.5-Flash can automatically generate code harnesses to prevent illegal moves in games, outperforming larger models like Gemini-2.5-Pro and GPT-5.2-High. The method eliminates 78% of failures attributed to illegal moves in chess competitions and demonstrates superior performance across 145 different games.

🧠 Gemini
AINeutralarXiv – CS AI · Mar 57/10
🧠

Certainty robustness: Evaluating LLM stability under self-challenging prompts

Researchers introduce the Certainty Robustness Benchmark, a new evaluation framework that tests how large language models handle challenges to their responses in interactive settings. The study reveals significant differences in how AI models balance confidence and adaptability when faced with prompts like "Are you sure?" or "You are wrong!", identifying a critical new dimension for AI evaluation.

AIBullisharXiv – CS AI · Mar 56/10
🧠

PulseLM: A Foundation Dataset and Benchmark for PPG-Text Learning

Researchers introduced PulseLM, a large-scale dataset combining PPG cardiovascular sensor data with natural language processing for multimodal AI models. The dataset contains 1.31 million PPG segments with 3.15 million question-answer pairs, designed to enable language-based physiological reasoning in healthcare AI applications.

AIBearisharXiv – CS AI · Mar 57/10
🧠

When Shallow Wins: Silent Failures and the Depth-Accuracy Paradox in Latent Reasoning

Research reveals that state-of-the-art AI mathematical reasoning models like Qwen2.5-Math-7B achieve 61% accuracy primarily through unreliable computational pathways, with only 18.4% using stable reasoning. The study exposes that 81.6% of correct predictions come from inconsistent methods and 8.8% are confident but incorrect outputs.

AINeutralarXiv – CS AI · Mar 57/10
🧠

On Google's SynthID-Text LLM Watermarking System: Theoretical Analysis and Empirical Validation

Researchers have conducted the first theoretical analysis of Google's SynthID-Text watermarking system, revealing vulnerabilities in its detection methods and proposing attacks that can break the system. The study identifies weaknesses in the mean score detection approach and demonstrates that the Bayesian score offers better robustness, while establishing optimal parameters for watermark detection.

AIBullisharXiv – CS AI · Mar 56/10
🧠

PRIVATEEDIT: A Privacy-Preserving Pipeline for Face-Centric Generative Image Editing

Researchers have developed PRIVATEEDIT, a privacy-preserving pipeline for face-centric image editing that keeps biometric data on-device rather than uploading to third-party services. The system uses local segmentation and masking to separate identity-sensitive regions from editable content, allowing high-quality editing while maintaining user control over facial data.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs

Researchers discovered that Large Language Models become increasingly sparse in their internal representations when handling more difficult or out-of-distribution tasks. This sparsity mechanism appears to be an adaptive response that helps stabilize reasoning under challenging conditions, leading to the development of a new learning strategy called Sparsity-Guided Curriculum In-Context Learning (SG-ICL).

AIBullisharXiv – CS AI · Mar 57/10
🧠

MemSifter: Offloading LLM Memory Retrieval via Outcome-Driven Proxy Reasoning

MemSifter is a new AI framework that uses smaller proxy models to handle memory retrieval for large language models, addressing computational costs in long-term memory tasks. The system uses reinforcement learning to optimize retrieval accuracy and has been open-sourced with demonstrated performance improvements on benchmark tests.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Parallel Test-Time Scaling with Multi-Sequence Verifiers

Researchers introduce Multi-Sequence Verifier (MSV), a new technique that improves large language model performance by jointly processing multiple candidate solutions rather than scoring them individually. The system achieves better accuracy while reducing inference latency by approximately half through improved calibration and early-stopping strategies.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Ethical and Explainable AI in Reusable MLOps Pipelines

Researchers developed a unified MLOps framework that integrates ethical AI principles, reducing demographic bias from 0.31 to 0.04 while maintaining predictive accuracy. The system automatically blocks deployments and triggers retraining based on fairness metrics, demonstrating practical implementation of ethical AI in production environments.

AINeutralarXiv – CS AI · Mar 56/10
🧠

Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations

Research reveals that Large Language Models show varying vulnerabilities to different types of Chain-of-Thought reasoning perturbations, with math errors causing 50-60% accuracy loss in small models while unit conversion issues remain challenging even for the largest models. The study tested 13 models across parameter ranges from 3B to 1.5T parameters, finding that scaling provides protection against some perturbations but limited defense against dimensional reasoning tasks.

AIBullisharXiv – CS AI · Mar 56/10
🧠

LiteVLA-Edge: Quantized On-Device Multimodal Control for Embedded Robotics

Researchers developed LiteVLA-Edge, a deployment-oriented Vision-Language-Action model pipeline that enables fully on-device inference on embedded robotics hardware like Jetson Orin. The system achieves 150.5ms latency (6.6Hz) through FP32 fine-tuning combined with 4-bit quantization and GPU-accelerated inference, operating entirely offline within a ROS 2 framework.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Phys4D: Fine-Grained Physics-Consistent 4D Modeling from Video Diffusion

Researchers have developed Phys4D, a new pipeline that enhances video diffusion models with physics-consistent 4D world representations through a three-stage training process. The system addresses current limitations where AI-generated videos often exhibit physically implausible dynamics, using pseudo-supervised pretraining, physics-grounded fine-tuning, and reinforcement learning to improve spatiotemporal consistency.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Perfect score on IPhO 2025 theory by Gemini agent

Google's Gemini 3.1 Pro Preview achieved a perfect score on IPhO 2025 theory problems across five runs, surpassing previous AI performance that fell behind top human contestants. However, the researchers acknowledge potential data contamination since the model was released after the competition.

🧠 Gemini
AIBearisharXiv – CS AI · Mar 57/10
🧠

Sleeper Cell: Injecting Latent Malice Temporal Backdoors into Tool-Using LLMs

Researchers demonstrate a novel backdoor attack method called 'SFT-then-GRPO' that can inject hidden malicious behavior into AI agents while maintaining their performance on standard benchmarks. The attack creates 'sleeper agents' that appear benign but can execute harmful actions under specific trigger conditions, highlighting critical security vulnerabilities in the adoption of third-party AI models.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Non-Invasive Reconstruction of Intracranial EEG Across the Deep Temporal Lobe from Scalp EEG based on Conditional Normalizing Flow

Researchers developed NeuroFlowNet, a novel AI framework using Conditional Normalizing Flow to reconstruct deep brain EEG signals from non-invasive scalp measurements. This breakthrough enables analysis of deep temporal lobe brain activity without requiring invasive electrode implantation, potentially transforming neuroscience research and clinical diagnosis.

← PrevPage 141 of 701Next →
◆ AI Mentions
🏢OpenAI
98×
🏢Nvidia
63×
🧠GPT-5
43×
🏢Anthropic
40×
🧠Claude
38×
🧠Gemini
38×
🧠ChatGPT
20×
🧠Llama
19×
🧠GPT-4
18×
🏢Meta
11×
🏢xAI
9×
🏢Perplexity
9×
🧠Sonnet
9×
🏢Google
8×
🧠Opus
8×
🏢Microsoft
7×
🧠Grok
6×
🏢Hugging Face
6×
🧠o1
2×
🏢Cohere
2×
▲ Trending Tags
1#ai5382#iran5363#market3894#geopolitical3565#trump1236#security1027#openai948#artificial-intelligence709#inflation6210#nvidia6211#geopolitics5612#machine-learning5613#fed5314#google4715#china41
Tag Sentiment
#ai538 articles
#iran536 articles
#market389 articles
#geopolitical356 articles
#trump123 articles
#security102 articles
#openai94 articles
#artificial-intelligence70 articles
#inflation62 articles
#nvidia62 articles
BullishNeutralBearish
Stay Updated
Models, papers, tools
Tag Connections
#geopolitical↔#iran
246
#iran↔#market
163
#geopolitical↔#market
135
#iran↔#trump
87
#ai↔#artificial-intelligence
58
#ai↔#market
56
#market↔#trump
46
#geopolitical↔#trump
46
#ai↔#openai
40
#market↔#nvidia
39
Filters
Sentiment
Importance
Sort
📡 See all 70+ sources
y0.exchange
Your AI agent for DeFi
Connect Claude or GPT to your wallet. AI reads balances, proposes swaps and bridges — you approve. Your keys never leave your device.
8 MCP tools · 15 chains · $0 fees
Connect Wallet to AI →How it works →
Viewing: AI Pulse feed
Filters
Sentiment
Importance
Sort
Stay Updated
Models, papers, tools
y0news
y0.exchangeLaunch AppDigestsSourcesAboutRSSAI NewsCrypto News
© 2026 y0.exchange