y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All26,359🧠AI11,672⛓️Crypto9,663💎DeFi989🤖AI × Crypto505📰General3,530
🧠

AI

11,672 AI articles curated from 50+ sources with AI-powered sentiment analysis, importance scoring, and key takeaways.

11672 articles
AIBullishAI News · Mar 57/10
🧠

JPMorgan expands AI investment as tech spending nears $20B

JPMorgan Chase is significantly expanding its AI investment, with technology spending projected to reach $19.8 billion by 2026. This reflects a broader enterprise trend where AI is transitioning from experimental pilot projects to core business infrastructure across large corporations.

AIBullishOpenAI News · Mar 56/10
🧠

Ensuring AI use in education leads to opportunity

OpenAI announces new educational tools, certifications, and measurement resources designed to help schools and universities address AI capability gaps. The initiative aims to expand educational opportunities by providing institutions with better resources to integrate AI into their curricula.

🏢 OpenAI
AINeutralarXiv – CS AI · Mar 57/10
🧠

Escaping the BLEU Trap: A Signal-Grounded Framework with Decoupled Semantic Guidance for EEG-to-Text Decoding

Researchers propose SemKey, a novel framework that addresses key limitations in EEG-to-text decoding by preventing hallucinations and improving semantic fidelity through decoupled guidance objectives. The system redesigns neural encoder-LLM interaction and introduces new evaluation metrics beyond BLEU scores to achieve state-of-the-art performance in brain-computer interfaces.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Towards Self-Robust LLMs: Intrinsic Prompt Noise Resistance via CoIPO

Researchers propose CoIPO (Contrastive Learning-based Inverse Direct Preference Optimization), a new method to improve Large Language Model robustness against noisy or imperfect user prompts. The approach enhances LLMs' intrinsic ability to handle prompt variations without relying on external preprocessing tools, showing significant accuracy improvements on benchmark tests.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Old Habits Die Hard: How Conversational History Geometrically Traps LLMs

Researchers introduce History-Echoes, a framework revealing how large language models become trapped by their conversational history, with past interactions creating geometric constraints in latent space that bias future responses. The study demonstrates that behavioral persistence in LLMs manifests as mathematical traps where previous hallucinations and responses influence subsequent model behavior across multiple model families and datasets.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Can Large Language Models Derive New Knowledge? A Dynamic Benchmark for Biological Knowledge Discovery

Researchers have developed DBench-Bio, a dynamic benchmark system that automatically evaluates AI's ability to discover new biological knowledge using a three-stage pipeline of data acquisition, question-answer extraction, and quality filtering. The benchmark addresses the critical problem of data contamination in static datasets and provides monthly updates across 12 biomedical domains, revealing current limitations in state-of-the-art AI models' knowledge discovery capabilities.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Quantum-Inspired Self-Attention in a Large Language Model

Researchers developed a quantum-inspired self-attention (QISA) mechanism and integrated it into GPT-1's language modeling pipeline, marking the first such integration in autoregressive language models. The QISA mechanism demonstrated significant performance improvements over standard self-attention, achieving 15.5x better character error rate and 13x better cross-entropy loss with only 2.6x longer inference time.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Controlling Chat Style in Language Models via Single-Direction Editing

Researchers developed a training-free method to control stylistic attributes in large language models by identifying that different styles are encoded as linear directions in the model's activation space. The approach enables precise style control while preserving core capabilities and supports linear style composition across over a dozen tested models.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Discern Truth from Falsehood: Reducing Over-Refusal via Contrastive Refinement

Researchers introduce DCR (Discernment via Contrastive Refinement), a new method to reduce over-refusal in safety-aligned large language models. The approach helps LLMs better distinguish between genuinely toxic and seemingly toxic prompts, maintaining safety while improving helpfulness without degrading general capabilities.

AINeutralarXiv – CS AI · Mar 56/10
🧠

Automated Concept Discovery for LLM-as-a-Judge Preference Analysis

Researchers developed automated methods to discover biases in Large Language Models when used as judges, analyzing over 27,000 paired responses. The study found LLMs exhibit systematic biases including preference for refusing sensitive requests more than humans, favoring concrete and empathetic responses, and showing bias against certain legal guidance.

AIBearisharXiv – CS AI · Mar 56/10
🧠

$\tau$-Knowledge: Evaluating Conversational Agents over Unstructured Knowledge

Researchers introduced τ-Knowledge, a new benchmark for evaluating AI conversational agents in knowledge-intensive environments, specifically testing their ability to retrieve and apply unstructured domain knowledge. Even frontier AI models achieved only 25.5% success rates when navigating complex fintech customer support scenarios with 700 interconnected knowledge documents.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Agentics 2.0: Logical Transduction Algebra for Agentic Data Workflows

Researchers have introduced Agentics 2.0, a Python framework for building enterprise-grade AI agent workflows using logical transduction algebra. The framework addresses reliability, scalability, and observability challenges in deploying agentic AI systems beyond research prototypes.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Certainty robustness: Evaluating LLM stability under self-challenging prompts

Researchers introduce the Certainty Robustness Benchmark, a new evaluation framework that tests how large language models handle challenges to their responses in interactive settings. The study reveals significant differences in how AI models balance confidence and adaptability when faced with prompts like "Are you sure?" or "You are wrong!", identifying a critical new dimension for AI evaluation.

AIBullisharXiv – CS AI · Mar 56/10
🧠

DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following

Researchers introduce DIALEVAL, a new automated framework that uses dual LLM agents to evaluate how well AI models follow instructions. The system achieves 90.38% accuracy by breaking down instructions into verifiable components and applying type-specific evaluation criteria, showing 26.45% error reduction over existing methods.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Phi-4-reasoning-vision-15B Technical Report

Researchers released Phi-4-reasoning-vision-15B, a compact open-weight multimodal AI model that combines vision and language capabilities with strong performance in scientific and mathematical reasoning. The model demonstrates that careful architecture design and high-quality data curation can enable smaller models to achieve competitive performance with less computational resources.

AIBullisharXiv – CS AI · Mar 57/10
🧠

AgentSelect: Benchmark for Narrative Query-to-Agent Recommendation

Researchers introduce AgentSelect, a comprehensive benchmark for recommending AI agent configurations based on narrative queries. The benchmark aggregates over 111,000 queries and 107,000 deployable agents from 40+ sources to address the critical gap in selecting optimal LLM agent setups for specific tasks.

AIBullisharXiv – CS AI · Mar 56/10
🧠

From Exact Hits to Close Enough: Semantic Caching for LLM Embeddings

Researchers propose semantic caching solutions for large language models to improve response times and reduce costs by reusing semantically similar requests. The study proves that optimal offline semantic caching is NP-hard and introduces polynomial-time heuristics and online policies combining recency, frequency, and locality factors.

AINeutralarXiv – CS AI · Mar 56/10
🧠

LifeBench: A Benchmark for Long-Horizon Multi-Source Memory

Researchers introduce LifeBench, a new AI benchmark that tests long-term memory systems by requiring integration of both declarative and non-declarative memory across extended timeframes. Current state-of-the-art memory systems achieve only 55.2% accuracy on this challenging benchmark, highlighting significant gaps in AI's ability to handle complex, multi-source memory tasks.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Draft-Conditioned Constrained Decoding for Structured Generation in LLMs

Researchers introduce Draft-Conditioned Constrained Decoding (DCCD), a training-free method that improves structured output generation in large language models by up to 24 percentage points. The technique uses a two-step process that first generates an unconstrained draft, then applies constraints to ensure valid outputs like JSON and API calls.

AIBullisharXiv – CS AI · Mar 57/10
🧠

HumanLM: Simulating Users with State Alignment Beats Response Imitation

Researchers introduce HumanLM, a novel AI training framework that creates user simulators by aligning psychological states rather than just imitating response patterns. The system achieved 16.3% improvement in alignment scores across six datasets with 26k users and 216k responses, demonstrating superior ability to simulate real human behavior.

AIBullisharXiv – CS AI · Mar 56/10
🧠

A Rubric-Supervised Critic from Sparse Real-World Outcomes

Researchers propose a new framework called Critic Rubrics to bridge the gap between academic coding agent benchmarks and real-world applications. The system learns from sparse, noisy human interaction data using 24 behavioral features and shows significant improvements in code generation tasks including 15.9% better reranking performance on SWE-bench.

AIBullisharXiv – CS AI · Mar 56/10
🧠

TATRA: Training-Free Instance-Adaptive Prompting Through Rephrasing and Aggregation

Researchers introduce TATRA, a training-free prompting method for Large Language Models that creates instance-specific few-shot prompts without requiring labeled training data. The method achieves state-of-the-art performance on mathematical reasoning benchmarks like GSM8K and DeepMath, matching or outperforming existing prompt optimization methods that rely on expensive training processes.

AIBullisharXiv – CS AI · Mar 56/10
🧠

A Dual-Helix Governance Approach Towards Reliable Agentic AI for WebGIS Development

Researchers propose a dual-helix governance framework to address AI agent reliability issues in WebGIS development, implementing a 3-track architecture that achieved 51% reduction in code complexity. The framework uses knowledge graphs and self-learning cycles to overcome LLM limitations like context constraints and instruction failures.

← PrevPage 58 of 467Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined