y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#language-models News & Analysis

350 articles tagged with #language-models. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

350 articles
AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

Bilinear representation mitigates reversal curse and enables consistent model editing

Researchers have identified that the 'reversal curse' in language models - their inability to infer 'B is A' from 'A is B' - can be overcome through bilinear representation structures. Training models on synthetic relational knowledge graphs creates internal geometries that enable consistent model editing and logical inference of reverse facts.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

A cross-species neural foundation model for end-to-end speech decoding

Researchers developed a new Brain-to-Text (BIT) framework that uses cross-species neural foundation models to decode speech from brain activity with significantly improved accuracy. The system reduces word error rates from 24.69% to 10.22% compared to previous methods and enables seamless translation of both attempted and imagined speech into text.

AIBearisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

On The Fragility of Benchmark Contamination Detection in Reasoning Models

New research reveals that benchmark contamination in language reasoning models (LRMs) is extremely difficult to detect, allowing developers to easily inflate performance scores on public leaderboards. The study shows that reinforcement learning methods like GRPO and PPO can effectively conceal contamination signals, undermining the integrity of AI model evaluations.

$NEAR
AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

On the Reasoning Abilities of Masked Diffusion Language Models

New research demonstrates that Masked Diffusion Models (MDMs) for text generation are computationally equivalent to chain-of-thought augmented transformers in finite-precision settings. The study proves MDMs can solve all reasoning problems that CoT transformers can, while being more efficient for certain problem classes due to parallel generation capabilities.

AINeutralarXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Steering Evaluation-Aware Language Models to Act Like They Are Deployed

Researchers demonstrate a technique using steering vectors to suppress evaluation-awareness in large language models, preventing them from adjusting their behavior during safety evaluations. The method makes models act as they would during actual deployment rather than performing differently when they detect they're being tested.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

ExGRPO: Learning to Reason from Experience

Researchers introduce ExGRPO, a new framework that improves AI reasoning by reusing and prioritizing valuable training experiences based on correctness and entropy. The method shows consistent performance gains of +3.5-7.6 points over standard approaches across multiple model sizes while providing more stable training.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Bridging Draft Policy Misalignment: Group Tree Optimization for Speculative Decoding

Researchers introduce Group Tree Optimization (GTO), a new training method that improves speculative decoding for large language models by aligning draft model training with actual decoding policies. GTO achieves 7.4% better acceptance length and 7.7% additional speedup over existing state-of-the-art methods across multiple benchmarks and LLMs.

AINeutralarXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

VeriTrail: Closed-Domain Hallucination Detection with Traceability

Researchers have developed VeriTrail, the first closed-domain hallucination detection method that can trace where AI-generated misinformation originates in multi-step processes. The system addresses a critical problem where language models generate unsubstantiated content even when instructed to stick to source material, with the risk being higher in complex multi-step generative processes.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

SPIRAL: Self-Play on Zero-Sum Games Incentivizes Reasoning via Multi-Agent Multi-Turn Reinforcement Learning

Researchers introduce SPIRAL, a self-play reinforcement learning framework that enables language models to develop reasoning capabilities by playing zero-sum games against themselves without human supervision. The system improves performance by up to 10% across 8 reasoning benchmarks on multiple model families including Qwen and Llama.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

mCLM: A Modular Chemical Language Model that Generates Functional and Makeable Molecules

Researchers developed mCLM, a 3-billion parameter modular Chemical Language Model that generates functional molecules compatible with automated synthesis by tokenizing at the building block level rather than individual atoms. The AI system outperformed larger models including GPT-5 in creating synthesizable drug candidates and can iteratively improve failed clinical trial compounds.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

Language Agents for Hypothesis-driven Clinical Decision Making with Reinforcement Learning

Researchers developed LA-CDM, a language agent that uses reinforcement learning to support clinical decision-making by iteratively requesting tests and generating hypotheses for diagnosis. The system was trained using a hybrid approach combining supervised and reinforcement learning, and tested on real-world data covering four abdominal diseases.

AIBullisharXiv โ€“ CS AI ยท Mar 37/105
๐Ÿง 

Elo-Evolve: A Co-evolutionary Framework for Language Model Alignment

Researchers introduce Elo-Evolve, a new framework for training AI language models using dynamic multi-agent competition instead of static reward functions. The method achieves 4.5x noise reduction and demonstrates superior performance compared to traditional alignment approaches when tested on Qwen2.5-7B models.

AINeutralarXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Optimal Sparsity of Mixture-of-Experts Language Models for Reasoning Tasks

Researchers analyzed Mixture-of-Experts (MoE) language models to determine optimal sparsity levels for different tasks. They found that reasoning tasks require balancing active compute (FLOPs) with optimal data-to-parameter ratios, while memorization tasks benefit from more parameters regardless of sparsity.

AINeutralarXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Control Tax: The Price of Keeping AI in Check

Researchers introduce 'Control Tax' - a framework to quantify the operational and financial costs of implementing AI safety oversight mechanisms. The study provides theoretical models and empirical cost estimates to help organizations balance AI safety measures with economic feasibility in real-world deployments.

AIBearisharXiv โ€“ CS AI ยท Feb 277/107
๐Ÿง 

Obscure but Effective: Classical Chinese Jailbreak Prompt Optimization via Bio-Inspired Search

Researchers developed CC-BOS, a framework that uses classical Chinese text to conduct more effective jailbreak attacks on Large Language Models. The method exploits the conciseness and obscurity of classical Chinese to bypass safety constraints, using bio-inspired optimization techniques to automatically generate adversarial prompts.

AIBullisharXiv โ€“ CS AI ยท Feb 277/106
๐Ÿง 

Intelligence per Watt: Measuring Intelligence Efficiency of Local AI

Researchers propose 'Intelligence per Watt' (IPW) as a metric to measure AI efficiency, finding that local AI models can handle 71.3% of queries while being 1.4x more energy efficient than cloud alternatives. The study demonstrates that smaller local language models (โ‰ค20B parameters) can redistribute computational demand from centralized cloud infrastructure.

AIBullisharXiv โ€“ CS AI ยท Feb 277/106
๐Ÿง 

Supervised Reinforcement Learning: From Expert Trajectories to Step-wise Reasoning

Researchers propose Supervised Reinforcement Learning (SRL), a new training framework that helps small-scale language models solve complex multi-step reasoning problems by generating internal reasoning monologues and providing step-wise rewards. SRL outperforms traditional Supervised Fine-Tuning and Reinforcement Learning approaches, enabling smaller models to tackle previously unlearnable problems.

AINeutralarXiv โ€“ CS AI ยท Feb 277/103
๐Ÿง 

The Tool Decathlon: Benchmarking Language Agents for Diverse, Realistic, and Long-Horizon Task Execution

Researchers introduce Tool Decathlon (Toolathlon), a comprehensive benchmark for evaluating AI language agents across 32 software applications and 604 tools in realistic, multi-step scenarios. The benchmark reveals significant limitations in current AI models, with the best performer (Claude-4.5-Sonnet) achieving only 38.6% success rate on complex, real-world tasks.

AIBullisharXiv โ€“ CS AI ยท Feb 277/106
๐Ÿง 

Affine-Scaled Attention: Towards Flexible and Stable Transformer Attention

Researchers propose Affine-Scaled Attention, a new mechanism that improves Transformer model training stability by introducing flexible scaling and bias terms to attention weights. The approach shows consistent improvements in optimization behavior and downstream task performance compared to standard softmax attention across multiple language model sizes.

AIBullisharXiv โ€“ CS AI ยท Feb 277/106
๐Ÿง 

Discovery of Interpretable Physical Laws in Materials via Language-Model-Guided Symbolic Regression

Researchers have developed a new framework that uses large language models to guide symbolic regression in discovering interpretable physical laws from high-dimensional materials data. The method reduces the search space by approximately 10^5 times compared to traditional approaches and successfully identified novel formulas for key properties of perovskite materials.

AIBullisharXiv โ€“ CS AI ยท Feb 277/106
๐Ÿง 

Predicting LLM Reasoning Performance with Small Proxy Model

Researchers introduce rBridge, a method that enables small AI models (โ‰ค1B parameters) to effectively predict the reasoning performance of much larger language models. This breakthrough could reduce dataset optimization costs by over 100x while maintaining strong correlation with large-model performance across reasoning benchmarks.

AIBullisharXiv โ€“ CS AI ยท Feb 277/105
๐Ÿง 

Cost-of-Pass: An Economic Framework for Evaluating Language Models

Researchers developed a new economic framework called 'cost-of-pass' to evaluate AI language models by combining accuracy with inference costs. The study found that lightweight models are most cost-effective for basic tasks while reasoning models excel at complex problems, with costs for complex quantitative tasks roughly halving every few months.

AINeutralarXiv โ€“ CS AI ยท Feb 277/105
๐Ÿง 

Transformers converge to invariant algorithmic cores

Researchers have discovered that transformer models, despite different training runs producing different weights, converge to the same compact 'algorithmic cores' - low-dimensional subspaces essential for task performance. The study shows these invariant structures persist across different scales and training runs, suggesting transformer computations are organized around shared algorithmic patterns rather than implementation-specific details.

AIBullishMIT News โ€“ AI ยท Dec 127/107
๐Ÿง 

Enabling small language models to solve complex reasoning tasks

The DisCIPL system represents a breakthrough in AI coordination, enabling small language models to collaborate on complex reasoning tasks like itinerary planning and budgeting. This 'self-steering' approach allows multiple smaller models to work together with constraints, potentially offering more efficient alternatives to large monolithic AI systems.