y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#mathematical-reasoning News & Analysis

52 articles tagged with #mathematical-reasoning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

52 articles
AIBullisharXiv โ€“ CS AI ยท 6d ago7/10
๐Ÿง 

The Master Key Hypothesis: Unlocking Cross-Model Capability Transfer via Linear Subspace Alignment

Researchers propose the Master Key Hypothesis, suggesting that AI model capabilities can be transferred across different model scales without retraining through linear subspace alignment. The UNLOCK framework demonstrates training-free capability transfer, achieving significant accuracy improvements such as 12.1% gains on mathematical reasoning tasks when transferring from larger to smaller models.

AIBearisharXiv โ€“ CS AI ยท 6d ago7/10
๐Ÿง 

Riemann-Bench: A Benchmark for Moonshot Mathematics

Researchers introduced Riemann-Bench, a private benchmark of 25 expert-curated mathematics problems designed to evaluate AI systems on research-level reasoning beyond competition mathematics. The benchmark reveals that all frontier AI models currently score below 10%, exposing a significant gap between olympiad-level problem solving and genuine mathematical research capabilities.

AIBullisharXiv โ€“ CS AI ยท Apr 77/10
๐Ÿง 

QED-Nano: Teaching a Tiny Model to Prove Hard Theorems

Researchers developed QED-Nano, a 4B parameter AI model that achieves competitive performance on Olympiad-level mathematical proofs despite being much smaller than proprietary systems. The model uses a three-stage training approach including supervised fine-tuning, reinforcement learning, and reasoning cache expansion to match larger models at a fraction of the inference cost.

๐Ÿง  Gemini
AIBullisharXiv โ€“ CS AI ยท Mar 267/10
๐Ÿง 

HDPO: Hybrid Distillation Policy Optimization via Privileged Self-Distillation

Researchers introduce Hybrid Distillation Policy Optimization (HDPO), a new method that improves large language model training for mathematical reasoning by addressing 'cliff prompts' where standard reinforcement learning fails. The technique uses privileged self-distillation to provide learning signals for previously unsolvable problems, showing measurable improvements in coverage metrics while maintaining accuracy.

AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

TTSR: Test-Time Self-Reflection for Continual Reasoning Improvement

Researchers introduce TTSR, a new framework that enables AI models to improve their reasoning abilities during test time by having a single model alternate between student and teacher roles. The system allows models to learn from their mistakes by analyzing failed reasoning attempts and generating targeted practice questions for continuous improvement.

AINeutralarXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations

Research reveals that Large Language Models show varying vulnerabilities to different types of Chain-of-Thought reasoning perturbations, with math errors causing 50-60% accuracy loss in small models while unit conversion issues remain challenging even for the largest models. The study tested 13 models across parameter ranges from 3B to 1.5T parameters, finding that scaling provides protection against some perturbations but limited defense against dimensional reasoning tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

Test-Time Meta-Adaptation with Self-Synthesis

Researchers introduce MASS, a meta-learning framework that enables large language models to self-adapt at test time by generating synthetic training data and performing targeted self-updates. The system uses bilevel optimization to meta-learn data-attribution signals and optimize synthetic data through scalable meta-gradients, showing effectiveness in mathematical reasoning tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

TATRA: Training-Free Instance-Adaptive Prompting Through Rephrasing and Aggregation

Researchers introduce TATRA, a training-free prompting method for Large Language Models that creates instance-specific few-shot prompts without requiring labeled training data. The method achieves state-of-the-art performance on mathematical reasoning benchmarks like GSM8K and DeepMath, matching or outperforming existing prompt optimization methods that rely on expensive training processes.

AIBearisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

When Shallow Wins: Silent Failures and the Depth-Accuracy Paradox in Latent Reasoning

Research reveals that state-of-the-art AI mathematical reasoning models like Qwen2.5-Math-7B achieve 61% accuracy primarily through unreliable computational pathways, with only 18.4% using stable reasoning. The study exposes that 81.6% of correct predictions come from inconsistent methods and 8.8% are confident but incorrect outputs.

AIBullisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

Phi-4-reasoning-vision-15B Technical Report

Researchers released Phi-4-reasoning-vision-15B, a compact open-weight multimodal AI model that combines vision and language capabilities with strong performance in scientific and mathematical reasoning. The model demonstrates that careful architecture design and high-quality data curation can enable smaller models to achieve competitive performance with less computational resources.

AIBullisharXiv โ€“ CS AI ยท Mar 46/105
๐Ÿง 

CORE: Concept-Oriented Reinforcement for Bridging the Definition-Application Gap in Mathematical Reasoning

Researchers introduce CORE (Concept-Oriented REinforcement), a new training framework that improves large language models' mathematical reasoning by bridging the gap between memorizing definitions and applying concepts. The method uses concept-aligned quizzes and concept-primed trajectories to provide fine-grained supervision, showing consistent improvements over traditional training approaches across multiple benchmarks.

AIBullisharXiv โ€“ CS AI ยท Mar 47/104
๐Ÿง 

PRISM: Pushing the Frontier of Deep Think via Process Reward Model-Guided Inference

Researchers introduce PRISM, a new AI inference algorithm that uses Process Reward Models to guide deep reasoning systems. The method significantly improves performance on mathematical and scientific benchmarks by treating candidate solutions as particles in an energy landscape and using score-guided refinement to concentrate on higher-quality reasoning paths.

AIBullisharXiv โ€“ CS AI ยท Mar 47/105
๐Ÿง 

NeuroProlog: Multi-Task Fine-Tuning for Neurosymbolic Mathematical Reasoning via the Cocktail Effect

Researchers introduce NeuroProlog, a neurosymbolic framework that improves mathematical reasoning in Large Language Models by converting math problems into executable Prolog programs. The multi-task 'Cocktail' training approach shows significant accuracy improvements of 3-5% across different model sizes, with larger models demonstrating better error correction capabilities.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

LEDOM: Reverse Language Model

Researchers have developed LEDOM, an open-source reverse autoregressive language model that trains right-to-left instead of the traditional left-to-right approach. The model demonstrates unique capabilities like abductive inference and question synthesis, and when combined with forward models through 'Reverse Reward' scoring, achieves significant performance gains of up to 15% on mathematical reasoning tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

LaDiR: Latent Diffusion Enhances LLMs for Text Reasoning

Researchers introduce LaDiR (Latent Diffusion Reasoner), a novel framework that combines continuous latent representation with iterative refinement capabilities to enhance Large Language Models' reasoning abilities. The system uses a Variational Autoencoder to encode reasoning steps and a latent diffusion model for parallel generation of diverse reasoning trajectories, showing improved accuracy and interpretability in mathematical reasoning benchmarks.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

AgentMath: Empowering Mathematical Reasoning for Large Language Models via Tool-Augmented Agent

Researchers introduced AgentMath, a new AI framework that combines language models with code interpreters to solve complex mathematical problems more efficiently than current Large Reasoning Models. The system achieves state-of-the-art performance on mathematical competition benchmarks, with AgentMath-30B-A3B reaching 90.6% accuracy on AIME24 while remaining competitive with much larger models like OpenAI-o3.

AINeutralarXiv โ€“ CS AI ยท Mar 37/105
๐Ÿง 

DAG-Math: Graph-of-Thought Guided Mathematical Reasoning in LLMs

Researchers introduce DAG-Math, a new framework for evaluating mathematical reasoning in Large Language Models that models Chain-of-Thought as rule-based processes over directed acyclic graphs. The framework includes a 'logical closeness' metric that reveals significant differences in reasoning quality between LLM families, even when final answer accuracy appears comparable.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Rewriting Pre-Training Data Boosts LLM Performance in Math and Code

Researchers released two open-source datasets, SwallowCode and SwallowMath, that significantly improve large language model performance in coding and mathematics through systematic data rewriting rather than filtering. The datasets boost Llama-3.1-8B performance by +17.0 on HumanEval for coding and +12.4 on GSM8K for math tasks.

AINeutralarXiv โ€“ CS AI ยท Feb 277/106
๐Ÿง 

Why Pass@k Optimization Can Degrade Pass@1: Prompt Interference in LLM Post-training

Researchers identify a critical trade-off in AI model training where optimizing for Pass@k metrics (multiple attempts) degrades Pass@1 performance (single attempt). The study reveals this occurs due to gradient conflicts when the training process reweights toward low-success prompts, creating interference that hurts single-shot performance.

AINeutralarXiv โ€“ CS AI ยท Feb 277/107
๐Ÿง 

LeanCat: A Benchmark Suite for Formal Category Theory in Lean (Part I: 1-Categories)

Researchers introduced LeanCat, a benchmark comprising 100 category-theory tasks in Lean to test AI's formal theorem proving capabilities. State-of-the-art models achieved only 12% success rates, revealing significant limitations in abstract mathematical reasoning, while a new retrieval-augmented approach doubled performance to 24%.

AIBullishOpenAI News ยท May 317/109
๐Ÿง 

Improving mathematical reasoning with process supervision

Researchers have developed a new AI training method called 'process supervision' that rewards each correct reasoning step rather than just the final answer, achieving state-of-the-art performance in mathematical problem solving. This approach not only improves performance but also ensures the AI's reasoning process aligns with human-endorsed thinking patterns.

AIBullisharXiv โ€“ CS AI ยท 6d ago6/10
๐Ÿง 

$S^3$: Stratified Scaling Search for Test-Time in Diffusion Language Models

Researchers introduce Sยณ (Stratified Scaling Search), a test-time scaling method for diffusion language models that improves output quality by reallocating compute during the denoising process rather than simple best-of-K sampling. The technique uses a lightweight verifier to evaluate and selectively resample candidate trajectories at each step, demonstrating consistent performance gains across mathematical reasoning and knowledge tasks without requiring model retraining.

AIBearisharXiv โ€“ CS AI ยท 6d ago6/10
๐Ÿง 

Lost in Cultural Translation: Do LLMs Struggle with Math Across Cultural Contexts?

Researchers found that large language models experience accuracy drops of 0.3% to 5.9% when math problems are presented in unfamiliar cultural contexts, even when the underlying mathematical logic remains identical. Testing 14 models across culturally adapted variants of the GSM8K benchmark reveals that LLM mathematical reasoning is not culturally neutral, with errors stemming from both reasoning failures and calculation mistakes.

๐Ÿข OpenAI๐Ÿข Anthropic๐Ÿง  Claude
AIBullisharXiv โ€“ CS AI ยท Apr 76/10
๐Ÿง 

Vocabulary Dropout for Curriculum Diversity in LLM Co-Evolution

Researchers introduce vocabulary dropout, a technique to prevent diversity collapse in co-evolutionary language model training where one model generates problems and another solves them. The method sustains proposer diversity and improves mathematical reasoning performance by +4.4 points on average in Qwen3 models.

AIBearisharXiv โ€“ CS AI ยท Apr 66/10
๐Ÿง 

From Abstract to Contextual: What LLMs Still Cannot Do in Mathematics

A new study reveals that large language models, despite excelling at benchmark math problems, struggle significantly with contextual mathematical reasoning where problems are embedded in real-world scenarios. The research shows performance drops of 13-34 points for open-source models and 13-20 points for proprietary models when abstract math problems are presented in contextual settings.

Page 1 of 3Next โ†’