25 articles tagged with #llm-reasoning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AINeutralarXiv – CS AI · Apr 157/10
🧠Researchers introduce REL, a benchmark framework that evaluates relational reasoning in large language models by measuring Relational Complexity (RC)—the number of entities that must be simultaneously bound to apply a relation. The study reveals that frontier LLMs consistently degrade in performance as RC increases, exposing a fundamental limitation in higher-arity reasoning that persists even with increased compute and in-context learning.
AIBearisharXiv – CS AI · Apr 147/10
🧠A new study reveals that large language models fail at counterfactual reasoning when policy findings contradict intuitive expectations, despite performing well on obvious cases. The research demonstrates that chain-of-thought prompting paradoxically worsens performance on counter-intuitive scenarios, suggesting current LLMs engage in 'slow talking' rather than genuine deliberative reasoning.
AIBullisharXiv – CS AI · Apr 147/10
🧠FACT-E is a new evaluation framework that uses controlled perturbations to assess the faithfulness of Chain-of-Thought reasoning in large language models, addressing the problem of models generating seemingly coherent explanations with invalid intermediate steps. By measuring both internal chain consistency and answer alignment, FACT-E enables more reliable detection of flawed reasoning and selection of trustworthy reasoning trajectories for in-context learning.
AINeutralarXiv – CS AI · Apr 147/10
🧠Researchers challenge the assumption that longer reasoning chains always improve LLM performance, discovering that extended test-time compute leads to diminishing returns and 'overthinking' where models abandon correct answers. The study demonstrates that optimal compute allocation varies by problem difficulty, enabling significant efficiency gains without sacrificing accuracy.
AINeutralarXiv – CS AI · Apr 147/10
🧠Researchers demonstrate that a large language model's diversity profile—how probability mass spreads across different solution approaches—should determine whether reasoning strategies prioritize breadth or depth exploration. Testing on Qwen and Olmo model families reveals that lightweight refinement signals work well for low-diversity aligned models but offer limited value for high-diversity base models, suggesting optimal inference strategies must be model-specific rather than universal.
AINeutralarXiv – CS AI · Apr 147/10
🧠Researchers developed the first real-world benchmark for evaluating whether large language models can infer causal relationships from complex academic texts. The study reveals that LLMs struggle significantly with this task, with the best models achieving only 0.535 F1 scores, highlighting a critical gap in AI reasoning capabilities needed for AGI advancement.
AIBearisharXiv – CS AI · Apr 137/10
🧠Researchers have developed a 14-technique perturbation pipeline to test the robustness of large language models' reasoning capabilities on mathematical problems. Testing reveals that while frontier models maintain resilience, open-weight models experience catastrophic accuracy collapses up to 55%, and all tested models degrade when solving sequential problems in a single context window, suggesting fundamental architectural limitations in current reasoning systems.
🧠 Claude🧠 Opus
AIBullisharXiv – CS AI · Mar 117/10
🧠Researchers have developed Zipage, a new high-concurrency inference engine for large language models that uses Compressed PagedAttention to solve memory bottlenecks. The system achieves 95% performance of full KV inference engines while delivering over 2.1x speedup on mathematical reasoning tasks.
AIBullisharXiv – CS AI · Mar 37/105
🧠Researchers provide mathematical proof that implicit models can achieve greater expressive power through increased test-time computation, explaining how these memory-efficient architectures can match larger explicit networks. The study validates this scaling property across image reconstruction, scientific computing, operations research, and LLM reasoning domains.
AIBullisharXiv – CS AI · Apr 156/10
🧠Researchers propose Heuristic Classification of Thoughts (HCoT), a novel prompting method that integrates expert system heuristics into large language models to improve structured reasoning on complex problems. The approach addresses LLMs' stochastic token generation and decoupled reasoning mechanisms by using heuristic classification to guide and optimize decision trajectories, demonstrating superior performance and token efficiency compared to existing methods like Chain-of-Thoughts and Tree-of-Thoughts prompting.
AIBullisharXiv – CS AI · Apr 156/10
🧠Researchers introduce KG-Reasoner, an end-to-end framework that uses reinforcement learning to train large language models to perform multi-hop reasoning over knowledge graphs without decomposing tasks into isolated pipeline steps. The approach demonstrates competitive or superior performance across eight reasoning benchmarks by enabling LLMs to dynamically explore reasoning paths and backtrack when necessary.
AINeutralarXiv – CS AI · Apr 156/10
🧠Researchers propose a graph-based soft prompting framework that enables LLMs to reason over incomplete knowledge graphs by processing subgraph structures rather than explicit node paths, achieving state-of-the-art results on multi-hop question-answering benchmarks while reducing computational costs through a two-stage inference approach.
AINeutralarXiv – CS AI · Apr 146/10
🧠Researchers introduce ConflictQA, a benchmark revealing that large language models struggle with conflicting information across different knowledge sources (text vs. knowledge graphs) in retrieval-augmented generation systems. The study proposes XoT, an explanation-based framework to improve faithful reasoning when LLMs encounter contradictory evidence.
AIBullisharXiv – CS AI · Apr 146/10
🧠Researchers introduce ILR, a novel multi-agent learning framework that enables Large Language Models to enhance their independent reasoning through interactive training with other LLMs, then solve problems autonomously without re-executing the multi-agent system. The approach combines dynamic interaction strategies and perception calibration, delivering up to 5% performance improvements across mathematical, coding, and reasoning benchmarks.
AIBullisharXiv – CS AI · Apr 146/10
🧠Researchers introduce PODS (Policy Optimization with Down-Sampling), a technique that accelerates reinforcement learning training for large language models by selectively training on high-variance rollouts rather than all generated data. The method achieves equivalent performance to standard approaches at 1.7x faster speeds, addressing computational bottlenecks in LLM reasoning optimization.
AINeutralarXiv – CS AI · Apr 146/10
🧠Researchers propose TokUR, a framework that enables large language models to estimate uncertainty at the token level during reasoning tasks, allowing LLMs to self-assess response quality and improve performance on mathematical problems. The approach uses low-rank random weight perturbation to generate predictive distributions, demonstrating strong correlation with answer correctness and potential for enhancing LLM reliability.
AINeutralarXiv – CS AI · Apr 146/10
🧠StyleBench is a new benchmark that evaluates how different reasoning structures (Chain-of-Thought, Tree-of-Thought, etc.) affect LLM performance across various tasks and model sizes. The research reveals that structural complexity only improves accuracy in specific scenarios, with simpler approaches often proving more efficient, and that learning adaptive reasoning strategies is itself a complex problem requiring advanced training methods.
AINeutralarXiv – CS AI · Apr 146/10
🧠Researchers introduce MERMAID, a memory-enhanced multi-agent framework for automated fact-checking that couples evidence retrieval with reasoning processes. The system achieves state-of-the-art performance on multiple benchmarks by reusing retrieved evidence across claims, reducing redundant searches and improving verification efficiency.
AINeutralarXiv – CS AI · Apr 106/10
🧠Researchers present ProofSketcher, a hybrid system combining large language models with lightweight proof verification to address mathematical reasoning errors in AI-generated proofs. The approach bridges the gap between LLM efficiency and the formal rigor of interactive theorem provers like Lean and Coq, enabling more reliable automated reasoning without requiring full formalization.
$AVAX
AINeutralarXiv – CS AI · Apr 106/10
🧠Researchers present CGD-PD, a test-time decoding method that improves large language models' performance on three-way logical question answering (True/False/Unknown) by enforcing negation consistency and resolving epistemic uncertainty through targeted entailment probes. The approach achieves up to 16% relative accuracy improvements on the FOLIO benchmark while reducing spurious Unknown predictions.
AIBullisharXiv – CS AI · Apr 106/10
🧠Researchers introduce S³ (Stratified Scaling Search), a test-time scaling method for diffusion language models that improves output quality by reallocating compute during the denoising process rather than simple best-of-K sampling. The technique uses a lightweight verifier to evaluate and selectively resample candidate trajectories at each step, demonstrating consistent performance gains across mathematical reasoning and knowledge tasks without requiring model retraining.
AINeutralarXiv – CS AI · Apr 106/10
🧠Researchers discovered that large language models have a fundamental limitation in latent reasoning: they can discover multi-step planning strategies without explicit supervision, but only up to depths of 3-7 steps depending on model size and training method. This finding suggests that complex reasoning tasks may require explicit chain-of-thought monitoring rather than relying on hidden internal computations.
🧠 GPT-4🧠 GPT-5
AINeutralarXiv – CS AI · Apr 66/10
🧠A replication study found that simple vocabulary constraints like banning filler words ('very', 'just') improved AI reasoning performance more than complex linguistic restrictions like E-Prime. The research suggests any constraint that disrupts default generation patterns acts as an output regularizer, with shallow constraints being most effective.
AIBullisharXiv – CS AI · Mar 176/10
🧠Researchers propose a new framework for large language models that separates planning from factual retrieval to improve reliability in fact-seeking question answering. The modular approach uses a lightweight student planner trained via teacher-student learning to generate structured reasoning steps, showing improved accuracy and speed on challenging benchmarks.
AINeutralarXiv – CS AI · Mar 35/104
🧠Researchers propose GHS-TDA, a new method to improve large language model reasoning by using global hypothesis graphs and topological data analysis. The approach addresses limitations in Chain-of-Thought reasoning by providing error correction mechanisms and filtering redundant reasoning paths.