y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#gsm8k News & Analysis

5 articles tagged with #gsm8k. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

5 articles
AINeutralarXiv – CS AI · 1d ago7/10
🧠

GSM-SEM: Benchmark and Framework for Generating Semantically Variant Augmentations

Researchers introduce GSM-SEM, a framework for generating semantically diverse variants of math benchmarks like GSM8K to combat memorization in LLM evaluations. Testing 14 state-of-the-art models reveals consistent performance drops averaging 28%, suggesting current leaderboard rankings may overstate true reasoning capabilities.

AIBullisharXiv – CS AI · 1d ago7/10
🧠

ESSAM: A Novel Competitive Evolution Strategies Approach to Reinforcement Learning for Memory Efficient LLMs Fine-Tuning

Researchers propose ESSAM, a novel training framework combining Evolution Strategies with Sharpness-Aware Maximization to fine-tune large language models for mathematical reasoning while dramatically reducing GPU memory requirements. The approach achieves comparable accuracy to reinforcement learning methods like PPO and GRPO while using 18-10× less memory, addressing a critical bottleneck in LLM development.

AIBearisharXiv – CS AI · Apr 137/10
🧠

On the Limits of Layer Pruning for Generative Reasoning in Large Language Models

Research demonstrates that layer pruning—a compression technique for large language models—effectively reduces model size while maintaining classification performance, but critically fails to preserve generative reasoning capabilities like arithmetic and code generation. Even with extensive post-training on 400B tokens, models cannot recover lost reasoning abilities, revealing fundamental limitations in current compression approaches.

AINeutralarXiv – CS AI · 3d ago6/10
🧠

More Is Not Always Better: Cross-Component Interference in LLM Agent Scaffolding

Researchers demonstrate that stacking more components into LLM agent systems doesn't improve performance and often degrades it due to cross-component interference. A comprehensive factorial study across 32 configurations shows optimal agent design is task-dependent and model-scale dependent, with the fully-equipped system consistently underperforming smaller, curated subsets by up to 79%.

🧠 Llama
AINeutralarXiv – CS AI · Mar 276/10
🧠

Efficient Detection of Bad Benchmark Items with Novel Scalability Coefficients

Researchers introduce a new nonparametric method called signed isotonic R² for efficiently detecting problematic items in AI benchmarks and assessments. The method outperforms traditional diagnostic techniques across major AI datasets including GSM8K and MMLU, offering a lightweight solution for improving evaluation quality.