y0news
← Feed
Back to feed
🧠 AI🔴 BearishImportance 6/10

Where does output diversity collapse in post-training?

arXiv – CS AI|Constantinos Karouzos, Xingwei Tan, Nikolaos Aletras|
🤖AI Summary

Researchers discover that post-trained language models experience systematic output diversity collapse, where fine-tuning methods reduce the variety of generated responses compared to base models. This collapse is determined during training by data composition choices and cannot be fixed through inference-time adjustments, with implications for scaling methods and creative AI applications.

Analysis

The research identifies a fundamental trade-off in language model post-training: while fine-tuning improves instruction-following and reduces incorrect outputs, it simultaneously narrows the range of valid responses the model can generate. This diversity collapse matters because sampling-based inference methods—crucial for scaling model reasoning—rely on generating multiple distinct candidate outputs to explore solution spaces effectively.

The study's analysis across three Olmo 3 post-training variants reveals that diversity loss correlates directly with training data composition rather than arising from any single post-training technique. Chain-of-thought distillation exhibits earlier diversity collapse at supervised fine-tuning stages, while broader instruction tuning shows larger effects from direct preference optimization. Critically, the researchers demonstrate that suppressing reasoning chains at inference does not recover diversity, proving the narrowing is embedded in model weights during training rather than a generation-time artifact.

For the AI development community, this finding suggests a genuine engineering constraint: achieving both instruction adherence and output diversity requires deliberate data curation during post-training. Removing low-quality outputs explains some diversity loss, but genuine narrowing among correct answers remains substantial and task-dependent. This challenges assumptions that inference-time techniques alone can compensate for training choices.

The implications extend to creative applications, value-laden decision-making tasks, and ensemble methods that depend on diversity. Developers must now consider diversity metrics alongside accuracy during data collection and curation phases. Future work should explore whether specific data composition strategies can preserve diversity while maintaining instruction-following, as this bottleneck may limit the effectiveness of uncertainty estimation and reasoning scaling approaches.

Key Takeaways
  • Output diversity collapse in post-trained models is determined by training data composition, not by post-training methods or generation formats
  • Chain-of-thought distillation exhibits earlier diversity loss than broad instruction tuning across comparable model variants
  • Inference-time modifications cannot recover diversity lost during training, requiring upstream solutions in data curation
  • Diversity loss splits between quality-control filtering and genuine narrowing, with varying proportions across different task types
  • Models retaining more correct-answer diversity may sacrifice aggregate diversity metrics, suggesting diversity metrics require task-specific design
Read Original →via arXiv – CS AI
Act on this with AI
Stay ahead of the market.
Connect your wallet to an AI agent. It reads balances, proposes swaps and bridges across 15 chains — you keep full control of your keys.
Connect Wallet to AI →How it works
Related Articles