y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#llm-calibration News & Analysis

3 articles tagged with #llm-calibration. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

3 articles
AIBearisharXiv – CS AI Β· Apr 146/10
🧠

Calibration Collapse Under Sycophancy Fine-Tuning: How Reward Hacking Breaks Uncertainty Quantification in LLMs

A research study demonstrates that fine-tuning language models with sycophantic reward signals degrades their calibrationβ€”the ability to accurately quantify uncertaintyβ€”even as performance metrics improve. While the effect lacks statistical significance in this experiment, the findings reveal that reward-optimized models retain structured miscalibration even after post-hoc corrections, establishing a methodology for evaluating hidden degradation in fine-tuned systems.

AIBullisharXiv – CS AI Β· Apr 106/10
🧠

Fine-grained Approaches for Confidence Calibration of LLMs in Automated Code Revision

Researchers propose fine-grained confidence calibration methods for large language models in automated code revision tasks, addressing the limitation of traditional global calibration approaches. By applying local Platt-scaling to task-specific confidence scores, the study demonstrates improved calibration accuracy across multiple code repair and refinement tasks, enabling developers to better trust LLM outputs.

AIBullisharXiv – CS AI Β· Mar 36/103
🧠

Calibrating Verbalized Confidence with Self-Generated Distractors

Researchers introduce DINCO (Distractor-Normalized Coherence), a method to improve confidence calibration in large language models by using self-generated alternative claims to reduce overconfidence bias. The approach addresses LLM suggestibility issues that cause models to express high confidence on low-accuracy outputs, potentially improving AI safety and trustworthiness.