โBack to feed
๐ง AIโช Neutral
Self-Anchoring Calibration Drift in Large Language Models: How Multi-Turn Conversations Reshape Model Confidence
๐คAI Summary
Researchers identified Self-Anchoring Calibration Drift (SACD), where large language models show systematic confidence changes when building on their own outputs in multi-turn conversations. Testing Claude Sonnet 4.6, Gemini 3.1 Pro, and GPT-5.2 revealed model-specific patterns, with Claude showing decreasing confidence and significant calibration errors, while GPT-5.2 exhibited opposite behavior in open-ended domains.
Key Takeaways
- โSelf-Anchoring Calibration Drift (SACD) causes LLMs to systematically alter confidence levels when referencing their own previous responses.
- โClaude Sonnet 4.6 showed decreasing confidence and significant calibration error drift in multi-turn conversations.
- โGPT-5.2 demonstrated increasing confidence in open-ended domains with escalating calibration errors by Turn 5.
- โGemini 3.1 Pro's natural calibration improvement was suppressed when engaging in self-anchoring behavior.
- โThe phenomenon varies significantly across different AI models, indicating heterogeneous response patterns to iterative self-referencing.
Read Original โvia arXiv โ CS AI
Act on this with AI
This article mentions $NEAR.
Let your AI agent check your portfolio, get quotes, and propose trades โ you review and approve from your device.
Related Articles