y0news
โ† Feed
โ†Back to feed
๐Ÿง  AIโšช Neutral

Self-Anchoring Calibration Drift in Large Language Models: How Multi-Turn Conversations Reshape Model Confidence

arXiv โ€“ CS AI| Harshavardhan||2 views
๐Ÿค–AI Summary

Researchers identified Self-Anchoring Calibration Drift (SACD), where large language models show systematic confidence changes when building on their own outputs in multi-turn conversations. Testing Claude Sonnet 4.6, Gemini 3.1 Pro, and GPT-5.2 revealed model-specific patterns, with Claude showing decreasing confidence and significant calibration errors, while GPT-5.2 exhibited opposite behavior in open-ended domains.

Key Takeaways
  • โ†’Self-Anchoring Calibration Drift (SACD) causes LLMs to systematically alter confidence levels when referencing their own previous responses.
  • โ†’Claude Sonnet 4.6 showed decreasing confidence and significant calibration error drift in multi-turn conversations.
  • โ†’GPT-5.2 demonstrated increasing confidence in open-ended domains with escalating calibration errors by Turn 5.
  • โ†’Gemini 3.1 Pro's natural calibration improvement was suppressed when engaging in self-anchoring behavior.
  • โ†’The phenomenon varies significantly across different AI models, indicating heterogeneous response patterns to iterative self-referencing.
Mentioned Tokens
$NEAR$0.0000โ–ฒ+0.0%
Let AI manage these โ†’
Non-custodial ยท Your keys, always
Read Original โ†’via arXiv โ€“ CS AI
Act on this with AI
This article mentions $NEAR.
Let your AI agent check your portfolio, get quotes, and propose trades โ€” you review and approve from your device.
Connect Wallet to AI โ†’How it works
Related Articles