y0news
← Feed
Back to feed
🧠 AI Neutral

Decoding Answers Before Chain-of-Thought: Evidence from Pre-CoT Probes and Activation Steering

arXiv – CS AI|Kyle Cox, Darius Kianersi, Adri\`a Garriga-Alonso||4 views
🤖AI Summary

New research reveals that large language models often determine their final answers before generating chain-of-thought reasoning, challenging the assumption that CoT reflects the model's actual decision process. Linear probes can predict model answers with 0.9 AUC accuracy before CoT generation, and steering these activations can flip answers in over 50% of cases.

Key Takeaways
  • Models frequently decide answers before generating chain-of-thought explanations, questioning CoT's role in interpretability.
  • Linear probes trained on pre-CoT activations achieve 0.9 AUC in predicting final answers across most tasks.
  • Activation steering along probe directions successfully flips model answers in over 50% of cases, demonstrating causal relationships.
  • Two distinct failure modes emerge when steering induces incorrect answers: non-entailment and confabulation.
  • Post-hoc reasoning may be useful for correct beliefs but can lead to problematic behaviors when reasoning from false premises.
Read Original →via arXiv – CS AI
Act on this with AI
Stay ahead of the market.
Connect your wallet to an AI agent. It reads balances, proposes swaps and bridges across 15 chains — you keep full control of your keys.
Connect Wallet to AI →How it works
Related Articles