AINeutralarXiv – CS AI · 9h ago7/10
🧠
Position: Mechanistic Interpretability Must Disclose Identification Assumptions for Causal Claims
A research paper argues that mechanistic interpretability studies increasingly make causal claims without explicitly stating their identification assumptions, creating a credibility gap in AI research. The authors audit 10 papers across multiple methodologies and find none contain dedicated identification-assumptions sections, proposing a new disclosure norm requiring researchers to clearly state causal claims, identification strategies, and the assumptions underpinning their conclusions.