y0news
AnalyticsDigestsRSSAICrypto
#attention-heads1 article
1 articles
AINeutralarXiv โ€“ CS AI ยท 5h ago
๐Ÿง 

Function Induction and Task Generalization: An Interpretability Study with Off-by-One Addition

Researchers studied how large language models generalize to new tasks through "off-by-one addition" experiments, discovering a "function induction" mechanism that operates at higher abstraction levels than previously known induction heads. The study reveals that multiple attention heads work in parallel to enable task-level generalization, with this mechanism being reusable across various synthetic and algorithmic tasks.