AINeutralarXiv โ CS AI ยท Feb 277/105
๐ง
Transformers converge to invariant algorithmic cores
Researchers have discovered that transformer models, despite different training runs producing different weights, converge to the same compact 'algorithmic cores' - low-dimensional subspaces essential for task performance. The study shows these invariant structures persist across different scales and training runs, suggesting transformer computations are organized around shared algorithmic patterns rather than implementation-specific details.