AINeutralarXiv โ CS AI ยท 6h ago2
๐ง
Scaling of learning time for high dimensional inputs
Researchers present theoretical analysis showing that neural network learning times scale supralinearly with input dimensionality, creating fundamental limitations for high-dimensional learning. The study uses Hebbian learning models to demonstrate that higher input dimensions result in smaller gradients and prohibitively long learning times.