y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#sgd News & Analysis

7 articles tagged with #sgd. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

7 articles
AINeutralarXiv โ€“ CS AI ยท Mar 46/102
๐Ÿง 

The Malignant Tail: Spectral Segregation of Label Noise in Over-Parameterized Networks

Researchers identify the 'Malignant Tail' phenomenon where over-parameterized neural networks segregate signal from noise during training, leading to harmful overfitting. They demonstrate that Stochastic Gradient Descent pushes label noise into high-frequency orthogonal subspaces while preserving semantic features in low-rank subspaces, and propose Explicit Spectral Truncation as a post-hoc solution to recover optimal generalization.

AINeutralarXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Loss Barcode: A Topological Measure of Escapability in Loss Landscapes

Researchers developed a new topological measure called the 'TO-score' to analyze neural network loss landscapes and understand how gradient descent optimization escapes local minima. Their findings show that deeper and wider networks have fewer topological obstructions to learning, and there's a connection between loss barcode characteristics and generalization performance.

AIBullisharXiv โ€“ CS AI ยท Mar 276/10
๐Ÿง 

Formal Semantics for Agentic Tool Protocols: A Process Calculus Approach

Researchers have developed the first formal mathematical framework for verifying AI agent protocols, specifically comparing Schema-Guided Dialogue (SGD) and Model Context Protocol (MCP). They proved these systems are structurally similar but identified critical gaps in MCP's capabilities, proposing MCP+ extensions to achieve full equivalence with SGD.

AINeutralarXiv โ€“ CS AI ยท Mar 45/103
๐Ÿง 

Why Adam Can Beat SGD: Second-Moment Normalization Yields Sharper Tails

Research paper establishes the first theoretical separation between Adam and SGD optimization algorithms, proving Adam achieves better high-probability convergence guarantees. The study provides mathematical backing for Adam's superior empirical performance through second-moment normalization analysis.

AINeutralarXiv โ€“ CS AI ยท Mar 34/104
๐Ÿง 

Scaling Laws of SignSGD in Linear Regression: When Does It Outperform SGD?

Researchers analyzed scaling laws for signSGD optimization in machine learning, comparing it to standard SGD under a power-law random features model. The study identifies unique effects in signSGD that can lead to steeper compute-optimal scaling laws than SGD in noise-dominant regimes.

AINeutralOpenAI News ยท Mar 74/105
๐Ÿง 

Reptile: A scalable meta-learning algorithm

Researchers have developed Reptile, a new meta-learning algorithm that improves machine learning efficiency by repeatedly sampling tasks and updating parameters through stochastic gradient descent. The algorithm is mathematically similar to first-order MAML but requires only black-box access to optimizers like SGD or Adam while maintaining similar performance and computational efficiency.

AINeutralarXiv โ€“ CS AI ยท Mar 24/105
๐Ÿง 

Optimizer-Induced Low-Dimensional Drift and Transverse Dynamics in Transformer Training

Researchers analyzed training trajectories in small transformer models, finding that parameter updates organize into a dominant drift direction with transverse dynamics. The study reveals that different optimizers (AdamW vs SGD) create substantially different trajectory geometries, with AdamW developing multi-dimensional structures while SGD produces more linear evolution.