y0news
AnalyticsDigestsSourcesRSSAICrypto
#hybrid-attention1 article
1 articles
AIBullisharXiv โ€“ CS AI ยท 6d ago7/102
๐Ÿง 

MiniCPM-SALA: Hybridizing Sparse and Linear Attention for Efficient Long-Context Modeling

MiniCPM-SALA introduces a 9B-parameter hybrid language model architecture that combines sparse and linear attention mechanisms to handle ultra-long contexts up to 1M tokens. The model achieves 3.5x faster inference than full-attention models while reducing training costs by 75% through a continual training framework that transforms existing Transformer models.