βBack to feed
π§ AIπ’ Bullish
LEDOM: Reverse Language Model
arXiv β CS AI|Xunjian Yin, Sitao Cheng, Yuxi Xie, Xinyu Hu, Li Lin, Xinyi Wang, Liangming Pan, William Yang Wang, Xiaojun Wan||1 views
π€AI Summary
Researchers have developed LEDOM, an open-source reverse autoregressive language model that trains right-to-left instead of the traditional left-to-right approach. The model demonstrates unique capabilities like abductive inference and question synthesis, and when combined with forward models through 'Reverse Reward' scoring, achieves significant performance gains of up to 15% on mathematical reasoning tasks.
Key Takeaways
- βLEDOM is the first large-scale reverse autoregressive language model trained right-to-left with 2B/7B parameters on 435B tokens.
- βReverse training enables unique reasoning capabilities including abductive inference, question synthesis, and resolution of the reversal curse.
- βThe Reverse Reward technique combines forward and reverse models to penalize hallucinated reasoning chains.
- βPerformance improvements of up to 6.6% on AIME 2024 and 15% on AMC 2023 mathematical benchmarks were achieved.
- βAll models, code, and training data have been released as open-source resources.
#language-models#reverse-training#autoregressive#open-source#mathematical-reasoning#ai-research#ledom#reasoning-capabilities
Read Original βvia arXiv β CS AI
Act on this with AI
Stay ahead of the market.
Connect your wallet to an AI agent. It reads balances, proposes swaps and bridges across 15 chains β you keep full control of your keys.
Related Articles