AIBullisharXiv โ CS AI ยท 8h ago7/10
๐ง
Explain in Your Own Words: Improving Reasoning via Token-Selective Dual Knowledge Distillation
Researchers developed Token-Selective Dual Knowledge Distillation (TSD-KD), a new framework that improves AI reasoning by allowing smaller models to learn from larger ones more effectively. The method achieved up to 54.4% better accuracy than baseline models on reasoning benchmarks, with student models sometimes outperforming their teachers by up to 20.3%.