AIBullisharXiv โ CS AI ยท 4h ago2
๐ง
Preference Packing: Efficient Preference Optimization for Large Language Models
Researchers propose 'preference packing,' a new optimization technique for training large language models that reduces training time by at least 37% through more efficient handling of duplicate input prompts. The method optimizes attention operations and KV cache memory usage in preference-based training methods like Direct Preference Optimization.