7 articles tagged with #peft. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท Mar 47/102
๐ง DiaBlo introduces a new Parameter-Efficient Fine-Tuning (PEFT) method that updates only diagonal blocks of weight matrices in large language models, offering better performance than LoRA while maintaining similar memory efficiency. The approach eliminates the need for low-rank matrix products and provides theoretical guarantees for convergence, showing competitive results across various AI tasks including reasoning and code generation.
AINeutralarXiv โ CS AI ยท Mar 37/104
๐ง Researchers introduce PsyAgent, a new AI framework that creates human-like agents by combining personality modeling based on Big Five traits with contextual social awareness. The system uses structured prompts and fine-tuning to produce AI agents that maintain stable personality traits while adapting appropriately to different social situations and roles.
AIBullisharXiv โ CS AI ยท Apr 146/10
๐ง Researchers propose a novel hybrid fine-tuning method for Large Language Models that combines full parameter updates with Parameter-Efficient Fine-Tuning (PEFT) modules using zeroth-order and first-order optimization. The approach addresses computational constraints of full fine-tuning while overcoming PEFT's limitations in knowledge acquisition, backed by theoretical convergence analysis and empirical validation across multiple tasks.
AIBullisharXiv โ CS AI ยท Apr 106/10
๐ง Researchers introduce LoRA-DA, a new initialization method for Low-Rank Adaptation that leverages target-domain data and theoretical optimization principles to improve fine-tuning performance. The method outperforms existing initialization approaches across multiple benchmarks while maintaining computational efficiency.
AIBullishHugging Face Blog ยท Jul 234/108
๐ง The article discusses technical improvements for Fast LoRA inference when working with Flux models using Diffusers and PEFT libraries. This represents an advancement in AI model optimization, specifically focusing on efficient fine-tuning and inference capabilities for diffusion models.
AINeutralHugging Face Blog ยท Feb 194/108
๐ง The article title suggests that PEFT (Parameter Efficient Fine-Tuning) has introduced new merging methods. However, the article body appears to be empty or unavailable, limiting detailed analysis of the specific technical developments or their implications.
AIBullishHugging Face Blog ยท Feb 105/104
๐ง The article discusses parameter-efficient fine-tuning methods using Hugging Face's PEFT library. PEFT enables efficient adaptation of large language models by updating only a small subset of parameters rather than full model retraining.