8 articles tagged with #unlearning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AINeutralarXiv โ CS AI ยท Mar 47/102
๐ง Researchers introduce WARP, a new defense mechanism for machine unlearning protocols that protects against privacy attacks where adversaries can exploit differences between pre- and post-unlearning AI models. The technique reduces attack success rates by up to 92% while maintaining model accuracy on retained data.
AINeutralarXiv โ CS AI ยท Mar 37/105
๐ง Researchers introduce 'agentic unlearning' through Synchronized Backflow Unlearning (SBU), a framework that removes sensitive information from both AI model parameters and persistent memory systems. The method addresses critical gaps in existing unlearning techniques by preventing cross-pathway recontamination between memory and parameters.
AINeutralarXiv โ CS AI ยท 1d ago6/10
๐ง Researchers introduce a sequential unlearning framework that enables Large Language Models to forget sensitive data while maintaining performance, addressing GDPR compliance and the Right to be Forgotten in politically sensitive deployments. The method stabilizes general capabilities through positive fine-tuning before selectively suppressing designated patterns, demonstrating effectiveness on the SemEval-2025 benchmark with minimal accuracy degradation.
AIBullisharXiv โ CS AI ยท Apr 76/10
๐ง Researchers introduce VLA-Forget, a new unlearning framework for vision-language-action (VLA) models used in robotic manipulation. The hybrid approach addresses the challenge of removing unsafe or unwanted behaviors from embodied AI foundation models while preserving their core perception, language, and action capabilities.
AINeutralarXiv โ CS AI ยท Mar 176/10
๐ง Researchers propose a new framework for improving safety in multimodal AI models by targeting unsafe relationships between objects rather than removing entire concepts. The approach uses parameter-efficient edits to suppress dangerous combinations while preserving benign uses of the same objects and relations.
AINeutralarXiv โ CS AI ยท Mar 37/107
๐ง Researchers introduce SurgUn, a surgical unlearning method for text-to-image diffusion models that enables precise removal of specific visual concepts while preserving other capabilities. The approach addresses challenges in copyright compliance and content policy enforcement by applying targeted weight-space updates based on retroactive interference theory.
AIBullisharXiv โ CS AI ยท Mar 37/105
๐ง Researchers introduce ALTER, a new framework for efficiently "unlearning" specific knowledge from large language models while preserving their overall utility. The system uses asymmetric LoRA architecture to selectively forget targeted information with 95% effectiveness while maintaining over 90% model utility, significantly outperforming existing methods.
AIBullisharXiv โ CS AI ยท Mar 27/1024
๐ง Researchers propose DUET, a new distillation-based method for LLM unlearning that removes undesirable knowledge from AI models without full retraining. The technique combines computational efficiency with security advantages, achieving better performance in both knowledge removal and utility preservation while being significantly more data-efficient than existing methods.