y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#unlearning News & Analysis

8 articles tagged with #unlearning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

8 articles
AINeutralarXiv โ€“ CS AI ยท Mar 47/102
๐Ÿง 

WARP: Weight Teleportation for Attack-Resilient Unlearning Protocols

Researchers introduce WARP, a new defense mechanism for machine unlearning protocols that protects against privacy attacks where adversaries can exploit differences between pre- and post-unlearning AI models. The technique reduces attack success rates by up to 92% while maintaining model accuracy on retained data.

AINeutralarXiv โ€“ CS AI ยท Mar 37/105
๐Ÿง 

Agentic Unlearning: When LLM Agent Meets Machine Unlearning

Researchers introduce 'agentic unlearning' through Synchronized Backflow Unlearning (SBU), a framework that removes sensitive information from both AI model parameters and persistent memory systems. The method addresses critical gaps in existing unlearning techniques by preventing cross-pathway recontamination between memory and parameters.

AINeutralarXiv โ€“ CS AI ยท 1d ago6/10
๐Ÿง 

Operationalising the Right to be Forgotten in LLMs: A Lightweight Sequential Unlearning Framework for Privacy-Aligned Deployment in Politically Sensitive Environments

Researchers introduce a sequential unlearning framework that enables Large Language Models to forget sensitive data while maintaining performance, addressing GDPR compliance and the Right to be Forgotten in politically sensitive deployments. The method stabilizes general capabilities through positive fine-tuning before selectively suppressing designated patterns, demonstrating effectiveness on the SemEval-2025 benchmark with minimal accuracy degradation.

AIBullisharXiv โ€“ CS AI ยท Apr 76/10
๐Ÿง 

VLA-Forget: Vision-Language-Action Unlearning for Embodied Foundation Models

Researchers introduce VLA-Forget, a new unlearning framework for vision-language-action (VLA) models used in robotic manipulation. The hybrid approach addresses the challenge of removing unsafe or unwanted behaviors from embodied AI foundation models while preserving their core perception, language, and action capabilities.

AINeutralarXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

Relationship-Aware Safety Unlearning for Multimodal LLMs

Researchers propose a new framework for improving safety in multimodal AI models by targeting unsafe relationships between objects rather than removing entire concepts. The approach uses parameter-efficient edits to suppress dangerous combinations while preserving benign uses of the same objects and relations.

AINeutralarXiv โ€“ CS AI ยท Mar 37/107
๐Ÿง 

Forgetting is Competition: Rethinking Unlearning as Representation Interference in Diffusion Models

Researchers introduce SurgUn, a surgical unlearning method for text-to-image diffusion models that enables precise removal of specific visual concepts while preserving other capabilities. The approach addresses challenges in copyright compliance and content policy enforcement by applying targeted weight-space updates based on retroactive interference theory.

AIBullisharXiv โ€“ CS AI ยท Mar 37/105
๐Ÿง 

ALTER: Asymmetric LoRA for Token-Entropy-Guided Unlearning of LLMs

Researchers introduce ALTER, a new framework for efficiently "unlearning" specific knowledge from large language models while preserving their overall utility. The system uses asymmetric LoRA architecture to selectively forget targeted information with 95% effectiveness while maintaining over 90% model utility, significantly outperforming existing methods.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1024
๐Ÿง 

DUET: Distilled LLM Unlearning from an Efficiently Contextualized Teacher

Researchers propose DUET, a new distillation-based method for LLM unlearning that removes undesirable knowledge from AI models without full retraining. The technique combines computational efficiency with security advantages, achieving better performance in both knowledge removal and utility preservation while being significantly more data-efficient than existing methods.