y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#privacy-preservation News & Analysis

5 articles tagged with #privacy-preservation. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

5 articles
AIBullisharXiv โ€“ CS AI ยท 2d ago7/10
๐Ÿง 

Private Seeds, Public LLMs: Realistic and Privacy-Preserving Synthetic Data Generation

Researchers propose RPSG, a novel method for generating synthetic data from private text using large language models while maintaining differential privacy protections. The approach uses private seeds and formal privacy mechanisms during candidate selection, achieving high fidelity synthetic data with stronger privacy guarantees than existing methods.

AIBullisharXiv โ€“ CS AI ยท Mar 167/10
๐Ÿง 

Learnability and Privacy Vulnerability are Entangled in a Few Critical Weights

Researchers discovered that privacy vulnerabilities in neural networks exist in only a small fraction of weights, but these same weights are critical for model performance. They developed a new approach that preserves privacy by rewinding and fine-tuning only these critical weights instead of retraining entire networks, maintaining utility while defending against membership inference attacks.

AINeutralarXiv โ€“ CS AI ยท 6d ago6/10
๐Ÿง 

AdaProb: Efficient Machine Unlearning via Adaptive Probability

Researchers propose AdaProb, a machine unlearning method that enables trained AI models to efficiently forget specific data while preserving privacy and complying with regulations like GDPR. The approach uses adaptive probability distributions and demonstrates 20% improvement in forgetting effectiveness with 50% less computational overhead compared to existing methods.

AIBullisharXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

Computation and Communication Efficient Federated Unlearning via On-server Gradient Conflict Mitigation and Expression

Researchers propose FOUL (Federated On-server Unlearning), a new framework for efficiently removing specific participants' data from federated learning models without accessing client data. The approach reduces computational and communication costs while maintaining privacy compliance through a two-stage process that performs unlearning operations on the server side.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1017
๐Ÿง 

Controllable Reasoning Models Are Private Thinkers

Researchers developed a method to train AI reasoning models to follow privacy instructions in their internal reasoning traces, not just final answers. The approach uses separate LoRA adapters and achieves up to 51.9% improvement on privacy benchmarks, though with some trade-offs in task performance.