y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#dataset-distillation News & Analysis

3 articles tagged with #dataset-distillation. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

3 articles
AIBearisharXiv โ€“ CS AI ยท Mar 37/106
๐Ÿง 

Turning Black Box into White Box: Dataset Distillation Leaks

Researchers discovered that dataset distillation, a technique for compressing large datasets into smaller synthetic ones, has serious privacy vulnerabilities. The study introduces an Information Revelation Attack (IRA) that can extract sensitive information from synthetic datasets, including predicting the distillation algorithm, model architecture, and recovering original training samples.

AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

Towards Principled Dataset Distillation: A Spectral Distribution Perspective

Researchers propose Class-Aware Spectral Distribution Matching (CSDM), a new dataset distillation method that addresses performance issues on imbalanced datasets. The technique achieves 14% improvement over existing methods on CIFAR-10-LT with enhanced stability on long-tailed data distributions.

AINeutralarXiv โ€“ CS AI ยท Mar 44/103
๐Ÿง 

From Fewer Samples to Fewer Bits: Reframing Dataset Distillation as Joint Optimization of Precision and Compactness

Researchers propose QuADD (Quantization-aware Dataset Distillation), a new framework that jointly optimizes dataset compression and precision to create more efficient synthetic training datasets. The method integrates differentiable quantization within the distillation process, achieving better accuracy per bit than existing approaches on image classification and 3GPP beam management tasks.