AIBullisharXiv โ CS AI ยท 4h ago2
๐ง Researchers propose FedRot-LoRA, a new framework that solves rotational misalignment issues in federated learning for large language models. The solution uses orthogonal transformations to align client updates before aggregation, improving training stability and performance without increasing communication costs.
AIBullisharXiv โ CS AI ยท 4h ago4
๐ง Researchers have developed MPU, a privacy-preserving framework that enables machine unlearning for large language models without requiring servers to share parameters or clients to share data. The framework uses perturbed model copies and harmonic denoising to achieve comparable performance to non-private methods, with most algorithms showing less than 1% performance degradation.
AIBullisharXiv โ CS AI ยท 4h ago3
๐ง Researchers propose FedNSAM, a new federated learning algorithm that improves global model performance by addressing the inconsistency between local and global flatness in distributed training environments. The algorithm uses global Nesterov momentum to harmonize local and global optimization, showing superior performance compared to existing FedSAM approaches.
AIBullisharXiv โ CS AI ยท 4h ago3
๐ง Researchers propose an efficient unsupervised federated learning framework for anomaly detection in heterogeneous IoT networks that preserves privacy while leveraging shared features from multiple datasets. The approach uses explainable AI techniques like SHAP for transparency and demonstrates superior performance compared to conventional federated learning methods on real-world IoT datasets.
AIBullisharXiv โ CS AI ยท 4h ago0
๐ง Researchers introduce FedDAG, a new clustered federated learning framework that improves AI model training across heterogeneous client environments. The system combines data and gradient similarity metrics for better client clustering and uses a dual-encoder architecture to enable knowledge sharing across clusters while maintaining specialization.
AIBullisharXiv โ CS AI ยท 4h ago0
๐ง Researchers have developed a new framework for privacy-preserving feature selection that uses permutation-invariant representation learning and federated learning techniques. The approach addresses data imbalance and privacy constraints in distributed scenarios while improving computational efficiency and downstream task performance.
AINeutralarXiv โ CS AI ยท 4h ago0
๐ง Researchers introduce FedVG, a new federated learning framework that uses gradient-guided aggregation and global validation sets to improve model performance in distributed training environments. The approach addresses client drift issues in heterogeneous data settings and can be integrated with existing federated learning algorithms.