AINeutralarXiv – CS AI · 9h ago6/10
🧠
The Safety-Aware Denoiser for Text Diffusion Models
Researchers propose Safety-Aware Denoiser (SAD), an inference-time safety framework that guides text diffusion models toward secure outputs during the denoising process without requiring model retraining. The method reduces unsafe text generation while maintaining output quality, offering a scalable alternative to post-hoc filtering approaches.