173 articles tagged with #diffusion-models. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv β CS AI Β· Mar 176/10
π§ Researchers introduce RAZOR, a new framework for efficiently removing sensitive information from AI models like CLIP and Stable Diffusion without requiring full retraining. The method selectively edits specific layers and attention heads in transformer models to achieve targeted 'unlearning' while preserving overall performance.
π§ Stable Diffusion
AINeutralarXiv β CS AI Β· Mar 176/10
π§ Researchers conducted the first systematic study on post-training quantization for diffusion large language models (dLLMs), identifying activation outliers as a key challenge for compression. The study evaluated state-of-the-art quantization methods across multiple dimensions to provide insights for efficient dLLM deployment on edge devices.
AIBullisharXiv β CS AI Β· Mar 176/10
π§ Researchers introduce Contrastive Noise Optimization, a new method that improves diversity in text-to-image AI generation by optimizing initial noise patterns rather than intermediate outputs. The technique uses contrastive loss to maximize diversity while preserving image quality, achieving superior results across multiple text-to-image model architectures.
AINeutralarXiv β CS AI Β· Mar 176/10
π§ EgoGrasp introduces the first method to reconstruct world-space hand-object interactions from egocentric videos using open-vocabulary objects. The multi-stage framework combines vision foundation models with body-guided diffusion models to achieve state-of-the-art performance in 3D scene reconstruction and hand pose estimation.
AIBullisharXiv β CS AI Β· Mar 176/10
π§ Researchers introduce Agentic Retoucher, a new AI framework that fixes common distortions in text-to-image generation through a three-agent system for perception, reasoning, and correction. The system outperformed existing methods on a new 27K-image dataset, potentially improving the quality and reliability of AI-generated images.
AIBullisharXiv β CS AI Β· Mar 166/10
π§ Researchers propose NaΓ―ve PAINE, a lightweight system that improves text-to-image generation quality by predicting which initial noise inputs will produce better results before running the full diffusion model. The approach reduces the need for multiple generation cycles to get satisfactory images by pre-selecting higher-quality noise patterns.
AIBullisharXiv β CS AI Β· Mar 116/10
π§ Researchers introduce Latent-DARM, a framework that bridges discrete diffusion language models and autoregressive models to improve multi-agent AI reasoning capabilities. The system achieved significant improvements on reasoning benchmarks, increasing accuracy from 27% to 36% on DART-5 while using less than 2.2% of the token budget of state-of-the-art models.
AINeutralarXiv β CS AI Β· Mar 116/10
π§ Researchers developed tunable-complexity priors for generative models (diffusion models, normalizing flows, and variational autoencoders) that can dynamically adjust complexity based on the specific inverse problem. The approach uses nested dropout and demonstrates superior performance across compressed sensing, inpainting, denoising, and phase retrieval tasks compared to fixed-complexity baselines.
AIBullisharXiv β CS AI Β· Mar 96/10
π§ Researchers introduce TempoSyncDiff, a new AI framework that uses distilled diffusion models to generate realistic talking head videos from audio with significantly reduced computational latency. The system addresses key challenges in AI-driven video synthesis including temporal instability, identity drift, and audio-visual alignment while enabling deployment on edge computing devices.
AIBullisharXiv β CS AI Β· Mar 96/10
π§ Researchers introduce Place-it-R1, an AI framework that uses Multimodal Large Language Models to insert objects into videos while maintaining physical realism. The system employs Chain-of-Thought reasoning to ensure inserted objects interact naturally with their environment, addressing the gap between visual quality and physical plausibility in video editing.
AINeutralarXiv β CS AI Β· Mar 96/10
π§ Researchers have developed ContextBench, a new benchmark for evaluating methods that generate targeted inputs to trigger specific behaviors in language models. The study introduces enhanced Evolutionary Prompt Optimization techniques that better balance effectiveness in activating AI model features while maintaining linguistic fluency.
AIBullisharXiv β CS AI Β· Mar 55/10
π§ Researchers have developed a new approach using multiplicative LoRA (Low-Rank Adaptation) weights for neural field representation learning, achieving improved quality in reconstruction, generation, and analysis tasks. The method constrains optimization space through pre-trained base models, creating structured weight representations that outperform existing weight-space methods when used with latent diffusion models.
AIBullishHugging Face Blog Β· Mar 56/10
π§ The article introduces Modular Diffusers, a new framework for building composable and flexible diffusion model pipelines. This development allows developers to create more modular AI systems by breaking down diffusion processes into reusable components.
AIBullisharXiv β CS AI Β· Mar 36/103
π§ Researchers introduce SHINE, a training-free framework that enables FLUX and other diffusion models to perform high-quality image composition without retraining. The framework addresses complex lighting scenarios like shadows and reflections, achieving state-of-the-art performance on new benchmark ComplexCompo.
AIBullisharXiv β CS AI Β· Mar 36/103
π§ Sketch2Colab is a new AI system that converts 2D sketches into realistic 3D multi-human animations with precise control over interactions and movements. The technology uses a novel approach combining sketch-driven diffusion with rectified-flow distillation for faster, more stable animation generation than existing methods.
AIBullisharXiv β CS AI Β· Mar 36/104
π§ Researchers developed MAP-Diff, a multi-anchor guided diffusion framework that improves 3D whole-body PET scan denoising by using intermediate-dose scans as trajectory anchors. The method achieves significant improvements in image quality metrics, increasing PSNR from 42.48 dB to 43.71 dB while reducing radiation exposure for patients.
AIBullisharXiv β CS AI Β· Mar 36/104
π§ DragFlow introduces the first framework to leverage FLUX's DiT priors for drag-based image editing, addressing distortion issues that plagued earlier Stable Diffusion-based approaches. The system uses region-based editing with affine transformations instead of point-based supervision, achieving state-of-the-art results on benchmarks.
AIBullisharXiv β CS AI Β· Mar 36/103
π§ Researchers propose EquiReg, a new framework that improves diffusion models for inverse problems like image restoration by keeping sampling trajectories on the data manifold. The method uses equivariance regularization to guide sampling toward symmetry-preserving regions, enabling high-quality reconstructions with fewer sampling steps.
AIBullisharXiv β CS AI Β· Mar 36/104
π§ LiftAvatar is a new AI system that enhances 3D avatar animation by completing sparse monocular video observations in kinematic space using expression-controlled video diffusion Transformers. The technology addresses limitations in 3D Gaussian Splatting-based avatars by generating high-quality, temporally coherent facial expressions from single or multiple reference images.
AIBullisharXiv β CS AI Β· Mar 37/104
π§ Researchers propose FreeAct, a new quantization framework for Large Language Models that improves efficiency by using dynamic transformation matrices for different token types. The method achieves up to 5.3% performance improvement over existing approaches by addressing the memory and computational overhead challenges in LLMs.
AIBullisharXiv β CS AI Β· Mar 37/108
π§ Researchers introduce State-Action Inpainting Diffuser (SAID), a new AI framework that addresses signal delay challenges in continuous control and reinforcement learning. SAID combines model-based and model-free approaches using a generative formulation that can be applied to both online and offline RL, demonstrating state-of-the-art performance on delayed control benchmarks.
AINeutralarXiv β CS AI Β· Mar 37/106
π§ Researchers introduce StaTS, a new diffusion model for time series forecasting that learns adaptive noise schedules and uses frequency-guided denoising. The model addresses limitations of fixed noise schedules in existing diffusion models by incorporating spectral regularization and data-adaptive scheduling for improved structural preservation.
$NEAR
AIBullisharXiv β CS AI Β· Mar 36/103
π§ Researchers introduce SVG, a new latent diffusion model that eliminates the need for variational autoencoders by using self-supervised representations. The approach leverages frozen DINO features to create semantically structured latent spaces, enabling faster training, fewer sampling steps, and better generative quality while maintaining semantic capabilities.
AIBullisharXiv β CS AI Β· Mar 36/104
π§ Researchers have developed DCDP, a Dynamic Closed-Loop Diffusion Policy framework that significantly improves robotic manipulation in dynamic environments. The system achieves 19% better adaptability without retraining while requiring only 5% additional computational overhead through real-time action correction and environmental dynamics integration.
AIBullisharXiv β CS AI Β· Mar 37/108
π§ Researchers introduce Coupled Discrete Diffusion (CoDD), a breakthrough framework that solves the "factorization barrier" in diffusion language models by enabling parallel token generation without sacrificing coherence. The approach uses a lightweight probabilistic inference layer to model complex joint dependencies while maintaining computational efficiency.