y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#diffusion-models News & Analysis

173 articles tagged with #diffusion-models. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

173 articles
AIBullisharXiv – CS AI · Mar 27/1016
🧠

DiffuMamba: High-Throughput Diffusion LMs with Mamba Backbone

Researchers introduce DiffuMamba, a new diffusion language model using Mamba backbone architecture that achieves up to 8.2x higher inference throughput than Transformer-based models while maintaining comparable performance. The model demonstrates linear scaling with sequence length and represents a significant advancement in efficient AI text generation systems.

AIBullisharXiv – CS AI · Mar 27/1017
🧠

SceneTok: A Compressed, Diffusable Token Space for 3D Scenes

SceneTok introduces a novel 3D scene tokenizer that compresses view sets into permutation-invariant tokens, achieving 1-3 orders of magnitude better compression than existing methods while maintaining state-of-the-art reconstruction quality. The system enables efficient 3D scene generation in 5 seconds using a lightweight decoder that can render novel viewpoints.

AIBullisharXiv – CS AI · Mar 27/1019
🧠

Provably Safe Generative Sampling with Constricting Barrier Functions

Researchers have developed a safety filtering framework that ensures AI generative models like diffusion models produce outputs that satisfy hard constraints without requiring model retraining. The approach uses Control Barrier Functions to create a 'constricting safety tube' that progressively tightens constraints during the generation process, achieving 100% constraint satisfaction across image generation, trajectory sampling, and robotic manipulation tasks.

AIBullisharXiv – CS AI · Feb 276/105
🧠

BetterScene: 3D Scene Synthesis with Representation-Aligned Generative Model

BetterScene is a new AI approach that enhances 3D scene synthesis and novel view generation from sparse photos by leveraging Stable Video Diffusion with improved regularization techniques. The method integrates 3D Gaussian Splatting and addresses consistency issues in existing diffusion-based solutions through temporal equivariance and vision foundation model alignment.

$RNDR
AIBullisharXiv – CS AI · Feb 276/105
🧠

dLLM: Simple Diffusion Language Modeling

Researchers introduce dLLM, an open-source framework that unifies core components of diffusion language modeling including training, inference, and evaluation. The framework enables users to reproduce, finetune, and deploy large diffusion language models like LLaDA and Dream while providing tools to build smaller models from scratch with accessible compute resources.

AIBullisharXiv – CS AI · Feb 276/108
🧠

Test-Time Scaling with Diffusion Language Models via Reward-Guided Stitching

Researchers developed a new framework called 'Stitching Noisy Diffusion Thoughts' that improves AI reasoning by combining the best parts of multiple solution attempts rather than just selecting complete answers. The method achieves up to 23.8% accuracy improvement on math and coding tasks while reducing computation time by 1.8x compared to existing approaches.

AINeutralarXiv – CS AI · Feb 276/1011
🧠

Why Diffusion Language Models Struggle with Truly Parallel (Non-Autoregressive) Decoding?

Researchers identify why Diffusion Language Models (DLMs) struggle with parallel token generation, finding that training data structure forces autoregressive-like behavior. They propose NAP, a data-centric approach using multiple independent reasoning trajectories that improves parallel decoding performance on math benchmarks.

AIBullisharXiv – CS AI · Feb 276/105
🧠

Diffusion Model in Latent Space for Medical Image Segmentation Task

Researchers developed MedSegLatDiff, a new AI framework combining variational autoencoders with diffusion models for medical image segmentation. The system operates in compressed latent space to reduce computational costs while generating multiple plausible segmentation masks, achieving state-of-the-art performance on skin lesion, polyp, and lung nodule datasets.

AIBullisharXiv – CS AI · Feb 276/106
🧠

ColoDiff: Integrating Dynamic Consistency With Content Awareness for Colonoscopy Video Generation

ColoDiff is a new AI framework that uses diffusion models to generate high-quality colonoscopy videos for medical training and diagnosis. The system addresses data scarcity in medical imaging by creating synthetic videos with temporal consistency and precise clinical attribute control, achieving 90% faster generation through optimized sampling.

AINeutralOpenAI News · Jun 206/106
🧠

Consistency Models

Diffusion models have made significant breakthroughs in generating images, audio, and video content. However, these models face a key limitation in their reliance on iterative sampling processes, which results in slower generation speeds.

AIBullishHugging Face Blog · May 236/105
🧠

Instruction-tuning Stable Diffusion with InstructPix2Pix

The article discusses InstructPix2Pix, a method for instruction-tuning Stable Diffusion models to enable text-guided image editing. This technique allows users to provide natural language instructions to modify existing images rather than generating new ones from scratch.

AINeutralLil'Log (Lilian Weng) · Jul 116/10
🧠

What are Diffusion Models?

Diffusion models are a new type of generative AI model that can learn complex data distributions and generate high-quality images competitive with state-of-the-art GANs. The article covers recent developments including classifier-free guidance, GLIDE, unCLIP, Imagen, latent diffusion models, and consistency models.

AINeutralarXiv – CS AI · Apr 75/10
🧠

BLK-Assist: A Methodological Framework for Artist-Led Co-Creation with Generative AI Models

Researchers have developed BLK-Assist, a modular framework that enables artists to fine-tune AI diffusion models using their own artwork while maintaining privacy and stylistic control. The system includes three components for concept generation, transparency-preserving assets, and high-resolution outputs, demonstrating a consent-based approach to human-AI collaboration in creative work.

AINeutralarXiv – CS AI · Apr 65/10
🧠

Adaptive Guidance for Retrieval-Augmented Masked Diffusion Models

Researchers introduce ARAM (Adaptive Retrieval-Augmented Masked Diffusion), a training-free framework that improves AI language generation by dynamically adjusting guidance based on retrieved context quality. The system addresses noise and conflicts in retrieval-augmented generation for diffusion-based language models, showing improved performance on knowledge-intensive QA benchmarks.

AINeutralarXiv – CS AI · Mar 164/10
🧠

Finite Difference Flow Optimization for RL Post-Training of Text-to-Image Models

Researchers propose a new online reinforcement learning method for improving text-to-image diffusion models that reduces variance by comparing paired trajectories and treating the entire sampling process as a single action. The approach demonstrates faster convergence and better image quality and prompt alignment compared to existing methods.

AINeutralarXiv – CS AI · Mar 124/10
🧠

PC-Diffuser: Path-Consistent Capsule CBF Safety Filtering for Diffusion-Based Trajectory Planner

Researchers developed PC-Diffuser, a safety framework for autonomous vehicle trajectory planning that integrates certifiable safety measures directly into diffusion-based planning models. The system addresses safety failures in AI-driven autonomous vehicles by embedding barrier functions into the denoising process rather than applying safety fixes after planning.

AINeutralarXiv – CS AI · Mar 54/10
🧠

Conjuring Semantic Similarity

Researchers propose a novel method for measuring semantic similarity between text by comparing the image distributions generated by AI models from textual prompts, rather than traditional text-based comparisons. The approach uses Jeffreys divergence between diffusion model outputs to quantify semantic distance, offering new evaluation methods for text-conditioned generative models.

AINeutralarXiv – CS AI · Mar 54/10
🧠

Fairness Begins with State: Purifying Latent Preferences for Hierarchical Reinforcement Learning in Interactive Recommendation

Researchers propose DSRM-HRL, a new framework that uses diffusion models to purify user preference data and hierarchical reinforcement learning to balance recommendation accuracy with fairness. The system addresses bias in interactive recommendation systems by separating state estimation from decision-making, achieving better outcomes on both utility and exposure equity.

AINeutralarXiv – CS AI · Mar 44/103
🧠

AnchorDrive: LLM Scenario Rollout with Anchor-Guided Diffusion Regeneration for Safety-Critical Scenario Generation

Researchers have developed AnchorDrive, a two-stage AI framework that combines large language models with diffusion models to generate realistic safety-critical scenarios for autonomous driving systems. The system uses LLMs for controllable scenario generation based on natural language instructions, then employs diffusion models to create realistic driving trajectories.

AINeutralarXiv – CS AI · Mar 44/102
🧠

Diffusion-MPC in Discrete Domains: Feasibility Constraints, Horizon Effects, and Critic Alignment: Case study with Tetris

Researchers studied diffusion-based model predictive control in discrete domains using Tetris, finding that feasibility constraints are necessary and shorter planning horizons outperform longer ones. The study reveals structural challenges with discrete diffusion planners, particularly misalignment issues with DQN critics that produce high decision regret.

← PrevPage 6 of 7Next →