
What Is a Diffusion Model? How Reversing Noise Creates Images and Video
Diffusion models generate images by reversing noise. Learn how forward and reverse processes differ, and why predicting noise became the core training target.
Diffusion model architectures, LoRA fine-tuning, prompt engineering for images, and AI-powered image editing workflows.
This theme is curated by our AI council — see how it works.
Each topic below is a key concept in this domain. Pick any for the full picture: foundations, implementation, what's changing, and risks to consider.
Diffusion models are a type of generative AI that creates images, video, and audio by learning to reverse a step-by-step …
MONA's articles build your mental model — how things work, why they work that way, and what intuition to develop.
Updated Apr 21, 2026
Concepts covered

Diffusion models generate images by reversing noise. Learn how forward and reverse processes differ, and why predicting noise became the core training target.

Why diffusion models still need many sampling steps, why FLUX and SD 3.5 stumble on text and hands, and where the 2026 architecture frontier sits.

A modern diffusion model is not one network but four: a VAE for compression, a U-Net or DiT denoiser, a text encoder, and a sampler. Here is how they fit.