Probability Path

Conditional Gaussian Probability Path

The Gaussian conditional probability path forms the foundation of denoising diffusion models and flow matching.

Definition: Let , be noise schedulers: two continuously differentiable, monotonic functions with boundary conditions and . We define the conditional probability path as a family of distribution over :

Boundary Conditions: The imposed conditions on and ensure:

where is a data point, is the Dirac delta “distribution” that sample from always returns .

Marginal Gaussian Probability Path

Sampling Procedure: For , , we can sample from marginal path:

This provides a tracable way to sample from the marginal distribution .

Vector Field

Conditinoal Gaussian Vector Field

Definition: Let and denote respective time derivatives of and . The conditional Gaussian vector field given by:

is a valid conditional vector field model.

Property: This vector field generates ODE trajectories that satisfy if .

Proof

Construct a conditional flow model by defining

If is the ODE trajectory of with , then

The conditional vector field is:

In (ii), we reparameterized .

Score Function for Conditional Gaussian Probability Paths

For the Gaussian path , we can use the form of the Gaussian probability density to get the conditional Gaussian score function, which is the derivetive of ,

This linear relationship is a unique property of Gaussian distributions and is fundamental to efficient training.

Flow Matching for Gaussian Conditional Probability Paths

The conditional flow matching loss is

In (i) we replace by .

Let us make even more concrete for the special case of , and . The corresponding conditional probability path is referred to as the (Gaussian) CondOT probability path. Then we have , so that

Models like Stable Diffusion 3, Meta’s Movie Gen Video are using this procedure.

Training Procedure

Given a dataset of samples , vector field network . For each batch of data:

  1. Sample a data example from the dataset,

  2. Sample a random time ,

  3. Sample noise

  4. Set

  5. Compute loss

  6. Update the model.

Score Matching for Gaussian Probability Paths

The conditional score matching loss is

Note that learns to predict the noise that was used to corrupt a data sample . Therefore, the above training loss is also called denoising score matching. It was soon realized that the above loss is numerically unstable for close to zero (i.e. denoising score matching only works if you add a sufficient amount of noise).

In Denosing Diffusion Probabilitic Models (DDPM), it was proposed to drop the constant in the loss, and reparameterize into a noise predictor network via:

thus,