PairFlow: Closed-Form Source-Target Coupling for Few-Step Generation in Discrete Flow Models

ICLR 2026

TL;DR

PairFlow is a teacher-free acceleration framework for Discrete Flow Models that builds source–target training pairs via a closed-form inversion/backward velocity, so the model learns straighter, few-step paths. It’s cheap (≈0.2–1.7% of full training compute) yet improves few-step sampling and can even strengthen the base model for later distillation.

PairFlow Teaser

Abstract

We introduce PairFlow, a lightweight preprocessing step for training Discrete Flow Models (DFMs) to achieve few-step sampling without requiring a pretrained teacher. DFMs have recently emerged as a new class of generative models for discrete data, offering strong performance. However, they suffer from slow sampling due to their iterative nature. Existing acceleration methods largely depend on finetuning, which introduces substantial additional training overhead. PairFlow addresses this issue with a lightweight preprocessing step. Inspired by ReFlow and its extension to DFMs, we train DFMs from coupled samples of source and target distributions, without requiring any pretrained teacher.

At the core of our approach is a closed-form inversion for DFMs, which allows efficient construction of paired source–target samples. Despite its extremely low cost, taking only up to 1.7% of the compute needed for full model training, PairFlow matches or even surpasses the performance of two-stage training involving finetuning. Furthermore, models trained with our framework provide stronger base models for subsequent distillation, yielding further acceleration after finetuning. Experiments on molecular data as well as binary and RGB images demonstrate the broad applicability and effectiveness of our approach.

Method

PairFlow is a training framework for DFMs that enables few-step sampling by constructing paired source–target samples using closed-form velocities. While inspired by ReDi-style coupling-driven training, our approach eliminates the need for a pretrained teacher by using closed-form formulations and achieves acceleration without finetuning. The algorithm for computing source–target pairs is fully parallelizable and requires at most 1.7% of compute needed for full model training.

PairFlow Method Overview

Figure 1: Illustrations of data inversion in PairFlow and the standard corruption process in UDLM. PairFlow achieves a lower average Hamming distance (6.47 vs. 9.0), promoting straighter paths during training.

Algorithm 1: PairFlow Source-Target Coupling

Algorithm 1: Procedure for generating source-target pairs using the closed-form backward velocity.

Closed-Form Velocity Fields

We derive the closed-form forward velocity field $\hat{v}_t(x^i, z)$ for discrete domains. For the empirical target distribution $\tilde{q}(x)$, the closed-form denoiser $p_{1|t}(x^i|z)$ and its associated velocity field $\hat{v}_t(x^i, z)$ are given by:

$$ \begin{align} p_{1|t}(x^i|z) = \cfrac{\sum_{m=1}^M \delta_{d_{m}^i}(x^i) \gamma^{- h(d_{m}, z)}}{ \sum_{m=1}^M \gamma^{- h(d_{m}, z)}}, \quad \Longrightarrow \quad \hat{v}_t(x^i, z) = \frac{\dot\kappa_t}{1 - \kappa_t} \left[p_{1|t}(x^i|z) - \delta_{z}(x^i) \right]. \tag{9} \end{align} $$

To effectively identify suitable source-target pairs, we derive the closed-form backward velocity. Specifically, we first derive the closed-form noise predictor $p_{0|t}(x^i|z)$:

$$ \begin{align} p_{0|t}(x^i|z) = \delta_{z}(x^i) - \frac{\kappa_t(K \delta_{x^i}(z^i) - 1)}{1 + (K-1)\kappa_t} \frac{\sum_{m=1}^M \delta_{d_m^i}(z^i)\gamma^{-h(d_m, z)}}{\sum_{m=1}^M \gamma^{-h(d_m, z)}}. \tag{11} \end{align} $$

Substituting this into the definition of the backward velocity field, we obtain the desired closed-form expression:

$$ \begin{align} \check{v}_t(x^i, z) = \frac{\dot{\kappa}_t(K \delta_{x^i}(z^i)-1)}{1 + (K-1)\kappa_t} \frac{\sum_{m=1}^M \delta_{d_m^i}(z^i)\gamma^{-h(d_m, z)}}{\sum_{m=1}^M \gamma^{-h(d_m, z)}}. \tag{12} \end{align} $$

PairFlow Algorithm

Using the derived closed-form backward velocity, we can efficiently compute source-target pairs. As outlined in Algorithm 1, starting from a given target data sample $x_1$, we iteratively apply the backward update rule using $\check{v}_t(x^i, z)$ to trace back the probability path. The final state $x_0$ reached by this iterative process serves as the source noise. This directly constructs a well-aligned source-target pair $(x_0, x_1)$ purely from data, effectively eliminating the need for generating samples from a pretrained teacher model.

Quantitative Results

We validate the effectiveness of the proposed method and the source–target pairs it discovers across several discrete generative modeling benchmarks involving molecular data and images.

QM9 Validity
QM9 Uniqueness
QM9 Novelty

Figure 2: Step-wise performance analysis on the QM9 dataset.

ZINC-250k Validity
ZINC-250k Uniqueness
ZINC-250k Novelty

Figure 3: Step-wise performance analysis on the ZINC-250k dataset.

MNIST-Binary FID
CIFAR-10 FID
CIFAR-10 IS

Figure 4: Step-wise performance analysis on discretized image datasets.

Qualitative Results

MNIST-Binary (1-Step)
MDLM
UDLM
PAIRFLOW
UDLM + DCD
PAIRFLOW + DCD
UDLM + ReDi
PAIRFLOW + ReDi
MDLM 0
UDLM 0
Ours 0
UDLM+DCD 0
Ours+DCD 0
UDLM+ReDi 0
Ours+ReDi 0
MDLM 1
UDLM 1
Ours 1
UDLM+DCD 1
Ours+DCD 1
UDLM+ReDi 1
Ours+ReDi 1
MDLM 2
UDLM 2
Ours 2
UDLM+DCD 2
Ours+DCD 2
UDLM+ReDi 2
Ours+ReDi 2
MDLM 3
UDLM 3
Ours 3
UDLM+DCD 3
Ours+DCD 3
UDLM+ReDi 3
Ours+ReDi 3
CIFAR-10 (64 Steps)
MDLM
MDLM 0 MDLM 1 MDLM 2 MDLM 3 MDLM 4
UDLM
UDLM 0 UDLM 1 UDLM 2 UDLM 3 UDLM 4
PAIRFLOW
Ours 0 Ours 1 Ours 2 Ours 3 Ours 4
CIFAR-10 (256 Steps)
MDLM
MDLM 0 MDLM 1 MDLM 2 MDLM 3 MDLM 4
UDLM
UDLM 0 UDLM 1 UDLM 2 UDLM 3 UDLM 4
PAIRFLOW
Ours 0 Ours 1 Ours 2 Ours 3 Ours 4

Citation

@inproceedings{park2026:pairflow,
  title={PairFlow: Closed-Form Source-Target Coupling for Few-Step Generation in Discrete Flow Models},
  author={Park, Mingue and Hwang, Jisung and Yoo, Seungwoo and Yeo, Kyeongmin and Sung, Minhyuk},
  booktitle={ICLR},
  year={2026}
}