ICLR 2026 Orals

GLASS Flows: Efficient Inference for Reward Alignment of Flow and Diffusion Models

Peter Holderrieth, Uriel Singer, Tommi Jaakkola, Ricky T. Q. Chen, Yaron Lipman, Brian Karrer

Diffusion & Flow Matching Thu, Apr 23 · 11:30 AM–11:40 AM · 201 A/B Avg rating: 7.00 (6–8)
Author-provided TL;DR

We improve inference-time reward alignment of flow matching and diffusion models by proposing a novel sampling paradigm that enables more efficient exploration.

Abstract

The performance of flow matching and diffusion models can be greatly improved at inference time using reward adaptation algorithms, yet efficiency remains a major limitation. While several algorithms were proposed, we demonstrate that a common bottleneck is the *sampling* method these algorithms rely on: many algorithms require to sample Markov transitions via SDE sampling, which is significantly less efficient and often less performant than ODE sampling. To remove this bottleneck, we introduce GLASS Flows, a new sampling paradigm that simulates a ''flow matching model within a flow matching model'' to sample Markov transitions. As we show in this work, this ''inner'' flow matching model can be retrieved from any pre-trained model without any re-training, effectively combining the efficiency of ODEs with the stochastic evolution of SDEs. On large-scale text-to-image models, we show that GLASS Flows eliminate the trade-off between stochastic evolution and efficiency. GLASS Flows improve state-of-the-art performance in text-to-image generation, making it a simple, drop-in solution for inference-time scaling of flow and diffusion models.

One-sentence summary·Auto-generated by claude-haiku-4-5-20251001(?)

GLASS Flows samples Markov transitions via inner flow matching models to improve inference-time reward alignment in flow and diffusion models.

Contributions·Auto-generated by claude-haiku-4-5-20251001(?)
  • Introduces GLASS Flows, sampling paradigm that simulates inner flow matching model within outer model
  • Retrieves inner flow matching model from pre-trained models without retraining using sufficient statistics
  • Combines efficiency of ODE sampling with stochasticity of SDE evolution for reward adaptation
  • Improves state-of-the-art performance in text-to-image generation as drop-in replacement for existing methods
Methods used·Auto-generated by claude-haiku-4-5-20251001(?)
  • Flow matching
  • ODE sampling
  • SDE sampling
  • Reward adaptation
Datasets used·Auto-generated by claude-haiku-4-5-20251001(?)
  • Text-to-image generation tasks
Limitations (author-stated)·Auto-generated by claude-haiku-4-5-20251001(?)

Authors did not state explicit limitations.

Future work (author-stated)·Auto-generated by claude-haiku-4-5-20251001(?)
  • Explore applying GLASS Flows to other methods relying on SDE sampling such as reward fine-tuning and image editing
    from the paper
  • Explore learning or dynamically adjusting the correlation parameter rho defining GLASS transitions
    from the paper

Author keywords

  • Flow Matching; Diffusion Models; Reward Alignment; Reward Adaptation; Inference-time scaling; Feynman-Kac Steering; Markov transitions; Sampling methods

Related orals

Generative Human Geometry Distribution

Introduces distribution-over-distribution model combining geometry distributions with two-stage flow matching for human 3D generation.

Avg rating: 5.50 (2–8) · Xiangjun Tang et al.
Something off? Let us know →