PulseAugur
LIVE 06:55:07
research · [6 sources] ·
1
research

New research explores one-step generative models via Wasserstein flows

Two new research papers explore novel approaches to generative modeling, aiming to significantly speed up the process. One paper introduces W-Flow, a framework that uses Wasserstein gradient flows to compress complex evolutionary paths into a single-step generation, achieving state-of-the-art results on ImageNet with drastically reduced sampling times. The second paper investigates the theoretical underpinnings of one-shot generative flows, characterizing when such direct transport maps exist and identifying obstructions for targets with well-separated modes, particularly for Gaussian distributions. AI

Summary written by gemini-2.5-flash-lite from 6 sources. How we write summaries →

IMPACT These papers propose faster, more efficient methods for generative modeling, potentially reducing computational costs and increasing accessibility.

RANK_REASON Two academic papers published on arXiv introducing new theoretical frameworks and empirical results for generative modeling.

Read on arXiv stat.ML →

COVERAGE [6]

  1. arXiv stat.ML TIER_1 · Yu-Jui Huang, Zachariah Malik ·

    Generative Modeling by Minimizing the Wasserstein-2 Loss

    arXiv:2406.13619v4 Announce Type: replace Abstract: This paper develops a generative model by minimizing the second-order Wasserstein loss (the $W_2$ loss) through a distribution-dependent ordinary differential equation (ODE), whose dynamics involves the Kantorovich potential ass…

  2. arXiv stat.ML TIER_1 · Samuel Willis, Paul Duckworth, Jack Simons, Aleksandra Kalisz, Krisztina Sinkovics, Noam Ghenassia, Shikha Surana, Henry T. Oldroyd, Alexandru I. Stere, Dragos D Margineantu, Carl Henrik Ek, Henry Moss, Erik Bodin ·

    Sample-Efficient Optimisation over the Outputs of Generative Models

    arXiv:2509.23800v3 Announce Type: replace Abstract: Modern generative AI models, such as diffusion and flow matching models, can sample from rich data distributions. However, many applications, especially in science and engineering, require more than drawing samples from the mode…

  3. arXiv stat.ML TIER_1 · Romann M. Weber ·

    The Score-Difference Flow for Implicit Generative Modeling

    arXiv:2304.12906v4 Announce Type: replace-cross Abstract: Implicit generative modeling (IGM) aims to produce samples of synthetic data matching the characteristics of a target data distribution. Recent work (e.g. score-matching networks, diffusion models) has approached the IGM p…

  4. arXiv stat.ML TIER_1 Deutsch(DE) · Jiaqi Han, Puheng Li, Qiushan Guo, Renyuan Xu, Stefano Ermon, Emmanuel J. Cand\`es ·

    One-Step Generative Modeling via Wasserstein Gradient Flows

    arXiv:2605.11755v1 Announce Type: cross Abstract: Diffusion models and flow-based methods have shown impressive generative capability, especially for images, but their sampling is expensive because it requires many iterative updates. We introduce W-Flow, a framework for training …

  5. arXiv stat.ML TIER_1 Deutsch(DE) · Emmanuel J. Candès ·

    One-Step Generative Modeling via Wasserstein Gradient Flows

    Diffusion models and flow-based methods have shown impressive generative capability, especially for images, but their sampling is expensive because it requires many iterative updates. We introduce W-Flow, a framework for training a generator that transforms samples from a simple …

  6. arXiv stat.ML TIER_1 · Panos Tsimpos, Daniel Sharp, Youssef Marzouk ·

    One-Shot Generative Flows: Existence and Obstructions

    arXiv:2604.15439v3 Announce Type: replace Abstract: We study dynamic measure transport for generative modeling, focusing on transport maps that connect a source measure $P_0$ to a target measure $P_1$ by integrating a velocity field of the form $v_t(x) = \mathbb{E}[\dot X_t \mid …