Compositional Diffusion with Guided search for Long-Horizon Planning
Introduces CDGS integrating compositional diffusion with guided search for coherent long-horizon plan generation.
Diffusion models, score-based generative models, flow matching, consistency models, and related generative techniques.
Introduces CDGS integrating compositional diffusion with guided search for coherent long-horizon plan generation.
Cross-domain lossy compression unifies rate and classification constraints via optimal transport framework.
Distills AlphaFold3 into single-step sampler with temporal geodesic matching achieving 15x inference acceleration.
Triple-BERT addresses order dispatching via centralized SARL with action decomposition and BERT-based attention.
EditVerse unifies image and video generation/editing via token sequences enabling cross-modal knowledge transfer.
Introduces distribution-over-distribution model combining geometry distributions with two-stage flow matching for human 3D generation.
GLASS Flows samples Markov transitions via inner flow matching models to improve inference-time reward alignment in flow and diffusion models.
Neon inverts model degradation from self-training by extrapolating away from it, improving generative models with minimal compute.
NextStep-1 achieves state-of-the-art autoregressive text-to-image generation by modeling continuous image tokens with lightweight flow matching instead of diffusion.
Pareto-Conditioned Diffusion formulates offline multi-objective optimization as conditional sampling problem avoiding explicit surrogate models.
Theoretical characterization shows MDMs are expressively equivalent to padded looped transformers, more efficient for parallel problems.
Quotient-space diffusion models reduce learning difficulty for molecular structure generation via SE(3) symmetry handling.
SGF unifies negative guidance in safe generation via MMD potentials and control barrier analysis with time-critical guidance windows.
Proteina-Complexa unifies generative modeling and hallucination for atomistic binder design via pretraining on Teddymer and test-time optimization.
Generates ultra-long videos by actively correcting self-generated errors through error-recycling fine-tuning.
Spacetime perspective views diffusion latent spaces as Fisher-Rao metric manifolds enabling efficient geodesic computation without simulation.
TRACE reveals diffusion models encode hidden instance boundary priors and leverages them for unsupervised instance segmentation without dense annotations.
RealUID provides universal distillation for matching models without GANs, incorporating real data into one-step generator training.