TY - UNPB
T1 - Inference-Time Alignment of Diffusion Models via Evolutionary Algorithms
AU - Jajal, Purvish
AU - Eliopoulos, Nick John
AU - Chou, Benjamin Shiue-Hal
AU - Thiruvathukal, George K.
AU - Davis, James C.
AU - Lu, Yung-Hsiang
N1 - P. Jajal and N. J. Eliopoulos contributed equally to this work
PY - 2025/5/30
Y1 - 2025/5/30
N2 - Diffusion models are state-of-the-art generative models, yet their samples often fail to satisfy application objectives such as safety constraints or domain-specific validity. Existing techniques for alignment require gradients, internal model access, or large computational budgets resulting in high compute demands, or lack of support for certain objectives. In response, we introduce an inference-time alignment framework based on evolutionary algorithms. We treat diffusion models as black boxes and search their latent space to maximize alignment objectives. Given equal or less running time, our method achieves 3-35% higher ImageReward scores than gradient-free and gradient-based methods. On the Open Image Preferences dataset, our method achieves competitive results across four popular alignment objectives. In terms of computational efficiency, we require 55% to 76% less GPU memory and are 72% to 80% faster than gradient-based methods.
AB - Diffusion models are state-of-the-art generative models, yet their samples often fail to satisfy application objectives such as safety constraints or domain-specific validity. Existing techniques for alignment require gradients, internal model access, or large computational budgets resulting in high compute demands, or lack of support for certain objectives. In response, we introduce an inference-time alignment framework based on evolutionary algorithms. We treat diffusion models as black boxes and search their latent space to maximize alignment objectives. Given equal or less running time, our method achieves 3-35% higher ImageReward scores than gradient-free and gradient-based methods. On the Open Image Preferences dataset, our method achieves competitive results across four popular alignment objectives. In terms of computational efficiency, we require 55% to 76% less GPU memory and are 72% to 80% faster than gradient-based methods.
KW - cs.LG
M3 - Preprint
BT - Inference-Time Alignment of Diffusion Models via Evolutionary Algorithms
ER -