Composition and Control with Distilled Energy Diffusion Models and Sequential Monte Carlo

James Thornton, Louis Béthune, Ruixiang ZHANG, Arwen Bradley, Preetum Nakkiran, Shuangfei Zhai
Proceedings of The 28th International Conference on Artificial Intelligence and Statistics, PMLR 258:3259-3267, 2025.

Abstract

Diffusion models may be formulated as a time-indexed sequence of energy-based models, where the score corresponds to the negative gradient of an energy function. As opposed to learning the score directly, an energy parameterization is attractive as the energy itself can be used to control generation via Monte Carlo samplers. Architectural constraints and training instability in energy parameterized models have so far yielded inferior performance compared to directly approximating the score or denoiser. We address these deficiencies by introducing a novel training regime for the energy function through distillation of pre-trained diffusion models, resembling a Helmholtz decomposition of the score vector field. We further showcase the synergies between energy and score by casting the diffusion sampling procedure as a Feynman Kac Model where sampling is controlled using potentials from the learnt energy functions. The Feynman Kac model formalism enables composition and low temperature sampling through sequential Monte Carlo.

Cite this Paper


BibTeX
@InProceedings{pmlr-v258-thornton25a, title = {Composition and Control with Distilled Energy Diffusion Models and Sequential Monte Carlo}, author = {Thornton, James and B{\'e}thune, Louis and ZHANG, Ruixiang and Bradley, Arwen and Nakkiran, Preetum and Zhai, Shuangfei}, booktitle = {Proceedings of The 28th International Conference on Artificial Intelligence and Statistics}, pages = {3259--3267}, year = {2025}, editor = {Li, Yingzhen and Mandt, Stephan and Agrawal, Shipra and Khan, Emtiyaz}, volume = {258}, series = {Proceedings of Machine Learning Research}, month = {03--05 May}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v258/main/assets/thornton25a/thornton25a.pdf}, url = {https://proceedings.mlr.press/v258/thornton25a.html}, abstract = {Diffusion models may be formulated as a time-indexed sequence of energy-based models, where the score corresponds to the negative gradient of an energy function. As opposed to learning the score directly, an energy parameterization is attractive as the energy itself can be used to control generation via Monte Carlo samplers. Architectural constraints and training instability in energy parameterized models have so far yielded inferior performance compared to directly approximating the score or denoiser. We address these deficiencies by introducing a novel training regime for the energy function through distillation of pre-trained diffusion models, resembling a Helmholtz decomposition of the score vector field. We further showcase the synergies between energy and score by casting the diffusion sampling procedure as a Feynman Kac Model where sampling is controlled using potentials from the learnt energy functions. The Feynman Kac model formalism enables composition and low temperature sampling through sequential Monte Carlo.} }
Endnote
%0 Conference Paper %T Composition and Control with Distilled Energy Diffusion Models and Sequential Monte Carlo %A James Thornton %A Louis Béthune %A Ruixiang ZHANG %A Arwen Bradley %A Preetum Nakkiran %A Shuangfei Zhai %B Proceedings of The 28th International Conference on Artificial Intelligence and Statistics %C Proceedings of Machine Learning Research %D 2025 %E Yingzhen Li %E Stephan Mandt %E Shipra Agrawal %E Emtiyaz Khan %F pmlr-v258-thornton25a %I PMLR %P 3259--3267 %U https://proceedings.mlr.press/v258/thornton25a.html %V 258 %X Diffusion models may be formulated as a time-indexed sequence of energy-based models, where the score corresponds to the negative gradient of an energy function. As opposed to learning the score directly, an energy parameterization is attractive as the energy itself can be used to control generation via Monte Carlo samplers. Architectural constraints and training instability in energy parameterized models have so far yielded inferior performance compared to directly approximating the score or denoiser. We address these deficiencies by introducing a novel training regime for the energy function through distillation of pre-trained diffusion models, resembling a Helmholtz decomposition of the score vector field. We further showcase the synergies between energy and score by casting the diffusion sampling procedure as a Feynman Kac Model where sampling is controlled using potentials from the learnt energy functions. The Feynman Kac model formalism enables composition and low temperature sampling through sequential Monte Carlo.
APA
Thornton, J., Béthune, L., ZHANG, R., Bradley, A., Nakkiran, P. & Zhai, S.. (2025). Composition and Control with Distilled Energy Diffusion Models and Sequential Monte Carlo. Proceedings of The 28th International Conference on Artificial Intelligence and Statistics, in Proceedings of Machine Learning Research 258:3259-3267 Available from https://proceedings.mlr.press/v258/thornton25a.html.

Related Material