Denoising Diffusion Implicit Models

0
Structured data

A study conducted in the UK from 2009 to 2010 by leading scientists explored neonatal resuscitation practices in various neonatal units, aiming to assess adherence to international guidelines and identify differences between tertiary and non-tertiary care providers...

Read on arXivCardiologyLorem ipsum dolor sit amet, consectetur adipiscing elit. Sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.

One Sentence Abstract

Denoising diffusion implicit models (DDIMs) are introduced, accelerating image sampling up to 50x faster than DDPMs, enabling flexible trade-offs for sample quality and high-quality interpolation and reconstruction in the latent space.

Simplified Abstract

Researchers have developed a new tool called Denoising Diffusion Implicit Models (DDIMs) to improve the way images are generated. This new tool works by speeding up the process of creating images, making it more efficient compared to the previous method called Denoising Diffusion Probabilistic Models (DDPMs).

To understand how DDPMs work, imagine a journey where a person moves step-by-step from a noisy, blurry image to a clear, beautiful one. DDPMs help the person take these steps, but it takes a long time. DDIMs, on the other hand, is like finding a shortcut in this journey, allowing the person to reach the final, clear image much faster.

DDIMs achieve this by changing the way the images are generated. They use a new type of process that can be deterministic, which means it follows a specific set of rules, making the journey more predictable and faster. By using DDIMs, the researchers found that they could create high-quality images up to 50 times faster than before.

This new method, DDIMs, is also useful for making images look more like a blend between two given images, and for fixing mistakes in images, making them very accurate. This research is significant because it provides a new and more efficient way to generate images, which will help scientists work together across countries and improve the quality of their work.

Study Fields

Main fields:

  • Image Generation
  • Diffusion Probabilistic Models (DDPMs)
  • Denoising Diffusion Implicit Models (DDIMs)

Subfields:

  • Generative Processes
  • Markov Chains
  • Non-Markovian Diffusion Processes
  • Implicit Models
  • Training Procedures
  • Sampling Efficiency
  • Wall-clock Time
  • Sample Quality
  • Semantic Image Interpolation
  • Latent Space
  • Observation Reconstruction

Study Objectives

  • Develop a more efficient class of iterative implicit probabilistic models called denoising diffusion implicit models (DDIMs)
  • Accelerate sampling by reducing the time required to produce a sample
  • Achieve high-quality image generation without adversarial training
  • Generalize denoising diffusion probabilistic models (DDPMs) by introducing non-Markovian diffusion processes
  • Demonstrate that DDIMs can produce high-quality samples faster than DDPMs (10× to 50× faster)
  • Allow for trade-offs between computation and sample quality
  • Enable semantically meaningful image interpolation directly in the latent space
  • Reconstruct observations with very low error

Conclusions

  • The researchers introduce denoising diffusion implicit models (DDIMs) as a more efficient class of iterative implicit probabilistic models, aimed at accelerating sampling in comparison to DDPMs.
  • DDIMs are trained using the same procedure as DDPMs, but they offer a significant speed advantage. Empirical results demonstrate that DDIMs can produce high-quality samples 10x to 50x faster than DDPMs in terms of wall-clock time.
  • DDIMs enable traders-off between computation and sample quality.
  • DDIMs allow for semantically meaningful image interpolation directly in the latent space.
  • DDIMs can reconstruct observations with very low error.

References

Tero. Karras, Samuli. Laine, Miika. Aittala, Janne. Hellsten, Jaakko. Lehtinen, Timo. Aila
Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  8110–8119, 2020.
Aaron. Van Den Oord, Sander. Dieleman, Heiga. Zen, Karen. Simonyan, Oriol. Vinyals, Alex. Graves, Nal. Kalchbrenner, Andrew. Senior, Koray. Kavukcuoglu
Aaron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal Kalchbrenner, Andrew Senior, and Koray Kavukcuoglu. WaveNet: A generative model for raw audio. arXiv preprint arXiv:1609.03499, September 2016a.
Ian. Goodfellow, Jean. Pouget-Abadie, Mehdi. Mirza, Bing. Xu, David. Warde-Farley, Sherjil. Ozair, Aaron. Courville, Yoshua. Bengio
Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in neural information processing systems, pp. 2672–2680, 2014.
P. Diederik, Max. Kingma, None. Welling
Diederik P Kingma and Max Welling. Auto-Encoding variational bayes. arXiv preprint arXiv:1312.6114v10, December 2013.
Aaron. Van Den Oord, Nal. Kalchbrenner, Koray. Kavukcuoglu
Aaron van den Oord, Nal Kalchbrenner, and Koray Kavukcuoglu. Pixel recurrent neural networks. arXiv preprint arXiv:1601.06759, January 2016b.
Danilo. Jimenez, Rezende. None, Shakir. Mohamed
Danilo Jimenez Rezende and Shakir Mohamed. Variational inference with normalizing flows. arXiv preprint arXiv:1505.05770, May 2015.
Laurent. Dinh, Jascha. Sohl-Dickstein, Samy. Bengio
Laurent Dinh, Jascha Sohl-Dickstein, and Samy Bengio. Density estimation using real NVP. arXiv preprint arXiv:1605.08803, May 2016.
Martin. Arjovsky, Soumith. Chintala, Léon. Bottou, Gan. Wasserstein
Martin Arjovsky, Soumith Chintala, and Léon Bottou. Wasserstein GAN. arXiv preprint arXiv:1701.07875, January 2017.
Ishaan. Gulrajani, Faruk. Ahmed, Martin. Arjovsky, Aaron.C. Vincent Dumoulin, None. Courville
Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, and Aaron C Courville. Improved training of wasserstein gans. In Advances in Neural Information Processing Systems, pp. 5769–5779, 2017.
Tero. Karras, Samuli. Laine, Timo. Aila
Tero Karras, Samuli Laine, and Timo Aila. A Style-Based generator architecture for generative adversarial networks. arXiv preprint arXiv:1812.04948, December 2018.
Andrew. Brock, Jeff. Donahue, Karen. Simonyan
Andrew Brock, Jeff Donahue, and Karen Simonyan. Large scale GAN training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096, September 2018.
Shengjia. Zhao, Hongyu. Ren, Arianna. Yuan, Jiaming. Song, Noah. Goodman, Stefano. Ermon
Shengjia Zhao, Hongyu Ren, Arianna Yuan, Jiaming Song, Noah Goodman, and Stefano Ermon. Bias and generalization in deep generative models: An empirical study. In Advances in Neural Information Processing Systems, pp. 10792–10801, 2018.
Yoshua. Bengio, Eric. Laufer, Guillaume. Alain, Jason. Yosinski
Yoshua Bengio, Eric Laufer, Guillaume Alain, and Jason Yosinski. Deep generative stochastic networks trainable by backprop. In International Conference on Machine Learning, pp. 226–234, January 2014.
Jonathan. Ho, Ajay. Jain, Pieter. Abbeel
Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. arXiv preprint arXiv:2006.11239, June 2020.
Yang. Song, Stefano. Ermon
Yang Song and Stefano Ermon. Generative modeling by estimating gradients of the data distribution. arXiv preprint arXiv:1907.05600, July 2019.
Jascha. Sohl-Dickstein, Eric.A. Weiss, Niru. Maheswaranathan, Surya. Ganguli
Jascha Sohl-Dickstein, Eric A Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. arXiv preprint arXiv:1503.03585, March 2015.
Shakir. Mohamed, Balaji. Lakshminarayanan
Shakir Mohamed and Balaji Lakshminarayanan. Learning in implicit generative models. arXiv preprint arXiv:1610.03483, October 2016.
Danilo. Jimenez Rezende, Shakir. Mohamed, Daan. Wierstra
Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra. Stochastic backpropagation and approximate inference in deep generative models. arXiv preprint arXiv:1401.4082, 2014.
Aapo. HyvärinenJournal of Machine Learning Researc h
Aapo Hyvärinen. Estimation of Non-Normalized statistical models by score matching. Journal of Machine Learning Researc h, 6:695–709, 2005.
Pascal. VincentNeural computation
Pascal Vincent. A connection between score matching and denoising autoencoders. Neural computation, 23(7):1661–1674, 2011.
Nanxin. Chen, Yu. Zhang, Heiga. Zen, Ron.J. Weiss, Mohammad. Norouzi, William. Chan
Nanxin Chen, Yu Zhang, Heiga Zen, Ron J Weiss, Mohammad Norouzi, and William Chan. WaveGrad: Estimating gradients for waveform generation. arXiv preprint arXiv:2009.00713, September 2020.
<FootnoteDefinition authors="Yang. Song, Jascha. Sohl-Dickstein, P. Diederik, Abhishek. Kingma, Stefano. Kumar, Ben. Ermon, None. Poole...

References

Unlock full article access by joining Solve