Obsidian/200. Library Papers/sohl-dicksteinDeepUnsupervisedLearning2015a.md

3.8 KiB
Executable File
Raw Blame History

status dateread title year authors DOI ISBN citekey itemType
false Deep Unsupervised Learning using Nonequilibrium Thermodynamics 2015
Sohl-Dickstein, Jascha
Weiss, Eric A.
Maheswaranathan, Niru
Ganguli, Surya
10.48550/arXiv.1503.03585 sohl-dicksteinDeepUnsupervisedLearning2015a preprint

DOI

10.48550/arXiv.1503.03585

ISBN

Tags:

#Computer-Science---Machine-Learning, #Quantitative-Biology---Neurons-and-Cognition, #Condensed-Matter---Disordered-Systems-and-Neural-Networks, #Statistics---Machine-Learning

Contribution::

Related::

[!LINK]

arXiv Fulltext PDF.

[!Abstract]

A central problem in machine learning involves modeling complex data-sets using highly flexible families of probability distributions in which learning, sampling, inference, and evaluation are still analytically or computationally tractable. Here, we develop an approach that simultaneously achieves both flexibility and tractability. The essential idea, inspired by non-equilibrium statistical physics, is to systematically and slowly destroy structure in a data distribution through an iterative forward diffusion process. We then learn a reverse diffusion process that restores structure in data, yielding a highly flexible and tractable generative model of the data. This approach allows us to rapidly learn, sample from, and evaluate probabilities in deep generative models with thousands of layers or time steps, as well as to compute conditional and posterior probabilities under the learned model. We additionally release an open source reference implementation of the algorithm. .

Notes

.

Annotations%% begin annotations %%

Imported: 2024-08-08 12:07 pm

Quote

systematically and slowly destroy structure in a data distribution through an iterative forward diffusion process. We then learn a reverse diffusion process that restores structure in data, yielding a highly flexible and tractable generative model of the data.

Quote

robabilistic models suffer from a tradeoff between two conflicting objectives: tractability and flexibility

Quote

Our goal is to define a forward (or inference) diffusion process which converts any complex data distribution into a simple, tractable, distribution, and then learn a finite-time reversal of this diffusion process which defines our generative model distribution (See Figure 1).

Quote

For both Gaussian and binomial diffusion, for continuous diffusion (limit of small step size β) the reversal of the diffusion process has the identical functional form as the forward process (Feller, 1949).

Quote

Since q (x(t)|x(t1)) is a Gaussian (binomial) distribution, and if βt is small, then q (x(t1)|x(t)) will also be a Gaussian (binomial) distribution.

Quote

= ∫ dx(1···T )q ( x(1···T )|x(0)) · p ( x(T )) T ∏ t=1 p (x(t1)|x(t)) q (x(t)|x(t1)) . (9)

Quote

Thus, the task of estimating a probability distribution has been reduced to the task of performing regression on the functions which set the mean and covariance of a sequence of Gaussians (or set the state flip probability for a sequence of Bernoulli trials).

%% end annotations %%

%% Import Date: 2024-08-08T12:07:42.961-04:00 %%