Skip to main content

Questions tagged [diffusion]

0 votes
0 answers
14 views

Title: Why does the DDPM noise predictor model require both the image and time step as input? Question: In DDPM (Denoising Diffusion Probabilistic Models), the model predicts noise in the denoising ...
459zyt's user avatar
  • 1
0 votes
0 answers
51 views

In diffusion models, the forward process is $q(x_t \mid x_{t-1}) = \mathcal{N}\big(x_{t};\sqrt{1-\beta_t} x_{t-1},\beta_t I\big)$, and the reverse model is parameterized as $p_\theta(x_{t-1}\mid x_t)=\...
user24200147's user avatar
7 votes
1 answer
182 views

Let $X$ follow an inverse Gaussian distribution, and $Y\mid X$ a Gaussian distribution. $$X \sim IG\left( \frac{\alpha}{v_X}, \frac{\alpha^2}{2D_X} \right)$$ $$Y_{\text{given $X=x$}} \sim \mathcal N(...
Sextus Empiricus's user avatar
1 vote
1 answer
80 views

When training consistency models with distillation, the loss is designed to drive the model to produce similar outputs on two consecutive points of the discretized probability flow ODE trajectory (eq. ...
Andrea Allais's user avatar
0 votes
1 answer
115 views

I'm trying to derive the ELBO (Evidence Lower Bound) based loss-function used for training Diffusion Models. The following equation(s) are from arXiv:2208.11970 Eq. 43 is written as follows: $$ \...
x.projekt's user avatar
  • 240
0 votes
0 answers
39 views

I've been reading about score matching and I have a very basic question about how one would (naively) implement the algorithm via gradient descent. Say I have some sort of neural network that that ...
Vasting's user avatar
  • 155
1 vote
1 answer
229 views

I am reading through the DDPM paper, and I am trying to understand the following. Imagine that $\epsilon_{\theta}(x_t,t)$ is our noise predictor. Further imagine that it is fully expressive, i.e., $\...
Max's user avatar
  • 11
0 votes
1 answer
309 views

I am trying to understand how the linear relationship between the diffusion noise prediction model $\epsilon_\theta(x_t)$ which predicts noise added to a sample and the score function is derived $$\...
JustBlaze's user avatar
4 votes
1 answer
303 views

I am going through the derivation for Denoising Diffusion Probabilistic Models (DDPMs) based on Calvin Luo's Diffusion tutorial, where he finally develops the reconstruction term, the prior matching ...
randomforest42's user avatar
2 votes
1 answer
231 views

Can anyone help me with understanding how the $\tilde{\beta}$ and ${\tilde\mu_t{(x_t, x_0)}}$ are derived? It seems to me that exponential term is a 2nd order polynomial term and it doesn't really ...
kaizerbox's user avatar
4 votes
2 answers
1k views

As pointed out by the DDPM paper, we can choose to reparameterize the prediction of the mean to prediction of the total noise "εθ is a function approximator intended to predict ε from x" (...
Daniel Mendoza's user avatar
2 votes
0 answers
163 views

I have read several papers about diffusion models in the context of deep learning. especially this one As explained in the paper, by learning the score function (∇log(𝑝𝑡(𝑥))) ,probability flow ode ...
saleh's user avatar
  • 121
2 votes
1 answer
334 views

In DDPM, ${\tilde\mu}_t$ is the mean of the conditional distribution $q(x_{t-1}|x_t,x_0)$ while the neural network $\mu_\theta$ is modeling a different conditional distribution $p_\theta(x_{t-1}|x_t)$....
Daniel Mendoza's user avatar
2 votes
1 answer
155 views

I've seen in many tutorials on diffusion models refer to the distribution of the latent variables induced by the forward process as "ground truth". I wonder why. What we can actually see is ...
Daniel Mendoza's user avatar
2 votes
2 answers
197 views

ELBO is a lower bound, and only matches the true likelihood when the q-distribution/encoder we choose equals to the true posterior distribution. Are there any guarantees that maximizing ELBO indeed ...
Daniel Mendoza's user avatar

15 30 50 per page