-
Notifications
You must be signed in to change notification settings - Fork 136
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About the tT_loss #63
Comments
Same question for me. The other loss terms written in training_losses_e2e are clear, which are also described in the paper (Equation 2). But I don't quite understand this tT_loss: why loss is calculated on each timestep of this forward diffusion process? |
Seems like Diffusion-LM/improved-diffusion/improved_diffusion/gaussian_diffusion.py Lines 252 to 259 in 759889d
The squared difference between Diffusion-LM/improved-diffusion/improved_diffusion/gaussian_diffusion.py Lines 1566 to 1567 in 759889d
|
Hi, Thanks for your excellent work, but I have a small question about the loss function. When I was reading the code, I found that tT_loss calculates the loss between X_t and 0. Is there any meaning to doing this?
The code in the gaussian_diffusion.py, the function training_losses_e2e in class GaussianDiffusion
out_mean, _, _ = self.q_mean_variance(x_start, torch.LongTensor([self.num_timesteps - 1]).to(x_start.device))
tT_loss = mean_flat(out_mean ** 2)
The text was updated successfully, but these errors were encountered: