Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

A problem about the weight λ of Lvlb #114

Open
yinguanchun opened this issue Sep 30, 2023 · 4 comments
Open

A problem about the weight λ of Lvlb #114

yinguanchun opened this issue Sep 30, 2023 · 4 comments

Comments

@yinguanchun
Copy link

In the paper, λ is 0.001. The code sets learn_sigma as True and rescale_learned_sigmas as False, so the loss type will be gd.LossType.MSE, in this loss type ,the Lvlb will not multply 0.001. Even if the loss type is gd.LossType.RESCALED_MSE, terms["vb"] *= self.num_timesteps / 1000.0, what is self.num_timesteps, and what is its effect?
Thank you .

@toyot-li
Copy link

@yinguanchun I am also confused about this scaling factor, have you understood that?

@Feynman1999
Copy link

I am also confused about this scaling factor, have you understood that?

@yhy258
Copy link

yhy258 commented Aug 1, 2024

In my opinion, authors define L_{vlb} = L_0 + ... + L_T, not L_t.
Thus, they may calculate the vlb loss with scale factor T (self.num_timestep).

@unl1002
Copy link

unl1002 commented Nov 23, 2024

@yhy258 Thank you for your answer, so, which means we use L_t * T (self. num_timestep) to approximate L_ {vlb}?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants