Hi, nice repo, really appreciate it.
One thing is that in the implementation of Song's consistency models, before inputting sigmas in to the network, there is a rescale:
rescaled_t = 1000 * 0.25 * th.log(sigmas + 1e-44)
You can chekck it here: https://github.com/openai/consistency_models/blob/e32b69ee436d518377db86fb2127a3972d0d8716/cm/karras_diffusion.py#L346C58-L346C58
Similarly, in EDM's implementation, there is also a rescale before inputting sigma to the network.
c_noise = sigma.log() / 4
The link: https://github.com/NVlabs/edm/blob/62072d2612c7da05165d6233d13d17d71f213fee/training/networks.py#L663C9-L663C34
But I did not find this rescaling in your implementation.
I am aware of that the code for improved consistency model has not released yet, so we really do not know if there is such an operation, what do you think?
Hi, nice repo, really appreciate it.
One thing is that in the implementation of Song's consistency models, before inputting sigmas in to the network, there is a rescale:
rescaled_t = 1000 * 0.25 * th.log(sigmas + 1e-44)You can chekck it here: https://github.com/openai/consistency_models/blob/e32b69ee436d518377db86fb2127a3972d0d8716/cm/karras_diffusion.py#L346C58-L346C58
Similarly, in EDM's implementation, there is also a rescale before inputting sigma to the network.
c_noise = sigma.log() / 4The link: https://github.com/NVlabs/edm/blob/62072d2612c7da05165d6233d13d17d71f213fee/training/networks.py#L663C9-L663C34
But I did not find this rescaling in your implementation.
I am aware of that the code for improved consistency model has not released yet, so we really do not know if there is such an operation, what do you think?