Inquiry about the Role of use_dynamic_rescale in Training vs. Inference
Thanks for the great work! I am currently delving into the functionality of the use_dynamic_rescale parameter in your project and have encountered a point of confusion that I hope you can clarify.
It appears that during the training phase, use_dynamic_rescale is applied to the input data xt (x = x * extract_into_tensor(self.scale_arr, t, x.shape)). However, during inference, it seems the rescaling is performed on the predicted x0(prev_scale_t = torch.full(size, self.ddim_scale_arr_prev[index], device=device)). This discrepancy—wherein the adjustment is made to the inputs during training but to the predictions at inference time—raises questions about the alignment between the training and inference processes. Is there any reference for this strategy?
Same here, hope @Doubiiu can share some insights
Same confusion
any update on this question? I'm also little confused and can't find related literature about this strategy.