-
Notifications
You must be signed in to change notification settings - Fork 26
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
NaNs during training #1
Comments
+1 |
@eps696 there is a bug if you are using pytorch>=1.3.0: y = F.grid_sample(frame1, flow.permute(0,2,3,1), padding_mode="border") with y = F.grid_sample(frame1, flow.permute(0,2,3,1), padding_mode="border", align_corners=True) |
@olegkhomenko thanks; i use pytorch <= 1.2.0, there's no such option |
Have you solved this problem? could you please tell me the solution? |
@HellwayXue for that time i just added some dumb logging after defining
you may want to use |
@eps696 Thank you for your reply! but how these nans come from? I've tried gradient clipping, optimizer weight decay and none of them work. Your suggestions are after defining codebook, but i get nans even if i don't save it. |
@HellwayXue no idea alas |
Thanks you for this very interesting repo.
I got some issues with training though: the parameters (codebook values for motion and total_loss for appearance) drop to NaN within few epochs (5-20) when trained with default hyperparameters.
Increase in learning rate to 0.5~1e-4 eliminates this behaviour, but doesn't look like a solution.
Did you encounter such issues in your practice and do you have any advice to sort this out?
The text was updated successfully, but these errors were encountered: