Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training loss increases with time #31

Open
rollovd opened this issue Sep 6, 2021 · 0 comments
Open

Training loss increases with time #31

rollovd opened this issue Sep 6, 2021 · 0 comments

Comments

@rollovd
Copy link

rollovd commented Sep 6, 2021

Hello there!

I have finetuned your model with 2 output classes (skin and background). As a backbone I took your pretrained (19 classes) model in order to speed-up the training process.
So, I just took context path, ffm, conv, conv16, conv32. Thus, only three convolutional layers (feat_out, feat_out16, feat_out32 respectively) are trained from scratch.

I have utilized all of your hyperparameters and all of your approaches to fit the model. But after some steps (after 250 steps) training loss increase (from 1.58) until 1.98 and afterwards remains there with small fluctuations.
What could be the problem? Any other ideas?

I tried using scheduler after 180-200 steps (initial lr=0.01, gamma=0.1), gradient clipping. Anyway, it is always the same picture - an increase in loss.
Thanks in advance for any ideas!

Batchsize = 64
Initial learning rate = 0.01
Oprimizer: SGD
Loss: OhemCELoss

P.S. I have been waiting for the 2500 steps (near 7 epochs) - there is not a single hint subsequent decreasing.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant