Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Hyper-parameters setting #162

Open
jie-li-hust opened this issue Feb 17, 2025 · 1 comment
Open

Hyper-parameters setting #162

jie-li-hust opened this issue Feb 17, 2025 · 1 comment

Comments

@jie-li-hust
Copy link

jie-li-hust commented Feb 17, 2025

Hi! Thank you for the great work.
I’m fine-tuning this model, and using the checkpoint trained with focal seems to lower performance on COCO and RefCOCOg. Could the batch size be influencing this as well?
I am training on 4 GPUs, the setting is "TRAIN.BATCH_SIZE_TOTAL 20 \ TRAIN.BATCH_SIZE_PER_GPU 5 \ DATALOADER_NUM_WORKERS 4" ,and results on RefCOCOg (~62-63% cIoU),COCO(~38.1% mAP, ~60.7% mIoU).
Would increasing the epoch parameter, reducing the LR_MULTIPLIER for the backbone (to 0.05), or lowering WARMUP_ITERS(to 5) be helpful?

@jie-li-hust
Copy link
Author

By the way, the demo code shows davit.py, but how can we calculate the parameters using davitd5_unicl_lang_v1.yaml? Also, I was wondering why the BACKBONE_NAME is set to 'davit' instead of 'davitd5'?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant