-
Notifications
You must be signed in to change notification settings - Fork 64
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix --output option to dump the model in a specific directory #124
Comments
My suggestion would the following:
With this approach the wrapper will just need to know the tmp directory and if something fails delft will not pollute the What do you think? |
I think to clarify, there is no problem currently with The Using the tmp folder will modify the saving mechanism to address #126, but the |
Maybe the current version is good now. When I opened this issue it was the version 0.2.6 I think and at that time it was not working only in the case of scibert + 10fold or training (I can't remember exactly) because only config and preprocessor were saved correctly in the output directory, while the rest was saved in the default directory Since there is a new version we can close it from my side 🎉 |
I've implemented this some time ago and it worked relatively well for normal models, however for transformers we realised it was not taken in account. With the new update to TF2 it seems to work fine for sequenceLabelling, except for the n-fold cross-validation.
In particular the problem seems to be related to the fact that the store of the model is hidden within the cross-validation. In my opinion we should call the model.save() after the n-fold cross-validation which will save either just one model (the best) or all of them (e.g. in case of ensemble).
My proposal is to give the wrapper a working/temporary directory and then explicitly save the model using model.save() and passing either the
--output
path or the default path within thedata/models
.The text was updated successfully, but these errors were encountered: