Replies: 2 comments
-
I don't think this is feasible, because nnUNet is designed to predict using an ensemble of models trained on different folds. As a consequence, each model is loaded (sequentially) to do inference on a single input file. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Okay, thank you very much for your reply. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
If nnUNet can be designed to run as a server, keep Pytorch waiting after loading the model, stay in memory, and use a queue to process the received segmentation tasks, this will greatly improve performance and avoid wasting time caused by repeated model loading.
Beta Was this translation helpful? Give feedback.
All reactions