You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Right now the model used is a fp32 model provided by Nvidia. This is the only trained model i can find for detecting faces right now other than MTCNN which is actually three models and harder to plug in the pipeline. In any case, any trained models available would requires a clusterf**k of conversions.
Around the same time I created this app, Nvidia seemed to be thinking the same thing, so they produced this tutorial which is fantastic, but they don't provide a trained model... which means I must train it myself.
The dataset required to train is over 1TB -- more than is available on my GPU box, meaning I would have to use my NAS and nfs or something for the actual storage. All this is doable, but it means tying up my GPUs for what would likely be days. According to Nvidia it takes 8 hours on a DGX-1, but two 1080s is proably quite a bit slower.
If anybody is willing to follow that tutorial, pay for the cloud time, and send me the trained model it would be greatly appreciated (and you'd get your name in the credits). Otherwise performance enhancements will have to wait until my GPU box has a week free, which might be a while since I'm using nvidia-docker on it all the time.
Alternatively, if somebody knows how to quantize the model I have, that works too. Most of the tools I have found expect an .onnx, but it looks like the code might be able to be modified to use a different parser. Something worth exploring.
The text was updated successfully, but these errors were encountered:
So it looks like DS 5.0 has a bundled model that will be perfect. At the moment the pre-release files needed are under a restrictive license, but I've been told the final will be a permissive license.
So. In the Dockerfile, I did a "FROM" Nvidia's latest DeepStream for the latest social distiancing release. An int8 model is now used for the social distancing support. I have no clue if this violates any license, but if you have an issue with it, Nvidia, lmk and I will delete the image on Docker hub. People will just have to build the Dockerfile manually.
Right now the model used is a fp32 model provided by Nvidia. This is the only trained model i can find for detecting faces right now other than MTCNN which is actually three models and harder to plug in the pipeline. In any case, any trained models available would requires a clusterf**k of conversions.
Around the same time I created this app, Nvidia seemed to be thinking the same thing, so they produced this tutorial which is fantastic, but they don't provide a trained model... which means I must train it myself.
The dataset required to train is over 1TB -- more than is available on my GPU box, meaning I would have to use my NAS and nfs or something for the actual storage. All this is doable, but it means tying up my GPUs for what would likely be days. According to Nvidia it takes 8 hours on a DGX-1, but two 1080s is proably quite a bit slower.
If anybody is willing to follow that tutorial, pay for the cloud time, and send me the trained model it would be greatly appreciated (and you'd get your name in the credits). Otherwise performance enhancements will have to wait until my GPU box has a week free, which might be a while since I'm using nvidia-docker on it all the time.
Alternatively, if somebody knows how to quantize the model I have, that works too. Most of the tools I have found expect an .onnx, but it looks like the code might be able to be modified to use a different parser. Something worth exploring.
The text was updated successfully, but these errors were encountered: