-
Notifications
You must be signed in to change notification settings - Fork 213
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error in Conv.pad after converting to half #905
Comments
I'm having a look. There are a couple of issue. Support for f16 is experimental, this is a good exercise. Congrats for venturing once again in an unbaked area of tract ;) |
Some fixes on the branch #907 . Also, you'll need to do a couple of fixes on your side. Gist of it: don't call the f16 translator /after/ optimize, just after pulsing or declutter, and cast some f16 outputs back to f32. Tell me how it goes.
|
Woops, I just found the option and wanted to try out.
Works now, thanks. But is approx 63 times slower than f32 on x86. |
Lol, good to know it's working now. No big surprise about the performance on intel, the only thing which has been optimised at this stage is matrix multiplication on armv8.2+ . |
@Rikorose might also be worth profiling it on your target, as the last time I tried to do a half-precision model (especially at something running on the order of 10ms), a major bottleneck was f32/f16 conversion (more-so than the “useful work”). |
I converted my model to half using:
Which results in the following error:
To reproduce, see: Rikorose/DeepFilterNet#211
The text was updated successfully, but these errors were encountered: