Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Introduce FP16 quantization #437

Open
LHT129 opened this issue Feb 24, 2025 · 0 comments
Open

Introduce FP16 quantization #437

LHT129 opened this issue Feb 24, 2025 · 0 comments
Assignees
Labels
kind/feature New feature or request version/0.14

Comments

@LHT129
Copy link
Collaborator

LHT129 commented Feb 24, 2025

FP16 is 16bit quantization shorter than float, but have high precise on ANN calculation, FP16 can be calculated with avx512 too.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/feature New feature or request version/0.14
Projects
None yet
Development

No branches or pull requests

2 participants