.bin .pt and .safetensors #895
Unanswered
arnoldtheaardvark
asked this question in
Q&A
Replies: 1 comment 4 replies
-
Based on the error it sounds like you are trying to load a full fp16 model with the 4bit command line arguments, which won't work since the files have to be pre-quantized before they can be used in 4bit mode. The main difference between safetensors and the .bin/.pt formats is that safetensors can't execute code so they are safer to distribute. There can also be some loading speed benefits but I don't know if this project takes advantage of those yet. I believe .bin and .pt are both pytorch checkpoints, just with different extensions. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
Very new to this so apologies for my lack of understanding.
When I try to use the built in models such as facebook-opt6.7 or others from hugging face which are .bin files.
I get en error saying "could not find the quantized model in .pt or .safetensors format"
Could someone please explain either what I need to do to get .bin files working or point me in the direction to learn about the difference please? I have searched a lot and I cant quite get the answer.
I imagine they are different formats.
So can .bin models be used in this tool?
Or is it possible to convert a .bin int a .pt?
Thanks
Beta Was this translation helpful? Give feedback.
All reactions