.bin .pt and .safetensors #895
Unanswered
arnoldtheaardvark
asked this question in
Q&A
Replies: 1 comment 4 replies
-
|
Based on the error it sounds like you are trying to load a full fp16 model with the 4bit command line arguments, which won't work since the files have to be pre-quantized before they can be used in 4bit mode. The main difference between safetensors and the .bin/.pt formats is that safetensors can't execute code so they are safer to distribute. There can also be some loading speed benefits but I don't know if this project takes advantage of those yet. I believe .bin and .pt are both pytorch checkpoints, just with different extensions. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
Very new to this so apologies for my lack of understanding.
When I try to use the built in models such as facebook-opt6.7 or others from hugging face which are .bin files.
I get en error saying "could not find the quantized model in .pt or .safetensors format"
Could someone please explain either what I need to do to get .bin files working or point me in the direction to learn about the difference please? I have searched a lot and I cant quite get the answer.
I imagine they are different formats.
So can .bin models be used in this tool?
Or is it possible to convert a .bin int a .pt?
Thanks
Beta Was this translation helpful? Give feedback.
All reactions