Encountering Unknown quantization type, got fp8 - supported types are: XXXXX

#1
by ivanmanu - opened

Dear developers and community users,

I tried to load this unsloth/DeepSeek-V3-bf16 via AutoModelForCausalLM.from_pretrained(" unsloth/DeepSeek-V3-bf16 ", trust_remote_code=True)

But encountered the below error:

Unknown quantization type, got fp8 - supported types are: ['awq', 'bitsandbytes_4bit', 'bitsandbytes_8bit', 'gptq', 'aqlm', 'quanto', 'eetq', 'hqq', 'compressed-tensors', 'fbgemm_fp8', 'torchao', 'bitnet']

I tried with different transformers versions from 4.33.1, to 4.55, but none worked.

The config file of your model showed

https://huggingface.co/unsloth/DeepSeek-V3-bf16/blob/main/config.json

"torch_dtype": "bfloat16",
"transformers_version": "4.33.1",

Please advise of any fixes? Thanks

Dear developers and community users,

I tried to load this unsloth/DeepSeek-V3-bf16 via AutoModelForCausalLM.from_pretrained(" unsloth/DeepSeek-V3-bf16 ", trust_remote_code=True)

But encountered the below error:

Unknown quantization type, got fp8 - supported types are: ['awq', 'bitsandbytes_4bit', 'bitsandbytes_8bit', 'gptq', 'aqlm', 'quanto', 'eetq', 'hqq', 'compressed-tensors', 'fbgemm_fp8', 'torchao', 'bitnet']

I tried with different transformers versions from 4.33.1, to 4.55, but none worked.

The config file of your model showed

https://huggingface.co/unsloth/DeepSeek-V3-bf16/blob/main/config.json

"torch_dtype": "bfloat16",
"transformers_version": "4.33.1",

Please advise of any fixes? Thanks

ohhh im not sure if hugging face implemented the support for it. :(

Dear developers and community users,

I tried to load this unsloth/DeepSeek-V3-bf16 via AutoModelForCausalLM.from_pretrained(" unsloth/DeepSeek-V3-bf16 ", trust_remote_code=True)

But encountered the below error:

Unknown quantization type, got fp8 - supported types are: ['awq', 'bitsandbytes_4bit', 'bitsandbytes_8bit', 'gptq', 'aqlm', 'quanto', 'eetq', 'hqq', 'compressed-tensors', 'fbgemm_fp8', 'torchao', 'bitnet']

I tried with different transformers versions from 4.33.1, to 4.55, but none worked.

The config file of your model showed

https://huggingface.co/unsloth/DeepSeek-V3-bf16/blob/main/config.json

"torch_dtype": "bfloat16",
"transformers_version": "4.33.1",

Please advise of any fixes? Thanks

ohhh im not sure if hugging face implemented the support for it. :(

Could you please advise how I can load the model and run it? If HF's AutoModelForCausalLM.from_pretrained() doesn't support it?
thanks.

Sign up or log in to comment