ValueError: not enough values to unpack (expected 3, got 2)
#2
by
rhamnett
- opened
Hello getting the following error when generating a response using the example in the Model card.
Fixed with inject_fused_attention=False ππ
it goes in the .from_quantized()
call
ctransformers? ctransformers is for GGML/GGUF models, not GPTQ models. You want normal Huggingface Transformers to load GPTQ models
but ctransformers also supports GPTQ via exallama