ValueError: Block pattern could not be match. Pass `block_name_to_quantize` argument in `quantize_model`
1
#36 opened 5 months ago
by
Gonzalomoreno01
Some weights of the model checkpoint at Llama-2-7B-Chat-GPTQ were not used when initializing LlamaForCausalLM
#35 opened 6 months ago
by
thlw
[AUTOMATED] Model Memory Requirements
#34 opened 8 months ago
by
model-sizer-bot
[AUTOMATED] Model Memory Requirements
#33 opened 8 months ago
by
model-sizer-bot
[AUTOMATED] Model Memory Requirements
#32 opened 8 months ago
by
model-sizer-bot
Fails with transformers==4.38.1
#30 opened 9 months ago
by
rohithkrn
Index out of range error: QAchain for pdf chatbot
#29 opened 10 months ago
by
Artemis3196
How to overcoming bad output for better results?
#28 opened 12 months ago
by
notmax123
Functional example of finetuning of Llama-2-7b-Chat-GPTQ
#26 opened about 1 year ago
by
echogit
AUTOGPTQ Error in Google Colab
6
#25 opened about 1 year ago
by
echogit
Does the model response correctly
1
#24 opened about 1 year ago
by
mnwato
Can't Load Model in Kubernetes but can in Docker
#23 opened about 1 year ago
by
jrsperry
TheBloke/Llama-2-7b-(Chat-)GPTQ repeats request
#22 opened about 1 year ago
by
hyzhak
Cannot run batch on transformer
3
#20 opened about 1 year ago
by
DatenlaborBerlin
The response is not formatted
#18 opened about 1 year ago
by
Octavian81
how to load the GPTQ model using any pipeline method
2
#17 opened about 1 year ago
by
harithushan
Error trying to run on a revision, tensors not conforming?
#16 opened about 1 year ago
by
JohnSnyderTC
for faster GPU inference
2
#15 opened about 1 year ago
by
harithushan
How to set it up in a way that it just returns output without the system message or query, basically the information after [/INST].
1
#14 opened about 1 year ago
by
Pavan-growexx
Update for Transformers GPTQ support
#13 opened about 1 year ago
by
TheBloke
LORA fine tuning error
#12 opened about 1 year ago
by
tongwuhugging
GPTQ bugging: Wondering if I'm loading the model correctly
4
#9 opened over 1 year ago
by
quantuan125
Please make this model quantised GPTQ
2
#7 opened over 1 year ago
by
AiModelsMarket
TGI error
#5 opened over 1 year ago
by
aiamateur101
Cannot use anything but what's in the main branch
14
#3 opened over 1 year ago
by
HAvietisov
How to use this GPTQ Model from Python code for continue conversation?
1
#2 opened over 1 year ago
by
shifa
"max_length": 4096, "max_position_embeddings": 4096,
1
#1 opened over 1 year ago
by
pseudotensor