best settings to load this model on oobabooga? in most cases i get errors
what are the best settings to load this model on oobabooga.
in most cases i get errors while loading or while doing inference
Model Loader: Transformers
Found modules on cpu/disk. Using Exllama backend requires all the modules to be on GPU.You can deactivate exllama backend by setting disable_exllama=True
in the quantization config object
Model Loader: ExLlamav2_HF
Load error
cannot open shared object file: No such file or directory
Model Loader: ExLlama_Hf
inference error
NotImplementedError: Cannot copy out of meta tensor; no data!
Model Loader: AutoGPT with disable exllama=true
Load error
[NotImplementedError: Cannot copy out of meta tensor; no data!](NotImplementedError: Cannot copy out of meta tensor; no data!)
What is your computer setup? GPUs VRAM and RAM?
It should be sufficient to load everything in a 24GB card. I mostly do ExLlamav2_HF for GPTQs as it is the fastest.
Would you also check integrity of model files downloaded ?
I have a nvidia h100 with 40gb of ram in a 32 core server. i don't think is a hardware issue