GPU VRAM requirements
#1
by
bbooth
- opened
Has anyone gotten this to run on an NVidia GPU?
I keep running out of memory on a 3090 24Gig.
I have not tried the 3090 jet, but it works fine on the H100 and A100 vhips. Maybe you can try quantization if the model doesn't fit.
Thanks for the quick reply! I thought the 7B-hf model should fit in the 3090-24GB using the supplied code, but will try 8-bit quantization.
LukasHug
changed discussion status to
closed