Weights are not present in the repo
#1
by
julien-c
HF staff
- opened
see title
We are aware of this issue. We currently can't distribute the quantized weights so we are working on a way to produce these weights automatically. Stay tuned! This is the highest priority for us.
exciting stuff!
Any news?
Targeting the next release. (We release roughly every 2 weeks)
Llama2 with pre-computed quantization encodings is released.
Now you can use AI Hub Llama2 model to export models to QNN context binary and verify with demo.py on-device.
Sorry for the late response. This page has all the information https://aihub.qualcomm.com/mobile/models/llama_v2_7b_chat_quantized?domain=Generative+AI