GGUF format?
#2
by
toranb
- opened
Anyone else interested in a f16 GGUF of this model?
With the latest llama cpp I was able to generate a GGUF quick with this command after downloading the files
git clone --depth=1 https://github.com/ggerganov/llama.cpp.git cpp
cd cpp
make clean && LLAMA_CUBLAS=1 make -j
python3 -m venv env
source env/bin/activate
pip install -r requirements/requirements-convert.txt
python3 convert.py Mistral_Pro_8B_v0.1 --outfile mistralpro.gguf --outtype f16
Here is a 3min video of the process end to end