daniellnichols
commited on
Commit
•
088e534
1
Parent(s):
963b22b
Update README.md
Browse files
README.md
CHANGED
@@ -45,6 +45,6 @@ Below is an instruction that describes a task. Write a response that appropriate
|
|
45 |
## Quantized Models
|
46 |
|
47 |
4 and 8 bit quantized weights are available in the GGUF format for use with [llama.cpp](https://github.com/ggerganov/llama.cpp).
|
48 |
-
The 4 bit model requires ~
|
49 |
-
The 8 bit model requires ~
|
50 |
Further information on how to use them with llama.cpp can be found in [its documentation](https://github.com/ggerganov/llama.cpp).
|
|
|
45 |
## Quantized Models
|
46 |
|
47 |
4 and 8 bit quantized weights are available in the GGUF format for use with [llama.cpp](https://github.com/ggerganov/llama.cpp).
|
48 |
+
The 4 bit model requires ~0.8 GB memory and can be found [here](https://huggingface.co/hpcgroup/hpc-coder-v2-1.3b-Q4_K_S-GGUF).
|
49 |
+
The 8 bit model requires ~1.4 GB memory and can be found [here](https://huggingface.co/hpcgroup/hpc-coder-v2-1.3b-Q8_0-GGUF).
|
50 |
Further information on how to use them with llama.cpp can be found in [its documentation](https://github.com/ggerganov/llama.cpp).
|