shimmyshimmer
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -12,10 +12,10 @@ tags:
|
|
12 |
- meta
|
13 |
---
|
14 |
|
15 |
-
## ***See [our collection](https://huggingface.co/collections/unsloth/deepseek-r1-all-versions-678e1c48f5d2fce87892ace5) for versions of Deepseek-R1 including GGUF and original formats.***
|
16 |
|
17 |
### Instructions to run this model in llama.cpp:
|
18 |
-
|
19 |
1. Do not forget about `<|User|>` and `<|Assistant|>` tokens! - Or use a chat template formatter
|
20 |
2. Obtain the latest `llama.cpp` at https://github.com/ggerganov/llama.cpp
|
21 |
3. Example with Q8_0 K quantized cache **Notice -no-cnv disables auto conversation mode**
|
|
|
12 |
- meta
|
13 |
---
|
14 |
|
15 |
+
## ***See [our collection](https://huggingface.co/collections/unsloth/deepseek-r1-all-versions-678e1c48f5d2fce87892ace5) for versions of Deepseek-R1 including GGUF, 4-bit and original formats.***
|
16 |
|
17 |
### Instructions to run this model in llama.cpp:
|
18 |
+
You can view more detailed instructions in our blog: [unsloth.ai/blog/deepseek-r1](https://unsloth.ai/blog/deepseek-r1)
|
19 |
1. Do not forget about `<|User|>` and `<|Assistant|>` tokens! - Or use a chat template formatter
|
20 |
2. Obtain the latest `llama.cpp` at https://github.com/ggerganov/llama.cpp
|
21 |
3. Example with Q8_0 K quantized cache **Notice -no-cnv disables auto conversation mode**
|