Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -172,12 +172,12 @@ Invoke the llama.cpp server or the CLI.
|
|
172 |
|
173 |
### CLI:
|
174 |
```bash
|
175 |
-
llama-cli --hf-repo matrixportal/Einstein-v7-Qwen2-7B-GGUF --hf-file einstein-v7-qwen2-7b-
|
176 |
```
|
177 |
|
178 |
### Server:
|
179 |
```bash
|
180 |
-
llama-server --hf-repo matrixportal/Einstein-v7-Qwen2-7B-GGUF --hf-file einstein-v7-qwen2-7b-
|
181 |
```
|
182 |
|
183 |
Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
|
@@ -194,9 +194,9 @@ cd llama.cpp && LLAMA_CURL=1 make
|
|
194 |
|
195 |
Step 3: Run inference through the main binary.
|
196 |
```
|
197 |
-
./llama-cli --hf-repo matrixportal/Einstein-v7-Qwen2-7B-GGUF --hf-file einstein-v7-qwen2-7b-
|
198 |
```
|
199 |
or
|
200 |
```
|
201 |
-
./llama-server --hf-repo matrixportal/Einstein-v7-Qwen2-7B-GGUF --hf-file einstein-v7-qwen2-7b-
|
202 |
```
|
|
|
172 |
|
173 |
### CLI:
|
174 |
```bash
|
175 |
+
llama-cli --hf-repo matrixportal/Einstein-v7-Qwen2-7B-GGUF --hf-file einstein-v7-qwen2-7b-q4_k_s.gguf -p "The meaning to life and the universe is"
|
176 |
```
|
177 |
|
178 |
### Server:
|
179 |
```bash
|
180 |
+
llama-server --hf-repo matrixportal/Einstein-v7-Qwen2-7B-GGUF --hf-file einstein-v7-qwen2-7b-q4_k_s.gguf -c 2048
|
181 |
```
|
182 |
|
183 |
Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
|
|
|
194 |
|
195 |
Step 3: Run inference through the main binary.
|
196 |
```
|
197 |
+
./llama-cli --hf-repo matrixportal/Einstein-v7-Qwen2-7B-GGUF --hf-file einstein-v7-qwen2-7b-q4_k_s.gguf -p "The meaning to life and the universe is"
|
198 |
```
|
199 |
or
|
200 |
```
|
201 |
+
./llama-server --hf-repo matrixportal/Einstein-v7-Qwen2-7B-GGUF --hf-file einstein-v7-qwen2-7b-q4_k_s.gguf -c 2048
|
202 |
```
|