apepkuss79
commited on
Commit
•
03fe989
1
Parent(s):
663c3f3
Update README.md
Browse files
README.md
CHANGED
@@ -43,7 +43,7 @@ license: apache-2.0
|
|
43 |
|
44 |
- Reverse prompt: `<|im_end|>`
|
45 |
|
46 |
-
- Context size: `
|
47 |
|
48 |
- Run as LlamaEdge service
|
49 |
|
@@ -52,7 +52,7 @@ license: apache-2.0
|
|
52 |
llama-api-server.wasm \
|
53 |
--prompt-template chatml \
|
54 |
--reverse-prompt "<|im_end|>" \
|
55 |
-
--ctx-size
|
56 |
--model-name Yi-Coder-1.5B-Chat
|
57 |
```
|
58 |
|
@@ -63,7 +63,7 @@ license: apache-2.0
|
|
63 |
llama-chat.wasm \
|
64 |
--prompt-template chatml \
|
65 |
--reverse-prompt "<|im_end|>" \
|
66 |
-
--ctx-size
|
67 |
```
|
68 |
|
69 |
## Quantized GGUF Models
|
@@ -84,4 +84,4 @@ license: apache-2.0
|
|
84 |
| [Yi-Coder-1.5B-Chat-Q8_0.gguf](https://huggingface.co/second-state/Yi-Coder-1.5B-Chat-GGUF/blob/main/Yi-Coder-1.5B-Chat-Q8_0.gguf) | Q8_0 | 8 | 1.57 GB| very large, extremely low quality loss - not recommended |
|
85 |
| [Yi-Coder-1.5B-Chat-f16.gguf](https://huggingface.co/second-state/Yi-Coder-1.5B-Chat-GGUF/blob/main/Yi-Coder-1.5B-Chat-f16.gguf) | f16 | 16 | 2.95 GB| |
|
86 |
|
87 |
-
*Quantized with llama.cpp
|
|
|
43 |
|
44 |
- Reverse prompt: `<|im_end|>`
|
45 |
|
46 |
+
- Context size: `128000`
|
47 |
|
48 |
- Run as LlamaEdge service
|
49 |
|
|
|
52 |
llama-api-server.wasm \
|
53 |
--prompt-template chatml \
|
54 |
--reverse-prompt "<|im_end|>" \
|
55 |
+
--ctx-size 128000 \
|
56 |
--model-name Yi-Coder-1.5B-Chat
|
57 |
```
|
58 |
|
|
|
63 |
llama-chat.wasm \
|
64 |
--prompt-template chatml \
|
65 |
--reverse-prompt "<|im_end|>" \
|
66 |
+
--ctx-size 128000
|
67 |
```
|
68 |
|
69 |
## Quantized GGUF Models
|
|
|
84 |
| [Yi-Coder-1.5B-Chat-Q8_0.gguf](https://huggingface.co/second-state/Yi-Coder-1.5B-Chat-GGUF/blob/main/Yi-Coder-1.5B-Chat-Q8_0.gguf) | Q8_0 | 8 | 1.57 GB| very large, extremely low quality loss - not recommended |
|
85 |
| [Yi-Coder-1.5B-Chat-f16.gguf](https://huggingface.co/second-state/Yi-Coder-1.5B-Chat-GGUF/blob/main/Yi-Coder-1.5B-Chat-f16.gguf) | f16 | 16 | 2.95 GB| |
|
86 |
|
87 |
+
*Quantized with llama.cpp b3664*
|