kaiokendev
commited on
Commit
•
f19f47d
1
Parent(s):
cbf4d7a
Update README.md
Browse files
README.md
CHANGED
@@ -26,6 +26,8 @@ It uses a mixture of the following datasets:
|
|
26 |
#### 13B
|
27 |
- 13B (unquantized): [https://huggingface.co/ausboss/llama-13b-supercot](https://huggingface.co/ausboss/llama-13b-supercot)
|
28 |
- 13B 4-bit 128g CUDA: [https://huggingface.co/ausboss/llama-13b-supercot-4bit-128g](https://huggingface.co/ausboss/llama-13b-supercot-4bit-128g)
|
|
|
|
|
29 |
|
30 |
(Thanks to all the awesome anons with supercomputers)
|
31 |
|
@@ -59,7 +61,7 @@ Remember that with lower parameter sizes, the structure of the prompt becomes mo
|
|
59 |
### Coming Soon
|
60 |
- 2048 7B version
|
61 |
- 512 variants of 13B and 7B
|
62 |
-
- merged ggml
|
63 |
- Tweet fix for 13B and 7B - lower model sizes seem to be extremely sensitive to hashtags at the end of training data responses, especially at longer cutoffs
|
64 |
|
65 |
### Citations
|
|
|
26 |
#### 13B
|
27 |
- 13B (unquantized): [https://huggingface.co/ausboss/llama-13b-supercot](https://huggingface.co/ausboss/llama-13b-supercot)
|
28 |
- 13B 4-bit 128g CUDA: [https://huggingface.co/ausboss/llama-13b-supercot-4bit-128g](https://huggingface.co/ausboss/llama-13b-supercot-4bit-128g)
|
29 |
+
- 13B 4-bit 128g TRITON: [TheYuriLover/llama-13b-SuperCOT-4bit-TRITON](TheYuriLover/llama-13b-SuperCOT-4bit-TRITON)
|
30 |
+
- GGML 13B 4-bit: [gozfarb/llama-13b-supercot-ggml](gozfarb/llama-13b-supercot-ggml)
|
31 |
|
32 |
(Thanks to all the awesome anons with supercomputers)
|
33 |
|
|
|
61 |
### Coming Soon
|
62 |
- 2048 7B version
|
63 |
- 512 variants of 13B and 7B
|
64 |
+
- merged ggml 7B?
|
65 |
- Tweet fix for 13B and 7B - lower model sizes seem to be extremely sensitive to hashtags at the end of training data responses, especially at longer cutoffs
|
66 |
|
67 |
### Citations
|