Update README.md
Browse files
README.md
CHANGED
@@ -39,13 +39,14 @@ This model is for personal use, made for myself as experiment. I would like to m
|
|
39 |
|
40 |
Setting from Midnight-Rose should work in SillyTavern. This is almost same what I use for testing.
|
41 |
|
42 |
-
I use max_seq_len 8K with alpha_value 2
|
43 |
|
44 |
### Quants
|
45 |
|
46 |
- [3.92bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3.92bpw-EXL2) --> 40GB VRAM
|
47 |
- [3.5bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3.5bpw-EXL2)
|
48 |
- [3bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3bpw-EXL2)
|
|
|
49 |
- [measurements](https://huggingface.co/altomek/measurements/resolve/main/CodeRosa-AB1_measurement.json) --> ExLlamav2 measurments
|
50 |
|
51 |
### PS
|
|
|
39 |
|
40 |
Setting from Midnight-Rose should work in SillyTavern. This is almost same what I use for testing.
|
41 |
|
42 |
+
I use max_seq_len 8K with alpha_value 2.65.
|
43 |
|
44 |
### Quants
|
45 |
|
46 |
- [3.92bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3.92bpw-EXL2) --> 40GB VRAM
|
47 |
- [3.5bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3.5bpw-EXL2)
|
48 |
- [3bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3bpw-EXL2)
|
49 |
+
- [2.4bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-2.4bpw-EXL2) --> 24GB VRAM
|
50 |
- [measurements](https://huggingface.co/altomek/measurements/resolve/main/CodeRosa-AB1_measurement.json) --> ExLlamav2 measurments
|
51 |
|
52 |
### PS
|