altomek commited on
Commit
ec5158b
1 Parent(s): 2ca2901

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -39,13 +39,14 @@ This model is for personal use, made for myself as experiment. I would like to m
39
 
40
  Setting from Midnight-Rose should work in SillyTavern. This is almost same what I use for testing.
41
 
42
- I use max_seq_len 8K with alpha_value 2,65.
43
 
44
  ### Quants
45
 
46
  - [3.92bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3.92bpw-EXL2) --> 40GB VRAM
47
  - [3.5bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3.5bpw-EXL2)
48
  - [3bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3bpw-EXL2)
 
49
  - [measurements](https://huggingface.co/altomek/measurements/resolve/main/CodeRosa-AB1_measurement.json) --> ExLlamav2 measurments
50
 
51
  ### PS
 
39
 
40
  Setting from Midnight-Rose should work in SillyTavern. This is almost same what I use for testing.
41
 
42
+ I use max_seq_len 8K with alpha_value 2.65.
43
 
44
  ### Quants
45
 
46
  - [3.92bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3.92bpw-EXL2) --> 40GB VRAM
47
  - [3.5bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3.5bpw-EXL2)
48
  - [3bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-3bpw-EXL2)
49
+ - [2.4bpw](https://huggingface.co/altomek/CodeRosa-70B-AB1-2.4bpw-EXL2) --> 24GB VRAM
50
  - [measurements](https://huggingface.co/altomek/measurements/resolve/main/CodeRosa-AB1_measurement.json) --> ExLlamav2 measurments
51
 
52
  ### PS