MetaIX commited on
Commit
4d4ffc8
1 Parent(s): 11c8d3f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -5,7 +5,7 @@ GPT4-X-Alpasta-30b working with Oobabooga's Text Generation Webui and KoboldAI.
5
  <p><strong>What's included</strong></p>
6
 
7
  <P>GPTQ: 2 quantized versions. One quantized --true-sequential and act-order optimizations, and the other was quantized using --true-sequential --groupsize 128 optimizations (coming soon)</P>
8
- <P>GGML: 1 quantized version. One quantized using q4_1</P>
9
 
10
  <p><strong>GPU/GPTQ Usage</strong></p>
11
  <p>To use with your GPU using GPTQ pick one of the .safetensors along with all of the .jsons and .model files.</p>
@@ -31,10 +31,10 @@ GPT4-X-Alpasta-30b working with Oobabooga's Text Generation Webui and KoboldAI.
31
 
32
  <p><strong><font size="4">--true-sequential --groupsize 128</font></strong></p>
33
 
34
- <strong>Wikitext2</strong>: TBD
35
 
36
- <strong>Ptb-New</strong>: TBD
37
 
38
- <strong>C4-New</strong>: TBD
39
 
40
  <strong>Note</strong>: This version uses <i>--groupsize 128</i>, resulting in better evaluations. However, it consumes more VRAM.
 
5
  <p><strong>What's included</strong></p>
6
 
7
  <P>GPTQ: 2 quantized versions. One quantized --true-sequential and act-order optimizations, and the other was quantized using --true-sequential --groupsize 128 optimizations (coming soon)</P>
8
+ <P>GGML: 2 quantized versions. One quantized using q4_1, and the other was quantized using q5_0.</P>
9
 
10
  <p><strong>GPU/GPTQ Usage</strong></p>
11
  <p>To use with your GPU using GPTQ pick one of the .safetensors along with all of the .jsons and .model files.</p>
 
31
 
32
  <p><strong><font size="4">--true-sequential --groupsize 128</font></strong></p>
33
 
34
+ <strong>Wikitext2</strong>: 4.70257568359375
35
 
36
+ <strong>Ptb-New</strong>: 9.323467254638672
37
 
38
+ <strong>C4-New</strong>: 7.041860580444336
39
 
40
  <strong>Note</strong>: This version uses <i>--groupsize 128</i>, resulting in better evaluations. However, it consumes more VRAM.