iproskurina commited on
Commit
1cda51a
1 Parent(s): 697e76c

AutoGPTQ model for mistralai/Mistral-7B-v0.3: 8bits, gr128, desc_act=False

Browse files
Files changed (3) hide show
  1. config.json +13 -0
  2. model.safetensors +2 -2
  3. quantize_config.json +4 -2
config.json CHANGED
@@ -16,6 +16,19 @@
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  "rms_norm_eps": 1e-05,
20
  "rope_theta": 1000000.0,
21
  "sliding_window": null,
 
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
19
+ "quantization_config": {
20
+ "bits": 8,
21
+ "checkpoint_format": "gptq",
22
+ "damp_percent": 0.01,
23
+ "desc_act": false,
24
+ "group_size": 128,
25
+ "model_file_base_name": "gptq_model-8bit-128g",
26
+ "model_name_or_path": "Mistral-7B-v0.3-gptq-4bit",
27
+ "quant_method": "gptq",
28
+ "static_groups": false,
29
+ "sym": true,
30
+ "true_sequential": true
31
+ },
32
  "rms_norm_eps": 1e-05,
33
  "rope_theta": 1000000.0,
34
  "sliding_window": null,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e31960a7b0b7febb8cf860e7f64423438e0bad97d04fffe0a6e44494cd09ffbf
3
- size 7688169544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5309e066356b08e9352f168f77d7cdcb4a7fae624f3af2d2f562085ad3135f4
3
+ size 7685392808
quantize_config.json CHANGED
@@ -6,6 +6,8 @@
6
  "static_groups": false,
7
  "sym": true,
8
  "true_sequential": true,
9
- "model_name_or_path": null,
10
- "model_file_base_name": null
 
 
11
  }
 
6
  "static_groups": false,
7
  "sym": true,
8
  "true_sequential": true,
9
+ "model_name_or_path": "Mistral-7B-v0.3-gptq-4bit",
10
+ "model_file_base_name": "gptq_model-8bit-128g",
11
+ "quant_method": "gptq",
12
+ "checkpoint_format": "gptq"
13
  }