sadkins65 commited on
Commit
7a68152
1 Parent(s): a8e06b9

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. config.json +3 -4
  2. model.safetensors +2 -2
  3. recipe.yaml +2 -2
config.json CHANGED
@@ -39,9 +39,8 @@
39
  }
40
  },
41
  "format": "int-quantized",
42
- "global_compression_ratio": 1.454005842904474,
43
  "ignore": [
44
- "model.layers.0.mlp.down_proj",
45
  "lm_head"
46
  ],
47
  "quant_method": "sparseml",
@@ -49,7 +48,7 @@
49
  },
50
  "sparsity_config": {
51
  "format": "dense",
52
- "global_sparsity": 7.713156824200198,
53
  "registry_requires_subclass": false,
54
  "sparsity_structure": "0:0"
55
  }
@@ -69,7 +68,7 @@
69
  "rope_scaling": null,
70
  "rope_theta": 10000.0,
71
  "tie_word_embeddings": false,
72
- "torch_dtype": "bfloat16",
73
  "transformers_version": "4.39.0",
74
  "use_cache": true,
75
  "vocab_size": 32000
 
39
  }
40
  },
41
  "format": "int-quantized",
42
+ "global_compression_ratio": 1.4583332397611055,
43
  "ignore": [
 
44
  "lm_head"
45
  ],
46
  "quant_method": "sparseml",
 
48
  },
49
  "sparsity_config": {
50
  "format": "dense",
51
+ "global_sparsity": 7.8259900429979625,
52
  "registry_requires_subclass": false,
53
  "sparsity_structure": "0:0"
54
  }
 
68
  "rope_scaling": null,
69
  "rope_theta": 10000.0,
70
  "tie_word_embeddings": false,
71
+ "torch_dtype": "float16",
72
  "transformers_version": "4.39.0",
73
  "use_cache": true,
74
  "vocab_size": 32000
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2904d1e37d759d31bbdec6c0b5c226d8ca1db1bc487dd69eb3f49c75ecdcb640
3
- size 1242839118
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e50b56978024a9aed7179f00aabd8689c07912534fa25058fd873c39f3e687bd
3
+ size 1231304860
recipe.yaml CHANGED
@@ -1,7 +1,7 @@
1
- quant_stage:
2
  quant_modifiers:
3
  vLLMQuantizationModifier:
4
- ignore: [lm_head, model.layers.0.mlp.down_proj]
5
  config_groups:
6
  group_0:
7
  weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor}
 
1
+ test_stage:
2
  quant_modifiers:
3
  vLLMQuantizationModifier:
4
+ ignore: [lm_head]
5
  config_groups:
6
  group_0:
7
  weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor}