kajdun commited on
Commit
231340f
·
1 Parent(s): 8e9b3d6

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "kajdun/iubaris-13b-v3_GPTQ",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -12,7 +12,6 @@
12
  "max_length": 4096,
13
  "max_position_embeddings": 4096,
14
  "model_type": "llama",
15
- "model_id": "iubaris-13b-v3_GPTQ",
16
  "num_attention_heads": 40,
17
  "num_hidden_layers": 40,
18
  "num_key_value_heads": 40,
 
1
  {
2
+ "_name_or_path": "kajdun/iubaris-13b-v3-GPTQ",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
12
  "max_length": 4096,
13
  "max_position_embeddings": 4096,
14
  "model_type": "llama",
 
15
  "num_attention_heads": 40,
16
  "num_hidden_layers": 40,
17
  "num_key_value_heads": 40,
gptq_model-4bit-128g.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22fc368b0f95f28d645019d5ac1273aa5d0b0ddca032fb36c1afac5d05015b93
3
+ size 7259449576
handler.py CHANGED
@@ -14,7 +14,7 @@ model_directory = "/repository/"
14
 
15
  tokenizer_path = f"{model_directory}tokenizer.model"
16
  model_config_path = f"{model_directory}config.json"
17
- model_path = f"{model_directory}model.safetensors"
18
 
19
  class EndpointHandler():
20
  def __init__(self, path=""):
 
14
 
15
  tokenizer_path = f"{model_directory}tokenizer.model"
16
  model_config_path = f"{model_directory}config.json"
17
+ model_path = f"{model_directory}gptq_model-4bit-128g.safetensors"
18
 
19
  class EndpointHandler():
20
  def __init__(self, path=""):
quantize_config.json CHANGED
@@ -7,5 +7,5 @@
7
  "sym": true,
8
  "true_sequential": true,
9
  "model_name_or_path": null,
10
- "model_file_base_name": "model"
11
  }
 
7
  "sym": true,
8
  "true_sequential": true,
9
  "model_name_or_path": null,
10
+ "model_file_base_name": null
11
  }