rodrigomasini commited on
Commit
a110fec
1 Parent(s): 060489c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -16,7 +16,7 @@ local_folder = cachedir + "/model"
16
 
17
  quantized_model_dir = "FPHam/Jackson_The_Formalizer_V2_13b_GPTQ"
18
 
19
- #snapshot_download(repo_id=quantized_model_dir, local_dir=local_folder, local_dir_use_symlinks=False)
20
 
21
  model_basename = cachedir + "/model/Jackson2-4bit-128g-GPTQ"
22
 
@@ -24,7 +24,7 @@ use_strict = False
24
 
25
  use_triton = False
26
 
27
- tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir, use_fast=False)
28
 
29
  quantize_config = BaseQuantizeConfig(
30
  bits=4,
@@ -32,7 +32,7 @@ quantize_config = BaseQuantizeConfig(
32
  desc_act=False
33
  )
34
 
35
- model = AutoGPTQForCausalLM.from_quantized(quantized_model_dir,
36
  use_safetensors=True,
37
  strict=use_strict,
38
  model_basename=model_basename,
 
16
 
17
  quantized_model_dir = "FPHam/Jackson_The_Formalizer_V2_13b_GPTQ"
18
 
19
+ snapshot_download(repo_id=quantized_model_dir, local_dir=local_folder, local_dir_use_symlinks=True)
20
 
21
  model_basename = cachedir + "/model/Jackson2-4bit-128g-GPTQ"
22
 
 
24
 
25
  use_triton = False
26
 
27
+ tokenizer = AutoTokenizer.from_pretrained(local_foler, use_fast=False)
28
 
29
  quantize_config = BaseQuantizeConfig(
30
  bits=4,
 
32
  desc_act=False
33
  )
34
 
35
+ model = AutoGPTQForCausalLM.from_quantized(local_folder,
36
  use_safetensors=True,
37
  strict=use_strict,
38
  model_basename=model_basename,