captainkyd commited on
Commit
30476bb
β€’
1 Parent(s): 555e2ea

Update app.py

Browse files

added 4bit double quant

Files changed (1) hide show
  1. app.py +10 -4
app.py CHANGED
@@ -24,6 +24,13 @@ Answer the Question by exploring multiple reasoning paths as follows:
24
  In summary, leverage a Tree of Thoughts approach to actively explore multiple reasoning paths, evaluate thoughts heuristically, and explain the process - with the goal of producing insightful answers.
25
  """
26
 
 
 
 
 
 
 
 
27
  model_path = "WhiteRabbitNeo/Trinity-13B"
28
 
29
  hf_token = os.getenv("HF_TOKEN")
@@ -32,10 +39,9 @@ if not hf_token:
32
 
33
  model = AutoModelForCausalLM.from_pretrained(
34
  model_path,
35
- torch_dtype=torch.float16,
36
- device_map="auto",
37
- load_in_8bit=True,
38
- trust_remote_code=True,
39
  )
40
 
41
  tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
 
24
  In summary, leverage a Tree of Thoughts approach to actively explore multiple reasoning paths, evaluate thoughts heuristically, and explain the process - with the goal of producing insightful answers.
25
  """
26
 
27
+ model = AutoModelForCausalLM.from_pretrained(
28
+ model_path,
29
+ device_map="auto",
30
+ trust_remote_code=True,
31
+ quantization_config=quantization_config,
32
+ )
33
+
34
  model_path = "WhiteRabbitNeo/Trinity-13B"
35
 
36
  hf_token = os.getenv("HF_TOKEN")
 
39
 
40
  model = AutoModelForCausalLM.from_pretrained(
41
  model_path,
42
+ device_map="auto",
43
+ trust_remote_code=True,
44
+ quantization_config=quantization_config
 
45
  )
46
 
47
  tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)