RamAnanth1 commited on
Commit
d2758fe
·
1 Parent(s): 68ae08c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -4
app.py CHANGED
@@ -1,6 +1,6 @@
1
  import gradio as gr
2
  from instruct_pipeline import InstructionTextGenerationPipeline
3
- from transformers import AutoModelForCausalLM, AutoTokenizer
4
 
5
  import torch
6
 
@@ -12,10 +12,12 @@ theme = gr.themes.Monochrome(
12
  font=[gr.themes.GoogleFont("Open Sans"), "ui-sans-serif", "system-ui", "sans-serif"],
13
  )
14
 
15
- tokenizer = AutoTokenizer.from_pretrained("databricks/dolly-v2-12b", padding_side="left")
16
- model = AutoModelForCausalLM.from_pretrained("databricks/dolly-v2-12b", device_map="auto")
17
 
18
- generate_text = InstructionTextGenerationPipeline(model=model, tokenizer=tokenizer)
 
 
19
 
20
  def generate(instruction):
21
  response = generate_text(instruction)
 
1
  import gradio as gr
2
  from instruct_pipeline import InstructionTextGenerationPipeline
3
+ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
4
 
5
  import torch
6
 
 
12
  font=[gr.themes.GoogleFont("Open Sans"), "ui-sans-serif", "system-ui", "sans-serif"],
13
  )
14
 
15
+ #tokenizer = AutoTokenizer.from_pretrained("databricks/dolly-v2-12b", padding_side="left")
16
+ #model = AutoModelForCausalLM.from_pretrained("databricks/dolly-v2-12b", device_map="auto")
17
 
18
+ #generate_text = InstructionTextGenerationPipeline(model=model, tokenizer=tokenizer)
19
+
20
+ generate_text = pipeline(model="databricks/dolly-v2-12b", torch_dtype=torch.bfloat16, trust_remote_code=True, device_map="auto")
21
 
22
  def generate(instruction):
23
  response = generate_text(instruction)