1024m commited on
Commit
54bbcd7
·
verified ·
1 Parent(s): 566a8fc

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -1,4 +1,4 @@
1
- import gradio as gr
2
  import torch
3
  import time
4
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
@@ -79,7 +79,7 @@ def generate_response(message, temperature, max_new_tokens, top_p, task):
79
  print(f"Prompt: {prompt}")
80
  start_time = time.time()
81
  inputs = tokenizer.encode(prompt, return_tensors="pt").to(model.device)
82
- outputs = model.generate(input_ids=inputs, do_sample=True, max_new_tokens=max_new_tokens, use_cache=True, temperature=temperature, min_p=top_p, pad_token_id=tokenizer.eos_token_id)
83
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
84
  processed_response = response.split("RESPONSE :")[-1].strip()
85
  end_time = time.time()
@@ -111,4 +111,4 @@ with gr.Blocks(theme='1024m/1024m-1') as demo:
111
  send_btn.click(fn=generate_response, inputs=[input_text, temperature, max_new_tokens, top_p, task_dropdown], outputs=output_text)
112
  clear_btn.click(fn=lambda: ("", ""), inputs=None, outputs=[input_text, output_text])
113
  if __name__ == "__main__":
114
- demo.queue().launch()"""
 
1
+ """import gradio as gr
2
  import torch
3
  import time
4
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
 
79
  print(f"Prompt: {prompt}")
80
  start_time = time.time()
81
  inputs = tokenizer.encode(prompt, return_tensors="pt").to(model.device)
82
+ outputs = model.generate(input_ids=inputs, max_new_tokens=max_new_tokens, use_cache=True, temperature=temperature, min_p=top_p, pad_token_id=tokenizer.eos_token_id)
83
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
84
  processed_response = response.split("RESPONSE :")[-1].strip()
85
  end_time = time.time()
 
111
  send_btn.click(fn=generate_response, inputs=[input_text, temperature, max_new_tokens, top_p, task_dropdown], outputs=output_text)
112
  clear_btn.click(fn=lambda: ("", ""), inputs=None, outputs=[input_text, output_text])
113
  if __name__ == "__main__":
114
+ demo.queue().launch()