FlawedLLM commited on
Commit
731e7a8
1 Parent(s): d2a5fcd

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -2
app.py CHANGED
@@ -60,12 +60,11 @@ from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
60
  # # torch_dtype =torch.float16
61
  # model = AutoModelForCausalLM.from_pretrained("FlawedLLM/Bhashini_9",config=config, ignore_mismatched_sizes=True).to('cuda')
62
  # Load model directly
63
- @spaces.GPU(duration=300)
64
 
65
  tokenizer = AutoTokenizer.from_pretrained("FlawedLLM/Bhashini_00")
66
  model = AutoModelForCausalLM.from_pretrained("FlawedLLM/Bhashini_00", load_in_4bit=True).to('cuda')
67
 
68
-
69
  def chunk_it(input_command):
70
  alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
71
 
 
60
  # # torch_dtype =torch.float16
61
  # model = AutoModelForCausalLM.from_pretrained("FlawedLLM/Bhashini_9",config=config, ignore_mismatched_sizes=True).to('cuda')
62
  # Load model directly
 
63
 
64
  tokenizer = AutoTokenizer.from_pretrained("FlawedLLM/Bhashini_00")
65
  model = AutoModelForCausalLM.from_pretrained("FlawedLLM/Bhashini_00", load_in_4bit=True).to('cuda')
66
 
67
+ @spaces.GPU(duration=300)
68
  def chunk_it(input_command):
69
  alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
70