kenchan0226 commited on
Commit
626df5d
1 Parent(s): db3ef81

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -1
app.py CHANGED
@@ -7,6 +7,8 @@ import spaces
7
  import torch
8
  from transformers import AutoModelForCausalLM, Qwen2TokenizerFast, TextIteratorStreamer
9
 
 
 
10
  DESCRIPTION = """\
11
  # SeaLLMs Med
12
 
@@ -20,7 +22,7 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
20
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
21
 
22
  #model_id = "SeaLLMs/SeaLLMs-v3-7B-Chat"
23
- model_id = "SeaLLMs/debug_v2.1"
24
  tokenizer = Qwen2TokenizerFast.from_pretrained(model_id)
25
  model = AutoModelForCausalLM.from_pretrained(
26
  model_id,
 
7
  import torch
8
  from transformers import AutoModelForCausalLM, Qwen2TokenizerFast, TextIteratorStreamer
9
 
10
+ # based on gemma demo code
11
+
12
  DESCRIPTION = """\
13
  # SeaLLMs Med
14
 
 
22
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
23
 
24
  #model_id = "SeaLLMs/SeaLLMs-v3-7B-Chat"
25
+ model_id = "SeaLLMs/debug_4"
26
  tokenizer = Qwen2TokenizerFast.from_pretrained(model_id)
27
  model = AutoModelForCausalLM.from_pretrained(
28
  model_id,