Spaces:
Sleeping
Sleeping
kenchan0226
commited on
Commit
•
626df5d
1
Parent(s):
db3ef81
Update app.py
Browse files
app.py
CHANGED
@@ -7,6 +7,8 @@ import spaces
|
|
7 |
import torch
|
8 |
from transformers import AutoModelForCausalLM, Qwen2TokenizerFast, TextIteratorStreamer
|
9 |
|
|
|
|
|
10 |
DESCRIPTION = """\
|
11 |
# SeaLLMs Med
|
12 |
|
@@ -20,7 +22,7 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
|
20 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
21 |
|
22 |
#model_id = "SeaLLMs/SeaLLMs-v3-7B-Chat"
|
23 |
-
model_id = "SeaLLMs/
|
24 |
tokenizer = Qwen2TokenizerFast.from_pretrained(model_id)
|
25 |
model = AutoModelForCausalLM.from_pretrained(
|
26 |
model_id,
|
|
|
7 |
import torch
|
8 |
from transformers import AutoModelForCausalLM, Qwen2TokenizerFast, TextIteratorStreamer
|
9 |
|
10 |
+
# based on gemma demo code
|
11 |
+
|
12 |
DESCRIPTION = """\
|
13 |
# SeaLLMs Med
|
14 |
|
|
|
22 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
23 |
|
24 |
#model_id = "SeaLLMs/SeaLLMs-v3-7B-Chat"
|
25 |
+
model_id = "SeaLLMs/debug_4"
|
26 |
tokenizer = Qwen2TokenizerFast.from_pretrained(model_id)
|
27 |
model = AutoModelForCausalLM.from_pretrained(
|
28 |
model_id,
|