bobber commited on
Commit
9a59698
·
verified ·
1 Parent(s): cb7408b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -11,7 +11,7 @@ HF_TOKEN = os.environ.get("HF_TOKEN", None)
11
 
12
  DESCRIPTION = '''
13
  <div>
14
- <h1 style="text-align: center;">deepseek-ai/DeepSeek-R1-Distill-Llama-8B</h1>
15
  </div>
16
  '''
17
 
@@ -23,7 +23,7 @@ LICENSE = """
23
 
24
  PLACEHOLDER = """
25
  <div style="padding: 30px; text-align: center; display: flex; flex-direction: column; align-items: center;">
26
- <h1 style="font-size: 28px; margin-bottom: 2px; opacity: 0.55;">DeepSeek-R1-Distill-Llama-8B</h1>
27
  <p style="font-size: 18px; margin-bottom: 2px; opacity: 0.65;">Ask me anything...</p>
28
  </div>
29
  """
@@ -44,8 +44,8 @@ h1 {
44
  """
45
 
46
  # Load the tokenizer and model
47
- tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/DeepSeek-R1-Distill-Llama-8B")
48
- model = AutoModelForCausalLM.from_pretrained("deepseek-ai/DeepSeek-R1-Distill-Llama-8B", device_map="auto") # to("cuda:0")
49
  terminators = [
50
  tokenizer.eos_token_id,
51
  tokenizer.convert_tokens_to_ids("<|eot_id|>")
@@ -117,7 +117,7 @@ with gr.Blocks(fill_height=True, css=css) as demo:
117
  label="Temperature",
118
  render=False),
119
  gr.Slider(minimum=128,
120
- maximum=4096,
121
  step=1,
122
  value=1024,
123
  label="Max new tokens",
 
11
 
12
  DESCRIPTION = '''
13
  <div>
14
+ <h1 style="text-align: center;">deepseek-ai/DeepSeek-R1-Distill-Llama-70B</h1>
15
  </div>
16
  '''
17
 
 
23
 
24
  PLACEHOLDER = """
25
  <div style="padding: 30px; text-align: center; display: flex; flex-direction: column; align-items: center;">
26
+ <h1 style="font-size: 28px; margin-bottom: 2px; opacity: 0.55;">DeepSeek-R1-Distill-Llama-70B</h1>
27
  <p style="font-size: 18px; margin-bottom: 2px; opacity: 0.65;">Ask me anything...</p>
28
  </div>
29
  """
 
44
  """
45
 
46
  # Load the tokenizer and model
47
+ tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/DeepSeek-R1-Distill-Llama-70B")
48
+ model = AutoModelForCausalLM.from_pretrained("deepseek-ai/DeepSeek-R1-Distill-Llama-70B", device_map="auto") # to("cuda:0")
49
  terminators = [
50
  tokenizer.eos_token_id,
51
  tokenizer.convert_tokens_to_ids("<|eot_id|>")
 
117
  label="Temperature",
118
  render=False),
119
  gr.Slider(minimum=128,
120
+ maximum=128024,
121
  step=1,
122
  value=1024,
123
  label="Max new tokens",