Spaces:
Running
on
T4
Running
on
T4
Update app.py
Browse files
app.py
CHANGED
@@ -465,8 +465,8 @@ if __name__ == '__main__':
|
|
465 |
|
466 |
if prompt := st.chat_input("Message Video Game Assistant"):
|
467 |
if st.session_state.chat_id not in conversations.keys():
|
468 |
-
sampling_params = vllm.SamplingParams(temperature=0.75, max_tokens=
|
469 |
-
prompt_conversation = [{"role": "user", "content": f"{prompt}\
|
470 |
inputs = tokenizer.apply_chat_template(prompt_conversation, tokenize=False, add_generation_prompt=True)
|
471 |
outputs = llm.generate(prompts=inputs, sampling_params=sampling_params)
|
472 |
|
|
|
465 |
|
466 |
if prompt := st.chat_input("Message Video Game Assistant"):
|
467 |
if st.session_state.chat_id not in conversations.keys():
|
468 |
+
sampling_params = vllm.SamplingParams(temperature=0.75, max_tokens=20)
|
469 |
+
prompt_conversation = [{"role": "user", "content": f"{prompt}\nResume the above in one sentence:"}]
|
470 |
inputs = tokenizer.apply_chat_template(prompt_conversation, tokenize=False, add_generation_prompt=True)
|
471 |
outputs = llm.generate(prompts=inputs, sampling_params=sampling_params)
|
472 |
|