Spaces:
Runtime error
Runtime error
giannantonio15
commited on
Commit
•
44e9e50
1
Parent(s):
c1da4e2
Update app.py
Browse files
app.py
CHANGED
@@ -85,11 +85,6 @@ css = """
|
|
85 |
box-sizing: border-box; /* Include padding and border in the element's total width and height */
|
86 |
}
|
87 |
|
88 |
-
#dropdown{
|
89 |
-
position:relative;
|
90 |
-
height: 10%;
|
91 |
-
}
|
92 |
-
|
93 |
#responseMode{
|
94 |
width: 5%;
|
95 |
}
|
@@ -370,14 +365,14 @@ def main():
|
|
370 |
|
371 |
set_chat_engine()
|
372 |
with gr.Blocks(css=css, head=head) as demo:
|
373 |
-
|
374 |
with gr.Row():
|
375 |
-
output = gr.HTML()
|
376 |
with gr.Row(elem_id="buttonChat"):
|
377 |
gr.Button("STANDARD", size="sm").click(fn=select_standard_mode, outputs=output)
|
378 |
gr.Button("CHAT",size="sm").click(fn=select_chat_mode, outputs=output)
|
379 |
|
380 |
chatbot = gr.Chatbot(elem_id="chatbot", container=False)
|
|
|
381 |
|
382 |
with gr.Column(elem_id="colonnaElementi"):
|
383 |
gr.Dropdown(
|
@@ -510,8 +505,6 @@ def main():
|
|
510 |
yield history, responseHTML
|
511 |
|
512 |
|
513 |
-
|
514 |
-
|
515 |
torch.cuda.empty_cache()
|
516 |
torch.cuda.reset_max_memory_allocated()
|
517 |
torch.cuda.reset_max_memory_cached()
|
|
|
85 |
box-sizing: border-box; /* Include padding and border in the element's total width and height */
|
86 |
}
|
87 |
|
|
|
|
|
|
|
|
|
|
|
88 |
#responseMode{
|
89 |
width: 5%;
|
90 |
}
|
|
|
365 |
|
366 |
set_chat_engine()
|
367 |
with gr.Blocks(css=css, head=head) as demo:
|
368 |
+
|
369 |
with gr.Row():
|
|
|
370 |
with gr.Row(elem_id="buttonChat"):
|
371 |
gr.Button("STANDARD", size="sm").click(fn=select_standard_mode, outputs=output)
|
372 |
gr.Button("CHAT",size="sm").click(fn=select_chat_mode, outputs=output)
|
373 |
|
374 |
chatbot = gr.Chatbot(elem_id="chatbot", container=False)
|
375 |
+
output = gr.HTML()
|
376 |
|
377 |
with gr.Column(elem_id="colonnaElementi"):
|
378 |
gr.Dropdown(
|
|
|
505 |
yield history, responseHTML
|
506 |
|
507 |
|
|
|
|
|
508 |
torch.cuda.empty_cache()
|
509 |
torch.cuda.reset_max_memory_allocated()
|
510 |
torch.cuda.reset_max_memory_cached()
|