Spaces:
Runtime error
Runtime error
fix submission info
Browse files
app.py
CHANGED
@@ -297,104 +297,105 @@ with demo:
|
|
297 |
gr.Markdown(FAQ_TEXT, elem_classes="markdown-text")
|
298 |
|
299 |
# with gr.TabItem("π Submit here! ", elem_id="llm-benchmark-tab-table", id=3):
|
300 |
-
|
301 |
-
|
302 |
-
|
303 |
-
|
304 |
-
|
305 |
-
|
306 |
-
|
307 |
-
|
308 |
-
|
309 |
-
|
310 |
-
|
311 |
-
|
312 |
-
|
313 |
-
|
314 |
-
|
315 |
-
|
316 |
-
|
317 |
-
|
318 |
-
|
319 |
-
|
320 |
-
|
321 |
-
|
322 |
-
|
323 |
-
|
324 |
-
|
325 |
-
|
326 |
-
|
327 |
-
|
328 |
-
|
329 |
-
|
330 |
-
|
331 |
-
|
332 |
-
|
333 |
-
|
334 |
-
|
335 |
-
|
336 |
-
|
337 |
-
|
338 |
-
|
339 |
-
|
340 |
-
|
341 |
-
|
342 |
-
|
343 |
-
|
344 |
-
|
345 |
-
|
346 |
-
|
347 |
-
|
348 |
-
|
349 |
-
|
350 |
-
|
351 |
-
|
352 |
-
|
353 |
-
|
354 |
-
|
355 |
-
|
356 |
-
|
357 |
-
|
358 |
-
|
359 |
-
|
360 |
-
|
361 |
-
|
362 |
-
|
363 |
-
|
364 |
-
|
365 |
-
|
366 |
-
|
367 |
-
|
368 |
-
|
369 |
-
|
370 |
-
|
371 |
-
|
372 |
-
|
373 |
-
|
374 |
-
|
375 |
-
|
376 |
-
|
377 |
-
|
378 |
-
|
379 |
-
|
380 |
-
|
381 |
-
|
382 |
-
|
383 |
-
|
384 |
-
|
385 |
-
|
386 |
-
|
387 |
-
|
388 |
-
|
389 |
-
|
390 |
-
|
391 |
-
|
392 |
-
|
393 |
-
|
394 |
-
|
395 |
-
|
396 |
-
|
397 |
-
|
|
|
398 |
|
399 |
with gr.Row():
|
400 |
with gr.Accordion("π Citation", open=False):
|
|
|
297 |
gr.Markdown(FAQ_TEXT, elem_classes="markdown-text")
|
298 |
|
299 |
# with gr.TabItem("π Submit here! ", elem_id="llm-benchmark-tab-table", id=3):
|
300 |
+
with gr.TabItem("Submission Info", elem_id="llm-benchmark-tab-table", id=3):
|
301 |
+
with gr.Column():
|
302 |
+
with gr.Row():
|
303 |
+
gr.Markdown(EVALUATION_QUEUE_TEXT, elem_classes="markdown-text")
|
304 |
+
|
305 |
+
with gr.Column():
|
306 |
+
with gr.Accordion(
|
307 |
+
f"β
Finished Evaluations ({len(finished_eval_queue_df)})",
|
308 |
+
open=False,
|
309 |
+
):
|
310 |
+
with gr.Row():
|
311 |
+
finished_eval_table = gr.components.Dataframe(
|
312 |
+
value=finished_eval_queue_df,
|
313 |
+
headers=EVAL_COLS,
|
314 |
+
datatype=EVAL_TYPES,
|
315 |
+
row_count=5,
|
316 |
+
)
|
317 |
+
with gr.Accordion(
|
318 |
+
f"π Running Evaluation Queue ({len(running_eval_queue_df)})",
|
319 |
+
open=False,
|
320 |
+
):
|
321 |
+
with gr.Row():
|
322 |
+
running_eval_table = gr.components.Dataframe(
|
323 |
+
value=running_eval_queue_df,
|
324 |
+
headers=EVAL_COLS,
|
325 |
+
datatype=EVAL_TYPES,
|
326 |
+
row_count=5,
|
327 |
+
)
|
328 |
+
|
329 |
+
with gr.Accordion(
|
330 |
+
f"β³ Pending Evaluation Queue ({len(pending_eval_queue_df)})",
|
331 |
+
open=False,
|
332 |
+
):
|
333 |
+
with gr.Row():
|
334 |
+
pending_eval_table = gr.components.Dataframe(
|
335 |
+
value=pending_eval_queue_df,
|
336 |
+
headers=EVAL_COLS,
|
337 |
+
datatype=EVAL_TYPES,
|
338 |
+
row_count=5,
|
339 |
+
)
|
340 |
+
with gr.Accordion(
|
341 |
+
f"β Failed Evaluations ({len(failed_eval_queue_df)})",
|
342 |
+
open=False,
|
343 |
+
):
|
344 |
+
with gr.Row():
|
345 |
+
pending_eval_table = gr.components.Dataframe(
|
346 |
+
value=failed_eval_queue_df,
|
347 |
+
headers=EVAL_COLS,
|
348 |
+
datatype=EVAL_TYPES,
|
349 |
+
row_count=5,
|
350 |
+
)
|
351 |
+
# with gr.Row():
|
352 |
+
# gr.Markdown("# βοΈβ¨ Submit your model here!", elem_classes="markdown-text")
|
353 |
+
|
354 |
+
# with gr.Row():
|
355 |
+
# with gr.Column():
|
356 |
+
# model_name_textbox = gr.Textbox(label="Model name")
|
357 |
+
# revision_name_textbox = gr.Textbox(label="Revision commit", placeholder="main")
|
358 |
+
# private = gr.Checkbox(False, label="Private", visible=not IS_PUBLIC)
|
359 |
+
# model_type = gr.Dropdown(
|
360 |
+
# choices=[t.to_str(" : ") for t in ModelType if t != ModelType.Unknown],
|
361 |
+
# label="Model type",
|
362 |
+
# multiselect=False,
|
363 |
+
# value=ModelType.FT.to_str(" : "),
|
364 |
+
# interactive=True,
|
365 |
+
# )
|
366 |
+
|
367 |
+
# with gr.Column():
|
368 |
+
# precision = gr.Dropdown(
|
369 |
+
# choices=[i.value.name for i in Precision if i != Precision.Unknown],
|
370 |
+
# label="Precision",
|
371 |
+
# multiselect=False,
|
372 |
+
# value="float16",
|
373 |
+
# interactive=True,
|
374 |
+
# )
|
375 |
+
# weight_type = gr.Dropdown(
|
376 |
+
# choices=[i.value.name for i in WeightType],
|
377 |
+
# label="Weights type",
|
378 |
+
# multiselect=False,
|
379 |
+
# value="Original",
|
380 |
+
# interactive=True,
|
381 |
+
# )
|
382 |
+
# base_model_name_textbox = gr.Textbox(label="Base model (for delta or adapter weights)")
|
383 |
+
|
384 |
+
# submit_button = gr.Button("Submit Evalulation!")
|
385 |
+
# submission_result = gr.Markdown()
|
386 |
+
# submit_button.click(
|
387 |
+
# add_new_eval,
|
388 |
+
# [
|
389 |
+
# model_name_textbox,
|
390 |
+
# base_model_name_textbox,
|
391 |
+
# revision_name_textbox,
|
392 |
+
# precision,
|
393 |
+
# private,
|
394 |
+
# weight_type,
|
395 |
+
# model_type,
|
396 |
+
# ],
|
397 |
+
# submission_result,
|
398 |
+
# )
|
399 |
|
400 |
with gr.Row():
|
401 |
with gr.Accordion("π Citation", open=False):
|