abdiharyadi
commited on
Commit
•
59d9ffe
1
Parent(s):
bc7ecdf
Training in progress, step 57600, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1575259780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7697c1f1adb8e015d5c3f3c0b2864dbad3627bcaec5427bbce6ed72735228e4a
|
3 |
size 1575259780
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3150397656
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9a34992540cb175216b01ab29ab220f0265b25d4bf6711ad7290a3faf7d8191f
|
3 |
size 3150397656
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:72102f722967ab077dc39614db80ae9cd5e3dd4217ccf449e4996fcbd9f54753
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0a887e4ba28c632b6567b2d7697a877a729f6f855a5dd3c18e0fb3a2c5e6def7
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 52.1498,
|
3 |
"best_model_checkpoint": "/kaggle/working/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted-amr-generation-v2-fted/checkpoint-14400",
|
4 |
-
"epoch":
|
5 |
"eval_steps": 3600,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -16363,6 +16363,1096 @@
|
|
16363 |
"eval_samples_per_second": 8.851,
|
16364 |
"eval_steps_per_second": 1.77,
|
16365 |
"step": 54000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16366 |
}
|
16367 |
],
|
16368 |
"logging_steps": 20,
|
@@ -16377,12 +17467,12 @@
|
|
16377 |
"should_evaluate": false,
|
16378 |
"should_log": false,
|
16379 |
"should_save": true,
|
16380 |
-
"should_training_stop":
|
16381 |
},
|
16382 |
"attributes": {}
|
16383 |
}
|
16384 |
},
|
16385 |
-
"total_flos": 1.
|
16386 |
"train_batch_size": 5,
|
16387 |
"trial_name": null,
|
16388 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 52.1498,
|
3 |
"best_model_checkpoint": "/kaggle/working/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted-amr-generation-v2-fted/checkpoint-14400",
|
4 |
+
"epoch": 635.7615894039735,
|
5 |
"eval_steps": 3600,
|
6 |
+
"global_step": 57600,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
16363 |
"eval_samples_per_second": 8.851,
|
16364 |
"eval_steps_per_second": 1.77,
|
16365 |
"step": 54000
|
16366 |
+
},
|
16367 |
+
{
|
16368 |
+
"epoch": 596.2472406181015,
|
16369 |
+
"learning_rate": 1.5613240418118463e-07,
|
16370 |
+
"loss": 1.0834,
|
16371 |
+
"step": 54020
|
16372 |
+
},
|
16373 |
+
{
|
16374 |
+
"epoch": 596.467991169978,
|
16375 |
+
"learning_rate": 1.5581881533101046e-07,
|
16376 |
+
"loss": 1.1256,
|
16377 |
+
"step": 54040
|
16378 |
+
},
|
16379 |
+
{
|
16380 |
+
"epoch": 596.6887417218543,
|
16381 |
+
"learning_rate": 1.555052264808362e-07,
|
16382 |
+
"loss": 1.0986,
|
16383 |
+
"step": 54060
|
16384 |
+
},
|
16385 |
+
{
|
16386 |
+
"epoch": 596.9094922737307,
|
16387 |
+
"learning_rate": 1.5519163763066202e-07,
|
16388 |
+
"loss": 1.1141,
|
16389 |
+
"step": 54080
|
16390 |
+
},
|
16391 |
+
{
|
16392 |
+
"epoch": 597.1302428256071,
|
16393 |
+
"learning_rate": 1.5487804878048785e-07,
|
16394 |
+
"loss": 1.1094,
|
16395 |
+
"step": 54100
|
16396 |
+
},
|
16397 |
+
{
|
16398 |
+
"epoch": 597.3509933774834,
|
16399 |
+
"learning_rate": 1.5456445993031357e-07,
|
16400 |
+
"loss": 1.0814,
|
16401 |
+
"step": 54120
|
16402 |
+
},
|
16403 |
+
{
|
16404 |
+
"epoch": 597.5717439293599,
|
16405 |
+
"learning_rate": 1.542508710801394e-07,
|
16406 |
+
"loss": 1.105,
|
16407 |
+
"step": 54140
|
16408 |
+
},
|
16409 |
+
{
|
16410 |
+
"epoch": 597.7924944812362,
|
16411 |
+
"learning_rate": 1.5393728222996513e-07,
|
16412 |
+
"loss": 1.1205,
|
16413 |
+
"step": 54160
|
16414 |
+
},
|
16415 |
+
{
|
16416 |
+
"epoch": 598.0132450331126,
|
16417 |
+
"learning_rate": 1.5362369337979096e-07,
|
16418 |
+
"loss": 1.0908,
|
16419 |
+
"step": 54180
|
16420 |
+
},
|
16421 |
+
{
|
16422 |
+
"epoch": 598.233995584989,
|
16423 |
+
"learning_rate": 1.533101045296167e-07,
|
16424 |
+
"loss": 1.0925,
|
16425 |
+
"step": 54200
|
16426 |
+
},
|
16427 |
+
{
|
16428 |
+
"epoch": 598.4547461368653,
|
16429 |
+
"learning_rate": 1.5299651567944252e-07,
|
16430 |
+
"loss": 1.1171,
|
16431 |
+
"step": 54220
|
16432 |
+
},
|
16433 |
+
{
|
16434 |
+
"epoch": 598.6754966887418,
|
16435 |
+
"learning_rate": 1.5268292682926835e-07,
|
16436 |
+
"loss": 1.1022,
|
16437 |
+
"step": 54240
|
16438 |
+
},
|
16439 |
+
{
|
16440 |
+
"epoch": 598.8962472406181,
|
16441 |
+
"learning_rate": 1.5236933797909405e-07,
|
16442 |
+
"loss": 1.1215,
|
16443 |
+
"step": 54260
|
16444 |
+
},
|
16445 |
+
{
|
16446 |
+
"epoch": 599.1169977924944,
|
16447 |
+
"learning_rate": 1.5205574912891988e-07,
|
16448 |
+
"loss": 1.0968,
|
16449 |
+
"step": 54280
|
16450 |
+
},
|
16451 |
+
{
|
16452 |
+
"epoch": 599.3377483443709,
|
16453 |
+
"learning_rate": 1.517421602787456e-07,
|
16454 |
+
"loss": 1.1202,
|
16455 |
+
"step": 54300
|
16456 |
+
},
|
16457 |
+
{
|
16458 |
+
"epoch": 599.5584988962472,
|
16459 |
+
"learning_rate": 1.5142857142857144e-07,
|
16460 |
+
"loss": 1.1196,
|
16461 |
+
"step": 54320
|
16462 |
+
},
|
16463 |
+
{
|
16464 |
+
"epoch": 599.7792494481237,
|
16465 |
+
"learning_rate": 1.5111498257839716e-07,
|
16466 |
+
"loss": 1.0487,
|
16467 |
+
"step": 54340
|
16468 |
+
},
|
16469 |
+
{
|
16470 |
+
"epoch": 600.0,
|
16471 |
+
"learning_rate": 1.5080139372822302e-07,
|
16472 |
+
"loss": 1.115,
|
16473 |
+
"step": 54360
|
16474 |
+
},
|
16475 |
+
{
|
16476 |
+
"epoch": 600.2207505518763,
|
16477 |
+
"learning_rate": 1.5048780487804883e-07,
|
16478 |
+
"loss": 1.112,
|
16479 |
+
"step": 54380
|
16480 |
+
},
|
16481 |
+
{
|
16482 |
+
"epoch": 600.4415011037528,
|
16483 |
+
"learning_rate": 1.5017421602787455e-07,
|
16484 |
+
"loss": 1.107,
|
16485 |
+
"step": 54400
|
16486 |
+
},
|
16487 |
+
{
|
16488 |
+
"epoch": 600.6622516556291,
|
16489 |
+
"learning_rate": 1.4986062717770036e-07,
|
16490 |
+
"loss": 1.0768,
|
16491 |
+
"step": 54420
|
16492 |
+
},
|
16493 |
+
{
|
16494 |
+
"epoch": 600.8830022075056,
|
16495 |
+
"learning_rate": 1.495470383275261e-07,
|
16496 |
+
"loss": 1.1118,
|
16497 |
+
"step": 54440
|
16498 |
+
},
|
16499 |
+
{
|
16500 |
+
"epoch": 601.1037527593819,
|
16501 |
+
"learning_rate": 1.4923344947735191e-07,
|
16502 |
+
"loss": 1.079,
|
16503 |
+
"step": 54460
|
16504 |
+
},
|
16505 |
+
{
|
16506 |
+
"epoch": 601.3245033112582,
|
16507 |
+
"learning_rate": 1.489198606271777e-07,
|
16508 |
+
"loss": 1.0904,
|
16509 |
+
"step": 54480
|
16510 |
+
},
|
16511 |
+
{
|
16512 |
+
"epoch": 601.5452538631347,
|
16513 |
+
"learning_rate": 1.4860627177700347e-07,
|
16514 |
+
"loss": 1.108,
|
16515 |
+
"step": 54500
|
16516 |
+
},
|
16517 |
+
{
|
16518 |
+
"epoch": 601.766004415011,
|
16519 |
+
"learning_rate": 1.482926829268293e-07,
|
16520 |
+
"loss": 1.0864,
|
16521 |
+
"step": 54520
|
16522 |
+
},
|
16523 |
+
{
|
16524 |
+
"epoch": 601.9867549668874,
|
16525 |
+
"learning_rate": 1.4797909407665503e-07,
|
16526 |
+
"loss": 1.0926,
|
16527 |
+
"step": 54540
|
16528 |
+
},
|
16529 |
+
{
|
16530 |
+
"epoch": 602.2075055187638,
|
16531 |
+
"learning_rate": 1.4766550522648086e-07,
|
16532 |
+
"loss": 1.0962,
|
16533 |
+
"step": 54560
|
16534 |
+
},
|
16535 |
+
{
|
16536 |
+
"epoch": 602.4282560706401,
|
16537 |
+
"learning_rate": 1.4735191637630658e-07,
|
16538 |
+
"loss": 1.1248,
|
16539 |
+
"step": 54580
|
16540 |
+
},
|
16541 |
+
{
|
16542 |
+
"epoch": 602.6490066225166,
|
16543 |
+
"learning_rate": 1.4703832752613242e-07,
|
16544 |
+
"loss": 1.1002,
|
16545 |
+
"step": 54600
|
16546 |
+
},
|
16547 |
+
{
|
16548 |
+
"epoch": 602.8697571743929,
|
16549 |
+
"learning_rate": 1.4672473867595814e-07,
|
16550 |
+
"loss": 1.1349,
|
16551 |
+
"step": 54620
|
16552 |
+
},
|
16553 |
+
{
|
16554 |
+
"epoch": 603.0905077262693,
|
16555 |
+
"learning_rate": 1.4641114982578397e-07,
|
16556 |
+
"loss": 1.1023,
|
16557 |
+
"step": 54640
|
16558 |
+
},
|
16559 |
+
{
|
16560 |
+
"epoch": 603.3112582781457,
|
16561 |
+
"learning_rate": 1.460975609756098e-07,
|
16562 |
+
"loss": 1.0891,
|
16563 |
+
"step": 54660
|
16564 |
+
},
|
16565 |
+
{
|
16566 |
+
"epoch": 603.532008830022,
|
16567 |
+
"learning_rate": 1.457839721254355e-07,
|
16568 |
+
"loss": 1.0676,
|
16569 |
+
"step": 54680
|
16570 |
+
},
|
16571 |
+
{
|
16572 |
+
"epoch": 603.7527593818985,
|
16573 |
+
"learning_rate": 1.4547038327526133e-07,
|
16574 |
+
"loss": 1.1309,
|
16575 |
+
"step": 54700
|
16576 |
+
},
|
16577 |
+
{
|
16578 |
+
"epoch": 603.9735099337748,
|
16579 |
+
"learning_rate": 1.4515679442508706e-07,
|
16580 |
+
"loss": 1.1159,
|
16581 |
+
"step": 54720
|
16582 |
+
},
|
16583 |
+
{
|
16584 |
+
"epoch": 604.1942604856512,
|
16585 |
+
"learning_rate": 1.448432055749129e-07,
|
16586 |
+
"loss": 1.0951,
|
16587 |
+
"step": 54740
|
16588 |
+
},
|
16589 |
+
{
|
16590 |
+
"epoch": 604.4150110375276,
|
16591 |
+
"learning_rate": 1.4452961672473862e-07,
|
16592 |
+
"loss": 1.1112,
|
16593 |
+
"step": 54760
|
16594 |
+
},
|
16595 |
+
{
|
16596 |
+
"epoch": 604.635761589404,
|
16597 |
+
"learning_rate": 1.4421602787456445e-07,
|
16598 |
+
"loss": 1.0841,
|
16599 |
+
"step": 54780
|
16600 |
+
},
|
16601 |
+
{
|
16602 |
+
"epoch": 604.8565121412804,
|
16603 |
+
"learning_rate": 1.4390243902439028e-07,
|
16604 |
+
"loss": 1.0821,
|
16605 |
+
"step": 54800
|
16606 |
+
},
|
16607 |
+
{
|
16608 |
+
"epoch": 605.0772626931567,
|
16609 |
+
"learning_rate": 1.43588850174216e-07,
|
16610 |
+
"loss": 1.0895,
|
16611 |
+
"step": 54820
|
16612 |
+
},
|
16613 |
+
{
|
16614 |
+
"epoch": 605.2980132450331,
|
16615 |
+
"learning_rate": 1.4327526132404184e-07,
|
16616 |
+
"loss": 1.1064,
|
16617 |
+
"step": 54840
|
16618 |
+
},
|
16619 |
+
{
|
16620 |
+
"epoch": 605.5187637969095,
|
16621 |
+
"learning_rate": 1.4296167247386756e-07,
|
16622 |
+
"loss": 1.1224,
|
16623 |
+
"step": 54860
|
16624 |
+
},
|
16625 |
+
{
|
16626 |
+
"epoch": 605.7395143487859,
|
16627 |
+
"learning_rate": 1.426480836236934e-07,
|
16628 |
+
"loss": 1.0574,
|
16629 |
+
"step": 54880
|
16630 |
+
},
|
16631 |
+
{
|
16632 |
+
"epoch": 605.9602649006622,
|
16633 |
+
"learning_rate": 1.4233449477351912e-07,
|
16634 |
+
"loss": 1.117,
|
16635 |
+
"step": 54900
|
16636 |
+
},
|
16637 |
+
{
|
16638 |
+
"epoch": 606.1810154525386,
|
16639 |
+
"learning_rate": 1.4202090592334495e-07,
|
16640 |
+
"loss": 1.1048,
|
16641 |
+
"step": 54920
|
16642 |
+
},
|
16643 |
+
{
|
16644 |
+
"epoch": 606.401766004415,
|
16645 |
+
"learning_rate": 1.4170731707317073e-07,
|
16646 |
+
"loss": 1.0805,
|
16647 |
+
"step": 54940
|
16648 |
+
},
|
16649 |
+
{
|
16650 |
+
"epoch": 606.6225165562914,
|
16651 |
+
"learning_rate": 1.4139372822299648e-07,
|
16652 |
+
"loss": 1.1119,
|
16653 |
+
"step": 54960
|
16654 |
+
},
|
16655 |
+
{
|
16656 |
+
"epoch": 606.8432671081678,
|
16657 |
+
"learning_rate": 1.410801393728223e-07,
|
16658 |
+
"loss": 1.1512,
|
16659 |
+
"step": 54980
|
16660 |
+
},
|
16661 |
+
{
|
16662 |
+
"epoch": 607.0640176600441,
|
16663 |
+
"learning_rate": 1.4076655052264804e-07,
|
16664 |
+
"loss": 1.0824,
|
16665 |
+
"step": 55000
|
16666 |
+
},
|
16667 |
+
{
|
16668 |
+
"epoch": 607.2847682119206,
|
16669 |
+
"learning_rate": 1.4045296167247387e-07,
|
16670 |
+
"loss": 1.0672,
|
16671 |
+
"step": 55020
|
16672 |
+
},
|
16673 |
+
{
|
16674 |
+
"epoch": 607.5055187637969,
|
16675 |
+
"learning_rate": 1.401393728222997e-07,
|
16676 |
+
"loss": 1.1384,
|
16677 |
+
"step": 55040
|
16678 |
+
},
|
16679 |
+
{
|
16680 |
+
"epoch": 607.7262693156733,
|
16681 |
+
"learning_rate": 1.398257839721254e-07,
|
16682 |
+
"loss": 1.1327,
|
16683 |
+
"step": 55060
|
16684 |
+
},
|
16685 |
+
{
|
16686 |
+
"epoch": 607.9470198675497,
|
16687 |
+
"learning_rate": 1.3951219512195126e-07,
|
16688 |
+
"loss": 1.081,
|
16689 |
+
"step": 55080
|
16690 |
+
},
|
16691 |
+
{
|
16692 |
+
"epoch": 608.167770419426,
|
16693 |
+
"learning_rate": 1.3919860627177698e-07,
|
16694 |
+
"loss": 1.121,
|
16695 |
+
"step": 55100
|
16696 |
+
},
|
16697 |
+
{
|
16698 |
+
"epoch": 608.3885209713025,
|
16699 |
+
"learning_rate": 1.3888501742160279e-07,
|
16700 |
+
"loss": 1.0568,
|
16701 |
+
"step": 55120
|
16702 |
+
},
|
16703 |
+
{
|
16704 |
+
"epoch": 608.6092715231788,
|
16705 |
+
"learning_rate": 1.3857142857142854e-07,
|
16706 |
+
"loss": 1.1017,
|
16707 |
+
"step": 55140
|
16708 |
+
},
|
16709 |
+
{
|
16710 |
+
"epoch": 608.8300220750552,
|
16711 |
+
"learning_rate": 1.3825783972125434e-07,
|
16712 |
+
"loss": 1.1331,
|
16713 |
+
"step": 55160
|
16714 |
+
},
|
16715 |
+
{
|
16716 |
+
"epoch": 609.0507726269316,
|
16717 |
+
"learning_rate": 1.3794425087108017e-07,
|
16718 |
+
"loss": 1.1321,
|
16719 |
+
"step": 55180
|
16720 |
+
},
|
16721 |
+
{
|
16722 |
+
"epoch": 609.2715231788079,
|
16723 |
+
"learning_rate": 1.376306620209059e-07,
|
16724 |
+
"loss": 1.1024,
|
16725 |
+
"step": 55200
|
16726 |
+
},
|
16727 |
+
{
|
16728 |
+
"epoch": 609.4922737306844,
|
16729 |
+
"learning_rate": 1.3731707317073173e-07,
|
16730 |
+
"loss": 1.0933,
|
16731 |
+
"step": 55220
|
16732 |
+
},
|
16733 |
+
{
|
16734 |
+
"epoch": 609.7130242825607,
|
16735 |
+
"learning_rate": 1.3700348432055746e-07,
|
16736 |
+
"loss": 1.1043,
|
16737 |
+
"step": 55240
|
16738 |
+
},
|
16739 |
+
{
|
16740 |
+
"epoch": 609.933774834437,
|
16741 |
+
"learning_rate": 1.366898954703833e-07,
|
16742 |
+
"loss": 1.0982,
|
16743 |
+
"step": 55260
|
16744 |
+
},
|
16745 |
+
{
|
16746 |
+
"epoch": 610.1545253863135,
|
16747 |
+
"learning_rate": 1.36376306620209e-07,
|
16748 |
+
"loss": 1.1275,
|
16749 |
+
"step": 55280
|
16750 |
+
},
|
16751 |
+
{
|
16752 |
+
"epoch": 610.3752759381898,
|
16753 |
+
"learning_rate": 1.3606271777003484e-07,
|
16754 |
+
"loss": 1.1074,
|
16755 |
+
"step": 55300
|
16756 |
+
},
|
16757 |
+
{
|
16758 |
+
"epoch": 610.5960264900663,
|
16759 |
+
"learning_rate": 1.3574912891986068e-07,
|
16760 |
+
"loss": 1.0888,
|
16761 |
+
"step": 55320
|
16762 |
+
},
|
16763 |
+
{
|
16764 |
+
"epoch": 610.8167770419426,
|
16765 |
+
"learning_rate": 1.354355400696864e-07,
|
16766 |
+
"loss": 1.101,
|
16767 |
+
"step": 55340
|
16768 |
+
},
|
16769 |
+
{
|
16770 |
+
"epoch": 611.0375275938189,
|
16771 |
+
"learning_rate": 1.3512195121951223e-07,
|
16772 |
+
"loss": 1.1021,
|
16773 |
+
"step": 55360
|
16774 |
+
},
|
16775 |
+
{
|
16776 |
+
"epoch": 611.2582781456954,
|
16777 |
+
"learning_rate": 1.3480836236933793e-07,
|
16778 |
+
"loss": 1.1006,
|
16779 |
+
"step": 55380
|
16780 |
+
},
|
16781 |
+
{
|
16782 |
+
"epoch": 611.4790286975717,
|
16783 |
+
"learning_rate": 1.344947735191638e-07,
|
16784 |
+
"loss": 1.1054,
|
16785 |
+
"step": 55400
|
16786 |
+
},
|
16787 |
+
{
|
16788 |
+
"epoch": 611.6997792494482,
|
16789 |
+
"learning_rate": 1.341811846689895e-07,
|
16790 |
+
"loss": 1.0861,
|
16791 |
+
"step": 55420
|
16792 |
+
},
|
16793 |
+
{
|
16794 |
+
"epoch": 611.9205298013245,
|
16795 |
+
"learning_rate": 1.3386759581881532e-07,
|
16796 |
+
"loss": 1.1017,
|
16797 |
+
"step": 55440
|
16798 |
+
},
|
16799 |
+
{
|
16800 |
+
"epoch": 612.1412803532008,
|
16801 |
+
"learning_rate": 1.3355400696864115e-07,
|
16802 |
+
"loss": 1.1222,
|
16803 |
+
"step": 55460
|
16804 |
+
},
|
16805 |
+
{
|
16806 |
+
"epoch": 612.3620309050773,
|
16807 |
+
"learning_rate": 1.3324041811846688e-07,
|
16808 |
+
"loss": 1.1538,
|
16809 |
+
"step": 55480
|
16810 |
+
},
|
16811 |
+
{
|
16812 |
+
"epoch": 612.5827814569536,
|
16813 |
+
"learning_rate": 1.329268292682927e-07,
|
16814 |
+
"loss": 1.0675,
|
16815 |
+
"step": 55500
|
16816 |
+
},
|
16817 |
+
{
|
16818 |
+
"epoch": 612.8035320088301,
|
16819 |
+
"learning_rate": 1.3261324041811843e-07,
|
16820 |
+
"loss": 1.1322,
|
16821 |
+
"step": 55520
|
16822 |
+
},
|
16823 |
+
{
|
16824 |
+
"epoch": 613.0242825607064,
|
16825 |
+
"learning_rate": 1.3229965156794426e-07,
|
16826 |
+
"loss": 1.0858,
|
16827 |
+
"step": 55540
|
16828 |
+
},
|
16829 |
+
{
|
16830 |
+
"epoch": 613.2450331125827,
|
16831 |
+
"learning_rate": 1.3198606271777e-07,
|
16832 |
+
"loss": 1.0964,
|
16833 |
+
"step": 55560
|
16834 |
+
},
|
16835 |
+
{
|
16836 |
+
"epoch": 613.4657836644592,
|
16837 |
+
"learning_rate": 1.3167247386759582e-07,
|
16838 |
+
"loss": 1.1252,
|
16839 |
+
"step": 55580
|
16840 |
+
},
|
16841 |
+
{
|
16842 |
+
"epoch": 613.6865342163355,
|
16843 |
+
"learning_rate": 1.3135888501742163e-07,
|
16844 |
+
"loss": 1.1413,
|
16845 |
+
"step": 55600
|
16846 |
+
},
|
16847 |
+
{
|
16848 |
+
"epoch": 613.9072847682119,
|
16849 |
+
"learning_rate": 1.3104529616724738e-07,
|
16850 |
+
"loss": 1.0924,
|
16851 |
+
"step": 55620
|
16852 |
+
},
|
16853 |
+
{
|
16854 |
+
"epoch": 614.1280353200883,
|
16855 |
+
"learning_rate": 1.3073170731707316e-07,
|
16856 |
+
"loss": 1.063,
|
16857 |
+
"step": 55640
|
16858 |
+
},
|
16859 |
+
{
|
16860 |
+
"epoch": 614.3487858719647,
|
16861 |
+
"learning_rate": 1.3041811846689894e-07,
|
16862 |
+
"loss": 1.094,
|
16863 |
+
"step": 55660
|
16864 |
+
},
|
16865 |
+
{
|
16866 |
+
"epoch": 614.5695364238411,
|
16867 |
+
"learning_rate": 1.3010452961672474e-07,
|
16868 |
+
"loss": 1.0872,
|
16869 |
+
"step": 55680
|
16870 |
+
},
|
16871 |
+
{
|
16872 |
+
"epoch": 614.7902869757174,
|
16873 |
+
"learning_rate": 1.2979094076655047e-07,
|
16874 |
+
"loss": 1.119,
|
16875 |
+
"step": 55700
|
16876 |
+
},
|
16877 |
+
{
|
16878 |
+
"epoch": 615.0110375275938,
|
16879 |
+
"learning_rate": 1.294773519163763e-07,
|
16880 |
+
"loss": 1.1066,
|
16881 |
+
"step": 55720
|
16882 |
+
},
|
16883 |
+
{
|
16884 |
+
"epoch": 615.2317880794702,
|
16885 |
+
"learning_rate": 1.2916376306620213e-07,
|
16886 |
+
"loss": 1.0862,
|
16887 |
+
"step": 55740
|
16888 |
+
},
|
16889 |
+
{
|
16890 |
+
"epoch": 615.4525386313466,
|
16891 |
+
"learning_rate": 1.2885017421602785e-07,
|
16892 |
+
"loss": 1.1227,
|
16893 |
+
"step": 55760
|
16894 |
+
},
|
16895 |
+
{
|
16896 |
+
"epoch": 615.673289183223,
|
16897 |
+
"learning_rate": 1.2853658536585368e-07,
|
16898 |
+
"loss": 1.1169,
|
16899 |
+
"step": 55780
|
16900 |
+
},
|
16901 |
+
{
|
16902 |
+
"epoch": 615.8940397350993,
|
16903 |
+
"learning_rate": 1.282229965156794e-07,
|
16904 |
+
"loss": 1.0992,
|
16905 |
+
"step": 55800
|
16906 |
+
},
|
16907 |
+
{
|
16908 |
+
"epoch": 616.1147902869757,
|
16909 |
+
"learning_rate": 1.2790940766550524e-07,
|
16910 |
+
"loss": 1.0949,
|
16911 |
+
"step": 55820
|
16912 |
+
},
|
16913 |
+
{
|
16914 |
+
"epoch": 616.3355408388521,
|
16915 |
+
"learning_rate": 1.2759581881533097e-07,
|
16916 |
+
"loss": 1.1333,
|
16917 |
+
"step": 55840
|
16918 |
+
},
|
16919 |
+
{
|
16920 |
+
"epoch": 616.5562913907285,
|
16921 |
+
"learning_rate": 1.2728222996515677e-07,
|
16922 |
+
"loss": 1.1113,
|
16923 |
+
"step": 55860
|
16924 |
+
},
|
16925 |
+
{
|
16926 |
+
"epoch": 616.7770419426049,
|
16927 |
+
"learning_rate": 1.269686411149826e-07,
|
16928 |
+
"loss": 1.0913,
|
16929 |
+
"step": 55880
|
16930 |
+
},
|
16931 |
+
{
|
16932 |
+
"epoch": 616.9977924944812,
|
16933 |
+
"learning_rate": 1.2665505226480833e-07,
|
16934 |
+
"loss": 1.1172,
|
16935 |
+
"step": 55900
|
16936 |
+
},
|
16937 |
+
{
|
16938 |
+
"epoch": 617.2185430463576,
|
16939 |
+
"learning_rate": 1.2634146341463416e-07,
|
16940 |
+
"loss": 1.1392,
|
16941 |
+
"step": 55920
|
16942 |
+
},
|
16943 |
+
{
|
16944 |
+
"epoch": 617.439293598234,
|
16945 |
+
"learning_rate": 1.2602787456445989e-07,
|
16946 |
+
"loss": 1.0899,
|
16947 |
+
"step": 55940
|
16948 |
+
},
|
16949 |
+
{
|
16950 |
+
"epoch": 617.6600441501104,
|
16951 |
+
"learning_rate": 1.2571428571428572e-07,
|
16952 |
+
"loss": 1.0693,
|
16953 |
+
"step": 55960
|
16954 |
+
},
|
16955 |
+
{
|
16956 |
+
"epoch": 617.8807947019867,
|
16957 |
+
"learning_rate": 1.2540069686411144e-07,
|
16958 |
+
"loss": 1.0865,
|
16959 |
+
"step": 55980
|
16960 |
+
},
|
16961 |
+
{
|
16962 |
+
"epoch": 618.1015452538632,
|
16963 |
+
"learning_rate": 1.2508710801393727e-07,
|
16964 |
+
"loss": 1.1116,
|
16965 |
+
"step": 56000
|
16966 |
+
},
|
16967 |
+
{
|
16968 |
+
"epoch": 618.3222958057395,
|
16969 |
+
"learning_rate": 1.247735191637631e-07,
|
16970 |
+
"loss": 1.0897,
|
16971 |
+
"step": 56020
|
16972 |
+
},
|
16973 |
+
{
|
16974 |
+
"epoch": 618.5430463576159,
|
16975 |
+
"learning_rate": 1.2445993031358883e-07,
|
16976 |
+
"loss": 1.104,
|
16977 |
+
"step": 56040
|
16978 |
+
},
|
16979 |
+
{
|
16980 |
+
"epoch": 618.7637969094923,
|
16981 |
+
"learning_rate": 1.2414634146341466e-07,
|
16982 |
+
"loss": 1.0987,
|
16983 |
+
"step": 56060
|
16984 |
+
},
|
16985 |
+
{
|
16986 |
+
"epoch": 618.9845474613686,
|
16987 |
+
"learning_rate": 1.238327526132404e-07,
|
16988 |
+
"loss": 1.094,
|
16989 |
+
"step": 56080
|
16990 |
+
},
|
16991 |
+
{
|
16992 |
+
"epoch": 619.205298013245,
|
16993 |
+
"learning_rate": 1.235191637630662e-07,
|
16994 |
+
"loss": 1.0933,
|
16995 |
+
"step": 56100
|
16996 |
+
},
|
16997 |
+
{
|
16998 |
+
"epoch": 619.4260485651214,
|
16999 |
+
"learning_rate": 1.2320557491289202e-07,
|
17000 |
+
"loss": 1.0618,
|
17001 |
+
"step": 56120
|
17002 |
+
},
|
17003 |
+
{
|
17004 |
+
"epoch": 619.6467991169978,
|
17005 |
+
"learning_rate": 1.2289198606271775e-07,
|
17006 |
+
"loss": 1.1347,
|
17007 |
+
"step": 56140
|
17008 |
+
},
|
17009 |
+
{
|
17010 |
+
"epoch": 619.8675496688742,
|
17011 |
+
"learning_rate": 1.2257839721254358e-07,
|
17012 |
+
"loss": 1.0924,
|
17013 |
+
"step": 56160
|
17014 |
+
},
|
17015 |
+
{
|
17016 |
+
"epoch": 620.0883002207505,
|
17017 |
+
"learning_rate": 1.222648083623693e-07,
|
17018 |
+
"loss": 1.0804,
|
17019 |
+
"step": 56180
|
17020 |
+
},
|
17021 |
+
{
|
17022 |
+
"epoch": 620.309050772627,
|
17023 |
+
"learning_rate": 1.2195121951219514e-07,
|
17024 |
+
"loss": 1.0635,
|
17025 |
+
"step": 56200
|
17026 |
+
},
|
17027 |
+
{
|
17028 |
+
"epoch": 620.5298013245033,
|
17029 |
+
"learning_rate": 1.2163763066202086e-07,
|
17030 |
+
"loss": 1.1245,
|
17031 |
+
"step": 56220
|
17032 |
+
},
|
17033 |
+
{
|
17034 |
+
"epoch": 620.7505518763797,
|
17035 |
+
"learning_rate": 1.213240418118467e-07,
|
17036 |
+
"loss": 1.095,
|
17037 |
+
"step": 56240
|
17038 |
+
},
|
17039 |
+
{
|
17040 |
+
"epoch": 620.9713024282561,
|
17041 |
+
"learning_rate": 1.2101045296167253e-07,
|
17042 |
+
"loss": 1.1048,
|
17043 |
+
"step": 56260
|
17044 |
+
},
|
17045 |
+
{
|
17046 |
+
"epoch": 621.1920529801324,
|
17047 |
+
"learning_rate": 1.2069686411149825e-07,
|
17048 |
+
"loss": 1.0711,
|
17049 |
+
"step": 56280
|
17050 |
+
},
|
17051 |
+
{
|
17052 |
+
"epoch": 621.4128035320089,
|
17053 |
+
"learning_rate": 1.2038327526132406e-07,
|
17054 |
+
"loss": 1.0852,
|
17055 |
+
"step": 56300
|
17056 |
+
},
|
17057 |
+
{
|
17058 |
+
"epoch": 621.6335540838852,
|
17059 |
+
"learning_rate": 1.200696864111498e-07,
|
17060 |
+
"loss": 1.1443,
|
17061 |
+
"step": 56320
|
17062 |
+
},
|
17063 |
+
{
|
17064 |
+
"epoch": 621.8543046357615,
|
17065 |
+
"learning_rate": 1.197560975609756e-07,
|
17066 |
+
"loss": 1.098,
|
17067 |
+
"step": 56340
|
17068 |
+
},
|
17069 |
+
{
|
17070 |
+
"epoch": 622.075055187638,
|
17071 |
+
"learning_rate": 1.1944250871080134e-07,
|
17072 |
+
"loss": 1.1561,
|
17073 |
+
"step": 56360
|
17074 |
+
},
|
17075 |
+
{
|
17076 |
+
"epoch": 622.2958057395143,
|
17077 |
+
"learning_rate": 1.1912891986062717e-07,
|
17078 |
+
"loss": 1.0916,
|
17079 |
+
"step": 56380
|
17080 |
+
},
|
17081 |
+
{
|
17082 |
+
"epoch": 622.5165562913908,
|
17083 |
+
"learning_rate": 1.18815331010453e-07,
|
17084 |
+
"loss": 1.1205,
|
17085 |
+
"step": 56400
|
17086 |
+
},
|
17087 |
+
{
|
17088 |
+
"epoch": 622.7373068432671,
|
17089 |
+
"learning_rate": 1.1850174216027873e-07,
|
17090 |
+
"loss": 1.066,
|
17091 |
+
"step": 56420
|
17092 |
+
},
|
17093 |
+
{
|
17094 |
+
"epoch": 622.9580573951434,
|
17095 |
+
"learning_rate": 1.1818815331010456e-07,
|
17096 |
+
"loss": 1.1122,
|
17097 |
+
"step": 56440
|
17098 |
+
},
|
17099 |
+
{
|
17100 |
+
"epoch": 623.1788079470199,
|
17101 |
+
"learning_rate": 1.1787456445993028e-07,
|
17102 |
+
"loss": 1.1093,
|
17103 |
+
"step": 56460
|
17104 |
+
},
|
17105 |
+
{
|
17106 |
+
"epoch": 623.3995584988962,
|
17107 |
+
"learning_rate": 1.1756097560975611e-07,
|
17108 |
+
"loss": 1.082,
|
17109 |
+
"step": 56480
|
17110 |
+
},
|
17111 |
+
{
|
17112 |
+
"epoch": 623.6203090507727,
|
17113 |
+
"learning_rate": 1.1724738675958184e-07,
|
17114 |
+
"loss": 1.1245,
|
17115 |
+
"step": 56500
|
17116 |
+
},
|
17117 |
+
{
|
17118 |
+
"epoch": 623.841059602649,
|
17119 |
+
"learning_rate": 1.1693379790940766e-07,
|
17120 |
+
"loss": 1.103,
|
17121 |
+
"step": 56520
|
17122 |
+
},
|
17123 |
+
{
|
17124 |
+
"epoch": 624.0618101545253,
|
17125 |
+
"learning_rate": 1.1662020905923349e-07,
|
17126 |
+
"loss": 1.1024,
|
17127 |
+
"step": 56540
|
17128 |
+
},
|
17129 |
+
{
|
17130 |
+
"epoch": 624.2825607064018,
|
17131 |
+
"learning_rate": 1.1630662020905921e-07,
|
17132 |
+
"loss": 1.1122,
|
17133 |
+
"step": 56560
|
17134 |
+
},
|
17135 |
+
{
|
17136 |
+
"epoch": 624.5033112582781,
|
17137 |
+
"learning_rate": 1.1599303135888503e-07,
|
17138 |
+
"loss": 1.0891,
|
17139 |
+
"step": 56580
|
17140 |
+
},
|
17141 |
+
{
|
17142 |
+
"epoch": 624.7240618101546,
|
17143 |
+
"learning_rate": 1.1567944250871077e-07,
|
17144 |
+
"loss": 1.1048,
|
17145 |
+
"step": 56600
|
17146 |
+
},
|
17147 |
+
{
|
17148 |
+
"epoch": 624.9448123620309,
|
17149 |
+
"learning_rate": 1.1536585365853659e-07,
|
17150 |
+
"loss": 1.0908,
|
17151 |
+
"step": 56620
|
17152 |
+
},
|
17153 |
+
{
|
17154 |
+
"epoch": 625.1655629139073,
|
17155 |
+
"learning_rate": 1.1505226480836231e-07,
|
17156 |
+
"loss": 1.0883,
|
17157 |
+
"step": 56640
|
17158 |
+
},
|
17159 |
+
{
|
17160 |
+
"epoch": 625.3863134657837,
|
17161 |
+
"learning_rate": 1.1473867595818815e-07,
|
17162 |
+
"loss": 1.1402,
|
17163 |
+
"step": 56660
|
17164 |
+
},
|
17165 |
+
{
|
17166 |
+
"epoch": 625.60706401766,
|
17167 |
+
"learning_rate": 1.1442508710801398e-07,
|
17168 |
+
"loss": 1.0784,
|
17169 |
+
"step": 56680
|
17170 |
+
},
|
17171 |
+
{
|
17172 |
+
"epoch": 625.8278145695364,
|
17173 |
+
"learning_rate": 1.141114982578397e-07,
|
17174 |
+
"loss": 1.1173,
|
17175 |
+
"step": 56700
|
17176 |
+
},
|
17177 |
+
{
|
17178 |
+
"epoch": 626.0485651214128,
|
17179 |
+
"learning_rate": 1.1379790940766552e-07,
|
17180 |
+
"loss": 1.106,
|
17181 |
+
"step": 56720
|
17182 |
+
},
|
17183 |
+
{
|
17184 |
+
"epoch": 626.2693156732892,
|
17185 |
+
"learning_rate": 1.1348432055749126e-07,
|
17186 |
+
"loss": 1.1004,
|
17187 |
+
"step": 56740
|
17188 |
+
},
|
17189 |
+
{
|
17190 |
+
"epoch": 626.4900662251656,
|
17191 |
+
"learning_rate": 1.1317073170731708e-07,
|
17192 |
+
"loss": 1.0922,
|
17193 |
+
"step": 56760
|
17194 |
+
},
|
17195 |
+
{
|
17196 |
+
"epoch": 626.7108167770419,
|
17197 |
+
"learning_rate": 1.128571428571428e-07,
|
17198 |
+
"loss": 1.0832,
|
17199 |
+
"step": 56780
|
17200 |
+
},
|
17201 |
+
{
|
17202 |
+
"epoch": 626.9315673289183,
|
17203 |
+
"learning_rate": 1.1254355400696863e-07,
|
17204 |
+
"loss": 1.1321,
|
17205 |
+
"step": 56800
|
17206 |
+
},
|
17207 |
+
{
|
17208 |
+
"epoch": 627.1523178807947,
|
17209 |
+
"learning_rate": 1.1222996515679445e-07,
|
17210 |
+
"loss": 1.1027,
|
17211 |
+
"step": 56820
|
17212 |
+
},
|
17213 |
+
{
|
17214 |
+
"epoch": 627.3730684326711,
|
17215 |
+
"learning_rate": 1.1191637630662019e-07,
|
17216 |
+
"loss": 1.1085,
|
17217 |
+
"step": 56840
|
17218 |
+
},
|
17219 |
+
{
|
17220 |
+
"epoch": 627.5938189845475,
|
17221 |
+
"learning_rate": 1.1160278745644601e-07,
|
17222 |
+
"loss": 1.1344,
|
17223 |
+
"step": 56860
|
17224 |
+
},
|
17225 |
+
{
|
17226 |
+
"epoch": 627.8145695364238,
|
17227 |
+
"learning_rate": 1.1128919860627172e-07,
|
17228 |
+
"loss": 1.1029,
|
17229 |
+
"step": 56880
|
17230 |
+
},
|
17231 |
+
{
|
17232 |
+
"epoch": 628.0353200883002,
|
17233 |
+
"learning_rate": 1.1097560975609757e-07,
|
17234 |
+
"loss": 1.0907,
|
17235 |
+
"step": 56900
|
17236 |
+
},
|
17237 |
+
{
|
17238 |
+
"epoch": 628.2560706401766,
|
17239 |
+
"learning_rate": 1.1066202090592329e-07,
|
17240 |
+
"loss": 1.088,
|
17241 |
+
"step": 56920
|
17242 |
+
},
|
17243 |
+
{
|
17244 |
+
"epoch": 628.476821192053,
|
17245 |
+
"learning_rate": 1.1034843205574912e-07,
|
17246 |
+
"loss": 1.0601,
|
17247 |
+
"step": 56940
|
17248 |
+
},
|
17249 |
+
{
|
17250 |
+
"epoch": 628.6975717439294,
|
17251 |
+
"learning_rate": 1.1003484320557493e-07,
|
17252 |
+
"loss": 1.1104,
|
17253 |
+
"step": 56960
|
17254 |
+
},
|
17255 |
+
{
|
17256 |
+
"epoch": 628.9183222958058,
|
17257 |
+
"learning_rate": 1.0972125435540067e-07,
|
17258 |
+
"loss": 1.0918,
|
17259 |
+
"step": 56980
|
17260 |
+
},
|
17261 |
+
{
|
17262 |
+
"epoch": 629.1390728476821,
|
17263 |
+
"learning_rate": 1.094076655052265e-07,
|
17264 |
+
"loss": 1.1119,
|
17265 |
+
"step": 57000
|
17266 |
+
},
|
17267 |
+
{
|
17268 |
+
"epoch": 629.3598233995585,
|
17269 |
+
"learning_rate": 1.0909407665505222e-07,
|
17270 |
+
"loss": 1.0701,
|
17271 |
+
"step": 57020
|
17272 |
+
},
|
17273 |
+
{
|
17274 |
+
"epoch": 629.5805739514349,
|
17275 |
+
"learning_rate": 1.0878048780487805e-07,
|
17276 |
+
"loss": 1.0944,
|
17277 |
+
"step": 57040
|
17278 |
+
},
|
17279 |
+
{
|
17280 |
+
"epoch": 629.8013245033112,
|
17281 |
+
"learning_rate": 1.0846689895470378e-07,
|
17282 |
+
"loss": 1.0935,
|
17283 |
+
"step": 57060
|
17284 |
+
},
|
17285 |
+
{
|
17286 |
+
"epoch": 630.0220750551877,
|
17287 |
+
"learning_rate": 1.081533101045296e-07,
|
17288 |
+
"loss": 1.1178,
|
17289 |
+
"step": 57080
|
17290 |
+
},
|
17291 |
+
{
|
17292 |
+
"epoch": 630.242825607064,
|
17293 |
+
"learning_rate": 1.0783972125435543e-07,
|
17294 |
+
"loss": 1.0934,
|
17295 |
+
"step": 57100
|
17296 |
+
},
|
17297 |
+
{
|
17298 |
+
"epoch": 630.4635761589404,
|
17299 |
+
"learning_rate": 1.0752613240418116e-07,
|
17300 |
+
"loss": 1.057,
|
17301 |
+
"step": 57120
|
17302 |
+
},
|
17303 |
+
{
|
17304 |
+
"epoch": 630.6843267108168,
|
17305 |
+
"learning_rate": 1.0721254355400699e-07,
|
17306 |
+
"loss": 1.1279,
|
17307 |
+
"step": 57140
|
17308 |
+
},
|
17309 |
+
{
|
17310 |
+
"epoch": 630.9050772626931,
|
17311 |
+
"learning_rate": 1.0689895470383271e-07,
|
17312 |
+
"loss": 1.093,
|
17313 |
+
"step": 57160
|
17314 |
+
},
|
17315 |
+
{
|
17316 |
+
"epoch": 631.1258278145696,
|
17317 |
+
"learning_rate": 1.0658536585365854e-07,
|
17318 |
+
"loss": 1.0932,
|
17319 |
+
"step": 57180
|
17320 |
+
},
|
17321 |
+
{
|
17322 |
+
"epoch": 631.3465783664459,
|
17323 |
+
"learning_rate": 1.0627177700348436e-07,
|
17324 |
+
"loss": 1.0942,
|
17325 |
+
"step": 57200
|
17326 |
+
},
|
17327 |
+
{
|
17328 |
+
"epoch": 631.5673289183223,
|
17329 |
+
"learning_rate": 1.0595818815331009e-07,
|
17330 |
+
"loss": 1.0939,
|
17331 |
+
"step": 57220
|
17332 |
+
},
|
17333 |
+
{
|
17334 |
+
"epoch": 631.7880794701987,
|
17335 |
+
"learning_rate": 1.0564459930313592e-07,
|
17336 |
+
"loss": 1.1492,
|
17337 |
+
"step": 57240
|
17338 |
+
},
|
17339 |
+
{
|
17340 |
+
"epoch": 632.008830022075,
|
17341 |
+
"learning_rate": 1.0533101045296164e-07,
|
17342 |
+
"loss": 1.0726,
|
17343 |
+
"step": 57260
|
17344 |
+
},
|
17345 |
+
{
|
17346 |
+
"epoch": 632.2295805739515,
|
17347 |
+
"learning_rate": 1.0501742160278747e-07,
|
17348 |
+
"loss": 1.0441,
|
17349 |
+
"step": 57280
|
17350 |
+
},
|
17351 |
+
{
|
17352 |
+
"epoch": 632.4503311258278,
|
17353 |
+
"learning_rate": 1.047038327526132e-07,
|
17354 |
+
"loss": 1.1222,
|
17355 |
+
"step": 57300
|
17356 |
+
},
|
17357 |
+
{
|
17358 |
+
"epoch": 632.6710816777043,
|
17359 |
+
"learning_rate": 1.0439024390243902e-07,
|
17360 |
+
"loss": 1.1118,
|
17361 |
+
"step": 57320
|
17362 |
+
},
|
17363 |
+
{
|
17364 |
+
"epoch": 632.8918322295806,
|
17365 |
+
"learning_rate": 1.0407665505226485e-07,
|
17366 |
+
"loss": 1.1005,
|
17367 |
+
"step": 57340
|
17368 |
+
},
|
17369 |
+
{
|
17370 |
+
"epoch": 633.1125827814569,
|
17371 |
+
"learning_rate": 1.0376306620209058e-07,
|
17372 |
+
"loss": 1.0524,
|
17373 |
+
"step": 57360
|
17374 |
+
},
|
17375 |
+
{
|
17376 |
+
"epoch": 633.3333333333334,
|
17377 |
+
"learning_rate": 1.0344947735191641e-07,
|
17378 |
+
"loss": 1.1023,
|
17379 |
+
"step": 57380
|
17380 |
+
},
|
17381 |
+
{
|
17382 |
+
"epoch": 633.5540838852097,
|
17383 |
+
"learning_rate": 1.0313588850174213e-07,
|
17384 |
+
"loss": 1.0984,
|
17385 |
+
"step": 57400
|
17386 |
+
},
|
17387 |
+
{
|
17388 |
+
"epoch": 633.774834437086,
|
17389 |
+
"learning_rate": 1.0282229965156795e-07,
|
17390 |
+
"loss": 1.1291,
|
17391 |
+
"step": 57420
|
17392 |
+
},
|
17393 |
+
{
|
17394 |
+
"epoch": 633.9955849889625,
|
17395 |
+
"learning_rate": 1.0250871080139369e-07,
|
17396 |
+
"loss": 1.0789,
|
17397 |
+
"step": 57440
|
17398 |
+
},
|
17399 |
+
{
|
17400 |
+
"epoch": 634.2163355408388,
|
17401 |
+
"learning_rate": 1.0219512195121951e-07,
|
17402 |
+
"loss": 1.0854,
|
17403 |
+
"step": 57460
|
17404 |
+
},
|
17405 |
+
{
|
17406 |
+
"epoch": 634.4370860927153,
|
17407 |
+
"learning_rate": 1.0188153310104534e-07,
|
17408 |
+
"loss": 1.059,
|
17409 |
+
"step": 57480
|
17410 |
+
},
|
17411 |
+
{
|
17412 |
+
"epoch": 634.6578366445916,
|
17413 |
+
"learning_rate": 1.0156794425087106e-07,
|
17414 |
+
"loss": 1.116,
|
17415 |
+
"step": 57500
|
17416 |
+
},
|
17417 |
+
{
|
17418 |
+
"epoch": 634.878587196468,
|
17419 |
+
"learning_rate": 1.012543554006969e-07,
|
17420 |
+
"loss": 1.0855,
|
17421 |
+
"step": 57520
|
17422 |
+
},
|
17423 |
+
{
|
17424 |
+
"epoch": 635.0993377483444,
|
17425 |
+
"learning_rate": 1.0094076655052262e-07,
|
17426 |
+
"loss": 1.1052,
|
17427 |
+
"step": 57540
|
17428 |
+
},
|
17429 |
+
{
|
17430 |
+
"epoch": 635.3200883002207,
|
17431 |
+
"learning_rate": 1.0062717770034844e-07,
|
17432 |
+
"loss": 1.1072,
|
17433 |
+
"step": 57560
|
17434 |
+
},
|
17435 |
+
{
|
17436 |
+
"epoch": 635.5408388520972,
|
17437 |
+
"learning_rate": 1.0031358885017416e-07,
|
17438 |
+
"loss": 1.0774,
|
17439 |
+
"step": 57580
|
17440 |
+
},
|
17441 |
+
{
|
17442 |
+
"epoch": 635.7615894039735,
|
17443 |
+
"learning_rate": 1e-07,
|
17444 |
+
"loss": 1.1211,
|
17445 |
+
"step": 57600
|
17446 |
+
},
|
17447 |
+
{
|
17448 |
+
"epoch": 635.7615894039735,
|
17449 |
+
"eval_bleu": 50.4231,
|
17450 |
+
"eval_gen_len": 8.2667,
|
17451 |
+
"eval_loss": 1.9816315174102783,
|
17452 |
+
"eval_runtime": 3.323,
|
17453 |
+
"eval_samples_per_second": 9.028,
|
17454 |
+
"eval_steps_per_second": 1.806,
|
17455 |
+
"step": 57600
|
17456 |
}
|
17457 |
],
|
17458 |
"logging_steps": 20,
|
|
|
17467 |
"should_evaluate": false,
|
17468 |
"should_log": false,
|
17469 |
"should_save": true,
|
17470 |
+
"should_training_stop": true
|
17471 |
},
|
17472 |
"attributes": {}
|
17473 |
}
|
17474 |
},
|
17475 |
+
"total_flos": 1.704834619477033e+17,
|
17476 |
"train_batch_size": 5,
|
17477 |
"trial_name": null,
|
17478 |
"trial_params": null
|