abdiharyadi commited on
Commit
59d9ffe
1 Parent(s): bc7ecdf

Training in progress, step 57600, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb33f747a9cc806ffd94e5b8ddb3630cdf4f31c9ef74ba4ecc37bde528864ab9
3
  size 1575259780
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7697c1f1adb8e015d5c3f3c0b2864dbad3627bcaec5427bbce6ed72735228e4a
3
  size 1575259780
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f96f6033efc4c94c0bb727d96ecb0d84fd3f89d1e2cbadaa770f1e6db505e5e3
3
  size 3150397656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a34992540cb175216b01ab29ab220f0265b25d4bf6711ad7290a3faf7d8191f
3
  size 3150397656
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fa1aab919f439ceebd5f5725f4f3dc23890f16b939d3466d3695c6a3efa7a68
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72102f722967ab077dc39614db80ae9cd5e3dd4217ccf449e4996fcbd9f54753
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4911df471240862394f4736ab0bf9836e72d5a17b8d8f097afe6df45f823e678
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a887e4ba28c632b6567b2d7697a877a729f6f855a5dd3c18e0fb3a2c5e6def7
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 52.1498,
3
  "best_model_checkpoint": "/kaggle/working/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted-amr-generation-v2-fted/checkpoint-14400",
4
- "epoch": 596.0264900662252,
5
  "eval_steps": 3600,
6
- "global_step": 54000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -16363,6 +16363,1096 @@
16363
  "eval_samples_per_second": 8.851,
16364
  "eval_steps_per_second": 1.77,
16365
  "step": 54000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16366
  }
16367
  ],
16368
  "logging_steps": 20,
@@ -16377,12 +17467,12 @@
16377
  "should_evaluate": false,
16378
  "should_log": false,
16379
  "should_save": true,
16380
- "should_training_stop": false
16381
  },
16382
  "attributes": {}
16383
  }
16384
  },
16385
- "total_flos": 1.5982235594052403e+17,
16386
  "train_batch_size": 5,
16387
  "trial_name": null,
16388
  "trial_params": null
 
1
  {
2
  "best_metric": 52.1498,
3
  "best_model_checkpoint": "/kaggle/working/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted-amr-generation-v2-fted/checkpoint-14400",
4
+ "epoch": 635.7615894039735,
5
  "eval_steps": 3600,
6
+ "global_step": 57600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
16363
  "eval_samples_per_second": 8.851,
16364
  "eval_steps_per_second": 1.77,
16365
  "step": 54000
16366
+ },
16367
+ {
16368
+ "epoch": 596.2472406181015,
16369
+ "learning_rate": 1.5613240418118463e-07,
16370
+ "loss": 1.0834,
16371
+ "step": 54020
16372
+ },
16373
+ {
16374
+ "epoch": 596.467991169978,
16375
+ "learning_rate": 1.5581881533101046e-07,
16376
+ "loss": 1.1256,
16377
+ "step": 54040
16378
+ },
16379
+ {
16380
+ "epoch": 596.6887417218543,
16381
+ "learning_rate": 1.555052264808362e-07,
16382
+ "loss": 1.0986,
16383
+ "step": 54060
16384
+ },
16385
+ {
16386
+ "epoch": 596.9094922737307,
16387
+ "learning_rate": 1.5519163763066202e-07,
16388
+ "loss": 1.1141,
16389
+ "step": 54080
16390
+ },
16391
+ {
16392
+ "epoch": 597.1302428256071,
16393
+ "learning_rate": 1.5487804878048785e-07,
16394
+ "loss": 1.1094,
16395
+ "step": 54100
16396
+ },
16397
+ {
16398
+ "epoch": 597.3509933774834,
16399
+ "learning_rate": 1.5456445993031357e-07,
16400
+ "loss": 1.0814,
16401
+ "step": 54120
16402
+ },
16403
+ {
16404
+ "epoch": 597.5717439293599,
16405
+ "learning_rate": 1.542508710801394e-07,
16406
+ "loss": 1.105,
16407
+ "step": 54140
16408
+ },
16409
+ {
16410
+ "epoch": 597.7924944812362,
16411
+ "learning_rate": 1.5393728222996513e-07,
16412
+ "loss": 1.1205,
16413
+ "step": 54160
16414
+ },
16415
+ {
16416
+ "epoch": 598.0132450331126,
16417
+ "learning_rate": 1.5362369337979096e-07,
16418
+ "loss": 1.0908,
16419
+ "step": 54180
16420
+ },
16421
+ {
16422
+ "epoch": 598.233995584989,
16423
+ "learning_rate": 1.533101045296167e-07,
16424
+ "loss": 1.0925,
16425
+ "step": 54200
16426
+ },
16427
+ {
16428
+ "epoch": 598.4547461368653,
16429
+ "learning_rate": 1.5299651567944252e-07,
16430
+ "loss": 1.1171,
16431
+ "step": 54220
16432
+ },
16433
+ {
16434
+ "epoch": 598.6754966887418,
16435
+ "learning_rate": 1.5268292682926835e-07,
16436
+ "loss": 1.1022,
16437
+ "step": 54240
16438
+ },
16439
+ {
16440
+ "epoch": 598.8962472406181,
16441
+ "learning_rate": 1.5236933797909405e-07,
16442
+ "loss": 1.1215,
16443
+ "step": 54260
16444
+ },
16445
+ {
16446
+ "epoch": 599.1169977924944,
16447
+ "learning_rate": 1.5205574912891988e-07,
16448
+ "loss": 1.0968,
16449
+ "step": 54280
16450
+ },
16451
+ {
16452
+ "epoch": 599.3377483443709,
16453
+ "learning_rate": 1.517421602787456e-07,
16454
+ "loss": 1.1202,
16455
+ "step": 54300
16456
+ },
16457
+ {
16458
+ "epoch": 599.5584988962472,
16459
+ "learning_rate": 1.5142857142857144e-07,
16460
+ "loss": 1.1196,
16461
+ "step": 54320
16462
+ },
16463
+ {
16464
+ "epoch": 599.7792494481237,
16465
+ "learning_rate": 1.5111498257839716e-07,
16466
+ "loss": 1.0487,
16467
+ "step": 54340
16468
+ },
16469
+ {
16470
+ "epoch": 600.0,
16471
+ "learning_rate": 1.5080139372822302e-07,
16472
+ "loss": 1.115,
16473
+ "step": 54360
16474
+ },
16475
+ {
16476
+ "epoch": 600.2207505518763,
16477
+ "learning_rate": 1.5048780487804883e-07,
16478
+ "loss": 1.112,
16479
+ "step": 54380
16480
+ },
16481
+ {
16482
+ "epoch": 600.4415011037528,
16483
+ "learning_rate": 1.5017421602787455e-07,
16484
+ "loss": 1.107,
16485
+ "step": 54400
16486
+ },
16487
+ {
16488
+ "epoch": 600.6622516556291,
16489
+ "learning_rate": 1.4986062717770036e-07,
16490
+ "loss": 1.0768,
16491
+ "step": 54420
16492
+ },
16493
+ {
16494
+ "epoch": 600.8830022075056,
16495
+ "learning_rate": 1.495470383275261e-07,
16496
+ "loss": 1.1118,
16497
+ "step": 54440
16498
+ },
16499
+ {
16500
+ "epoch": 601.1037527593819,
16501
+ "learning_rate": 1.4923344947735191e-07,
16502
+ "loss": 1.079,
16503
+ "step": 54460
16504
+ },
16505
+ {
16506
+ "epoch": 601.3245033112582,
16507
+ "learning_rate": 1.489198606271777e-07,
16508
+ "loss": 1.0904,
16509
+ "step": 54480
16510
+ },
16511
+ {
16512
+ "epoch": 601.5452538631347,
16513
+ "learning_rate": 1.4860627177700347e-07,
16514
+ "loss": 1.108,
16515
+ "step": 54500
16516
+ },
16517
+ {
16518
+ "epoch": 601.766004415011,
16519
+ "learning_rate": 1.482926829268293e-07,
16520
+ "loss": 1.0864,
16521
+ "step": 54520
16522
+ },
16523
+ {
16524
+ "epoch": 601.9867549668874,
16525
+ "learning_rate": 1.4797909407665503e-07,
16526
+ "loss": 1.0926,
16527
+ "step": 54540
16528
+ },
16529
+ {
16530
+ "epoch": 602.2075055187638,
16531
+ "learning_rate": 1.4766550522648086e-07,
16532
+ "loss": 1.0962,
16533
+ "step": 54560
16534
+ },
16535
+ {
16536
+ "epoch": 602.4282560706401,
16537
+ "learning_rate": 1.4735191637630658e-07,
16538
+ "loss": 1.1248,
16539
+ "step": 54580
16540
+ },
16541
+ {
16542
+ "epoch": 602.6490066225166,
16543
+ "learning_rate": 1.4703832752613242e-07,
16544
+ "loss": 1.1002,
16545
+ "step": 54600
16546
+ },
16547
+ {
16548
+ "epoch": 602.8697571743929,
16549
+ "learning_rate": 1.4672473867595814e-07,
16550
+ "loss": 1.1349,
16551
+ "step": 54620
16552
+ },
16553
+ {
16554
+ "epoch": 603.0905077262693,
16555
+ "learning_rate": 1.4641114982578397e-07,
16556
+ "loss": 1.1023,
16557
+ "step": 54640
16558
+ },
16559
+ {
16560
+ "epoch": 603.3112582781457,
16561
+ "learning_rate": 1.460975609756098e-07,
16562
+ "loss": 1.0891,
16563
+ "step": 54660
16564
+ },
16565
+ {
16566
+ "epoch": 603.532008830022,
16567
+ "learning_rate": 1.457839721254355e-07,
16568
+ "loss": 1.0676,
16569
+ "step": 54680
16570
+ },
16571
+ {
16572
+ "epoch": 603.7527593818985,
16573
+ "learning_rate": 1.4547038327526133e-07,
16574
+ "loss": 1.1309,
16575
+ "step": 54700
16576
+ },
16577
+ {
16578
+ "epoch": 603.9735099337748,
16579
+ "learning_rate": 1.4515679442508706e-07,
16580
+ "loss": 1.1159,
16581
+ "step": 54720
16582
+ },
16583
+ {
16584
+ "epoch": 604.1942604856512,
16585
+ "learning_rate": 1.448432055749129e-07,
16586
+ "loss": 1.0951,
16587
+ "step": 54740
16588
+ },
16589
+ {
16590
+ "epoch": 604.4150110375276,
16591
+ "learning_rate": 1.4452961672473862e-07,
16592
+ "loss": 1.1112,
16593
+ "step": 54760
16594
+ },
16595
+ {
16596
+ "epoch": 604.635761589404,
16597
+ "learning_rate": 1.4421602787456445e-07,
16598
+ "loss": 1.0841,
16599
+ "step": 54780
16600
+ },
16601
+ {
16602
+ "epoch": 604.8565121412804,
16603
+ "learning_rate": 1.4390243902439028e-07,
16604
+ "loss": 1.0821,
16605
+ "step": 54800
16606
+ },
16607
+ {
16608
+ "epoch": 605.0772626931567,
16609
+ "learning_rate": 1.43588850174216e-07,
16610
+ "loss": 1.0895,
16611
+ "step": 54820
16612
+ },
16613
+ {
16614
+ "epoch": 605.2980132450331,
16615
+ "learning_rate": 1.4327526132404184e-07,
16616
+ "loss": 1.1064,
16617
+ "step": 54840
16618
+ },
16619
+ {
16620
+ "epoch": 605.5187637969095,
16621
+ "learning_rate": 1.4296167247386756e-07,
16622
+ "loss": 1.1224,
16623
+ "step": 54860
16624
+ },
16625
+ {
16626
+ "epoch": 605.7395143487859,
16627
+ "learning_rate": 1.426480836236934e-07,
16628
+ "loss": 1.0574,
16629
+ "step": 54880
16630
+ },
16631
+ {
16632
+ "epoch": 605.9602649006622,
16633
+ "learning_rate": 1.4233449477351912e-07,
16634
+ "loss": 1.117,
16635
+ "step": 54900
16636
+ },
16637
+ {
16638
+ "epoch": 606.1810154525386,
16639
+ "learning_rate": 1.4202090592334495e-07,
16640
+ "loss": 1.1048,
16641
+ "step": 54920
16642
+ },
16643
+ {
16644
+ "epoch": 606.401766004415,
16645
+ "learning_rate": 1.4170731707317073e-07,
16646
+ "loss": 1.0805,
16647
+ "step": 54940
16648
+ },
16649
+ {
16650
+ "epoch": 606.6225165562914,
16651
+ "learning_rate": 1.4139372822299648e-07,
16652
+ "loss": 1.1119,
16653
+ "step": 54960
16654
+ },
16655
+ {
16656
+ "epoch": 606.8432671081678,
16657
+ "learning_rate": 1.410801393728223e-07,
16658
+ "loss": 1.1512,
16659
+ "step": 54980
16660
+ },
16661
+ {
16662
+ "epoch": 607.0640176600441,
16663
+ "learning_rate": 1.4076655052264804e-07,
16664
+ "loss": 1.0824,
16665
+ "step": 55000
16666
+ },
16667
+ {
16668
+ "epoch": 607.2847682119206,
16669
+ "learning_rate": 1.4045296167247387e-07,
16670
+ "loss": 1.0672,
16671
+ "step": 55020
16672
+ },
16673
+ {
16674
+ "epoch": 607.5055187637969,
16675
+ "learning_rate": 1.401393728222997e-07,
16676
+ "loss": 1.1384,
16677
+ "step": 55040
16678
+ },
16679
+ {
16680
+ "epoch": 607.7262693156733,
16681
+ "learning_rate": 1.398257839721254e-07,
16682
+ "loss": 1.1327,
16683
+ "step": 55060
16684
+ },
16685
+ {
16686
+ "epoch": 607.9470198675497,
16687
+ "learning_rate": 1.3951219512195126e-07,
16688
+ "loss": 1.081,
16689
+ "step": 55080
16690
+ },
16691
+ {
16692
+ "epoch": 608.167770419426,
16693
+ "learning_rate": 1.3919860627177698e-07,
16694
+ "loss": 1.121,
16695
+ "step": 55100
16696
+ },
16697
+ {
16698
+ "epoch": 608.3885209713025,
16699
+ "learning_rate": 1.3888501742160279e-07,
16700
+ "loss": 1.0568,
16701
+ "step": 55120
16702
+ },
16703
+ {
16704
+ "epoch": 608.6092715231788,
16705
+ "learning_rate": 1.3857142857142854e-07,
16706
+ "loss": 1.1017,
16707
+ "step": 55140
16708
+ },
16709
+ {
16710
+ "epoch": 608.8300220750552,
16711
+ "learning_rate": 1.3825783972125434e-07,
16712
+ "loss": 1.1331,
16713
+ "step": 55160
16714
+ },
16715
+ {
16716
+ "epoch": 609.0507726269316,
16717
+ "learning_rate": 1.3794425087108017e-07,
16718
+ "loss": 1.1321,
16719
+ "step": 55180
16720
+ },
16721
+ {
16722
+ "epoch": 609.2715231788079,
16723
+ "learning_rate": 1.376306620209059e-07,
16724
+ "loss": 1.1024,
16725
+ "step": 55200
16726
+ },
16727
+ {
16728
+ "epoch": 609.4922737306844,
16729
+ "learning_rate": 1.3731707317073173e-07,
16730
+ "loss": 1.0933,
16731
+ "step": 55220
16732
+ },
16733
+ {
16734
+ "epoch": 609.7130242825607,
16735
+ "learning_rate": 1.3700348432055746e-07,
16736
+ "loss": 1.1043,
16737
+ "step": 55240
16738
+ },
16739
+ {
16740
+ "epoch": 609.933774834437,
16741
+ "learning_rate": 1.366898954703833e-07,
16742
+ "loss": 1.0982,
16743
+ "step": 55260
16744
+ },
16745
+ {
16746
+ "epoch": 610.1545253863135,
16747
+ "learning_rate": 1.36376306620209e-07,
16748
+ "loss": 1.1275,
16749
+ "step": 55280
16750
+ },
16751
+ {
16752
+ "epoch": 610.3752759381898,
16753
+ "learning_rate": 1.3606271777003484e-07,
16754
+ "loss": 1.1074,
16755
+ "step": 55300
16756
+ },
16757
+ {
16758
+ "epoch": 610.5960264900663,
16759
+ "learning_rate": 1.3574912891986068e-07,
16760
+ "loss": 1.0888,
16761
+ "step": 55320
16762
+ },
16763
+ {
16764
+ "epoch": 610.8167770419426,
16765
+ "learning_rate": 1.354355400696864e-07,
16766
+ "loss": 1.101,
16767
+ "step": 55340
16768
+ },
16769
+ {
16770
+ "epoch": 611.0375275938189,
16771
+ "learning_rate": 1.3512195121951223e-07,
16772
+ "loss": 1.1021,
16773
+ "step": 55360
16774
+ },
16775
+ {
16776
+ "epoch": 611.2582781456954,
16777
+ "learning_rate": 1.3480836236933793e-07,
16778
+ "loss": 1.1006,
16779
+ "step": 55380
16780
+ },
16781
+ {
16782
+ "epoch": 611.4790286975717,
16783
+ "learning_rate": 1.344947735191638e-07,
16784
+ "loss": 1.1054,
16785
+ "step": 55400
16786
+ },
16787
+ {
16788
+ "epoch": 611.6997792494482,
16789
+ "learning_rate": 1.341811846689895e-07,
16790
+ "loss": 1.0861,
16791
+ "step": 55420
16792
+ },
16793
+ {
16794
+ "epoch": 611.9205298013245,
16795
+ "learning_rate": 1.3386759581881532e-07,
16796
+ "loss": 1.1017,
16797
+ "step": 55440
16798
+ },
16799
+ {
16800
+ "epoch": 612.1412803532008,
16801
+ "learning_rate": 1.3355400696864115e-07,
16802
+ "loss": 1.1222,
16803
+ "step": 55460
16804
+ },
16805
+ {
16806
+ "epoch": 612.3620309050773,
16807
+ "learning_rate": 1.3324041811846688e-07,
16808
+ "loss": 1.1538,
16809
+ "step": 55480
16810
+ },
16811
+ {
16812
+ "epoch": 612.5827814569536,
16813
+ "learning_rate": 1.329268292682927e-07,
16814
+ "loss": 1.0675,
16815
+ "step": 55500
16816
+ },
16817
+ {
16818
+ "epoch": 612.8035320088301,
16819
+ "learning_rate": 1.3261324041811843e-07,
16820
+ "loss": 1.1322,
16821
+ "step": 55520
16822
+ },
16823
+ {
16824
+ "epoch": 613.0242825607064,
16825
+ "learning_rate": 1.3229965156794426e-07,
16826
+ "loss": 1.0858,
16827
+ "step": 55540
16828
+ },
16829
+ {
16830
+ "epoch": 613.2450331125827,
16831
+ "learning_rate": 1.3198606271777e-07,
16832
+ "loss": 1.0964,
16833
+ "step": 55560
16834
+ },
16835
+ {
16836
+ "epoch": 613.4657836644592,
16837
+ "learning_rate": 1.3167247386759582e-07,
16838
+ "loss": 1.1252,
16839
+ "step": 55580
16840
+ },
16841
+ {
16842
+ "epoch": 613.6865342163355,
16843
+ "learning_rate": 1.3135888501742163e-07,
16844
+ "loss": 1.1413,
16845
+ "step": 55600
16846
+ },
16847
+ {
16848
+ "epoch": 613.9072847682119,
16849
+ "learning_rate": 1.3104529616724738e-07,
16850
+ "loss": 1.0924,
16851
+ "step": 55620
16852
+ },
16853
+ {
16854
+ "epoch": 614.1280353200883,
16855
+ "learning_rate": 1.3073170731707316e-07,
16856
+ "loss": 1.063,
16857
+ "step": 55640
16858
+ },
16859
+ {
16860
+ "epoch": 614.3487858719647,
16861
+ "learning_rate": 1.3041811846689894e-07,
16862
+ "loss": 1.094,
16863
+ "step": 55660
16864
+ },
16865
+ {
16866
+ "epoch": 614.5695364238411,
16867
+ "learning_rate": 1.3010452961672474e-07,
16868
+ "loss": 1.0872,
16869
+ "step": 55680
16870
+ },
16871
+ {
16872
+ "epoch": 614.7902869757174,
16873
+ "learning_rate": 1.2979094076655047e-07,
16874
+ "loss": 1.119,
16875
+ "step": 55700
16876
+ },
16877
+ {
16878
+ "epoch": 615.0110375275938,
16879
+ "learning_rate": 1.294773519163763e-07,
16880
+ "loss": 1.1066,
16881
+ "step": 55720
16882
+ },
16883
+ {
16884
+ "epoch": 615.2317880794702,
16885
+ "learning_rate": 1.2916376306620213e-07,
16886
+ "loss": 1.0862,
16887
+ "step": 55740
16888
+ },
16889
+ {
16890
+ "epoch": 615.4525386313466,
16891
+ "learning_rate": 1.2885017421602785e-07,
16892
+ "loss": 1.1227,
16893
+ "step": 55760
16894
+ },
16895
+ {
16896
+ "epoch": 615.673289183223,
16897
+ "learning_rate": 1.2853658536585368e-07,
16898
+ "loss": 1.1169,
16899
+ "step": 55780
16900
+ },
16901
+ {
16902
+ "epoch": 615.8940397350993,
16903
+ "learning_rate": 1.282229965156794e-07,
16904
+ "loss": 1.0992,
16905
+ "step": 55800
16906
+ },
16907
+ {
16908
+ "epoch": 616.1147902869757,
16909
+ "learning_rate": 1.2790940766550524e-07,
16910
+ "loss": 1.0949,
16911
+ "step": 55820
16912
+ },
16913
+ {
16914
+ "epoch": 616.3355408388521,
16915
+ "learning_rate": 1.2759581881533097e-07,
16916
+ "loss": 1.1333,
16917
+ "step": 55840
16918
+ },
16919
+ {
16920
+ "epoch": 616.5562913907285,
16921
+ "learning_rate": 1.2728222996515677e-07,
16922
+ "loss": 1.1113,
16923
+ "step": 55860
16924
+ },
16925
+ {
16926
+ "epoch": 616.7770419426049,
16927
+ "learning_rate": 1.269686411149826e-07,
16928
+ "loss": 1.0913,
16929
+ "step": 55880
16930
+ },
16931
+ {
16932
+ "epoch": 616.9977924944812,
16933
+ "learning_rate": 1.2665505226480833e-07,
16934
+ "loss": 1.1172,
16935
+ "step": 55900
16936
+ },
16937
+ {
16938
+ "epoch": 617.2185430463576,
16939
+ "learning_rate": 1.2634146341463416e-07,
16940
+ "loss": 1.1392,
16941
+ "step": 55920
16942
+ },
16943
+ {
16944
+ "epoch": 617.439293598234,
16945
+ "learning_rate": 1.2602787456445989e-07,
16946
+ "loss": 1.0899,
16947
+ "step": 55940
16948
+ },
16949
+ {
16950
+ "epoch": 617.6600441501104,
16951
+ "learning_rate": 1.2571428571428572e-07,
16952
+ "loss": 1.0693,
16953
+ "step": 55960
16954
+ },
16955
+ {
16956
+ "epoch": 617.8807947019867,
16957
+ "learning_rate": 1.2540069686411144e-07,
16958
+ "loss": 1.0865,
16959
+ "step": 55980
16960
+ },
16961
+ {
16962
+ "epoch": 618.1015452538632,
16963
+ "learning_rate": 1.2508710801393727e-07,
16964
+ "loss": 1.1116,
16965
+ "step": 56000
16966
+ },
16967
+ {
16968
+ "epoch": 618.3222958057395,
16969
+ "learning_rate": 1.247735191637631e-07,
16970
+ "loss": 1.0897,
16971
+ "step": 56020
16972
+ },
16973
+ {
16974
+ "epoch": 618.5430463576159,
16975
+ "learning_rate": 1.2445993031358883e-07,
16976
+ "loss": 1.104,
16977
+ "step": 56040
16978
+ },
16979
+ {
16980
+ "epoch": 618.7637969094923,
16981
+ "learning_rate": 1.2414634146341466e-07,
16982
+ "loss": 1.0987,
16983
+ "step": 56060
16984
+ },
16985
+ {
16986
+ "epoch": 618.9845474613686,
16987
+ "learning_rate": 1.238327526132404e-07,
16988
+ "loss": 1.094,
16989
+ "step": 56080
16990
+ },
16991
+ {
16992
+ "epoch": 619.205298013245,
16993
+ "learning_rate": 1.235191637630662e-07,
16994
+ "loss": 1.0933,
16995
+ "step": 56100
16996
+ },
16997
+ {
16998
+ "epoch": 619.4260485651214,
16999
+ "learning_rate": 1.2320557491289202e-07,
17000
+ "loss": 1.0618,
17001
+ "step": 56120
17002
+ },
17003
+ {
17004
+ "epoch": 619.6467991169978,
17005
+ "learning_rate": 1.2289198606271775e-07,
17006
+ "loss": 1.1347,
17007
+ "step": 56140
17008
+ },
17009
+ {
17010
+ "epoch": 619.8675496688742,
17011
+ "learning_rate": 1.2257839721254358e-07,
17012
+ "loss": 1.0924,
17013
+ "step": 56160
17014
+ },
17015
+ {
17016
+ "epoch": 620.0883002207505,
17017
+ "learning_rate": 1.222648083623693e-07,
17018
+ "loss": 1.0804,
17019
+ "step": 56180
17020
+ },
17021
+ {
17022
+ "epoch": 620.309050772627,
17023
+ "learning_rate": 1.2195121951219514e-07,
17024
+ "loss": 1.0635,
17025
+ "step": 56200
17026
+ },
17027
+ {
17028
+ "epoch": 620.5298013245033,
17029
+ "learning_rate": 1.2163763066202086e-07,
17030
+ "loss": 1.1245,
17031
+ "step": 56220
17032
+ },
17033
+ {
17034
+ "epoch": 620.7505518763797,
17035
+ "learning_rate": 1.213240418118467e-07,
17036
+ "loss": 1.095,
17037
+ "step": 56240
17038
+ },
17039
+ {
17040
+ "epoch": 620.9713024282561,
17041
+ "learning_rate": 1.2101045296167253e-07,
17042
+ "loss": 1.1048,
17043
+ "step": 56260
17044
+ },
17045
+ {
17046
+ "epoch": 621.1920529801324,
17047
+ "learning_rate": 1.2069686411149825e-07,
17048
+ "loss": 1.0711,
17049
+ "step": 56280
17050
+ },
17051
+ {
17052
+ "epoch": 621.4128035320089,
17053
+ "learning_rate": 1.2038327526132406e-07,
17054
+ "loss": 1.0852,
17055
+ "step": 56300
17056
+ },
17057
+ {
17058
+ "epoch": 621.6335540838852,
17059
+ "learning_rate": 1.200696864111498e-07,
17060
+ "loss": 1.1443,
17061
+ "step": 56320
17062
+ },
17063
+ {
17064
+ "epoch": 621.8543046357615,
17065
+ "learning_rate": 1.197560975609756e-07,
17066
+ "loss": 1.098,
17067
+ "step": 56340
17068
+ },
17069
+ {
17070
+ "epoch": 622.075055187638,
17071
+ "learning_rate": 1.1944250871080134e-07,
17072
+ "loss": 1.1561,
17073
+ "step": 56360
17074
+ },
17075
+ {
17076
+ "epoch": 622.2958057395143,
17077
+ "learning_rate": 1.1912891986062717e-07,
17078
+ "loss": 1.0916,
17079
+ "step": 56380
17080
+ },
17081
+ {
17082
+ "epoch": 622.5165562913908,
17083
+ "learning_rate": 1.18815331010453e-07,
17084
+ "loss": 1.1205,
17085
+ "step": 56400
17086
+ },
17087
+ {
17088
+ "epoch": 622.7373068432671,
17089
+ "learning_rate": 1.1850174216027873e-07,
17090
+ "loss": 1.066,
17091
+ "step": 56420
17092
+ },
17093
+ {
17094
+ "epoch": 622.9580573951434,
17095
+ "learning_rate": 1.1818815331010456e-07,
17096
+ "loss": 1.1122,
17097
+ "step": 56440
17098
+ },
17099
+ {
17100
+ "epoch": 623.1788079470199,
17101
+ "learning_rate": 1.1787456445993028e-07,
17102
+ "loss": 1.1093,
17103
+ "step": 56460
17104
+ },
17105
+ {
17106
+ "epoch": 623.3995584988962,
17107
+ "learning_rate": 1.1756097560975611e-07,
17108
+ "loss": 1.082,
17109
+ "step": 56480
17110
+ },
17111
+ {
17112
+ "epoch": 623.6203090507727,
17113
+ "learning_rate": 1.1724738675958184e-07,
17114
+ "loss": 1.1245,
17115
+ "step": 56500
17116
+ },
17117
+ {
17118
+ "epoch": 623.841059602649,
17119
+ "learning_rate": 1.1693379790940766e-07,
17120
+ "loss": 1.103,
17121
+ "step": 56520
17122
+ },
17123
+ {
17124
+ "epoch": 624.0618101545253,
17125
+ "learning_rate": 1.1662020905923349e-07,
17126
+ "loss": 1.1024,
17127
+ "step": 56540
17128
+ },
17129
+ {
17130
+ "epoch": 624.2825607064018,
17131
+ "learning_rate": 1.1630662020905921e-07,
17132
+ "loss": 1.1122,
17133
+ "step": 56560
17134
+ },
17135
+ {
17136
+ "epoch": 624.5033112582781,
17137
+ "learning_rate": 1.1599303135888503e-07,
17138
+ "loss": 1.0891,
17139
+ "step": 56580
17140
+ },
17141
+ {
17142
+ "epoch": 624.7240618101546,
17143
+ "learning_rate": 1.1567944250871077e-07,
17144
+ "loss": 1.1048,
17145
+ "step": 56600
17146
+ },
17147
+ {
17148
+ "epoch": 624.9448123620309,
17149
+ "learning_rate": 1.1536585365853659e-07,
17150
+ "loss": 1.0908,
17151
+ "step": 56620
17152
+ },
17153
+ {
17154
+ "epoch": 625.1655629139073,
17155
+ "learning_rate": 1.1505226480836231e-07,
17156
+ "loss": 1.0883,
17157
+ "step": 56640
17158
+ },
17159
+ {
17160
+ "epoch": 625.3863134657837,
17161
+ "learning_rate": 1.1473867595818815e-07,
17162
+ "loss": 1.1402,
17163
+ "step": 56660
17164
+ },
17165
+ {
17166
+ "epoch": 625.60706401766,
17167
+ "learning_rate": 1.1442508710801398e-07,
17168
+ "loss": 1.0784,
17169
+ "step": 56680
17170
+ },
17171
+ {
17172
+ "epoch": 625.8278145695364,
17173
+ "learning_rate": 1.141114982578397e-07,
17174
+ "loss": 1.1173,
17175
+ "step": 56700
17176
+ },
17177
+ {
17178
+ "epoch": 626.0485651214128,
17179
+ "learning_rate": 1.1379790940766552e-07,
17180
+ "loss": 1.106,
17181
+ "step": 56720
17182
+ },
17183
+ {
17184
+ "epoch": 626.2693156732892,
17185
+ "learning_rate": 1.1348432055749126e-07,
17186
+ "loss": 1.1004,
17187
+ "step": 56740
17188
+ },
17189
+ {
17190
+ "epoch": 626.4900662251656,
17191
+ "learning_rate": 1.1317073170731708e-07,
17192
+ "loss": 1.0922,
17193
+ "step": 56760
17194
+ },
17195
+ {
17196
+ "epoch": 626.7108167770419,
17197
+ "learning_rate": 1.128571428571428e-07,
17198
+ "loss": 1.0832,
17199
+ "step": 56780
17200
+ },
17201
+ {
17202
+ "epoch": 626.9315673289183,
17203
+ "learning_rate": 1.1254355400696863e-07,
17204
+ "loss": 1.1321,
17205
+ "step": 56800
17206
+ },
17207
+ {
17208
+ "epoch": 627.1523178807947,
17209
+ "learning_rate": 1.1222996515679445e-07,
17210
+ "loss": 1.1027,
17211
+ "step": 56820
17212
+ },
17213
+ {
17214
+ "epoch": 627.3730684326711,
17215
+ "learning_rate": 1.1191637630662019e-07,
17216
+ "loss": 1.1085,
17217
+ "step": 56840
17218
+ },
17219
+ {
17220
+ "epoch": 627.5938189845475,
17221
+ "learning_rate": 1.1160278745644601e-07,
17222
+ "loss": 1.1344,
17223
+ "step": 56860
17224
+ },
17225
+ {
17226
+ "epoch": 627.8145695364238,
17227
+ "learning_rate": 1.1128919860627172e-07,
17228
+ "loss": 1.1029,
17229
+ "step": 56880
17230
+ },
17231
+ {
17232
+ "epoch": 628.0353200883002,
17233
+ "learning_rate": 1.1097560975609757e-07,
17234
+ "loss": 1.0907,
17235
+ "step": 56900
17236
+ },
17237
+ {
17238
+ "epoch": 628.2560706401766,
17239
+ "learning_rate": 1.1066202090592329e-07,
17240
+ "loss": 1.088,
17241
+ "step": 56920
17242
+ },
17243
+ {
17244
+ "epoch": 628.476821192053,
17245
+ "learning_rate": 1.1034843205574912e-07,
17246
+ "loss": 1.0601,
17247
+ "step": 56940
17248
+ },
17249
+ {
17250
+ "epoch": 628.6975717439294,
17251
+ "learning_rate": 1.1003484320557493e-07,
17252
+ "loss": 1.1104,
17253
+ "step": 56960
17254
+ },
17255
+ {
17256
+ "epoch": 628.9183222958058,
17257
+ "learning_rate": 1.0972125435540067e-07,
17258
+ "loss": 1.0918,
17259
+ "step": 56980
17260
+ },
17261
+ {
17262
+ "epoch": 629.1390728476821,
17263
+ "learning_rate": 1.094076655052265e-07,
17264
+ "loss": 1.1119,
17265
+ "step": 57000
17266
+ },
17267
+ {
17268
+ "epoch": 629.3598233995585,
17269
+ "learning_rate": 1.0909407665505222e-07,
17270
+ "loss": 1.0701,
17271
+ "step": 57020
17272
+ },
17273
+ {
17274
+ "epoch": 629.5805739514349,
17275
+ "learning_rate": 1.0878048780487805e-07,
17276
+ "loss": 1.0944,
17277
+ "step": 57040
17278
+ },
17279
+ {
17280
+ "epoch": 629.8013245033112,
17281
+ "learning_rate": 1.0846689895470378e-07,
17282
+ "loss": 1.0935,
17283
+ "step": 57060
17284
+ },
17285
+ {
17286
+ "epoch": 630.0220750551877,
17287
+ "learning_rate": 1.081533101045296e-07,
17288
+ "loss": 1.1178,
17289
+ "step": 57080
17290
+ },
17291
+ {
17292
+ "epoch": 630.242825607064,
17293
+ "learning_rate": 1.0783972125435543e-07,
17294
+ "loss": 1.0934,
17295
+ "step": 57100
17296
+ },
17297
+ {
17298
+ "epoch": 630.4635761589404,
17299
+ "learning_rate": 1.0752613240418116e-07,
17300
+ "loss": 1.057,
17301
+ "step": 57120
17302
+ },
17303
+ {
17304
+ "epoch": 630.6843267108168,
17305
+ "learning_rate": 1.0721254355400699e-07,
17306
+ "loss": 1.1279,
17307
+ "step": 57140
17308
+ },
17309
+ {
17310
+ "epoch": 630.9050772626931,
17311
+ "learning_rate": 1.0689895470383271e-07,
17312
+ "loss": 1.093,
17313
+ "step": 57160
17314
+ },
17315
+ {
17316
+ "epoch": 631.1258278145696,
17317
+ "learning_rate": 1.0658536585365854e-07,
17318
+ "loss": 1.0932,
17319
+ "step": 57180
17320
+ },
17321
+ {
17322
+ "epoch": 631.3465783664459,
17323
+ "learning_rate": 1.0627177700348436e-07,
17324
+ "loss": 1.0942,
17325
+ "step": 57200
17326
+ },
17327
+ {
17328
+ "epoch": 631.5673289183223,
17329
+ "learning_rate": 1.0595818815331009e-07,
17330
+ "loss": 1.0939,
17331
+ "step": 57220
17332
+ },
17333
+ {
17334
+ "epoch": 631.7880794701987,
17335
+ "learning_rate": 1.0564459930313592e-07,
17336
+ "loss": 1.1492,
17337
+ "step": 57240
17338
+ },
17339
+ {
17340
+ "epoch": 632.008830022075,
17341
+ "learning_rate": 1.0533101045296164e-07,
17342
+ "loss": 1.0726,
17343
+ "step": 57260
17344
+ },
17345
+ {
17346
+ "epoch": 632.2295805739515,
17347
+ "learning_rate": 1.0501742160278747e-07,
17348
+ "loss": 1.0441,
17349
+ "step": 57280
17350
+ },
17351
+ {
17352
+ "epoch": 632.4503311258278,
17353
+ "learning_rate": 1.047038327526132e-07,
17354
+ "loss": 1.1222,
17355
+ "step": 57300
17356
+ },
17357
+ {
17358
+ "epoch": 632.6710816777043,
17359
+ "learning_rate": 1.0439024390243902e-07,
17360
+ "loss": 1.1118,
17361
+ "step": 57320
17362
+ },
17363
+ {
17364
+ "epoch": 632.8918322295806,
17365
+ "learning_rate": 1.0407665505226485e-07,
17366
+ "loss": 1.1005,
17367
+ "step": 57340
17368
+ },
17369
+ {
17370
+ "epoch": 633.1125827814569,
17371
+ "learning_rate": 1.0376306620209058e-07,
17372
+ "loss": 1.0524,
17373
+ "step": 57360
17374
+ },
17375
+ {
17376
+ "epoch": 633.3333333333334,
17377
+ "learning_rate": 1.0344947735191641e-07,
17378
+ "loss": 1.1023,
17379
+ "step": 57380
17380
+ },
17381
+ {
17382
+ "epoch": 633.5540838852097,
17383
+ "learning_rate": 1.0313588850174213e-07,
17384
+ "loss": 1.0984,
17385
+ "step": 57400
17386
+ },
17387
+ {
17388
+ "epoch": 633.774834437086,
17389
+ "learning_rate": 1.0282229965156795e-07,
17390
+ "loss": 1.1291,
17391
+ "step": 57420
17392
+ },
17393
+ {
17394
+ "epoch": 633.9955849889625,
17395
+ "learning_rate": 1.0250871080139369e-07,
17396
+ "loss": 1.0789,
17397
+ "step": 57440
17398
+ },
17399
+ {
17400
+ "epoch": 634.2163355408388,
17401
+ "learning_rate": 1.0219512195121951e-07,
17402
+ "loss": 1.0854,
17403
+ "step": 57460
17404
+ },
17405
+ {
17406
+ "epoch": 634.4370860927153,
17407
+ "learning_rate": 1.0188153310104534e-07,
17408
+ "loss": 1.059,
17409
+ "step": 57480
17410
+ },
17411
+ {
17412
+ "epoch": 634.6578366445916,
17413
+ "learning_rate": 1.0156794425087106e-07,
17414
+ "loss": 1.116,
17415
+ "step": 57500
17416
+ },
17417
+ {
17418
+ "epoch": 634.878587196468,
17419
+ "learning_rate": 1.012543554006969e-07,
17420
+ "loss": 1.0855,
17421
+ "step": 57520
17422
+ },
17423
+ {
17424
+ "epoch": 635.0993377483444,
17425
+ "learning_rate": 1.0094076655052262e-07,
17426
+ "loss": 1.1052,
17427
+ "step": 57540
17428
+ },
17429
+ {
17430
+ "epoch": 635.3200883002207,
17431
+ "learning_rate": 1.0062717770034844e-07,
17432
+ "loss": 1.1072,
17433
+ "step": 57560
17434
+ },
17435
+ {
17436
+ "epoch": 635.5408388520972,
17437
+ "learning_rate": 1.0031358885017416e-07,
17438
+ "loss": 1.0774,
17439
+ "step": 57580
17440
+ },
17441
+ {
17442
+ "epoch": 635.7615894039735,
17443
+ "learning_rate": 1e-07,
17444
+ "loss": 1.1211,
17445
+ "step": 57600
17446
+ },
17447
+ {
17448
+ "epoch": 635.7615894039735,
17449
+ "eval_bleu": 50.4231,
17450
+ "eval_gen_len": 8.2667,
17451
+ "eval_loss": 1.9816315174102783,
17452
+ "eval_runtime": 3.323,
17453
+ "eval_samples_per_second": 9.028,
17454
+ "eval_steps_per_second": 1.806,
17455
+ "step": 57600
17456
  }
17457
  ],
17458
  "logging_steps": 20,
 
17467
  "should_evaluate": false,
17468
  "should_log": false,
17469
  "should_save": true,
17470
+ "should_training_stop": true
17471
  },
17472
  "attributes": {}
17473
  }
17474
  },
17475
+ "total_flos": 1.704834619477033e+17,
17476
  "train_batch_size": 5,
17477
  "trial_name": null,
17478
  "trial_params": null