|
{ |
|
"best_metric": 0.6819923371647509, |
|
"best_model_checkpoint": "beit-base-patch16-224-pt22k-ft22k-finetuned-FER2013/checkpoint-606", |
|
"epoch": 3.0, |
|
"global_step": 606, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.196721311475409e-06, |
|
"loss": 1.9392, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.6393442622950818e-05, |
|
"loss": 1.7671, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 2.459016393442623e-05, |
|
"loss": 1.6727, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 3.2786885245901635e-05, |
|
"loss": 1.4688, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.098360655737705e-05, |
|
"loss": 1.4389, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.918032786885246e-05, |
|
"loss": 1.3738, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.917431192660551e-05, |
|
"loss": 1.3653, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.8256880733944956e-05, |
|
"loss": 1.2999, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.733944954128441e-05, |
|
"loss": 1.301, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.642201834862386e-05, |
|
"loss": 1.3416, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.5504587155963305e-05, |
|
"loss": 1.2908, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.458715596330276e-05, |
|
"loss": 1.2205, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.366972477064221e-05, |
|
"loss": 1.2779, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.2752293577981654e-05, |
|
"loss": 1.2563, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 4.1834862385321106e-05, |
|
"loss": 1.2122, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.091743119266056e-05, |
|
"loss": 1.2289, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4e-05, |
|
"loss": 1.1613, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.9082568807339455e-05, |
|
"loss": 1.1646, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 3.81651376146789e-05, |
|
"loss": 1.2029, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.724770642201835e-05, |
|
"loss": 1.1701, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.6179031696273075, |
|
"eval_loss": 1.0163276195526123, |
|
"eval_runtime": 25.3453, |
|
"eval_samples_per_second": 113.276, |
|
"eval_steps_per_second": 3.551, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.6330275229357804e-05, |
|
"loss": 1.1281, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 3.541284403669725e-05, |
|
"loss": 1.1595, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.44954128440367e-05, |
|
"loss": 1.1224, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.3577981651376154e-05, |
|
"loss": 1.0808, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 3.26605504587156e-05, |
|
"loss": 1.1312, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 3.174311926605505e-05, |
|
"loss": 1.1423, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 3.0825688073394496e-05, |
|
"loss": 1.1882, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 2.9908256880733948e-05, |
|
"loss": 1.1086, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 2.8990825688073397e-05, |
|
"loss": 1.0772, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 2.807339449541285e-05, |
|
"loss": 1.1303, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.7155963302752297e-05, |
|
"loss": 1.1025, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 2.6238532110091746e-05, |
|
"loss": 1.1117, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 2.5321100917431194e-05, |
|
"loss": 1.0372, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 2.4403669724770646e-05, |
|
"loss": 1.0228, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.3486238532110095e-05, |
|
"loss": 1.0936, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.2568807339449544e-05, |
|
"loss": 1.0889, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.1651376146788992e-05, |
|
"loss": 1.0765, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 2.0733944954128444e-05, |
|
"loss": 1.0954, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.9816513761467893e-05, |
|
"loss": 1.1196, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.889908256880734e-05, |
|
"loss": 1.0447, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.6569139672587948, |
|
"eval_loss": 0.9236605763435364, |
|
"eval_runtime": 25.7906, |
|
"eval_samples_per_second": 111.32, |
|
"eval_steps_per_second": 3.49, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.798165137614679e-05, |
|
"loss": 1.0341, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.7064220183486242e-05, |
|
"loss": 0.9925, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.614678899082569e-05, |
|
"loss": 1.0571, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.5229357798165139e-05, |
|
"loss": 1.0625, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.431192660550459e-05, |
|
"loss": 1.0117, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.3394495412844038e-05, |
|
"loss": 0.9992, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.2477064220183488e-05, |
|
"loss": 1.0723, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.1559633027522937e-05, |
|
"loss": 1.0284, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 1.0642201834862387e-05, |
|
"loss": 0.9993, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 9.724770642201836e-06, |
|
"loss": 1.023, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 8.807339449541286e-06, |
|
"loss": 1.0313, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.889908256880735e-06, |
|
"loss": 1.0584, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 6.972477064220184e-06, |
|
"loss": 1.003, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 6.0550458715596335e-06, |
|
"loss": 1.0541, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 5.137614678899083e-06, |
|
"loss": 1.0189, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 4.220183486238532e-06, |
|
"loss": 0.9714, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.302752293577982e-06, |
|
"loss": 1.0046, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.3853211009174313e-06, |
|
"loss": 1.0082, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.467889908256881e-06, |
|
"loss": 1.0499, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 5.504587155963304e-07, |
|
"loss": 0.9712, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.6819923371647509, |
|
"eval_loss": 0.8754438161849976, |
|
"eval_runtime": 25.5354, |
|
"eval_samples_per_second": 112.432, |
|
"eval_steps_per_second": 3.525, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 606, |
|
"total_flos": 6.004415924974301e+18, |
|
"train_loss": 1.159771179602091, |
|
"train_runtime": 1779.2361, |
|
"train_samples_per_second": 43.566, |
|
"train_steps_per_second": 0.341 |
|
} |
|
], |
|
"max_steps": 606, |
|
"num_train_epochs": 3, |
|
"total_flos": 6.004415924974301e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|