abdiharyadi
commited on
Commit
•
c866537
1
Parent(s):
831ac70
Training in progress, step 50400, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1575259780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95d906efb2b8e32b4af8946368ae4884f80834ba95df64d70ef9ba7c594e017e
|
3 |
size 1575259780
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3150397656
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01dded4a5fd45068b856578e5730da012c7f12cd433405d8b81c9a6c3515b125
|
3 |
size 3150397656
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cd4f885adc31b81066d02d9b338d7e294e07d28f065bb3c8e772a4ad63ca9b90
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d7cfef00d5b235f498d57bdb2b7f6471a99754d27f7c93c39c1cca8137d5f26f
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 52.1498,
|
3 |
"best_model_checkpoint": "/kaggle/working/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted-amr-generation-v2-fted/checkpoint-14400",
|
4 |
-
"epoch":
|
5 |
"eval_steps": 3600,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -14183,6 +14183,1096 @@
|
|
14183 |
"eval_samples_per_second": 9.15,
|
14184 |
"eval_steps_per_second": 1.83,
|
14185 |
"step": 46800
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14186 |
}
|
14187 |
],
|
14188 |
"logging_steps": 20,
|
@@ -14202,7 +15292,7 @@
|
|
14202 |
"attributes": {}
|
14203 |
}
|
14204 |
},
|
14205 |
-
"total_flos": 1.
|
14206 |
"train_batch_size": 5,
|
14207 |
"trial_name": null,
|
14208 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 52.1498,
|
3 |
"best_model_checkpoint": "/kaggle/working/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted-amr-generation-v2-fted/checkpoint-14400",
|
4 |
+
"epoch": 556.2913907284768,
|
5 |
"eval_steps": 3600,
|
6 |
+
"global_step": 50400,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
14183 |
"eval_samples_per_second": 9.15,
|
14184 |
"eval_steps_per_second": 1.83,
|
14185 |
"step": 46800
|
14186 |
+
},
|
14187 |
+
{
|
14188 |
+
"epoch": 516.7770419426049,
|
14189 |
+
"learning_rate": 2.6902439024390234e-07,
|
14190 |
+
"loss": 1.0884,
|
14191 |
+
"step": 46820
|
14192 |
+
},
|
14193 |
+
{
|
14194 |
+
"epoch": 516.9977924944812,
|
14195 |
+
"learning_rate": 2.687108013937282e-07,
|
14196 |
+
"loss": 1.0961,
|
14197 |
+
"step": 46840
|
14198 |
+
},
|
14199 |
+
{
|
14200 |
+
"epoch": 517.2185430463576,
|
14201 |
+
"learning_rate": 2.6839721254355406e-07,
|
14202 |
+
"loss": 1.0813,
|
14203 |
+
"step": 46860
|
14204 |
+
},
|
14205 |
+
{
|
14206 |
+
"epoch": 517.439293598234,
|
14207 |
+
"learning_rate": 2.6808362369337976e-07,
|
14208 |
+
"loss": 1.1569,
|
14209 |
+
"step": 46880
|
14210 |
+
},
|
14211 |
+
{
|
14212 |
+
"epoch": 517.6600441501104,
|
14213 |
+
"learning_rate": 2.6777003484320556e-07,
|
14214 |
+
"loss": 1.1273,
|
14215 |
+
"step": 46900
|
14216 |
+
},
|
14217 |
+
{
|
14218 |
+
"epoch": 517.8807947019867,
|
14219 |
+
"learning_rate": 2.674564459930313e-07,
|
14220 |
+
"loss": 1.1095,
|
14221 |
+
"step": 46920
|
14222 |
+
},
|
14223 |
+
{
|
14224 |
+
"epoch": 518.1015452538632,
|
14225 |
+
"learning_rate": 2.6714285714285717e-07,
|
14226 |
+
"loss": 1.1044,
|
14227 |
+
"step": 46940
|
14228 |
+
},
|
14229 |
+
{
|
14230 |
+
"epoch": 518.3222958057395,
|
14231 |
+
"learning_rate": 2.668292682926829e-07,
|
14232 |
+
"loss": 1.123,
|
14233 |
+
"step": 46960
|
14234 |
+
},
|
14235 |
+
{
|
14236 |
+
"epoch": 518.5430463576159,
|
14237 |
+
"learning_rate": 2.665156794425087e-07,
|
14238 |
+
"loss": 1.1064,
|
14239 |
+
"step": 46980
|
14240 |
+
},
|
14241 |
+
{
|
14242 |
+
"epoch": 518.7637969094923,
|
14243 |
+
"learning_rate": 2.662020905923346e-07,
|
14244 |
+
"loss": 1.1221,
|
14245 |
+
"step": 47000
|
14246 |
+
},
|
14247 |
+
{
|
14248 |
+
"epoch": 518.9845474613686,
|
14249 |
+
"learning_rate": 2.6588850174216023e-07,
|
14250 |
+
"loss": 1.1345,
|
14251 |
+
"step": 47020
|
14252 |
+
},
|
14253 |
+
{
|
14254 |
+
"epoch": 519.205298013245,
|
14255 |
+
"learning_rate": 2.6557491289198604e-07,
|
14256 |
+
"loss": 1.1308,
|
14257 |
+
"step": 47040
|
14258 |
+
},
|
14259 |
+
{
|
14260 |
+
"epoch": 519.4260485651214,
|
14261 |
+
"learning_rate": 2.652613240418118e-07,
|
14262 |
+
"loss": 1.1336,
|
14263 |
+
"step": 47060
|
14264 |
+
},
|
14265 |
+
{
|
14266 |
+
"epoch": 519.6467991169978,
|
14267 |
+
"learning_rate": 2.6494773519163765e-07,
|
14268 |
+
"loss": 1.1642,
|
14269 |
+
"step": 47080
|
14270 |
+
},
|
14271 |
+
{
|
14272 |
+
"epoch": 519.8675496688742,
|
14273 |
+
"learning_rate": 2.6463414634146345e-07,
|
14274 |
+
"loss": 1.0879,
|
14275 |
+
"step": 47100
|
14276 |
+
},
|
14277 |
+
{
|
14278 |
+
"epoch": 520.0883002207505,
|
14279 |
+
"learning_rate": 2.6432055749128915e-07,
|
14280 |
+
"loss": 1.1269,
|
14281 |
+
"step": 47120
|
14282 |
+
},
|
14283 |
+
{
|
14284 |
+
"epoch": 520.309050772627,
|
14285 |
+
"learning_rate": 2.64006968641115e-07,
|
14286 |
+
"loss": 1.1539,
|
14287 |
+
"step": 47140
|
14288 |
+
},
|
14289 |
+
{
|
14290 |
+
"epoch": 520.5298013245033,
|
14291 |
+
"learning_rate": 2.6369337979094076e-07,
|
14292 |
+
"loss": 1.1108,
|
14293 |
+
"step": 47160
|
14294 |
+
},
|
14295 |
+
{
|
14296 |
+
"epoch": 520.7505518763797,
|
14297 |
+
"learning_rate": 2.6337979094076657e-07,
|
14298 |
+
"loss": 1.1258,
|
14299 |
+
"step": 47180
|
14300 |
+
},
|
14301 |
+
{
|
14302 |
+
"epoch": 520.9713024282561,
|
14303 |
+
"learning_rate": 2.6306620209059226e-07,
|
14304 |
+
"loss": 1.1445,
|
14305 |
+
"step": 47200
|
14306 |
+
},
|
14307 |
+
{
|
14308 |
+
"epoch": 521.1920529801324,
|
14309 |
+
"learning_rate": 2.627526132404181e-07,
|
14310 |
+
"loss": 1.1007,
|
14311 |
+
"step": 47220
|
14312 |
+
},
|
14313 |
+
{
|
14314 |
+
"epoch": 521.4128035320089,
|
14315 |
+
"learning_rate": 2.6243902439024393e-07,
|
14316 |
+
"loss": 1.1411,
|
14317 |
+
"step": 47240
|
14318 |
+
},
|
14319 |
+
{
|
14320 |
+
"epoch": 521.6335540838852,
|
14321 |
+
"learning_rate": 2.621254355400696e-07,
|
14322 |
+
"loss": 1.0919,
|
14323 |
+
"step": 47260
|
14324 |
+
},
|
14325 |
+
{
|
14326 |
+
"epoch": 521.8543046357615,
|
14327 |
+
"learning_rate": 2.618118466898955e-07,
|
14328 |
+
"loss": 1.1141,
|
14329 |
+
"step": 47280
|
14330 |
+
},
|
14331 |
+
{
|
14332 |
+
"epoch": 522.075055187638,
|
14333 |
+
"learning_rate": 2.6149825783972124e-07,
|
14334 |
+
"loss": 1.1354,
|
14335 |
+
"step": 47300
|
14336 |
+
},
|
14337 |
+
{
|
14338 |
+
"epoch": 522.2958057395143,
|
14339 |
+
"learning_rate": 2.6118466898954704e-07,
|
14340 |
+
"loss": 1.1543,
|
14341 |
+
"step": 47320
|
14342 |
+
},
|
14343 |
+
{
|
14344 |
+
"epoch": 522.5165562913908,
|
14345 |
+
"learning_rate": 2.6087108013937274e-07,
|
14346 |
+
"loss": 1.1414,
|
14347 |
+
"step": 47340
|
14348 |
+
},
|
14349 |
+
{
|
14350 |
+
"epoch": 522.7373068432671,
|
14351 |
+
"learning_rate": 2.605574912891986e-07,
|
14352 |
+
"loss": 1.1354,
|
14353 |
+
"step": 47360
|
14354 |
+
},
|
14355 |
+
{
|
14356 |
+
"epoch": 522.9580573951434,
|
14357 |
+
"learning_rate": 2.6024390243902446e-07,
|
14358 |
+
"loss": 1.1349,
|
14359 |
+
"step": 47380
|
14360 |
+
},
|
14361 |
+
{
|
14362 |
+
"epoch": 523.1788079470199,
|
14363 |
+
"learning_rate": 2.5993031358885015e-07,
|
14364 |
+
"loss": 1.0901,
|
14365 |
+
"step": 47400
|
14366 |
+
},
|
14367 |
+
{
|
14368 |
+
"epoch": 523.3995584988962,
|
14369 |
+
"learning_rate": 2.5961672473867596e-07,
|
14370 |
+
"loss": 1.1694,
|
14371 |
+
"step": 47420
|
14372 |
+
},
|
14373 |
+
{
|
14374 |
+
"epoch": 523.6203090507727,
|
14375 |
+
"learning_rate": 2.593031358885017e-07,
|
14376 |
+
"loss": 1.1072,
|
14377 |
+
"step": 47440
|
14378 |
+
},
|
14379 |
+
{
|
14380 |
+
"epoch": 523.841059602649,
|
14381 |
+
"learning_rate": 2.589895470383275e-07,
|
14382 |
+
"loss": 1.1324,
|
14383 |
+
"step": 47460
|
14384 |
+
},
|
14385 |
+
{
|
14386 |
+
"epoch": 524.0618101545253,
|
14387 |
+
"learning_rate": 2.5867595818815316e-07,
|
14388 |
+
"loss": 1.1341,
|
14389 |
+
"step": 47480
|
14390 |
+
},
|
14391 |
+
{
|
14392 |
+
"epoch": 524.2825607064018,
|
14393 |
+
"learning_rate": 2.5836236933797907e-07,
|
14394 |
+
"loss": 1.1216,
|
14395 |
+
"step": 47500
|
14396 |
+
},
|
14397 |
+
{
|
14398 |
+
"epoch": 524.5033112582781,
|
14399 |
+
"learning_rate": 2.5804878048780493e-07,
|
14400 |
+
"loss": 1.1209,
|
14401 |
+
"step": 47520
|
14402 |
+
},
|
14403 |
+
{
|
14404 |
+
"epoch": 524.7240618101546,
|
14405 |
+
"learning_rate": 2.5773519163763063e-07,
|
14406 |
+
"loss": 1.141,
|
14407 |
+
"step": 47540
|
14408 |
+
},
|
14409 |
+
{
|
14410 |
+
"epoch": 524.9448123620309,
|
14411 |
+
"learning_rate": 2.5742160278745643e-07,
|
14412 |
+
"loss": 1.1072,
|
14413 |
+
"step": 47560
|
14414 |
+
},
|
14415 |
+
{
|
14416 |
+
"epoch": 525.1655629139073,
|
14417 |
+
"learning_rate": 2.571080139372822e-07,
|
14418 |
+
"loss": 1.1676,
|
14419 |
+
"step": 47580
|
14420 |
+
},
|
14421 |
+
{
|
14422 |
+
"epoch": 525.3863134657837,
|
14423 |
+
"learning_rate": 2.5679442508710804e-07,
|
14424 |
+
"loss": 1.1765,
|
14425 |
+
"step": 47600
|
14426 |
+
},
|
14427 |
+
{
|
14428 |
+
"epoch": 525.60706401766,
|
14429 |
+
"learning_rate": 2.5648083623693374e-07,
|
14430 |
+
"loss": 1.1005,
|
14431 |
+
"step": 47620
|
14432 |
+
},
|
14433 |
+
{
|
14434 |
+
"epoch": 525.8278145695364,
|
14435 |
+
"learning_rate": 2.5616724738675955e-07,
|
14436 |
+
"loss": 1.0937,
|
14437 |
+
"step": 47640
|
14438 |
+
},
|
14439 |
+
{
|
14440 |
+
"epoch": 526.0485651214128,
|
14441 |
+
"learning_rate": 2.558536585365854e-07,
|
14442 |
+
"loss": 1.0996,
|
14443 |
+
"step": 47660
|
14444 |
+
},
|
14445 |
+
{
|
14446 |
+
"epoch": 526.2693156732892,
|
14447 |
+
"learning_rate": 2.5554006968641116e-07,
|
14448 |
+
"loss": 1.1112,
|
14449 |
+
"step": 47680
|
14450 |
+
},
|
14451 |
+
{
|
14452 |
+
"epoch": 526.4900662251656,
|
14453 |
+
"learning_rate": 2.552264808362369e-07,
|
14454 |
+
"loss": 1.1295,
|
14455 |
+
"step": 47700
|
14456 |
+
},
|
14457 |
+
{
|
14458 |
+
"epoch": 526.7108167770419,
|
14459 |
+
"learning_rate": 2.5491289198606266e-07,
|
14460 |
+
"loss": 1.0978,
|
14461 |
+
"step": 47720
|
14462 |
+
},
|
14463 |
+
{
|
14464 |
+
"epoch": 526.9315673289183,
|
14465 |
+
"learning_rate": 2.545993031358885e-07,
|
14466 |
+
"loss": 1.0924,
|
14467 |
+
"step": 47740
|
14468 |
+
},
|
14469 |
+
{
|
14470 |
+
"epoch": 527.1523178807947,
|
14471 |
+
"learning_rate": 2.542857142857142e-07,
|
14472 |
+
"loss": 1.184,
|
14473 |
+
"step": 47760
|
14474 |
+
},
|
14475 |
+
{
|
14476 |
+
"epoch": 527.3730684326711,
|
14477 |
+
"learning_rate": 2.5397212543554e-07,
|
14478 |
+
"loss": 1.1071,
|
14479 |
+
"step": 47780
|
14480 |
+
},
|
14481 |
+
{
|
14482 |
+
"epoch": 527.5938189845475,
|
14483 |
+
"learning_rate": 2.536585365853659e-07,
|
14484 |
+
"loss": 1.1094,
|
14485 |
+
"step": 47800
|
14486 |
+
},
|
14487 |
+
{
|
14488 |
+
"epoch": 527.8145695364238,
|
14489 |
+
"learning_rate": 2.533449477351916e-07,
|
14490 |
+
"loss": 1.1137,
|
14491 |
+
"step": 47820
|
14492 |
+
},
|
14493 |
+
{
|
14494 |
+
"epoch": 528.0353200883002,
|
14495 |
+
"learning_rate": 2.5303135888501744e-07,
|
14496 |
+
"loss": 1.1011,
|
14497 |
+
"step": 47840
|
14498 |
+
},
|
14499 |
+
{
|
14500 |
+
"epoch": 528.2560706401766,
|
14501 |
+
"learning_rate": 2.5271777003484314e-07,
|
14502 |
+
"loss": 1.129,
|
14503 |
+
"step": 47860
|
14504 |
+
},
|
14505 |
+
{
|
14506 |
+
"epoch": 528.476821192053,
|
14507 |
+
"learning_rate": 2.52404181184669e-07,
|
14508 |
+
"loss": 1.1166,
|
14509 |
+
"step": 47880
|
14510 |
+
},
|
14511 |
+
{
|
14512 |
+
"epoch": 528.6975717439294,
|
14513 |
+
"learning_rate": 2.520905923344948e-07,
|
14514 |
+
"loss": 1.153,
|
14515 |
+
"step": 47900
|
14516 |
+
},
|
14517 |
+
{
|
14518 |
+
"epoch": 528.9183222958058,
|
14519 |
+
"learning_rate": 2.5177700348432055e-07,
|
14520 |
+
"loss": 1.1238,
|
14521 |
+
"step": 47920
|
14522 |
+
},
|
14523 |
+
{
|
14524 |
+
"epoch": 529.1390728476821,
|
14525 |
+
"learning_rate": 2.5146341463414636e-07,
|
14526 |
+
"loss": 1.1209,
|
14527 |
+
"step": 47940
|
14528 |
+
},
|
14529 |
+
{
|
14530 |
+
"epoch": 529.3598233995585,
|
14531 |
+
"learning_rate": 2.511498257839721e-07,
|
14532 |
+
"loss": 1.1242,
|
14533 |
+
"step": 47960
|
14534 |
+
},
|
14535 |
+
{
|
14536 |
+
"epoch": 529.5805739514349,
|
14537 |
+
"learning_rate": 2.508362369337979e-07,
|
14538 |
+
"loss": 1.1133,
|
14539 |
+
"step": 47980
|
14540 |
+
},
|
14541 |
+
{
|
14542 |
+
"epoch": 529.8013245033112,
|
14543 |
+
"learning_rate": 2.505226480836236e-07,
|
14544 |
+
"loss": 1.13,
|
14545 |
+
"step": 48000
|
14546 |
+
},
|
14547 |
+
{
|
14548 |
+
"epoch": 530.0220750551877,
|
14549 |
+
"learning_rate": 2.5020905923344947e-07,
|
14550 |
+
"loss": 1.1395,
|
14551 |
+
"step": 48020
|
14552 |
+
},
|
14553 |
+
{
|
14554 |
+
"epoch": 530.242825607064,
|
14555 |
+
"learning_rate": 2.4989547038327533e-07,
|
14556 |
+
"loss": 1.0835,
|
14557 |
+
"step": 48040
|
14558 |
+
},
|
14559 |
+
{
|
14560 |
+
"epoch": 530.4635761589404,
|
14561 |
+
"learning_rate": 2.4958188153310103e-07,
|
14562 |
+
"loss": 1.1185,
|
14563 |
+
"step": 48060
|
14564 |
+
},
|
14565 |
+
{
|
14566 |
+
"epoch": 530.6843267108168,
|
14567 |
+
"learning_rate": 2.4926829268292683e-07,
|
14568 |
+
"loss": 1.157,
|
14569 |
+
"step": 48080
|
14570 |
+
},
|
14571 |
+
{
|
14572 |
+
"epoch": 530.9050772626931,
|
14573 |
+
"learning_rate": 2.489547038327526e-07,
|
14574 |
+
"loss": 1.1264,
|
14575 |
+
"step": 48100
|
14576 |
+
},
|
14577 |
+
{
|
14578 |
+
"epoch": 531.1258278145696,
|
14579 |
+
"learning_rate": 2.4864111498257844e-07,
|
14580 |
+
"loss": 1.0726,
|
14581 |
+
"step": 48120
|
14582 |
+
},
|
14583 |
+
{
|
14584 |
+
"epoch": 531.3465783664459,
|
14585 |
+
"learning_rate": 2.4832752613240414e-07,
|
14586 |
+
"loss": 1.1548,
|
14587 |
+
"step": 48140
|
14588 |
+
},
|
14589 |
+
{
|
14590 |
+
"epoch": 531.5673289183223,
|
14591 |
+
"learning_rate": 2.4801393728222995e-07,
|
14592 |
+
"loss": 1.1144,
|
14593 |
+
"step": 48160
|
14594 |
+
},
|
14595 |
+
{
|
14596 |
+
"epoch": 531.7880794701987,
|
14597 |
+
"learning_rate": 2.477003484320558e-07,
|
14598 |
+
"loss": 1.1134,
|
14599 |
+
"step": 48180
|
14600 |
+
},
|
14601 |
+
{
|
14602 |
+
"epoch": 532.008830022075,
|
14603 |
+
"learning_rate": 2.473867595818815e-07,
|
14604 |
+
"loss": 1.1017,
|
14605 |
+
"step": 48200
|
14606 |
+
},
|
14607 |
+
{
|
14608 |
+
"epoch": 532.2295805739515,
|
14609 |
+
"learning_rate": 2.470731707317073e-07,
|
14610 |
+
"loss": 1.1188,
|
14611 |
+
"step": 48220
|
14612 |
+
},
|
14613 |
+
{
|
14614 |
+
"epoch": 532.4503311258278,
|
14615 |
+
"learning_rate": 2.4675958188153306e-07,
|
14616 |
+
"loss": 1.1032,
|
14617 |
+
"step": 48240
|
14618 |
+
},
|
14619 |
+
{
|
14620 |
+
"epoch": 532.6710816777043,
|
14621 |
+
"learning_rate": 2.464459930313589e-07,
|
14622 |
+
"loss": 1.1564,
|
14623 |
+
"step": 48260
|
14624 |
+
},
|
14625 |
+
{
|
14626 |
+
"epoch": 532.8918322295806,
|
14627 |
+
"learning_rate": 2.461324041811846e-07,
|
14628 |
+
"loss": 1.1129,
|
14629 |
+
"step": 48280
|
14630 |
+
},
|
14631 |
+
{
|
14632 |
+
"epoch": 533.1125827814569,
|
14633 |
+
"learning_rate": 2.458188153310104e-07,
|
14634 |
+
"loss": 1.0803,
|
14635 |
+
"step": 48300
|
14636 |
+
},
|
14637 |
+
{
|
14638 |
+
"epoch": 533.3333333333334,
|
14639 |
+
"learning_rate": 2.455052264808363e-07,
|
14640 |
+
"loss": 1.1,
|
14641 |
+
"step": 48320
|
14642 |
+
},
|
14643 |
+
{
|
14644 |
+
"epoch": 533.5540838852097,
|
14645 |
+
"learning_rate": 2.4519163763066203e-07,
|
14646 |
+
"loss": 1.1149,
|
14647 |
+
"step": 48340
|
14648 |
+
},
|
14649 |
+
{
|
14650 |
+
"epoch": 533.774834437086,
|
14651 |
+
"learning_rate": 2.4487804878048783e-07,
|
14652 |
+
"loss": 1.1505,
|
14653 |
+
"step": 48360
|
14654 |
+
},
|
14655 |
+
{
|
14656 |
+
"epoch": 533.9955849889625,
|
14657 |
+
"learning_rate": 2.4456445993031353e-07,
|
14658 |
+
"loss": 1.1273,
|
14659 |
+
"step": 48380
|
14660 |
+
},
|
14661 |
+
{
|
14662 |
+
"epoch": 534.2163355408388,
|
14663 |
+
"learning_rate": 2.442508710801394e-07,
|
14664 |
+
"loss": 1.1458,
|
14665 |
+
"step": 48400
|
14666 |
+
},
|
14667 |
+
{
|
14668 |
+
"epoch": 534.4370860927153,
|
14669 |
+
"learning_rate": 2.439372822299651e-07,
|
14670 |
+
"loss": 1.0934,
|
14671 |
+
"step": 48420
|
14672 |
+
},
|
14673 |
+
{
|
14674 |
+
"epoch": 534.6578366445916,
|
14675 |
+
"learning_rate": 2.436236933797909e-07,
|
14676 |
+
"loss": 1.0883,
|
14677 |
+
"step": 48440
|
14678 |
+
},
|
14679 |
+
{
|
14680 |
+
"epoch": 534.878587196468,
|
14681 |
+
"learning_rate": 2.4331010452961675e-07,
|
14682 |
+
"loss": 1.1175,
|
14683 |
+
"step": 48460
|
14684 |
+
},
|
14685 |
+
{
|
14686 |
+
"epoch": 535.0993377483444,
|
14687 |
+
"learning_rate": 2.429965156794425e-07,
|
14688 |
+
"loss": 1.1428,
|
14689 |
+
"step": 48480
|
14690 |
+
},
|
14691 |
+
{
|
14692 |
+
"epoch": 535.3200883002207,
|
14693 |
+
"learning_rate": 2.426829268292683e-07,
|
14694 |
+
"loss": 1.1469,
|
14695 |
+
"step": 48500
|
14696 |
+
},
|
14697 |
+
{
|
14698 |
+
"epoch": 535.5408388520972,
|
14699 |
+
"learning_rate": 2.42369337979094e-07,
|
14700 |
+
"loss": 1.0978,
|
14701 |
+
"step": 48520
|
14702 |
+
},
|
14703 |
+
{
|
14704 |
+
"epoch": 535.7615894039735,
|
14705 |
+
"learning_rate": 2.4205574912891987e-07,
|
14706 |
+
"loss": 1.1107,
|
14707 |
+
"step": 48540
|
14708 |
+
},
|
14709 |
+
{
|
14710 |
+
"epoch": 535.9823399558499,
|
14711 |
+
"learning_rate": 2.417421602787456e-07,
|
14712 |
+
"loss": 1.1489,
|
14713 |
+
"step": 48560
|
14714 |
+
},
|
14715 |
+
{
|
14716 |
+
"epoch": 536.2030905077263,
|
14717 |
+
"learning_rate": 2.414285714285714e-07,
|
14718 |
+
"loss": 1.1286,
|
14719 |
+
"step": 48580
|
14720 |
+
},
|
14721 |
+
{
|
14722 |
+
"epoch": 536.4238410596026,
|
14723 |
+
"learning_rate": 2.4111498257839723e-07,
|
14724 |
+
"loss": 1.132,
|
14725 |
+
"step": 48600
|
14726 |
+
},
|
14727 |
+
{
|
14728 |
+
"epoch": 536.644591611479,
|
14729 |
+
"learning_rate": 2.40801393728223e-07,
|
14730 |
+
"loss": 1.076,
|
14731 |
+
"step": 48620
|
14732 |
+
},
|
14733 |
+
{
|
14734 |
+
"epoch": 536.8653421633554,
|
14735 |
+
"learning_rate": 2.404878048780488e-07,
|
14736 |
+
"loss": 1.1494,
|
14737 |
+
"step": 48640
|
14738 |
+
},
|
14739 |
+
{
|
14740 |
+
"epoch": 537.0860927152318,
|
14741 |
+
"learning_rate": 2.401742160278745e-07,
|
14742 |
+
"loss": 1.1247,
|
14743 |
+
"step": 48660
|
14744 |
+
},
|
14745 |
+
{
|
14746 |
+
"epoch": 537.3068432671082,
|
14747 |
+
"learning_rate": 2.3986062717770034e-07,
|
14748 |
+
"loss": 1.1095,
|
14749 |
+
"step": 48680
|
14750 |
+
},
|
14751 |
+
{
|
14752 |
+
"epoch": 537.5275938189845,
|
14753 |
+
"learning_rate": 2.395470383275261e-07,
|
14754 |
+
"loss": 1.1362,
|
14755 |
+
"step": 48700
|
14756 |
+
},
|
14757 |
+
{
|
14758 |
+
"epoch": 537.7483443708609,
|
14759 |
+
"learning_rate": 2.392334494773519e-07,
|
14760 |
+
"loss": 1.1193,
|
14761 |
+
"step": 48720
|
14762 |
+
},
|
14763 |
+
{
|
14764 |
+
"epoch": 537.9690949227373,
|
14765 |
+
"learning_rate": 2.389198606271777e-07,
|
14766 |
+
"loss": 1.1645,
|
14767 |
+
"step": 48740
|
14768 |
+
},
|
14769 |
+
{
|
14770 |
+
"epoch": 538.1898454746137,
|
14771 |
+
"learning_rate": 2.3860627177700346e-07,
|
14772 |
+
"loss": 1.1336,
|
14773 |
+
"step": 48760
|
14774 |
+
},
|
14775 |
+
{
|
14776 |
+
"epoch": 538.4105960264901,
|
14777 |
+
"learning_rate": 2.382926829268293e-07,
|
14778 |
+
"loss": 1.1218,
|
14779 |
+
"step": 48780
|
14780 |
+
},
|
14781 |
+
{
|
14782 |
+
"epoch": 538.6313465783664,
|
14783 |
+
"learning_rate": 2.37979094076655e-07,
|
14784 |
+
"loss": 1.104,
|
14785 |
+
"step": 48800
|
14786 |
+
},
|
14787 |
+
{
|
14788 |
+
"epoch": 538.8520971302428,
|
14789 |
+
"learning_rate": 2.3766550522648084e-07,
|
14790 |
+
"loss": 1.1036,
|
14791 |
+
"step": 48820
|
14792 |
+
},
|
14793 |
+
{
|
14794 |
+
"epoch": 539.0728476821192,
|
14795 |
+
"learning_rate": 2.3735191637630657e-07,
|
14796 |
+
"loss": 1.1054,
|
14797 |
+
"step": 48840
|
14798 |
+
},
|
14799 |
+
{
|
14800 |
+
"epoch": 539.2935982339956,
|
14801 |
+
"learning_rate": 2.3703832752613237e-07,
|
14802 |
+
"loss": 1.1439,
|
14803 |
+
"step": 48860
|
14804 |
+
},
|
14805 |
+
{
|
14806 |
+
"epoch": 539.514348785872,
|
14807 |
+
"learning_rate": 2.367247386759582e-07,
|
14808 |
+
"loss": 1.08,
|
14809 |
+
"step": 48880
|
14810 |
+
},
|
14811 |
+
{
|
14812 |
+
"epoch": 539.7350993377484,
|
14813 |
+
"learning_rate": 2.364111498257839e-07,
|
14814 |
+
"loss": 1.1405,
|
14815 |
+
"step": 48900
|
14816 |
+
},
|
14817 |
+
{
|
14818 |
+
"epoch": 539.9558498896247,
|
14819 |
+
"learning_rate": 2.3609756097560976e-07,
|
14820 |
+
"loss": 1.0862,
|
14821 |
+
"step": 48920
|
14822 |
+
},
|
14823 |
+
{
|
14824 |
+
"epoch": 540.1766004415011,
|
14825 |
+
"learning_rate": 2.357839721254355e-07,
|
14826 |
+
"loss": 1.1428,
|
14827 |
+
"step": 48940
|
14828 |
+
},
|
14829 |
+
{
|
14830 |
+
"epoch": 540.3973509933775,
|
14831 |
+
"learning_rate": 2.3547038327526132e-07,
|
14832 |
+
"loss": 1.1379,
|
14833 |
+
"step": 48960
|
14834 |
+
},
|
14835 |
+
{
|
14836 |
+
"epoch": 540.6181015452538,
|
14837 |
+
"learning_rate": 2.3515679442508715e-07,
|
14838 |
+
"loss": 1.1155,
|
14839 |
+
"step": 48980
|
14840 |
+
},
|
14841 |
+
{
|
14842 |
+
"epoch": 540.8388520971303,
|
14843 |
+
"learning_rate": 2.3484320557491288e-07,
|
14844 |
+
"loss": 1.0901,
|
14845 |
+
"step": 49000
|
14846 |
+
},
|
14847 |
+
{
|
14848 |
+
"epoch": 541.0596026490066,
|
14849 |
+
"learning_rate": 2.345296167247387e-07,
|
14850 |
+
"loss": 1.1415,
|
14851 |
+
"step": 49020
|
14852 |
+
},
|
14853 |
+
{
|
14854 |
+
"epoch": 541.280353200883,
|
14855 |
+
"learning_rate": 2.3421602787456443e-07,
|
14856 |
+
"loss": 1.1008,
|
14857 |
+
"step": 49040
|
14858 |
+
},
|
14859 |
+
{
|
14860 |
+
"epoch": 541.5011037527594,
|
14861 |
+
"learning_rate": 2.3390243902439026e-07,
|
14862 |
+
"loss": 1.1034,
|
14863 |
+
"step": 49060
|
14864 |
+
},
|
14865 |
+
{
|
14866 |
+
"epoch": 541.7218543046357,
|
14867 |
+
"learning_rate": 2.33588850174216e-07,
|
14868 |
+
"loss": 1.0984,
|
14869 |
+
"step": 49080
|
14870 |
+
},
|
14871 |
+
{
|
14872 |
+
"epoch": 541.9426048565122,
|
14873 |
+
"learning_rate": 2.332752613240418e-07,
|
14874 |
+
"loss": 1.1269,
|
14875 |
+
"step": 49100
|
14876 |
+
},
|
14877 |
+
{
|
14878 |
+
"epoch": 542.1633554083885,
|
14879 |
+
"learning_rate": 2.3296167247386763e-07,
|
14880 |
+
"loss": 1.1275,
|
14881 |
+
"step": 49120
|
14882 |
+
},
|
14883 |
+
{
|
14884 |
+
"epoch": 542.384105960265,
|
14885 |
+
"learning_rate": 2.3264808362369335e-07,
|
14886 |
+
"loss": 1.0807,
|
14887 |
+
"step": 49140
|
14888 |
+
},
|
14889 |
+
{
|
14890 |
+
"epoch": 542.6048565121413,
|
14891 |
+
"learning_rate": 2.3233449477351918e-07,
|
14892 |
+
"loss": 1.1037,
|
14893 |
+
"step": 49160
|
14894 |
+
},
|
14895 |
+
{
|
14896 |
+
"epoch": 542.8256070640176,
|
14897 |
+
"learning_rate": 2.320209059233449e-07,
|
14898 |
+
"loss": 1.1319,
|
14899 |
+
"step": 49180
|
14900 |
+
},
|
14901 |
+
{
|
14902 |
+
"epoch": 543.0463576158941,
|
14903 |
+
"learning_rate": 2.3170731707317074e-07,
|
14904 |
+
"loss": 1.1208,
|
14905 |
+
"step": 49200
|
14906 |
+
},
|
14907 |
+
{
|
14908 |
+
"epoch": 543.2671081677704,
|
14909 |
+
"learning_rate": 2.3139372822299646e-07,
|
14910 |
+
"loss": 1.126,
|
14911 |
+
"step": 49220
|
14912 |
+
},
|
14913 |
+
{
|
14914 |
+
"epoch": 543.4878587196469,
|
14915 |
+
"learning_rate": 2.310801393728223e-07,
|
14916 |
+
"loss": 1.0764,
|
14917 |
+
"step": 49240
|
14918 |
+
},
|
14919 |
+
{
|
14920 |
+
"epoch": 543.7086092715232,
|
14921 |
+
"learning_rate": 2.3076655052264813e-07,
|
14922 |
+
"loss": 1.104,
|
14923 |
+
"step": 49260
|
14924 |
+
},
|
14925 |
+
{
|
14926 |
+
"epoch": 543.9293598233995,
|
14927 |
+
"learning_rate": 2.3045296167247385e-07,
|
14928 |
+
"loss": 1.1319,
|
14929 |
+
"step": 49280
|
14930 |
+
},
|
14931 |
+
{
|
14932 |
+
"epoch": 544.150110375276,
|
14933 |
+
"learning_rate": 2.3013937282229968e-07,
|
14934 |
+
"loss": 1.1423,
|
14935 |
+
"step": 49300
|
14936 |
+
},
|
14937 |
+
{
|
14938 |
+
"epoch": 544.3708609271523,
|
14939 |
+
"learning_rate": 2.298257839721254e-07,
|
14940 |
+
"loss": 1.1174,
|
14941 |
+
"step": 49320
|
14942 |
+
},
|
14943 |
+
{
|
14944 |
+
"epoch": 544.5916114790286,
|
14945 |
+
"learning_rate": 2.2951219512195121e-07,
|
14946 |
+
"loss": 1.0878,
|
14947 |
+
"step": 49340
|
14948 |
+
},
|
14949 |
+
{
|
14950 |
+
"epoch": 544.8123620309051,
|
14951 |
+
"learning_rate": 2.2919860627177694e-07,
|
14952 |
+
"loss": 1.1226,
|
14953 |
+
"step": 49360
|
14954 |
+
},
|
14955 |
+
{
|
14956 |
+
"epoch": 545.0331125827814,
|
14957 |
+
"learning_rate": 2.2888501742160277e-07,
|
14958 |
+
"loss": 1.1167,
|
14959 |
+
"step": 49380
|
14960 |
+
},
|
14961 |
+
{
|
14962 |
+
"epoch": 545.2538631346579,
|
14963 |
+
"learning_rate": 2.285714285714286e-07,
|
14964 |
+
"loss": 1.074,
|
14965 |
+
"step": 49400
|
14966 |
+
},
|
14967 |
+
{
|
14968 |
+
"epoch": 545.4746136865342,
|
14969 |
+
"learning_rate": 2.2825783972125435e-07,
|
14970 |
+
"loss": 1.1233,
|
14971 |
+
"step": 49420
|
14972 |
+
},
|
14973 |
+
{
|
14974 |
+
"epoch": 545.6953642384105,
|
14975 |
+
"learning_rate": 2.2794425087108016e-07,
|
14976 |
+
"loss": 1.1342,
|
14977 |
+
"step": 49440
|
14978 |
+
},
|
14979 |
+
{
|
14980 |
+
"epoch": 545.916114790287,
|
14981 |
+
"learning_rate": 2.2763066202090588e-07,
|
14982 |
+
"loss": 1.1092,
|
14983 |
+
"step": 49460
|
14984 |
+
},
|
14985 |
+
{
|
14986 |
+
"epoch": 546.1368653421633,
|
14987 |
+
"learning_rate": 2.2731707317073172e-07,
|
14988 |
+
"loss": 1.151,
|
14989 |
+
"step": 49480
|
14990 |
+
},
|
14991 |
+
{
|
14992 |
+
"epoch": 546.3576158940398,
|
14993 |
+
"learning_rate": 2.2700348432055744e-07,
|
14994 |
+
"loss": 1.1341,
|
14995 |
+
"step": 49500
|
14996 |
+
},
|
14997 |
+
{
|
14998 |
+
"epoch": 546.5783664459161,
|
14999 |
+
"learning_rate": 2.2668989547038327e-07,
|
15000 |
+
"loss": 1.1053,
|
15001 |
+
"step": 49520
|
15002 |
+
},
|
15003 |
+
{
|
15004 |
+
"epoch": 546.7991169977925,
|
15005 |
+
"learning_rate": 2.2637630662020908e-07,
|
15006 |
+
"loss": 1.1347,
|
15007 |
+
"step": 49540
|
15008 |
+
},
|
15009 |
+
{
|
15010 |
+
"epoch": 547.0198675496689,
|
15011 |
+
"learning_rate": 2.260627177700348e-07,
|
15012 |
+
"loss": 1.1326,
|
15013 |
+
"step": 49560
|
15014 |
+
},
|
15015 |
+
{
|
15016 |
+
"epoch": 547.2406181015452,
|
15017 |
+
"learning_rate": 2.2574912891986063e-07,
|
15018 |
+
"loss": 1.1167,
|
15019 |
+
"step": 49580
|
15020 |
+
},
|
15021 |
+
{
|
15022 |
+
"epoch": 547.4613686534217,
|
15023 |
+
"learning_rate": 2.2543554006968636e-07,
|
15024 |
+
"loss": 1.1193,
|
15025 |
+
"step": 49600
|
15026 |
+
},
|
15027 |
+
{
|
15028 |
+
"epoch": 547.682119205298,
|
15029 |
+
"learning_rate": 2.251219512195122e-07,
|
15030 |
+
"loss": 1.0693,
|
15031 |
+
"step": 49620
|
15032 |
+
},
|
15033 |
+
{
|
15034 |
+
"epoch": 547.9028697571744,
|
15035 |
+
"learning_rate": 2.2480836236933792e-07,
|
15036 |
+
"loss": 1.1476,
|
15037 |
+
"step": 49640
|
15038 |
+
},
|
15039 |
+
{
|
15040 |
+
"epoch": 548.1236203090508,
|
15041 |
+
"learning_rate": 2.2449477351916375e-07,
|
15042 |
+
"loss": 1.0626,
|
15043 |
+
"step": 49660
|
15044 |
+
},
|
15045 |
+
{
|
15046 |
+
"epoch": 548.3443708609271,
|
15047 |
+
"learning_rate": 2.2418118466898958e-07,
|
15048 |
+
"loss": 1.116,
|
15049 |
+
"step": 49680
|
15050 |
+
},
|
15051 |
+
{
|
15052 |
+
"epoch": 548.5651214128035,
|
15053 |
+
"learning_rate": 2.238675958188153e-07,
|
15054 |
+
"loss": 1.1424,
|
15055 |
+
"step": 49700
|
15056 |
+
},
|
15057 |
+
{
|
15058 |
+
"epoch": 548.7858719646799,
|
15059 |
+
"learning_rate": 2.2355400696864114e-07,
|
15060 |
+
"loss": 1.1374,
|
15061 |
+
"step": 49720
|
15062 |
+
},
|
15063 |
+
{
|
15064 |
+
"epoch": 549.0066225165563,
|
15065 |
+
"learning_rate": 2.2324041811846686e-07,
|
15066 |
+
"loss": 1.1343,
|
15067 |
+
"step": 49740
|
15068 |
+
},
|
15069 |
+
{
|
15070 |
+
"epoch": 549.2273730684327,
|
15071 |
+
"learning_rate": 2.229268292682927e-07,
|
15072 |
+
"loss": 1.1125,
|
15073 |
+
"step": 49760
|
15074 |
+
},
|
15075 |
+
{
|
15076 |
+
"epoch": 549.448123620309,
|
15077 |
+
"learning_rate": 2.2261324041811842e-07,
|
15078 |
+
"loss": 1.1081,
|
15079 |
+
"step": 49780
|
15080 |
+
},
|
15081 |
+
{
|
15082 |
+
"epoch": 549.6688741721854,
|
15083 |
+
"learning_rate": 2.2229965156794422e-07,
|
15084 |
+
"loss": 1.1289,
|
15085 |
+
"step": 49800
|
15086 |
+
},
|
15087 |
+
{
|
15088 |
+
"epoch": 549.8896247240618,
|
15089 |
+
"learning_rate": 2.2198606271777005e-07,
|
15090 |
+
"loss": 1.0951,
|
15091 |
+
"step": 49820
|
15092 |
+
},
|
15093 |
+
{
|
15094 |
+
"epoch": 550.1103752759382,
|
15095 |
+
"learning_rate": 2.2167247386759578e-07,
|
15096 |
+
"loss": 1.1194,
|
15097 |
+
"step": 49840
|
15098 |
+
},
|
15099 |
+
{
|
15100 |
+
"epoch": 550.3311258278146,
|
15101 |
+
"learning_rate": 2.213588850174216e-07,
|
15102 |
+
"loss": 1.0951,
|
15103 |
+
"step": 49860
|
15104 |
+
},
|
15105 |
+
{
|
15106 |
+
"epoch": 550.551876379691,
|
15107 |
+
"learning_rate": 2.2104529616724734e-07,
|
15108 |
+
"loss": 1.103,
|
15109 |
+
"step": 49880
|
15110 |
+
},
|
15111 |
+
{
|
15112 |
+
"epoch": 550.7726269315673,
|
15113 |
+
"learning_rate": 2.2073170731707317e-07,
|
15114 |
+
"loss": 1.1474,
|
15115 |
+
"step": 49900
|
15116 |
+
},
|
15117 |
+
{
|
15118 |
+
"epoch": 550.9933774834437,
|
15119 |
+
"learning_rate": 2.20418118466899e-07,
|
15120 |
+
"loss": 1.1227,
|
15121 |
+
"step": 49920
|
15122 |
+
},
|
15123 |
+
{
|
15124 |
+
"epoch": 551.2141280353201,
|
15125 |
+
"learning_rate": 2.2010452961672473e-07,
|
15126 |
+
"loss": 1.1563,
|
15127 |
+
"step": 49940
|
15128 |
+
},
|
15129 |
+
{
|
15130 |
+
"epoch": 551.4348785871965,
|
15131 |
+
"learning_rate": 2.1979094076655056e-07,
|
15132 |
+
"loss": 1.1112,
|
15133 |
+
"step": 49960
|
15134 |
+
},
|
15135 |
+
{
|
15136 |
+
"epoch": 551.6556291390729,
|
15137 |
+
"learning_rate": 2.1947735191637628e-07,
|
15138 |
+
"loss": 1.0947,
|
15139 |
+
"step": 49980
|
15140 |
+
},
|
15141 |
+
{
|
15142 |
+
"epoch": 551.8763796909492,
|
15143 |
+
"learning_rate": 2.1916376306620206e-07,
|
15144 |
+
"loss": 1.0799,
|
15145 |
+
"step": 50000
|
15146 |
+
},
|
15147 |
+
{
|
15148 |
+
"epoch": 552.0971302428256,
|
15149 |
+
"learning_rate": 2.1885017421602784e-07,
|
15150 |
+
"loss": 1.0955,
|
15151 |
+
"step": 50020
|
15152 |
+
},
|
15153 |
+
{
|
15154 |
+
"epoch": 552.317880794702,
|
15155 |
+
"learning_rate": 2.1853658536585364e-07,
|
15156 |
+
"loss": 1.1181,
|
15157 |
+
"step": 50040
|
15158 |
+
},
|
15159 |
+
{
|
15160 |
+
"epoch": 552.5386313465783,
|
15161 |
+
"learning_rate": 2.1822299651567948e-07,
|
15162 |
+
"loss": 1.1288,
|
15163 |
+
"step": 50060
|
15164 |
+
},
|
15165 |
+
{
|
15166 |
+
"epoch": 552.7593818984548,
|
15167 |
+
"learning_rate": 2.179094076655052e-07,
|
15168 |
+
"loss": 1.1118,
|
15169 |
+
"step": 50080
|
15170 |
+
},
|
15171 |
+
{
|
15172 |
+
"epoch": 552.9801324503311,
|
15173 |
+
"learning_rate": 2.1759581881533103e-07,
|
15174 |
+
"loss": 1.0934,
|
15175 |
+
"step": 50100
|
15176 |
+
},
|
15177 |
+
{
|
15178 |
+
"epoch": 553.2008830022075,
|
15179 |
+
"learning_rate": 2.1728222996515673e-07,
|
15180 |
+
"loss": 1.0997,
|
15181 |
+
"step": 50120
|
15182 |
+
},
|
15183 |
+
{
|
15184 |
+
"epoch": 553.4216335540839,
|
15185 |
+
"learning_rate": 2.169686411149826e-07,
|
15186 |
+
"loss": 1.1146,
|
15187 |
+
"step": 50140
|
15188 |
+
},
|
15189 |
+
{
|
15190 |
+
"epoch": 553.6423841059602,
|
15191 |
+
"learning_rate": 2.1665505226480831e-07,
|
15192 |
+
"loss": 1.1399,
|
15193 |
+
"step": 50160
|
15194 |
+
},
|
15195 |
+
{
|
15196 |
+
"epoch": 553.8631346578367,
|
15197 |
+
"learning_rate": 2.1634146341463415e-07,
|
15198 |
+
"loss": 1.105,
|
15199 |
+
"step": 50180
|
15200 |
+
},
|
15201 |
+
{
|
15202 |
+
"epoch": 554.083885209713,
|
15203 |
+
"learning_rate": 2.1602787456445998e-07,
|
15204 |
+
"loss": 1.1155,
|
15205 |
+
"step": 50200
|
15206 |
+
},
|
15207 |
+
{
|
15208 |
+
"epoch": 554.3046357615895,
|
15209 |
+
"learning_rate": 2.1571428571428568e-07,
|
15210 |
+
"loss": 1.1422,
|
15211 |
+
"step": 50220
|
15212 |
+
},
|
15213 |
+
{
|
15214 |
+
"epoch": 554.5253863134658,
|
15215 |
+
"learning_rate": 2.1540069686411153e-07,
|
15216 |
+
"loss": 1.0936,
|
15217 |
+
"step": 50240
|
15218 |
+
},
|
15219 |
+
{
|
15220 |
+
"epoch": 554.7461368653421,
|
15221 |
+
"learning_rate": 2.1508710801393723e-07,
|
15222 |
+
"loss": 1.103,
|
15223 |
+
"step": 50260
|
15224 |
+
},
|
15225 |
+
{
|
15226 |
+
"epoch": 554.9668874172186,
|
15227 |
+
"learning_rate": 2.147735191637631e-07,
|
15228 |
+
"loss": 1.1244,
|
15229 |
+
"step": 50280
|
15230 |
+
},
|
15231 |
+
{
|
15232 |
+
"epoch": 555.1876379690949,
|
15233 |
+
"learning_rate": 2.144599303135888e-07,
|
15234 |
+
"loss": 1.119,
|
15235 |
+
"step": 50300
|
15236 |
+
},
|
15237 |
+
{
|
15238 |
+
"epoch": 555.4083885209714,
|
15239 |
+
"learning_rate": 2.1414634146341462e-07,
|
15240 |
+
"loss": 1.1021,
|
15241 |
+
"step": 50320
|
15242 |
+
},
|
15243 |
+
{
|
15244 |
+
"epoch": 555.6291390728477,
|
15245 |
+
"learning_rate": 2.1383275261324045e-07,
|
15246 |
+
"loss": 1.1232,
|
15247 |
+
"step": 50340
|
15248 |
+
},
|
15249 |
+
{
|
15250 |
+
"epoch": 555.849889624724,
|
15251 |
+
"learning_rate": 2.1351916376306618e-07,
|
15252 |
+
"loss": 1.0915,
|
15253 |
+
"step": 50360
|
15254 |
+
},
|
15255 |
+
{
|
15256 |
+
"epoch": 556.0706401766005,
|
15257 |
+
"learning_rate": 2.13205574912892e-07,
|
15258 |
+
"loss": 1.141,
|
15259 |
+
"step": 50380
|
15260 |
+
},
|
15261 |
+
{
|
15262 |
+
"epoch": 556.2913907284768,
|
15263 |
+
"learning_rate": 2.1289198606271773e-07,
|
15264 |
+
"loss": 1.1132,
|
15265 |
+
"step": 50400
|
15266 |
+
},
|
15267 |
+
{
|
15268 |
+
"epoch": 556.2913907284768,
|
15269 |
+
"eval_bleu": 48.9071,
|
15270 |
+
"eval_gen_len": 8.3667,
|
15271 |
+
"eval_loss": 1.9735721349716187,
|
15272 |
+
"eval_runtime": 3.3922,
|
15273 |
+
"eval_samples_per_second": 8.844,
|
15274 |
+
"eval_steps_per_second": 1.769,
|
15275 |
+
"step": 50400
|
15276 |
}
|
15277 |
],
|
15278 |
"logging_steps": 20,
|
|
|
15292 |
"attributes": {}
|
15293 |
}
|
15294 |
},
|
15295 |
+
"total_flos": 1.49160549425578e+17,
|
15296 |
"train_batch_size": 5,
|
15297 |
"trial_name": null,
|
15298 |
"trial_params": null
|