LLaMA-3.1-8B-Infinity3M-Kobo / all_results.json
g4rg's picture
Upload folder using huggingface_hub
2fe264d verified
raw
history blame contribute delete
402 Bytes
{
"epoch": 2.9974825174825175,
"eval_loss": 0.7186228632926941,
"eval_runtime": 247.0208,
"eval_samples_per_second": 2.364,
"eval_steps_per_second": 0.296,
"num_input_tokens_seen": 2809135104,
"total_flos": 4487006021222400.0,
"train_loss": 0.719823204545487,
"train_runtime": 157993.0137,
"train_samples_per_second": 0.543,
"train_steps_per_second": 0.017
}