--- library_name: transformers tags: - generated_from_trainer metrics: - bleu model-index: - name: mbart-large-50-many-to-many-mmt results: [] --- # mbart-large-50-many-to-many-mmt This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.3912 - Bleu: 12.0674 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 2.1894 | 0.8 | 1000 | 2.3682 | 10.7429 | | 1.7455 | 1.6 | 2000 | 2.3495 | 12.2553 | | 1.3893 | 2.4 | 3000 | 2.3956 | 11.2554 | ### Framework versions - Transformers 4.45.1 - Pytorch 2.4.0 - Datasets 3.0.1 - Tokenizers 0.20.0