metadata
library_name: transformers
language:
- en
- ur
metrics:
- bleu
##Fine-tuned mBART Model for English to Urdu Translation This repository contains a fine-tuned mBART model for English to Urdu translation. The model has been trained on a custom dataset and evaluated on test data.
Model Information
- Model Name:
abdulwaheed63/mbart_en_ur_finetuned
- Base Model:
facebook/mbart-large-50
- Tokenizer:
facebook/mbart-large-50
- Source Language: English (
en
) - Target Language: Urdu (
ur
)
Usage
```python
from transformers import MBart50TokenizerFast, MBartForConditionalGeneration
# Load the fine-tuned model
model_name = "abdulwaheed63/mbart_en_ur_finetuned"
tokenizer = MBart50TokenizerFast.from_pretrained(model_name, src_lang="en_XX", tgt_lang="ur_PK")
model = MBartForConditionalGeneration.from_pretrained(model_name)
Evaluation
The model has been evaluated on a test dataset, and the following metrics were obtained:
- BLEU Score: 35.87
- Generation Length: 42.56
- Meteor Score: 0.60
Training Details
The model was trained using the transformers
library with the following configuration:
- Training Loss: 1.5697
- Validation Loss: 1.1256
Dataset
The model was fine-tuned on a custom English-Urdu translation dataset. If you wish to use the same dataset, you can find the preprocessing script and dataset files in the data
directory.
Acknowledgments
The fine-tuning process and code were inspired by the Hugging Face Transformers library.