Fine-tuned mBART Model for Urdu to English Translation
This repository contains a fine-tuned mBART model for Urdu to English translation. The model has been trained on a custom dataset and evaluated on test data.
Model Information
- Model Name:
abdulwaheed1/urdu_to_english_translation_mbart
- Base Model:
facebook/mbart-large-50
- Tokenizer:
facebook/mbart-large-50
- Source Language: Urdu (
ur
) - Target Language: English (
en
)
Usage
python
from transformers import MBart50TokenizerFast, MBartForConditionalGeneration
# Load the fine-tuned model
model_name = "abdulwaheed1/urdu_to_english_translation_mbart"
tokenizer = MBart50TokenizerFast.from_pretrained(model_name, src_lang="ur_PK", tgt_lang="en_XX")
model = MBartForConditionalGeneration.from_pretrained(model_name)
Evaluation
The model has been evaluated on a test dataset, and the following metrics were obtained:
- BLEU Score: 30.95
- Generation Length: 35.90
- Meteor Score: 0.58
Training Details
The model was trained using the transformers
library with the following configuration:
- Training Loss: 1.450300
- Validation Loss: 1.281330
Dataset
The model was fine-tuned on a custom Urdu-English translation dataset. If you wish to use the same dataset, you can find the preprocessing script and dataset files in the data
directory.
Acknowledgments
The fine-tuning process and code were inspired by the Hugging Face Transformers library.
- Downloads last month
- 208
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.