Fine-tuned mBART Model for English to Urdu Translation
This repository contains a fine-tuned mBART model for English to Urdu translation. The model has been trained on a custom dataset and evaluated on test data.
Model Information
- Model Name:
abdulwaheed1/english-to-urdu-translation-mbart
- Base Model:
facebook/mbart-large-50
- Tokenizer:
facebook/mbart-large-50
- Source Language: English (
en
) - Target Language: Urdu (
ur
)
Usage
python
from transformers import MBart50TokenizerFast, MBartForConditionalGeneration
# Load the fine-tuned model
model_name = "abdulwaheed1/english-to-urdu-translation-mbart"
tokenizer = MBart50TokenizerFast.from_pretrained(model_name, src_lang="en_XX", tgt_lang="ur_PK")
model = MBartForConditionalGeneration.from_pretrained(model_name)
Evaluation
The model has been evaluated on a test dataset, and the following metrics were obtained:
- BLEU Score: 35.87
- Generation Length: 42.56
- Meteor Score: 0.60
Training Details
The model was trained using the transformers
library with the following configuration:
- Training Loss: 1.5697
- Validation Loss: 1.1256
Dataset
The model was fine-tuned on a custom English-Urdu translation dataset. If you wish to use the same dataset, you can find the preprocessing script and dataset files in the data
directory.
Acknowledgments
The fine-tuning process and code were inspired by the Hugging Face Transformers library.
- Downloads last month
- 106
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.