ashaduzzaman commited on
Commit
60c8c59
1 Parent(s): 017fd6c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +71 -27
README.md CHANGED
@@ -24,6 +24,9 @@ model-index:
24
  - name: Bleu
25
  type: bleu
26
  value: 49.64800786424299
 
 
 
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -31,44 +34,85 @@ should probably proofread and complete it, then remove this comment. -->
31
 
32
  # marian-finetuned-kde4-en-to-fr
33
 
34
- This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
35
- It achieves the following results on the evaluation set:
36
- - Loss: 1.0371
37
- - Bleu: 49.6480
38
 
39
- ## Model description
40
 
41
- More information needed
42
 
43
- ## Intended uses & limitations
 
 
 
44
 
45
- More information needed
46
 
47
- ## Training and evaluation data
 
 
 
48
 
49
- More information needed
 
 
 
50
 
51
- ## Training procedure
52
 
53
- ### Training hyperparameters
 
 
54
 
55
- The following hyperparameters were used during training:
56
- - learning_rate: 2e-05
57
- - train_batch_size: 32
58
- - eval_batch_size: 64
59
- - seed: 42
60
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
- - lr_scheduler_type: linear
62
- - num_epochs: 3
63
- - mixed_precision_training: Native AMP
64
 
65
- ### Training results
66
 
 
 
 
 
 
 
 
 
 
67
 
 
68
 
69
- ### Framework versions
 
 
 
 
70
 
71
- - Transformers 4.42.4
72
- - Pytorch 2.3.1+cu121
73
- - Datasets 2.21.0
74
- - Tokenizers 0.19.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  - name: Bleu
25
  type: bleu
26
  value: 49.64800786424299
27
+ language:
28
+ - en
29
+ pipeline_tag: translation
30
  ---
31
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
34
 
35
  # marian-finetuned-kde4-en-to-fr
36
 
37
+ This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr), specifically tailored for English-to-French translation tasks. It was trained on the `kde4` dataset, which consists of parallel texts from the KDE project, making it highly specialized in technical and software documentation translation.
 
 
 
38
 
39
+ ## Model Description
40
 
41
+ MarianMT is a neural machine translation model based on the Marian framework, designed for rapid training and inference. This particular model, `marian-finetuned-kde4-en-to-fr`, leverages the capabilities of the pre-trained `opus-mt-en-fr` model and further enhances its performance on the KDE4 dataset, which is focused on the translation of software and technical documentation.
42
 
43
+ ### Key Features:
44
+ - **Base Model**: [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr), a robust English-to-French translation model.
45
+ - **Fine-Tuned For**: Specialized translation of technical and software documentation.
46
+ - **Architecture**: Transformer-based MarianMT, known for efficient and scalable translation capabilities.
47
 
48
+ ## Intended Uses & Limitations
49
 
50
+ ### Intended Uses:
51
+ - **Technical Documentation Translation**: Translate software documentation, user manuals, and other technical texts from English to French.
52
+ - **Software Localization**: Aid in the localization process by translating software interfaces and messages.
53
+ - **General English-to-French Translation**: While specialized for technical texts, it can also handle general translation tasks.
54
 
55
+ ### Limitations:
56
+ - **Domain-Specific Performance**: The model's fine-tuning on technical texts means it excels in those areas but may not perform as well with colloquial language or literary texts.
57
+ - **Biases**: The model may reflect biases present in the training data, particularly around technical jargon and software terminology.
58
+ - **Limited Language Support**: This model is designed specifically for English-to-French translation. It is not suitable for other language pairs without further fine-tuning.
59
 
60
+ ## Training and Evaluation Data
61
 
62
+ ### Dataset:
63
+ - **Training Data**: The `kde4` dataset, which includes parallel English-French sentences derived from the KDE project. This dataset primarily consists of translations relevant to software documentation, user interfaces, and related technical content.
64
+ - **Evaluation Data**: A subset of the `kde4` dataset was used for evaluation to ensure the model's effectiveness in the same domain it was trained on.
65
 
66
+ ### Data Characteristics:
67
+ - **Domain**: Technical documentation, software localization.
68
+ - **Language**: Primarily English and French.
69
+ - **Dataset Size**: Contains thousands of sentence pairs, providing a robust dataset for fine-tuning in the technical domain.
 
 
 
 
 
70
 
71
+ ## Training Procedure
72
 
73
+ ### Training Hyperparameters:
74
+ - **Learning Rate**: 2e-05
75
+ - **Train Batch Size**: 32
76
+ - **Eval Batch Size**: 64
77
+ - **Seed**: 42
78
+ - **Optimizer**: Adam with betas=(0.9, 0.999) and epsilon=1e-08
79
+ - **Learning Rate Scheduler Type**: Linear
80
+ - **Number of Epochs**: 3
81
+ - **Mixed Precision Training**: Native AMP (Automatic Mixed Precision) to optimize training time and memory usage.
82
 
83
+ ### Training Results:
84
 
85
+ | Metric | Value |
86
+ |---------------|----------|
87
+ | Training Loss | 1.0371 |
88
+ | Evaluation Loss | 1.0371 |
89
+ | BLEU Score | 49.6480 |
90
 
91
+ - **Final Evaluation Loss**: 1.0371
92
+ - **BLEU Score**: 49.6480, indicating a high level of accuracy in translation.
93
+
94
+ ### Framework Versions
95
+ - **Transformers**: 4.42.4
96
+ - **PyTorch**: 2.3.1+cu121
97
+ - **Datasets**: 2.21.0
98
+ - **Tokenizers**: 0.19.1
99
+
100
+ ## Usage
101
+
102
+ You can use this model in a Hugging Face pipeline for translation tasks:
103
+
104
+ ```python
105
+ from transformers import pipeline
106
+
107
+ model_checkpoint = "ashaduzzaman/marian-finetuned-kde4-en-to-fr"
108
+ translator = pipeline("translation", model=model_checkpoint)
109
+
110
+ # Example usage
111
+ input_text = "The user manual provides detailed instructions on how to use the software."
112
+ translation = translator(input_text)
113
+ print(translation)
114
+ ```
115
+
116
+ ## Acknowledgments
117
+
118
+ This model was developed using the [Hugging Face Transformers](https://huggingface.co/transformers) library and fine-tuned using the `kde4` dataset. Special thanks to the contributors of the KDE project for providing a rich source of multilingual technical content.