Update README.md
Browse files
README.md
CHANGED
@@ -24,6 +24,9 @@ model-index:
|
|
24 |
- name: Bleu
|
25 |
type: bleu
|
26 |
value: 49.64800786424299
|
|
|
|
|
|
|
27 |
---
|
28 |
|
29 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -31,44 +34,85 @@ should probably proofread and complete it, then remove this comment. -->
|
|
31 |
|
32 |
# marian-finetuned-kde4-en-to-fr
|
33 |
|
34 |
-
This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
|
35 |
-
It achieves the following results on the evaluation set:
|
36 |
-
- Loss: 1.0371
|
37 |
-
- Bleu: 49.6480
|
38 |
|
39 |
-
## Model
|
40 |
|
41 |
-
|
42 |
|
43 |
-
|
|
|
|
|
|
|
44 |
|
45 |
-
|
46 |
|
47 |
-
|
|
|
|
|
|
|
48 |
|
49 |
-
|
|
|
|
|
|
|
50 |
|
51 |
-
## Training
|
52 |
|
53 |
-
###
|
|
|
|
|
54 |
|
55 |
-
|
56 |
-
-
|
57 |
-
-
|
58 |
-
-
|
59 |
-
- seed: 42
|
60 |
-
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
61 |
-
- lr_scheduler_type: linear
|
62 |
-
- num_epochs: 3
|
63 |
-
- mixed_precision_training: Native AMP
|
64 |
|
65 |
-
|
66 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
67 |
|
|
|
68 |
|
69 |
-
|
|
|
|
|
|
|
|
|
70 |
|
71 |
-
-
|
72 |
-
-
|
73 |
-
|
74 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
- name: Bleu
|
25 |
type: bleu
|
26 |
value: 49.64800786424299
|
27 |
+
language:
|
28 |
+
- en
|
29 |
+
pipeline_tag: translation
|
30 |
---
|
31 |
|
32 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
34 |
|
35 |
# marian-finetuned-kde4-en-to-fr
|
36 |
|
37 |
+
This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr), specifically tailored for English-to-French translation tasks. It was trained on the `kde4` dataset, which consists of parallel texts from the KDE project, making it highly specialized in technical and software documentation translation.
|
|
|
|
|
|
|
38 |
|
39 |
+
## Model Description
|
40 |
|
41 |
+
MarianMT is a neural machine translation model based on the Marian framework, designed for rapid training and inference. This particular model, `marian-finetuned-kde4-en-to-fr`, leverages the capabilities of the pre-trained `opus-mt-en-fr` model and further enhances its performance on the KDE4 dataset, which is focused on the translation of software and technical documentation.
|
42 |
|
43 |
+
### Key Features:
|
44 |
+
- **Base Model**: [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr), a robust English-to-French translation model.
|
45 |
+
- **Fine-Tuned For**: Specialized translation of technical and software documentation.
|
46 |
+
- **Architecture**: Transformer-based MarianMT, known for efficient and scalable translation capabilities.
|
47 |
|
48 |
+
## Intended Uses & Limitations
|
49 |
|
50 |
+
### Intended Uses:
|
51 |
+
- **Technical Documentation Translation**: Translate software documentation, user manuals, and other technical texts from English to French.
|
52 |
+
- **Software Localization**: Aid in the localization process by translating software interfaces and messages.
|
53 |
+
- **General English-to-French Translation**: While specialized for technical texts, it can also handle general translation tasks.
|
54 |
|
55 |
+
### Limitations:
|
56 |
+
- **Domain-Specific Performance**: The model's fine-tuning on technical texts means it excels in those areas but may not perform as well with colloquial language or literary texts.
|
57 |
+
- **Biases**: The model may reflect biases present in the training data, particularly around technical jargon and software terminology.
|
58 |
+
- **Limited Language Support**: This model is designed specifically for English-to-French translation. It is not suitable for other language pairs without further fine-tuning.
|
59 |
|
60 |
+
## Training and Evaluation Data
|
61 |
|
62 |
+
### Dataset:
|
63 |
+
- **Training Data**: The `kde4` dataset, which includes parallel English-French sentences derived from the KDE project. This dataset primarily consists of translations relevant to software documentation, user interfaces, and related technical content.
|
64 |
+
- **Evaluation Data**: A subset of the `kde4` dataset was used for evaluation to ensure the model's effectiveness in the same domain it was trained on.
|
65 |
|
66 |
+
### Data Characteristics:
|
67 |
+
- **Domain**: Technical documentation, software localization.
|
68 |
+
- **Language**: Primarily English and French.
|
69 |
+
- **Dataset Size**: Contains thousands of sentence pairs, providing a robust dataset for fine-tuning in the technical domain.
|
|
|
|
|
|
|
|
|
|
|
70 |
|
71 |
+
## Training Procedure
|
72 |
|
73 |
+
### Training Hyperparameters:
|
74 |
+
- **Learning Rate**: 2e-05
|
75 |
+
- **Train Batch Size**: 32
|
76 |
+
- **Eval Batch Size**: 64
|
77 |
+
- **Seed**: 42
|
78 |
+
- **Optimizer**: Adam with betas=(0.9, 0.999) and epsilon=1e-08
|
79 |
+
- **Learning Rate Scheduler Type**: Linear
|
80 |
+
- **Number of Epochs**: 3
|
81 |
+
- **Mixed Precision Training**: Native AMP (Automatic Mixed Precision) to optimize training time and memory usage.
|
82 |
|
83 |
+
### Training Results:
|
84 |
|
85 |
+
| Metric | Value |
|
86 |
+
|---------------|----------|
|
87 |
+
| Training Loss | 1.0371 |
|
88 |
+
| Evaluation Loss | 1.0371 |
|
89 |
+
| BLEU Score | 49.6480 |
|
90 |
|
91 |
+
- **Final Evaluation Loss**: 1.0371
|
92 |
+
- **BLEU Score**: 49.6480, indicating a high level of accuracy in translation.
|
93 |
+
|
94 |
+
### Framework Versions
|
95 |
+
- **Transformers**: 4.42.4
|
96 |
+
- **PyTorch**: 2.3.1+cu121
|
97 |
+
- **Datasets**: 2.21.0
|
98 |
+
- **Tokenizers**: 0.19.1
|
99 |
+
|
100 |
+
## Usage
|
101 |
+
|
102 |
+
You can use this model in a Hugging Face pipeline for translation tasks:
|
103 |
+
|
104 |
+
```python
|
105 |
+
from transformers import pipeline
|
106 |
+
|
107 |
+
model_checkpoint = "ashaduzzaman/marian-finetuned-kde4-en-to-fr"
|
108 |
+
translator = pipeline("translation", model=model_checkpoint)
|
109 |
+
|
110 |
+
# Example usage
|
111 |
+
input_text = "The user manual provides detailed instructions on how to use the software."
|
112 |
+
translation = translator(input_text)
|
113 |
+
print(translation)
|
114 |
+
```
|
115 |
+
|
116 |
+
## Acknowledgments
|
117 |
+
|
118 |
+
This model was developed using the [Hugging Face Transformers](https://huggingface.co/transformers) library and fine-tuned using the `kde4` dataset. Special thanks to the contributors of the KDE project for providing a rich source of multilingual technical content.
|