Amirmarshal
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -3,19 +3,26 @@ library_name: transformers
|
|
3 |
tags:
|
4 |
- trl
|
5 |
- dpo
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
---
|
7 |
|
8 |
# Model Card for Model ID
|
9 |
|
10 |
-
<!-- Provide a quick summary of what the model is/does. -->
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
## Model Details
|
|
|
|
|
15 |
|
16 |
### Model Description
|
17 |
|
18 |
-
|
19 |
|
20 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
21 |
|
|
|
3 |
tags:
|
4 |
- trl
|
5 |
- dpo
|
6 |
+
- text-generation-inference
|
7 |
+
- poem-generation-inference
|
8 |
+
datasets:
|
9 |
+
- DataAnalyticsLab/Persian-Poems
|
10 |
+
language:
|
11 |
+
- fa
|
12 |
+
base_model:
|
13 |
+
- openai-community/gpt2
|
14 |
+
pipeline_tag: text-generation
|
15 |
---
|
16 |
|
17 |
# Model Card for Model ID
|
18 |
|
|
|
|
|
|
|
|
|
19 |
## Model Details
|
20 |
+
- The model was trained using the new DPO training method using the trl.DPOTrainer and trl.DPOConfig packages. For this task, we created a dataset using the original dataset comprising positive and negative examples and by feeding them to the DPOTrainer, the model was trained. In this version, we used poems of length 1,2 and 3 Beyt shuffled randomly to train the model. The model was trained for one epoch.
|
21 |
+
|
22 |
|
23 |
### Model Description
|
24 |
|
25 |
+
<- The model was trained using the new DPO training method using the trl.DPOTrainer and trl.DPOConfig packages. For this task, we created a dataset using the original dataset comprising positive and negative examples and by feeding them to the DPOTrainer, the model was trained. In this version, we used poems of length 1,2 and 3 Beyt shuffled randomly to train the model. The model was trained for one epoch.->
|
26 |
|
27 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
28 |
|