abdiharyadi commited on
Commit
25ffa48
1 Parent(s): 59d9ffe

Model save

Browse files
README.md ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - generated_from_trainer
4
+ datasets:
5
+ - data
6
+ metrics:
7
+ - bleu
8
+ model-index:
9
+ - name: mbart-en-id-smaller-indo-amr-generation-translated-nafkhan
10
+ results:
11
+ - task:
12
+ name: Sequence-to-sequence Language Modeling
13
+ type: text2text-generation
14
+ dataset:
15
+ name: data
16
+ type: data
17
+ config: default
18
+ split: validation
19
+ args: default
20
+ metrics:
21
+ - name: Bleu
22
+ type: bleu
23
+ value: 50.4231
24
+ ---
25
+
26
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
27
+ should probably proofread and complete it, then remove this comment. -->
28
+
29
+ # mbart-en-id-smaller-indo-amr-generation-translated-nafkhan
30
+
31
+ This model was trained from scratch on the data dataset.
32
+ It achieves the following results on the evaluation set:
33
+ - Loss: 1.9816
34
+ - Bleu: 50.4231
35
+ - Gen Len: 8.2667
36
+
37
+ ## Model description
38
+
39
+ More information needed
40
+
41
+ ## Intended uses & limitations
42
+
43
+ More information needed
44
+
45
+ ## Training and evaluation data
46
+
47
+ More information needed
48
+
49
+ ## Training procedure
50
+
51
+ ### Training hyperparameters
52
+
53
+ The following hyperparameters were used during training:
54
+ - learning_rate: 1e-06
55
+ - train_batch_size: 5
56
+ - eval_batch_size: 5
57
+ - seed: 42
58
+ - gradient_accumulation_steps: 5
59
+ - total_train_batch_size: 25
60
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
+ - lr_scheduler_type: polynomial
62
+ - lr_scheduler_warmup_steps: 200
63
+ - num_epochs: 640.0
64
+ - label_smoothing_factor: 0.1
65
+
66
+ ### Training results
67
+
68
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
69
+ |:-------------:|:--------:|:-----:|:---------------:|:-------:|:-------:|
70
+ | 1.5364 | 39.7351 | 3600 | 2.0474 | 41.1359 | 9.5667 |
71
+ | 1.4794 | 79.4702 | 7200 | 2.0256 | 44.4125 | 9.0667 |
72
+ | 1.3621 | 119.2053 | 10800 | 2.0037 | 49.121 | 8.8 |
73
+ | 1.3278 | 158.9404 | 14400 | 1.9857 | 52.1498 | 8.5 |
74
+ | 1.2906 | 198.6755 | 18000 | 2.0048 | 48.3044 | 8.5333 |
75
+ | 1.2272 | 238.4106 | 21600 | 2.0017 | 47.3367 | 8.6 |
76
+ | 1.2339 | 278.1457 | 25200 | 1.9705 | 46.5655 | 8.5667 |
77
+ | 1.194 | 317.8808 | 28800 | 1.9818 | 51.4066 | 8.5 |
78
+ | 1.1416 | 357.6159 | 32400 | 1.9699 | 45.7022 | 8.4333 |
79
+ | 1.1437 | 397.3510 | 36000 | 1.9692 | 46.8726 | 8.2 |
80
+ | 1.156 | 437.0861 | 39600 | 1.9549 | 48.7386 | 8.4333 |
81
+ | 1.1355 | 476.8212 | 43200 | 1.9726 | 48.3929 | 8.5667 |
82
+ | 1.1246 | 516.5563 | 46800 | 1.9701 | 47.8897 | 8.4667 |
83
+ | 1.1132 | 556.2914 | 50400 | 1.9736 | 48.9071 | 8.3667 |
84
+ | 1.0843 | 596.0265 | 54000 | 1.9765 | 49.9897 | 8.4333 |
85
+ | 1.1211 | 635.7616 | 57600 | 1.9816 | 50.4231 | 8.2667 |
86
+
87
+
88
+ ### Framework versions
89
+
90
+ - Transformers 4.44.0
91
+ - Pytorch 2.4.0
92
+ - Datasets 2.20.0
93
+ - Tokenizers 0.19.1
generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "decoder_start_token_id": 2,
5
+ "early_stopping": true,
6
+ "eos_token_id": 2,
7
+ "forced_eos_token_id": 2,
8
+ "max_length": 200,
9
+ "num_beams": 5,
10
+ "pad_token_id": 1,
11
+ "transformers_version": "4.44.0"
12
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7697c1f1adb8e015d5c3f3c0b2864dbad3627bcaec5427bbce6ed72735228e4a
3
  size 1575259780
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a9d10b584cec8273eca5d86cc6016f2c25559a476d52f85256fa198af4ac99a
3
  size 1575259780
run-2024-10-02T06:53:25+00:00.log CHANGED
@@ -3167,3 +3167,10 @@ Non-default generation parameters: {'max_length': 200, 'early_stopping': True, '
3167
 
3168
 
3169
  [WARNING|configuration_utils.py:448] 2024-10-02 17:02:24,528 >> Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.
3170
  Non-default generation parameters: {'max_length': 200, 'early_stopping': True, 'num_beams': 5, 'forced_eos_token_id': 2}
 
 
3171
 
 
 
 
 
 
 
3167
 
3168
 
3169
  [WARNING|configuration_utils.py:448] 2024-10-02 17:02:24,528 >> Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.
3170
  Non-default generation parameters: {'max_length': 200, 'early_stopping': True, 'num_beams': 5, 'forced_eos_token_id': 2}
3171
+ [WARNING|trainer.py:2764] 2024-10-02 17:03:31,472 >> There were missing keys in the checkpoint model loaded: ['model.encoder.embed_tokens.weight', 'model.decoder.embed_tokens.weight', 'lm_head.weight'].
3172
+
3173
 
3174
+
3175
+ [WARNING|configuration_utils.py:448] 2024-10-02 17:04:57,059 >> Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.
3176
+ Non-default generation parameters: {'max_length': 200, 'early_stopping': True, 'num_beams': 5, 'forced_eos_token_id': 2}
3177
+ [WARNING|configuration_utils.py:448] 2024-10-02 17:05:03,905 >> Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.
3178
+ Non-default generation parameters: {'max_length': 200, 'early_stopping': True, 'num_beams': 5, 'forced_eos_token_id': 2}