hoang.dang1 commited on
Commit
d9f1042
1 Parent(s): 96239b9

inital commit

Browse files
README.md CHANGED
@@ -1,3 +1,76 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ base_model: microsoft/deberta-v3-large
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - precision
9
+ - recall
10
+ - f1
11
+ model-index:
12
+ - name: BBC_CLS_deberta_v3_large_v2
13
+ results: []
14
  ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # BBC_CLS_deberta_v3_large_v2
20
+
21
+ This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.0857
24
+ - Accuracy: 0.9866
25
+ - Precision: 0.9723
26
+ - Recall: 0.9780
27
+ - F1: 0.9751
28
+
29
+ ## Model description
30
+
31
+ More information needed
32
+
33
+ ## Intended uses & limitations
34
+
35
+ More information needed
36
+
37
+ ## Training and evaluation data
38
+
39
+ More information needed
40
+
41
+ ## Training procedure
42
+
43
+ ### Training hyperparameters
44
+
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 5e-05
47
+ - train_batch_size: 32
48
+ - eval_batch_size: 32
49
+ - seed: 42
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - lr_scheduler_warmup_steps: 500
53
+ - num_epochs: 10
54
+
55
+ ### Training results
56
+
57
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
58
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
59
+ | 1.235 | 1.0 | 66 | 0.6331 | 0.7964 | 0.4047 | 0.4873 | 0.4418 |
60
+ | 0.4336 | 2.0 | 132 | 0.2201 | 0.8971 | 0.6754 | 0.7091 | 0.6910 |
61
+ | 0.2133 | 3.0 | 198 | 0.0990 | 0.9776 | 0.9476 | 0.9786 | 0.9602 |
62
+ | 0.1083 | 4.0 | 264 | 0.1038 | 0.9821 | 0.9656 | 0.9651 | 0.9653 |
63
+ | 0.0848 | 5.0 | 330 | 0.0907 | 0.9866 | 0.9782 | 0.9714 | 0.9747 |
64
+ | 0.1087 | 6.0 | 396 | 0.1270 | 0.9799 | 0.9672 | 0.9689 | 0.9671 |
65
+ | 0.1011 | 7.0 | 462 | 0.1289 | 0.9754 | 0.9677 | 0.9660 | 0.9667 |
66
+ | 0.0827 | 8.0 | 528 | 0.0990 | 0.9799 | 0.9818 | 0.9479 | 0.9632 |
67
+ | 0.0621 | 9.0 | 594 | 0.0857 | 0.9866 | 0.9723 | 0.9780 | 0.9751 |
68
+ | 0.0444 | 10.0 | 660 | 0.1071 | 0.9843 | 0.9769 | 0.9663 | 0.9715 |
69
+
70
+
71
+ ### Framework versions
72
+
73
+ - Transformers 4.35.0.dev0
74
+ - Pytorch 1.13.1
75
+ - Datasets 2.13.0
76
+ - Tokenizers 0.14.1
added_tokens.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "[CLS]": 1,
3
+ "[MASK]": 128000,
4
+ "[PAD]": 0,
5
+ "[SEP]": 2,
6
+ "[UNK]": 3
7
+ }
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "id2label": {
11
+ "0": "Cargo",
12
+ "1": "Open Pos",
13
+ "2": "Time Charter",
14
+ "3": "Spam"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 4096,
18
+ "layer_norm_eps": 1e-07,
19
+ "max_position_embeddings": 512,
20
+ "max_relative_positions": -1,
21
+ "model_type": "deberta-v2",
22
+ "norm_rel_ebd": "layer_norm",
23
+ "num_attention_heads": 16,
24
+ "num_hidden_layers": 24,
25
+ "pad_token_id": 0,
26
+ "pooler_dropout": 0,
27
+ "pooler_hidden_act": "gelu",
28
+ "pooler_hidden_size": 1024,
29
+ "pos_att_type": [
30
+ "p2c",
31
+ "c2p"
32
+ ],
33
+ "position_biased_input": false,
34
+ "position_buckets": 256,
35
+ "relative_attention": true,
36
+ "share_att_key": true,
37
+ "torch_dtype": "float32",
38
+ "transformers_version": "4.35.0.dev0",
39
+ "type_vocab_size": 0,
40
+ "vocab_size": 128100
41
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b02e60ea41c996960ec95a8d582f6cf7d12b880695e85ce3665792a4117b879a
3
+ size 3480856123
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8ed3ca0df041bd4fbccb3cc002bd68656570223139d7b5016821575c728f691
3
+ size 1740399989
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b93e85b854d09c59c61f6928e21a8b24c26c39f575da928524dc8b935ca5f35b
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b44221305308757e0422afef9c458c76bfdd35eac0f0a0b8dbc4a6f6580308d
3
+ size 627
special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "additional_special_tokens": [],
45
+ "bos_token": "[CLS]",
46
+ "clean_up_tokenization_spaces": true,
47
+ "cls_token": "[CLS]",
48
+ "do_lower_case": false,
49
+ "eos_token": "[SEP]",
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 1000000000000000019884624838656,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "sp_model_kwargs": {},
55
+ "split_by_punct": false,
56
+ "tokenizer_class": "DebertaV2Tokenizer",
57
+ "unk_token": "[UNK]",
58
+ "vocab_type": "spm"
59
+ }
trainer_state.json ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.08565760403871536,
3
+ "best_model_checkpoint": "BBC_CLS_deberta_v3_large_v2/checkpoint-594",
4
+ "epoch": 10.0,
5
+ "eval_steps": 500,
6
+ "global_step": 660,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "learning_rate": 6.6e-06,
14
+ "loss": 1.235,
15
+ "step": 66
16
+ },
17
+ {
18
+ "epoch": 1.0,
19
+ "eval_accuracy": 0.796420581655481,
20
+ "eval_f1": 0.4417921999471439,
21
+ "eval_loss": 0.633094072341919,
22
+ "eval_precision": 0.4046829607174435,
23
+ "eval_recall": 0.4873423547822885,
24
+ "eval_runtime": 6.7539,
25
+ "eval_samples_per_second": 66.184,
26
+ "eval_steps_per_second": 2.073,
27
+ "step": 66
28
+ },
29
+ {
30
+ "epoch": 2.0,
31
+ "learning_rate": 1.32e-05,
32
+ "loss": 0.4336,
33
+ "step": 132
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_accuracy": 0.8970917225950783,
38
+ "eval_f1": 0.6910171764080433,
39
+ "eval_loss": 0.22008153796195984,
40
+ "eval_precision": 0.6753658536585365,
41
+ "eval_recall": 0.7090775277247049,
42
+ "eval_runtime": 6.7716,
43
+ "eval_samples_per_second": 66.011,
44
+ "eval_steps_per_second": 2.067,
45
+ "step": 132
46
+ },
47
+ {
48
+ "epoch": 3.0,
49
+ "learning_rate": 1.9800000000000004e-05,
50
+ "loss": 0.2133,
51
+ "step": 198
52
+ },
53
+ {
54
+ "epoch": 3.0,
55
+ "eval_accuracy": 0.9776286353467561,
56
+ "eval_f1": 0.9601648947616865,
57
+ "eval_loss": 0.0990455150604248,
58
+ "eval_precision": 0.9476303317535546,
59
+ "eval_recall": 0.9785651349741774,
60
+ "eval_runtime": 6.7752,
61
+ "eval_samples_per_second": 65.976,
62
+ "eval_steps_per_second": 2.066,
63
+ "step": 198
64
+ },
65
+ {
66
+ "epoch": 4.0,
67
+ "learning_rate": 2.64e-05,
68
+ "loss": 0.1083,
69
+ "step": 264
70
+ },
71
+ {
72
+ "epoch": 4.0,
73
+ "eval_accuracy": 0.9821029082774049,
74
+ "eval_f1": 0.9653218049514888,
75
+ "eval_loss": 0.10380033403635025,
76
+ "eval_precision": 0.9655642046538242,
77
+ "eval_recall": 0.9650874727825403,
78
+ "eval_runtime": 6.7715,
79
+ "eval_samples_per_second": 66.012,
80
+ "eval_steps_per_second": 2.067,
81
+ "step": 264
82
+ },
83
+ {
84
+ "epoch": 5.0,
85
+ "learning_rate": 3.3e-05,
86
+ "loss": 0.0848,
87
+ "step": 330
88
+ },
89
+ {
90
+ "epoch": 5.0,
91
+ "eval_accuracy": 0.9865771812080537,
92
+ "eval_f1": 0.9747350357333067,
93
+ "eval_loss": 0.09066177904605865,
94
+ "eval_precision": 0.9782430301452915,
95
+ "eval_recall": 0.9713632225190546,
96
+ "eval_runtime": 6.7664,
97
+ "eval_samples_per_second": 66.062,
98
+ "eval_steps_per_second": 2.069,
99
+ "step": 330
100
+ },
101
+ {
102
+ "epoch": 6.0,
103
+ "learning_rate": 3.960000000000001e-05,
104
+ "loss": 0.1087,
105
+ "step": 396
106
+ },
107
+ {
108
+ "epoch": 6.0,
109
+ "eval_accuracy": 0.9798657718120806,
110
+ "eval_f1": 0.9671245324193838,
111
+ "eval_loss": 0.12698371708393097,
112
+ "eval_precision": 0.9672288359788359,
113
+ "eval_recall": 0.9688744042217826,
114
+ "eval_runtime": 6.7679,
115
+ "eval_samples_per_second": 66.047,
116
+ "eval_steps_per_second": 2.069,
117
+ "step": 396
118
+ },
119
+ {
120
+ "epoch": 7.0,
121
+ "learning_rate": 4.6200000000000005e-05,
122
+ "loss": 0.1011,
123
+ "step": 462
124
+ },
125
+ {
126
+ "epoch": 7.0,
127
+ "eval_accuracy": 0.9753914988814317,
128
+ "eval_f1": 0.9666995780164981,
129
+ "eval_loss": 0.12892137467861176,
130
+ "eval_precision": 0.9676667575151919,
131
+ "eval_recall": 0.9659549559260323,
132
+ "eval_runtime": 6.7567,
133
+ "eval_samples_per_second": 66.157,
134
+ "eval_steps_per_second": 2.072,
135
+ "step": 462
136
+ },
137
+ {
138
+ "epoch": 8.0,
139
+ "learning_rate": 4.125e-05,
140
+ "loss": 0.0827,
141
+ "step": 528
142
+ },
143
+ {
144
+ "epoch": 8.0,
145
+ "eval_accuracy": 0.9798657718120806,
146
+ "eval_f1": 0.9631601759286037,
147
+ "eval_loss": 0.09904194623231888,
148
+ "eval_precision": 0.9818050507874383,
149
+ "eval_recall": 0.9479257225190546,
150
+ "eval_runtime": 6.7541,
151
+ "eval_samples_per_second": 66.182,
152
+ "eval_steps_per_second": 2.073,
153
+ "step": 528
154
+ },
155
+ {
156
+ "epoch": 9.0,
157
+ "learning_rate": 2.0625e-05,
158
+ "loss": 0.0621,
159
+ "step": 594
160
+ },
161
+ {
162
+ "epoch": 9.0,
163
+ "eval_accuracy": 0.9865771812080537,
164
+ "eval_f1": 0.9750903668573623,
165
+ "eval_loss": 0.08565760403871536,
166
+ "eval_precision": 0.9722969853754562,
167
+ "eval_recall": 0.9780020135988667,
168
+ "eval_runtime": 6.7572,
169
+ "eval_samples_per_second": 66.152,
170
+ "eval_steps_per_second": 2.072,
171
+ "step": 594
172
+ },
173
+ {
174
+ "epoch": 10.0,
175
+ "learning_rate": 0.0,
176
+ "loss": 0.0444,
177
+ "step": 660
178
+ },
179
+ {
180
+ "epoch": 10.0,
181
+ "eval_accuracy": 0.9843400447427293,
182
+ "eval_f1": 0.9714752151121933,
183
+ "eval_loss": 0.10706986486911774,
184
+ "eval_precision": 0.9768891875746715,
185
+ "eval_recall": 0.9662611817027281,
186
+ "eval_runtime": 6.7612,
187
+ "eval_samples_per_second": 66.112,
188
+ "eval_steps_per_second": 2.071,
189
+ "step": 660
190
+ }
191
+ ],
192
+ "logging_steps": 10,
193
+ "max_steps": 660,
194
+ "num_train_epochs": 10,
195
+ "save_steps": 500,
196
+ "total_flos": 9724835143127040.0,
197
+ "trial_name": null,
198
+ "trial_params": null
199
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27bea8c60e03ea217df3db5019fef3497c5c9ac1c3196db388e5f56c754a7b61
3
+ size 4027