scottgdaniel commited on
Commit
324b6c2
1 Parent(s): 8b13ee2

Model save

Browse files
README.md ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: scales-okn/docket-language-model
3
+ tags:
4
+ - generated_from_trainer
5
+ metrics:
6
+ - accuracy
7
+ - f1
8
+ - precision
9
+ - recall
10
+ model-index:
11
+ - name: ontology-answer-test
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # ontology-answer-test
19
+
20
+ This model is a fine-tuned version of [scales-okn/docket-language-model](https://huggingface.co/scales-okn/docket-language-model) on the None dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: 0.0009
23
+ - Accuracy: 1.0
24
+ - F1: 1.0
25
+ - Precision: 1.0
26
+ - Recall: 1.0
27
+
28
+ ## Model description
29
+
30
+ More information needed
31
+
32
+ ## Intended uses & limitations
33
+
34
+ More information needed
35
+
36
+ ## Training and evaluation data
37
+
38
+ More information needed
39
+
40
+ ## Training procedure
41
+
42
+ ### Training hyperparameters
43
+
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 3e-05
46
+ - train_batch_size: 8
47
+ - eval_batch_size: 16
48
+ - seed: 42
49
+ - gradient_accumulation_steps: 2
50
+ - total_train_batch_size: 16
51
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
+ - lr_scheduler_type: linear
53
+ - lr_scheduler_warmup_ratio: 0.06
54
+ - num_epochs: 5
55
+
56
+ ### Training results
57
+
58
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
59
+ |:-------------:|:------:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
60
+ | 0.0233 | 1.2903 | 100 | 0.0434 | 0.9862 | 0.9720 | 0.9630 | 0.9811 |
61
+ | 0.0007 | 2.5806 | 200 | 0.0072 | 0.9954 | 0.9905 | 1.0 | 0.9811 |
62
+ | 0.0003 | 3.8710 | 300 | 0.0009 | 1.0 | 1.0 | 1.0 | 1.0 |
63
+
64
+
65
+ ### Framework versions
66
+
67
+ - Transformers 4.41.2
68
+ - Pytorch 2.1.2+cu121
69
+ - Datasets 2.19.2
70
+ - Tokenizers 0.19.1
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "scales-okn/docket-language-model",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_eps": 1e-07,
13
+ "max_position_embeddings": 512,
14
+ "max_relative_positions": -1,
15
+ "model_type": "deberta-v2",
16
+ "norm_rel_ebd": "layer_norm",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
+ "pad_token_id": 0,
20
+ "pooler_dropout": 0,
21
+ "pooler_hidden_act": "gelu",
22
+ "pooler_hidden_size": 1024,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "position_buckets": 256,
29
+ "relative_attention": true,
30
+ "share_att_key": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.41.2",
33
+ "type_vocab_size": 0,
34
+ "vocab_size": 128001
35
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d73de8e9e8fbb84901b725ece1bb92cc851d64037a8f591f3086c2ba392b3832
3
+ size 1739898936
runs/events.out.tfevents.1718056601.delilah.39183.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca6c0e09317e41c2e5971315213a2997489e3fac06515359bf55e27a4da2dce5
3
+ size 22886
runs/events.out.tfevents.1718056997.delilah.39183.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0e18ae640e58a4e9e1661a6906177d73b5cc5438d72d9ee6eb786eda2f0db50
3
+ size 512
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "[CLS]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "[SEP]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "[MASK]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "[SEP]",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "max_length": 256,
51
+ "model_max_length": 1000000000000000019884624838656,
52
+ "pad_to_multiple_of": null,
53
+ "pad_token": "[PAD]",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
+ "sep_token": "[SEP]",
57
+ "sp_model_kwargs": {},
58
+ "split_by_punct": false,
59
+ "stride": 0,
60
+ "tokenizer_class": "DebertaV2Tokenizer",
61
+ "truncation_side": "right",
62
+ "truncation_strategy": "longest_first",
63
+ "unk_token": "[UNK]",
64
+ "vocab_type": "spm"
65
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a58e6b4fe46438892548857b39aae1d27703c6e32b0b6bc13989226a177399c4
3
+ size 5112
training_progress.csv ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ loss,grad_norm,learning_rate,epoch,steps,eval_loss,eval_accuracy,eval_f1,eval_precision,eval_recall,eval_runtime,eval_samples_per_second,eval_steps_per_second,train_runtime,train_samples_per_second,train_steps_per_second,total_flos,train_loss
2
+ 0.6909,4.090293884277344,6.25e-06,0.06451612903225806,5,,,,,,,,,,,,,
3
+ 0.6278,5.705702781677246,1.25e-05,0.12903225806451613,10,,,,,,,,,,,,,
4
+ 0.549,2.5465235710144043,1.8750000000000002e-05,0.1935483870967742,15,,,,,,,,,,,,,
5
+ 0.4981,3.545476198196411,2.5e-05,0.25806451612903225,20,,,,,,,,,,,,,
6
+ 0.3127,1.7096995115280151,2.991689750692521e-05,0.3225806451612903,25,,,,,,,,,,,,,
7
+ 0.2114,3.098984479904175,2.950138504155125e-05,0.3870967741935484,30,,,,,,,,,,,,,
8
+ 0.1548,2.043656826019287,2.9085872576177285e-05,0.45161290322580644,35,,,,,,,,,,,,,
9
+ 0.0987,0.37046054005622864,2.8670360110803324e-05,0.5161290322580645,40,,,,,,,,,,,,,
10
+ 0.1391,2.608640670776367,2.8254847645429363e-05,0.5806451612903226,45,,,,,,,,,,,,,
11
+ 0.0606,0.4053419530391693,2.78393351800554e-05,0.6451612903225806,50,,,,,,,,,,,,,
12
+ 0.0275,4.057948112487793,2.7423822714681444e-05,0.7096774193548387,55,,,,,,,,,,,,,
13
+ 0.0928,0.23072686791419983,2.700831024930748e-05,0.7741935483870968,60,,,,,,,,,,,,,
14
+ 0.0733,0.05615957826375961,2.6592797783933518e-05,0.8387096774193549,65,,,,,,,,,,,,,
15
+ 0.1347,2.636467456817627,2.6177285318559557e-05,0.9032258064516129,70,,,,,,,,,,,,,
16
+ 0.0766,0.11293677985668182,2.5761772853185596e-05,0.967741935483871,75,,,,,,,,,,,,,
17
+ 0.0055,0.1519118696451187,2.5346260387811635e-05,1.032258064516129,80,,,,,,,,,,,,,
18
+ 0.0052,0.2113182097673416,2.4930747922437674e-05,1.096774193548387,85,,,,,,,,,,,,,
19
+ 0.0395,2.404287099838257,2.4515235457063713e-05,1.1612903225806452,90,,,,,,,,,,,,,
20
+ 0.0084,0.08060836046934128,2.409972299168975e-05,1.2258064516129032,95,,,,,,,,,,,,,
21
+ 0.0233,0.7740814685821533,2.368421052631579e-05,1.2903225806451613,100,,,,,,,,,,,,,
22
+ ,,,1.2903225806451613,100,0.04341869056224823,0.9862385321100917,0.9719626168224299,0.9629629629629629,0.9811320754716981,4.3861,49.702,3.192,,,,,
23
+ 0.0204,1.7034190893173218,2.326869806094183e-05,1.3548387096774195,105,,,,,,,,,,,,,
24
+ 0.0664,0.06643515080213547,2.2853185595567868e-05,1.4193548387096775,110,,,,,,,,,,,,,
25
+ 0.0126,0.048984918743371964,2.2437673130193904e-05,1.4838709677419355,115,,,,,,,,,,,,,
26
+ 0.0027,0.42834651470184326,2.2022160664819946e-05,1.5483870967741935,120,,,,,,,,,,,,,
27
+ 0.0291,0.06453526765108109,2.1606648199445985e-05,1.6129032258064515,125,,,,,,,,,,,,,
28
+ 0.2114,7.842092990875244,2.1191135734072024e-05,1.6774193548387095,130,,,,,,,,,,,,,
29
+ 0.0494,0.04162972792983055,2.0775623268698063e-05,1.7419354838709677,135,,,,,,,,,,,,,
30
+ 0.0017,0.06891004741191864,2.0360110803324098e-05,1.8064516129032258,140,,,,,,,,,,,,,
31
+ 0.0021,0.040090594440698624,1.9944598337950137e-05,1.870967741935484,145,,,,,,,,,,,,,
32
+ 0.0017,0.07595518976449966,1.952908587257618e-05,1.935483870967742,150,,,,,,,,,,,,,
33
+ 0.0311,10.209275245666504,1.9113573407202218e-05,2.0,155,,,,,,,,,,,,,
34
+ 0.0027,0.14497926831245422,1.8698060941828257e-05,2.064516129032258,160,,,,,,,,,,,,,
35
+ 0.0024,0.40518391132354736,1.8282548476454293e-05,2.129032258064516,165,,,,,,,,,,,,,
36
+ 0.0403,0.01545148529112339,1.786703601108033e-05,2.193548387096774,170,,,,,,,,,,,,,
37
+ 0.0008,0.020430177450180054,1.745152354570637e-05,2.258064516129032,175,,,,,,,,,,,,,
38
+ 0.0008,0.01945902779698372,1.7036011080332413e-05,2.3225806451612905,180,,,,,,,,,,,,,
39
+ 0.0007,0.01845143549144268,1.662049861495845e-05,2.3870967741935485,185,,,,,,,,,,,,,
40
+ 0.0007,0.018286943435668945,1.6204986149584487e-05,2.4516129032258065,190,,,,,,,,,,,,,
41
+ 0.0024,0.014586242847144604,1.5789473684210526e-05,2.5161290322580645,195,,,,,,,,,,,,,
42
+ 0.0007,0.012451103888452053,1.5373961218836565e-05,2.5806451612903225,200,,,,,,,,,,,,,
43
+ ,,,2.5806451612903225,200,0.007189903408288956,0.9954128440366973,0.9904761904761905,1.0,0.9811320754716981,4.502,48.422,3.11,,,,,
44
+ 0.0009,0.011877791956067085,1.4958448753462605e-05,2.6451612903225805,205,,,,,,,,,,,,,
45
+ 0.0005,0.012689828872680664,1.4542936288088642e-05,2.709677419354839,210,,,,,,,,,,,,,
46
+ 0.0007,0.017396604642271996,1.4127423822714681e-05,2.774193548387097,215,,,,,,,,,,,,,
47
+ 0.0006,0.01691826991736889,1.3711911357340722e-05,2.838709677419355,220,,,,,,,,,,,,,
48
+ 0.0007,0.026222320273518562,1.3296398891966759e-05,2.903225806451613,225,,,,,,,,,,,,,
49
+ 0.0017,0.011283783242106438,1.2880886426592798e-05,2.967741935483871,230,,,,,,,,,,,,,
50
+ 0.0006,0.016978178173303604,1.2465373961218837e-05,3.032258064516129,235,,,,,,,,,,,,,
51
+ 0.0007,0.008052507415413857,1.2049861495844876e-05,3.096774193548387,240,,,,,,,,,,,,,
52
+ 0.0004,0.007332273758947849,1.1634349030470915e-05,3.161290322580645,245,,,,,,,,,,,,,
53
+ 0.0005,0.0079089580103755,1.1218836565096952e-05,3.225806451612903,250,,,,,,,,,,,,,
54
+ 0.0005,0.009510311298072338,1.0803324099722992e-05,3.2903225806451615,255,,,,,,,,,,,,,
55
+ 0.0005,0.024516815319657326,1.0387811634349031e-05,3.3548387096774195,260,,,,,,,,,,,,,
56
+ 0.0005,0.021235521882772446,9.972299168975068e-06,3.4193548387096775,265,,,,,,,,,,,,,
57
+ 0.0004,0.008695265278220177,9.556786703601109e-06,3.4838709677419355,270,,,,,,,,,,,,,
58
+ 0.0004,0.008880405686795712,9.141274238227146e-06,3.5483870967741935,275,,,,,,,,,,,,,
59
+ 0.0003,0.007136228494346142,8.725761772853185e-06,3.6129032258064515,280,,,,,,,,,,,,,
60
+ 0.0004,0.008296678774058819,8.310249307479226e-06,3.6774193548387095,285,,,,,,,,,,,,,
61
+ 0.0003,0.008309051394462585,7.894736842105263e-06,3.741935483870968,290,,,,,,,,,,,,,
62
+ 0.0003,0.006046871189028025,7.479224376731303e-06,3.806451612903226,295,,,,,,,,,,,,,
63
+ 0.0003,0.009278281591832638,7.063711911357341e-06,3.870967741935484,300,,,,,,,,,,,,,
64
+ ,,,3.870967741935484,300,0.0009307524305768311,1.0,1.0,1.0,1.0,4.3119,50.557,3.247,,,,,
65
+ 0.0036,0.008976064622402191,6.6481994459833796e-06,3.935483870967742,305,,,,,,,,,,,,,
66
+ 0.0006,0.009819263592362404,6.2326869806094184e-06,4.0,310,,,,,,,,,,,,,
67
+ 0.0004,0.007393244653940201,5.817174515235457e-06,4.064516129032258,315,,,,,,,,,,,,,
68
+ 0.0004,0.006700852420181036,5.401662049861496e-06,4.129032258064516,320,,,,,,,,,,,,,
69
+ 0.0004,0.009492475539445877,4.986149584487534e-06,4.193548387096774,325,,,,,,,,,,,,,
70
+ 0.0004,0.014097731560468674,4.570637119113573e-06,4.258064516129032,330,,,,,,,,,,,,,
71
+ 0.0003,0.0077942414209246635,4.155124653739613e-06,4.32258064516129,335,,,,,,,,,,,,,
72
+ 0.0003,0.007693646941334009,3.7396121883656513e-06,4.387096774193548,340,,,,,,,,,,,,,
73
+ 0.0003,0.009858954697847366,3.3240997229916898e-06,4.451612903225806,345,,,,,,,,,,,,,
74
+ 0.0003,0.005932302679866552,2.9085872576177287e-06,4.516129032258064,350,,,,,,,,,,,,,
75
+ 0.0003,0.010720741003751755,2.493074792243767e-06,4.580645161290323,355,,,,,,,,,,,,,
76
+ 0.0004,0.007814132608473301,2.0775623268698064e-06,4.645161290322581,360,,,,,,,,,,,,,
77
+ 0.0003,0.007145343814045191,1.6620498614958449e-06,4.709677419354839,365,,,,,,,,,,,,,
78
+ 0.0003,0.0064388057217001915,1.2465373961218836e-06,4.774193548387097,370,,,,,,,,,,,,,
79
+ 0.0003,0.008449292741715908,8.310249307479224e-07,4.838709677419355,375,,,,,,,,,,,,,
80
+ 0.0004,0.006520669441670179,4.155124653739612e-07,4.903225806451613,380,,,,,,,,,,,,,
81
+ 0.0004,0.008102640509605408,0.0,4.967741935483871,385,,,,,,,,,,,,,
82
+ ,,,4.967741935483871,385,,,,,,,,,391.5134,15.823,0.983,2868504101425152.0,0.056238630435477895