tjasad/prompt_fine_tuned_boolq_bert
Browse files- README.md +14 -10
- adapter_config.json +17 -0
- adapter_model.safetensors +3 -0
- runs/May03_08-24-34_53ae7b07d069/events.out.tfevents.1714724708.53ae7b07d069.855.2 +3 -0
- runs/May03_08-26-23_53ae7b07d069/events.out.tfevents.1714724789.53ae7b07d069.855.3 +3 -0
- runs/May03_08-27-10_53ae7b07d069/events.out.tfevents.1714724836.53ae7b07d069.855.4 +3 -0
- runs/May03_08-29-15_53ae7b07d069/events.out.tfevents.1714724964.53ae7b07d069.855.5 +3 -0
- runs/May03_08-30-29_53ae7b07d069/events.out.tfevents.1714725035.53ae7b07d069.855.6 +3 -0
- runs/May03_08-31-48_53ae7b07d069/events.out.tfevents.1714725114.53ae7b07d069.855.7 +3 -0
- runs/May03_08-33-51_53ae7b07d069/events.out.tfevents.1714725238.53ae7b07d069.855.8 +3 -0
- tokenizer.json +1 -1
- training_args.bin +1 -1
README.md
CHANGED
@@ -1,8 +1,9 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
-
|
4 |
tags:
|
5 |
- generated_from_trainer
|
|
|
6 |
metrics:
|
7 |
- accuracy
|
8 |
- f1
|
@@ -18,7 +19,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
18 |
|
19 |
This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on an unknown dataset.
|
20 |
It achieves the following results on the evaluation set:
|
21 |
-
- Loss: 0.
|
22 |
- Accuracy: 0.7778
|
23 |
- F1: 0.7481
|
24 |
|
@@ -45,23 +46,26 @@ The following hyperparameters were used during training:
|
|
45 |
- seed: 42
|
46 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
47 |
- lr_scheduler_type: linear
|
48 |
-
- num_epochs:
|
49 |
|
50 |
### Training results
|
51 |
|
52 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|
53 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
|
54 |
-
| No log | 1.0 | 12 | 0.
|
55 |
-
| No log | 2.0 | 24 | 0.
|
56 |
-
| No log | 3.0 | 36 | 0.
|
57 |
-
| No log | 4.0 | 48 | 0.
|
58 |
-
| No log | 5.0 | 60 | 0.
|
59 |
-
| No log | 6.0 | 72 | 0.
|
|
|
|
|
60 |
|
61 |
|
62 |
### Framework versions
|
63 |
|
|
|
64 |
- Transformers 4.40.1
|
65 |
- Pytorch 2.2.1+cu121
|
66 |
- Datasets 2.19.0
|
67 |
-
- Tokenizers 0.19.1
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
library_name: peft
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
+
base_model: google-bert/bert-base-uncased
|
7 |
metrics:
|
8 |
- accuracy
|
9 |
- f1
|
|
|
19 |
|
20 |
This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on an unknown dataset.
|
21 |
It achieves the following results on the evaluation set:
|
22 |
+
- Loss: 0.6710
|
23 |
- Accuracy: 0.7778
|
24 |
- F1: 0.7481
|
25 |
|
|
|
46 |
- seed: 42
|
47 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
48 |
- lr_scheduler_type: linear
|
49 |
+
- num_epochs: 8
|
50 |
|
51 |
### Training results
|
52 |
|
53 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|
54 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
|
55 |
+
| No log | 1.0 | 12 | 0.7544 | 0.2222 | 0.0808 |
|
56 |
+
| No log | 2.0 | 24 | 0.7285 | 0.2222 | 0.0808 |
|
57 |
+
| No log | 3.0 | 36 | 0.7061 | 0.3333 | 0.2833 |
|
58 |
+
| No log | 4.0 | 48 | 0.6909 | 0.5556 | 0.5926 |
|
59 |
+
| No log | 5.0 | 60 | 0.6817 | 0.7222 | 0.7325 |
|
60 |
+
| No log | 6.0 | 72 | 0.6760 | 0.7778 | 0.7778 |
|
61 |
+
| No log | 7.0 | 84 | 0.6722 | 0.7778 | 0.7481 |
|
62 |
+
| No log | 8.0 | 96 | 0.6710 | 0.7778 | 0.7481 |
|
63 |
|
64 |
|
65 |
### Framework versions
|
66 |
|
67 |
+
- PEFT 0.10.0
|
68 |
- Transformers 4.40.1
|
69 |
- Pytorch 2.2.1+cu121
|
70 |
- Datasets 2.19.0
|
71 |
+
- Tokenizers 0.19.1
|
adapter_config.json
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"auto_mapping": null,
|
3 |
+
"base_model_name_or_path": "google-bert/bert-base-uncased",
|
4 |
+
"inference_mode": true,
|
5 |
+
"num_attention_heads": 12,
|
6 |
+
"num_layers": 6,
|
7 |
+
"num_transformer_submodules": 1,
|
8 |
+
"num_virtual_tokens": 12,
|
9 |
+
"peft_type": "PROMPT_TUNING",
|
10 |
+
"prompt_tuning_init": "RANDOM",
|
11 |
+
"prompt_tuning_init_text": null,
|
12 |
+
"revision": null,
|
13 |
+
"task_type": "SEQ_CLS",
|
14 |
+
"token_dim": 768,
|
15 |
+
"tokenizer_kwargs": null,
|
16 |
+
"tokenizer_name_or_path": "google-bert/bert-base-uncased"
|
17 |
+
}
|
adapter_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9566620eb09d4c59c59014c76917f52915e0d1bc62894ff54957f8c97dde39aa
|
3 |
+
size 43304
|
runs/May03_08-24-34_53ae7b07d069/events.out.tfevents.1714724708.53ae7b07d069.855.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6fe34d10e9a5dfa02b4ed912dfeae5c531803237a53f14b89cd661dbf12cf551
|
3 |
+
size 4916
|
runs/May03_08-26-23_53ae7b07d069/events.out.tfevents.1714724789.53ae7b07d069.855.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:483b380b420d5a219ed44fc769bf889c2571f1acd1af78d73d9066b7ba48f40e
|
3 |
+
size 4825
|
runs/May03_08-27-10_53ae7b07d069/events.out.tfevents.1714724836.53ae7b07d069.855.4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:04c1419ae27945f14ae16efd293f31906177fff8f30f768483c60c3880d5c014
|
3 |
+
size 7345
|
runs/May03_08-29-15_53ae7b07d069/events.out.tfevents.1714724964.53ae7b07d069.855.5
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:589720b05b799987f06ce042abdb7bf6b5fb7c368cff72b01640d0e4160256ab
|
3 |
+
size 7394
|
runs/May03_08-30-29_53ae7b07d069/events.out.tfevents.1714725035.53ae7b07d069.855.6
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b5c5952fe1db4c7f02cee7840da81123c6d2cd203f638fc69de8fb5db6d1c8c3
|
3 |
+
size 7394
|
runs/May03_08-31-48_53ae7b07d069/events.out.tfevents.1714725114.53ae7b07d069.855.7
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:da4cd28aad8747c07657c11119ef223643914fd90be392768eb14123d63ea841
|
3 |
+
size 8843
|
runs/May03_08-33-51_53ae7b07d069/events.out.tfevents.1714725238.53ae7b07d069.855.8
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cc35c0b7e4244a0c59c105e803c24278f6f2755c7d0db0aecc912fe21dd61b0c
|
3 |
+
size 8069
|
tokenizer.json
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
"version": "1.0",
|
3 |
"truncation": {
|
4 |
"direction": "Right",
|
5 |
-
"max_length":
|
6 |
"strategy": "LongestFirst",
|
7 |
"stride": 0
|
8 |
},
|
|
|
2 |
"version": "1.0",
|
3 |
"truncation": {
|
4 |
"direction": "Right",
|
5 |
+
"max_length": 400,
|
6 |
"strategy": "LongestFirst",
|
7 |
"stride": 0
|
8 |
},
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5048
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46251bb5e0fabba4683f0085b4d463983230c20c15dc8103910ef3a76639abef
|
3 |
size 5048
|