End of training
Browse files- README.md +15 -12
- model.safetensors +1 -1
- training_args.bin +1 -1
README.md
CHANGED
@@ -20,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
|
|
20 |
|
21 |
This model is a fine-tuned version of [ml6team/keyphrase-extraction-distilbert-inspec](https://huggingface.co/ml6team/keyphrase-extraction-distilbert-inspec) on an unknown dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
-
- Loss:
|
24 |
-
- Precision: 0.
|
25 |
-
- Recall: 0.
|
26 |
-
- F1: 0.
|
27 |
-
- Accuracy: 0.
|
28 |
|
29 |
## Model description
|
30 |
|
@@ -43,12 +43,10 @@ More information needed
|
|
43 |
### Training hyperparameters
|
44 |
|
45 |
The following hyperparameters were used during training:
|
46 |
-
- learning_rate:
|
47 |
- train_batch_size: 64
|
48 |
- eval_batch_size: 4
|
49 |
- seed: 42
|
50 |
-
- gradient_accumulation_steps: 4
|
51 |
-
- total_train_batch_size: 256
|
52 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
53 |
- lr_scheduler_type: linear
|
54 |
- num_epochs: 10
|
@@ -58,10 +56,15 @@ The following hyperparameters were used during training:
|
|
58 |
|
59 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
60 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
61 |
-
| No log |
|
62 |
-
| No log |
|
63 |
-
| No log |
|
64 |
-
| No log |
|
|
|
|
|
|
|
|
|
|
|
65 |
|
66 |
|
67 |
### Framework versions
|
|
|
20 |
|
21 |
This model is a fine-tuned version of [ml6team/keyphrase-extraction-distilbert-inspec](https://huggingface.co/ml6team/keyphrase-extraction-distilbert-inspec) on an unknown dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
+
- Loss: 0.3161
|
24 |
+
- Precision: 0.9068
|
25 |
+
- Recall: 0.9119
|
26 |
+
- F1: 0.9094
|
27 |
+
- Accuracy: 0.9015
|
28 |
|
29 |
## Model description
|
30 |
|
|
|
43 |
### Training hyperparameters
|
44 |
|
45 |
The following hyperparameters were used during training:
|
46 |
+
- learning_rate: 0.0002
|
47 |
- train_batch_size: 64
|
48 |
- eval_batch_size: 4
|
49 |
- seed: 42
|
|
|
|
|
50 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
51 |
- lr_scheduler_type: linear
|
52 |
- num_epochs: 10
|
|
|
56 |
|
57 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
58 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
59 |
+
| No log | 0.59 | 10 | 0.4403 | 0.8234 | 0.8643 | 0.8433 | 0.8282 |
|
60 |
+
| No log | 1.18 | 20 | 0.3432 | 0.8806 | 0.8796 | 0.8801 | 0.8714 |
|
61 |
+
| No log | 1.76 | 30 | 0.3102 | 0.8818 | 0.9022 | 0.8919 | 0.8838 |
|
62 |
+
| No log | 2.35 | 40 | 0.3139 | 0.9009 | 0.8939 | 0.8974 | 0.8907 |
|
63 |
+
| No log | 2.94 | 50 | 0.2956 | 0.9063 | 0.8953 | 0.9008 | 0.8922 |
|
64 |
+
| No log | 3.53 | 60 | 0.3161 | 0.9068 | 0.9119 | 0.9094 | 0.9015 |
|
65 |
+
| No log | 4.12 | 70 | 0.3459 | 0.9019 | 0.9154 | 0.9086 | 0.9006 |
|
66 |
+
| No log | 4.71 | 80 | 0.3387 | 0.9078 | 0.9099 | 0.9089 | 0.9017 |
|
67 |
+
| No log | 5.29 | 90 | 0.3658 | 0.9081 | 0.8982 | 0.9032 | 0.8942 |
|
68 |
|
69 |
|
70 |
### Framework versions
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 265476168
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:034bdc010e9653203930571b7f9e78a576d1344170ee359f8a4d9ff58a6c3012
|
3 |
size 265476168
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4984
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11e4ec6176b240c4c2c6a90105bc9183cfaa56661c346a9115e753cb5dfb2b2d
|
3 |
size 4984
|