devhem commited on
Commit
d9acf58
·
verified ·
1 Parent(s): 747cb5d

Model save

Browse files
Files changed (1) hide show
  1. README.md +13 -39
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 1.3081
22
- - Accuracy: 0.7230
23
 
24
  ## Model description
25
 
@@ -38,52 +38,26 @@ More information needed
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 1e-05
42
  - train_batch_size: 32
43
  - eval_batch_size: 8
44
  - seed: 42
45
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
- - lr_scheduler_warmup_steps: 500
48
- - num_epochs: 32
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
52
 
53
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
54
- |:-------------:|:-----:|:-----:|:---------------:|:--------:|
55
- | 1.5456 | 1.0 | 876 | 0.6853 | 0.7600 |
56
- | 0.6286 | 2.0 | 1752 | 0.6386 | 0.7755 |
57
- | 0.5239 | 3.0 | 2628 | 0.6216 | 0.7690 |
58
- | 0.4708 | 4.0 | 3504 | 0.6294 | 0.7684 |
59
- | 0.4224 | 5.0 | 4380 | 0.6622 | 0.7690 |
60
- | 0.3963 | 6.0 | 5256 | 0.6938 | 0.7558 |
61
- | 0.3677 | 7.0 | 6132 | 0.7313 | 0.7575 |
62
- | 0.3486 | 8.0 | 7008 | 0.7668 | 0.7508 |
63
- | 0.3192 | 9.0 | 7884 | 0.7856 | 0.7488 |
64
- | 0.3119 | 10.0 | 8760 | 0.8193 | 0.7413 |
65
- | 0.297 | 11.0 | 9636 | 0.8250 | 0.7385 |
66
- | 0.2826 | 12.0 | 10512 | 0.8906 | 0.7264 |
67
- | 0.2664 | 13.0 | 11388 | 0.8942 | 0.7301 |
68
- | 0.2614 | 14.0 | 12264 | 0.9402 | 0.7281 |
69
- | 0.2585 | 15.0 | 13140 | 0.9722 | 0.7361 |
70
- | 0.2603 | 16.0 | 14016 | 1.0199 | 0.7285 |
71
- | 0.2366 | 17.0 | 14892 | 1.0044 | 0.7290 |
72
- | 0.2406 | 18.0 | 15768 | 1.0022 | 0.7237 |
73
- | 0.2341 | 19.0 | 16644 | 1.0498 | 0.7233 |
74
- | 0.2392 | 20.0 | 17520 | 1.0741 | 0.7258 |
75
- | 0.2135 | 21.0 | 18396 | 1.1113 | 0.7233 |
76
- | 0.2166 | 22.0 | 19272 | 1.1229 | 0.7288 |
77
- | 0.216 | 23.0 | 20148 | 1.1429 | 0.7204 |
78
- | 0.2164 | 24.0 | 21024 | 1.1872 | 0.7261 |
79
- | 0.2079 | 25.0 | 21900 | 1.2140 | 0.7161 |
80
- | 0.2027 | 26.0 | 22776 | 1.2285 | 0.7240 |
81
- | 0.201 | 27.0 | 23652 | 1.2247 | 0.7250 |
82
- | 0.2052 | 28.0 | 24528 | 1.2356 | 0.7207 |
83
- | 0.1959 | 29.0 | 25404 | 1.2895 | 0.7217 |
84
- | 0.1975 | 30.0 | 26280 | 1.2861 | 0.7245 |
85
- | 0.1937 | 31.0 | 27156 | 1.3034 | 0.7211 |
86
- | 0.194 | 32.0 | 28032 | 1.3081 | 0.7230 |
87
 
88
 
89
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.6824
22
+ - Accuracy: 0.7670
23
 
24
  ## Model description
25
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 2e-06
42
  - train_batch_size: 32
43
  - eval_batch_size: 8
44
  - seed: 42
45
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
+ - lr_scheduler_warmup_ratio: 0.1
48
+ - num_epochs: 6
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
52
 
53
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
54
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
55
+ | 1.9725 | 1.0 | 876 | 1.0953 | 0.6834 |
56
+ | 0.9389 | 2.0 | 1752 | 0.7982 | 0.7514 |
57
+ | 0.7549 | 3.0 | 2628 | 0.7214 | 0.7647 |
58
+ | 0.6895 | 4.0 | 3504 | 0.6963 | 0.7680 |
59
+ | 0.6712 | 5.0 | 4380 | 0.6856 | 0.7664 |
60
+ | 0.653 | 6.0 | 5256 | 0.6824 | 0.7670 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
61
 
62
 
63
  ### Framework versions