File size: 9,604 Bytes
b3fd691 bcb4564 b3fd691 693ebf3 b3fd691 ac822b6 693ebf3 b3fd691 ac822b6 693ebf3 b3fd691 bc85fee b3fd691 ac822b6 693ebf3 b3fd691 693ebf3 b3fd691 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 |
---
license: other
tags:
- generated_from_trainer
model-index:
- name: segformer-b0-finetuned-segments-toolwear
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# segformer-b0-finetuned-segments-toolwear
This model is a fine-tuned version of [nvidia/mit-b0](https://huggingface.co./nvidia/mit-b0) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0354
- Mean Iou: 0.3022
- Mean Accuracy: 0.6045
- Overall Accuracy: 0.6045
- Accuracy Unlabeled: nan
- Accuracy Tool: nan
- Accuracy Wear: 0.6045
- Iou Unlabeled: 0.0
- Iou Tool: nan
- Iou Wear: 0.6045
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-05
- train_batch_size: 2
- eval_batch_size: 2
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Accuracy Unlabeled | Accuracy Tool | Accuracy Wear | Iou Unlabeled | Iou Tool | Iou Wear |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:------------------:|:-------------:|:-------------:|:-------------:|:--------:|:--------:|
| 0.8671 | 1.18 | 20 | 0.9263 | 0.4061 | 0.8122 | 0.8122 | nan | nan | 0.8122 | 0.0 | nan | 0.8122 |
| 0.5691 | 2.35 | 40 | 0.5998 | 0.2895 | 0.5790 | 0.5790 | nan | nan | 0.5790 | 0.0 | nan | 0.5790 |
| 0.4378 | 3.53 | 60 | 0.3948 | 0.3106 | 0.6213 | 0.6213 | nan | nan | 0.6213 | 0.0 | nan | 0.6213 |
| 0.3842 | 4.71 | 80 | 0.3190 | 0.2679 | 0.5357 | 0.5357 | nan | nan | 0.5357 | 0.0 | nan | 0.5357 |
| 0.3234 | 5.88 | 100 | 0.2883 | 0.3574 | 0.7148 | 0.7148 | nan | nan | 0.7148 | 0.0 | nan | 0.7148 |
| 0.2731 | 7.06 | 120 | 0.2392 | 0.3456 | 0.6911 | 0.6911 | nan | nan | 0.6911 | 0.0 | nan | 0.6911 |
| 0.2137 | 8.24 | 140 | 0.1850 | 0.1844 | 0.3688 | 0.3688 | nan | nan | 0.3688 | 0.0 | nan | 0.3688 |
| 0.1798 | 9.41 | 160 | 0.1692 | 0.2757 | 0.5515 | 0.5515 | nan | nan | 0.5515 | 0.0 | nan | 0.5515 |
| 0.1607 | 10.59 | 180 | 0.1338 | 0.2978 | 0.5956 | 0.5956 | nan | nan | 0.5956 | 0.0 | nan | 0.5956 |
| 0.1399 | 11.76 | 200 | 0.1218 | 0.2906 | 0.5811 | 0.5811 | nan | nan | 0.5811 | 0.0 | nan | 0.5811 |
| 0.1173 | 12.94 | 220 | 0.1030 | 0.2612 | 0.5224 | 0.5224 | nan | nan | 0.5224 | 0.0 | nan | 0.5224 |
| 0.0922 | 14.12 | 240 | 0.0976 | 0.2817 | 0.5633 | 0.5633 | nan | nan | 0.5633 | 0.0 | nan | 0.5633 |
| 0.081 | 15.29 | 260 | 0.0795 | 0.3154 | 0.6308 | 0.6308 | nan | nan | 0.6308 | 0.0 | nan | 0.6308 |
| 0.0852 | 16.47 | 280 | 0.0716 | 0.2188 | 0.4377 | 0.4377 | nan | nan | 0.4377 | 0.0 | nan | 0.4377 |
| 0.0709 | 17.65 | 300 | 0.0680 | 0.2691 | 0.5382 | 0.5382 | nan | nan | 0.5382 | 0.0 | nan | 0.5382 |
| 0.073 | 18.82 | 320 | 0.0611 | 0.2830 | 0.5660 | 0.5660 | nan | nan | 0.5660 | 0.0 | nan | 0.5660 |
| 0.0602 | 20.0 | 340 | 0.0592 | 0.2829 | 0.5657 | 0.5657 | nan | nan | 0.5657 | 0.0 | nan | 0.5657 |
| 0.0547 | 21.18 | 360 | 0.0577 | 0.2842 | 0.5684 | 0.5684 | nan | nan | 0.5684 | 0.0 | nan | 0.5684 |
| 0.0554 | 22.35 | 380 | 0.0537 | 0.2613 | 0.5226 | 0.5226 | nan | nan | 0.5226 | 0.0 | nan | 0.5226 |
| 0.0515 | 23.53 | 400 | 0.0523 | 0.3076 | 0.6152 | 0.6152 | nan | nan | 0.6152 | 0.0 | nan | 0.6152 |
| 0.0444 | 24.71 | 420 | 0.0487 | 0.3063 | 0.6126 | 0.6126 | nan | nan | 0.6126 | 0.0 | nan | 0.6126 |
| 0.088 | 25.88 | 440 | 0.0467 | 0.3041 | 0.6082 | 0.6082 | nan | nan | 0.6082 | 0.0 | nan | 0.6082 |
| 0.0472 | 27.06 | 460 | 0.0437 | 0.2623 | 0.5245 | 0.5245 | nan | nan | 0.5245 | 0.0 | nan | 0.5245 |
| 0.0396 | 28.24 | 480 | 0.0474 | 0.3352 | 0.6704 | 0.6704 | nan | nan | 0.6704 | 0.0 | nan | 0.6704 |
| 0.0351 | 29.41 | 500 | 0.0436 | 0.3060 | 0.6120 | 0.6120 | nan | nan | 0.6120 | 0.0 | nan | 0.6120 |
| 0.0392 | 30.59 | 520 | 0.0428 | 0.2975 | 0.5951 | 0.5951 | nan | nan | 0.5951 | 0.0 | nan | 0.5951 |
| 0.0317 | 31.76 | 540 | 0.0431 | 0.3253 | 0.6507 | 0.6507 | nan | nan | 0.6507 | 0.0 | nan | 0.6507 |
| 0.0391 | 32.94 | 560 | 0.0404 | 0.2863 | 0.5726 | 0.5726 | nan | nan | 0.5726 | 0.0 | nan | 0.5726 |
| 0.0309 | 34.12 | 580 | 0.0408 | 0.3215 | 0.6429 | 0.6429 | nan | nan | 0.6429 | 0.0 | nan | 0.6429 |
| 0.0493 | 35.29 | 600 | 0.0381 | 0.2581 | 0.5162 | 0.5162 | nan | nan | 0.5162 | 0.0 | nan | 0.5162 |
| 0.0321 | 36.47 | 620 | 0.0376 | 0.3147 | 0.6293 | 0.6293 | nan | nan | 0.6293 | 0.0 | nan | 0.6293 |
| 0.0333 | 37.65 | 640 | 0.0372 | 0.3118 | 0.6236 | 0.6236 | nan | nan | 0.6236 | 0.0 | nan | 0.6236 |
| 0.0295 | 38.82 | 660 | 0.0362 | 0.3036 | 0.6072 | 0.6072 | nan | nan | 0.6072 | 0.0 | nan | 0.6072 |
| 0.0302 | 40.0 | 680 | 0.0365 | 0.3157 | 0.6314 | 0.6314 | nan | nan | 0.6314 | 0.0 | nan | 0.6314 |
| 0.0272 | 41.18 | 700 | 0.0367 | 0.3012 | 0.6024 | 0.6024 | nan | nan | 0.6024 | 0.0 | nan | 0.6024 |
| 0.0278 | 42.35 | 720 | 0.0353 | 0.2935 | 0.5870 | 0.5870 | nan | nan | 0.5870 | 0.0 | nan | 0.5870 |
| 0.0283 | 43.53 | 740 | 0.0353 | 0.2970 | 0.5940 | 0.5940 | nan | nan | 0.5940 | 0.0 | nan | 0.5940 |
| 0.0256 | 44.71 | 760 | 0.0355 | 0.3090 | 0.6181 | 0.6181 | nan | nan | 0.6181 | 0.0 | nan | 0.6181 |
| 0.0365 | 45.88 | 780 | 0.0358 | 0.3008 | 0.6015 | 0.6015 | nan | nan | 0.6015 | 0.0 | nan | 0.6015 |
| 0.025 | 47.06 | 800 | 0.0353 | 0.2965 | 0.5930 | 0.5930 | nan | nan | 0.5930 | 0.0 | nan | 0.5930 |
| 0.0299 | 48.24 | 820 | 0.0361 | 0.3109 | 0.6219 | 0.6219 | nan | nan | 0.6219 | 0.0 | nan | 0.6219 |
| 0.0239 | 49.41 | 840 | 0.0354 | 0.3022 | 0.6045 | 0.6045 | nan | nan | 0.6045 | 0.0 | nan | 0.6045 |
### Framework versions
- Transformers 4.28.0
- Pytorch 2.1.0+cu121
- Datasets 2.16.0
- Tokenizers 0.13.3
|