|
--- |
|
license: other |
|
tags: |
|
- generated_from_trainer |
|
model-index: |
|
- name: segformer-b0-finetuned-segments-toolwear |
|
results: [] |
|
--- |
|
|
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You |
|
should probably proofread and complete it, then remove this comment. --> |
|
|
|
# segformer-b0-finetuned-segments-toolwear |
|
|
|
This model is a fine-tuned version of [nvidia/mit-b0](https://huggingface.co./nvidia/mit-b0) on an unknown dataset. |
|
It achieves the following results on the evaluation set: |
|
- Loss: 0.7517 |
|
- Mean Iou: 0.3530 |
|
- Mean Accuracy: 0.7066 |
|
- Overall Accuracy: 0.7444 |
|
- Accuracy Unlabeled: nan |
|
- Accuracy Tool: 0.6653 |
|
- Accuracy Wear: 0.7480 |
|
- Iou Unlabeled: 0.0 |
|
- Iou Tool: 0.3188 |
|
- Iou Wear: 0.7403 |
|
|
|
## Model description |
|
|
|
More information needed |
|
|
|
## Intended uses & limitations |
|
|
|
More information needed |
|
|
|
## Training and evaluation data |
|
|
|
More information needed |
|
|
|
## Training procedure |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- learning_rate: 6e-05 |
|
- train_batch_size: 2 |
|
- eval_batch_size: 2 |
|
- seed: 42 |
|
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 |
|
- lr_scheduler_type: linear |
|
- num_epochs: 50 |
|
|
|
### Training results |
|
|
|
| Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Accuracy Unlabeled | Accuracy Tool | Accuracy Wear | Iou Unlabeled | Iou Tool | Iou Wear | |
|
|:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:------------------:|:-------------:|:-------------:|:-------------:|:--------:|:--------:| |
|
| 0.8264 | 1.82 | 20 | 0.9929 | 0.3016 | 0.5119 | 0.6940 | nan | 0.3130 | 0.7109 | 0.0 | 0.2149 | 0.6899 | |
|
| 0.566 | 3.64 | 40 | 0.8390 | 0.3172 | 0.6658 | 0.6561 | nan | 0.6765 | 0.6552 | 0.0 | 0.3052 | 0.6466 | |
|
| 0.5515 | 5.45 | 60 | 0.7996 | 0.3015 | 0.7085 | 0.5831 | nan | 0.8455 | 0.5715 | 0.0 | 0.3365 | 0.5680 | |
|
| 0.496 | 7.27 | 80 | 0.7495 | 0.3370 | 0.7783 | 0.6771 | nan | 0.8889 | 0.6676 | 0.0 | 0.3465 | 0.6645 | |
|
| 0.4945 | 9.09 | 100 | 0.7214 | 0.3106 | 0.6966 | 0.6150 | nan | 0.7858 | 0.6074 | 0.0 | 0.3294 | 0.6025 | |
|
| 0.4392 | 10.91 | 120 | 0.7105 | 0.3012 | 0.7519 | 0.5990 | nan | 0.9191 | 0.5848 | 0.0 | 0.3198 | 0.5839 | |
|
| 0.3211 | 12.73 | 140 | 0.7570 | 0.3470 | 0.7008 | 0.7352 | nan | 0.6632 | 0.7384 | 0.0 | 0.3116 | 0.7292 | |
|
| 0.2289 | 14.55 | 160 | 0.9477 | 0.3748 | 0.7214 | 0.7566 | nan | 0.6830 | 0.7598 | 0.0 | 0.3718 | 0.7527 | |
|
| 0.4674 | 16.36 | 180 | 0.8172 | 0.3637 | 0.7442 | 0.7533 | nan | 0.7344 | 0.7541 | 0.0 | 0.3437 | 0.7476 | |
|
| 0.3226 | 18.18 | 200 | 0.8199 | 0.3238 | 0.7286 | 0.6845 | nan | 0.7769 | 0.6804 | 0.0 | 0.2939 | 0.6777 | |
|
| 0.1706 | 20.0 | 220 | 0.7336 | 0.3410 | 0.6894 | 0.7096 | nan | 0.6673 | 0.7115 | 0.0 | 0.3185 | 0.7044 | |
|
| 0.2786 | 21.82 | 240 | 0.9254 | 0.3662 | 0.7577 | 0.7864 | nan | 0.7264 | 0.7891 | 0.0 | 0.3164 | 0.7821 | |
|
| 0.1685 | 23.64 | 260 | 0.8291 | 0.3435 | 0.7685 | 0.7294 | nan | 0.8113 | 0.7258 | 0.0 | 0.3082 | 0.7224 | |
|
| 0.1649 | 25.45 | 280 | 0.7200 | 0.3303 | 0.7133 | 0.6593 | nan | 0.7723 | 0.6543 | 0.0 | 0.3394 | 0.6516 | |
|
| 0.1481 | 27.27 | 300 | 0.8155 | 0.3531 | 0.7558 | 0.7434 | nan | 0.7695 | 0.7422 | 0.0 | 0.3206 | 0.7385 | |
|
| 0.1476 | 29.09 | 320 | 0.7374 | 0.3455 | 0.6734 | 0.7252 | nan | 0.6169 | 0.7300 | 0.0 | 0.3153 | 0.7211 | |
|
| 0.2284 | 30.91 | 340 | 0.7254 | 0.3265 | 0.6989 | 0.6766 | nan | 0.7233 | 0.6745 | 0.0 | 0.3099 | 0.6695 | |
|
| 0.1212 | 32.73 | 360 | 0.8022 | 0.3591 | 0.7252 | 0.7662 | nan | 0.6804 | 0.7700 | 0.0 | 0.3153 | 0.7620 | |
|
| 0.1284 | 34.55 | 380 | 0.7345 | 0.3449 | 0.7044 | 0.7331 | nan | 0.6731 | 0.7357 | 0.0 | 0.3062 | 0.7284 | |
|
| 0.1685 | 36.36 | 400 | 0.7581 | 0.3275 | 0.7357 | 0.6991 | nan | 0.7757 | 0.6957 | 0.0 | 0.2910 | 0.6915 | |
|
| 0.1018 | 38.18 | 420 | 0.7303 | 0.3401 | 0.6575 | 0.7173 | nan | 0.5921 | 0.7228 | 0.0 | 0.3069 | 0.7133 | |
|
| 0.1405 | 40.0 | 440 | 0.7375 | 0.3555 | 0.7301 | 0.7475 | nan | 0.7111 | 0.7491 | 0.0 | 0.3234 | 0.7431 | |
|
| 0.08 | 41.82 | 460 | 0.7449 | 0.3561 | 0.7047 | 0.7457 | nan | 0.6598 | 0.7495 | 0.0 | 0.3265 | 0.7417 | |
|
| 0.1311 | 43.64 | 480 | 0.7680 | 0.3552 | 0.7205 | 0.7444 | nan | 0.6945 | 0.7466 | 0.0 | 0.3257 | 0.7398 | |
|
| 0.1235 | 45.45 | 500 | 0.7589 | 0.3523 | 0.7117 | 0.7398 | nan | 0.6811 | 0.7424 | 0.0 | 0.3218 | 0.7352 | |
|
| 0.1169 | 47.27 | 520 | 0.7676 | 0.3535 | 0.6952 | 0.7529 | nan | 0.6320 | 0.7583 | 0.0 | 0.3110 | 0.7494 | |
|
| 0.14 | 49.09 | 540 | 0.7517 | 0.3530 | 0.7066 | 0.7444 | nan | 0.6653 | 0.7480 | 0.0 | 0.3188 | 0.7403 | |
|
|
|
|
|
### Framework versions |
|
|
|
- Transformers 4.28.0 |
|
- Pytorch 2.0.1+cu118 |
|
- Datasets 2.14.5 |
|
- Tokenizers 0.13.3 |
|
|