google-vit-base-patch16-224-batch64-lr0.005-standford-dogs
This model is a fine-tuned version of google/vit-base-patch16-224 on the stanford-dogs dataset. It achieves the following results on the evaluation set:
- Loss: 0.4279
- Accuracy: 0.8827
- F1: 0.8784
- Precision: 0.8844
- Recall: 0.8791
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 256
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 1000
Training results
Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
---|---|---|---|---|---|---|---|
4.7972 | 0.1550 | 10 | 4.5522 | 0.0510 | 0.0368 | 0.0394 | 0.0471 |
4.4634 | 0.3101 | 20 | 4.1231 | 0.1919 | 0.1378 | 0.1493 | 0.1771 |
4.0593 | 0.4651 | 30 | 3.6920 | 0.4014 | 0.3301 | 0.3884 | 0.3787 |
3.6865 | 0.6202 | 40 | 3.2802 | 0.5620 | 0.5020 | 0.5568 | 0.5395 |
3.3661 | 0.7752 | 50 | 2.9159 | 0.6489 | 0.6004 | 0.6552 | 0.6310 |
3.0631 | 0.9302 | 60 | 2.5874 | 0.7065 | 0.6721 | 0.7353 | 0.6925 |
2.7493 | 1.0853 | 70 | 2.3189 | 0.7320 | 0.7025 | 0.7660 | 0.7177 |
2.5223 | 1.2403 | 80 | 2.0780 | 0.7621 | 0.7376 | 0.7863 | 0.7497 |
2.3107 | 1.3953 | 90 | 1.8651 | 0.7760 | 0.7547 | 0.8037 | 0.7643 |
2.079 | 1.5504 | 100 | 1.6706 | 0.7952 | 0.7776 | 0.8150 | 0.7850 |
2.0001 | 1.7054 | 110 | 1.5130 | 0.8044 | 0.7880 | 0.8117 | 0.7951 |
1.8082 | 1.8605 | 120 | 1.3746 | 0.8144 | 0.8036 | 0.8295 | 0.8068 |
1.6836 | 2.0155 | 130 | 1.2598 | 0.8275 | 0.8146 | 0.8381 | 0.8200 |
1.5852 | 2.1705 | 140 | 1.1557 | 0.8311 | 0.8203 | 0.8400 | 0.8235 |
1.4695 | 2.3256 | 150 | 1.0706 | 0.8377 | 0.8290 | 0.8492 | 0.8303 |
1.3991 | 2.4806 | 160 | 1.0125 | 0.8426 | 0.8327 | 0.8526 | 0.8357 |
1.3486 | 2.6357 | 170 | 0.9519 | 0.8423 | 0.8331 | 0.8464 | 0.8364 |
1.3257 | 2.7907 | 180 | 0.9015 | 0.8467 | 0.8365 | 0.8517 | 0.8404 |
1.3175 | 2.9457 | 190 | 0.8607 | 0.8482 | 0.8403 | 0.8545 | 0.8424 |
1.2188 | 3.1008 | 200 | 0.8220 | 0.8494 | 0.8400 | 0.8561 | 0.8432 |
1.1733 | 3.2558 | 210 | 0.7847 | 0.8535 | 0.8471 | 0.8594 | 0.8483 |
1.1245 | 3.4109 | 220 | 0.7571 | 0.8523 | 0.8467 | 0.8586 | 0.8466 |
1.0503 | 3.5659 | 230 | 0.7358 | 0.8545 | 0.8459 | 0.8617 | 0.8492 |
1.0812 | 3.7209 | 240 | 0.7087 | 0.8554 | 0.8455 | 0.8622 | 0.8491 |
1.1002 | 3.8760 | 250 | 0.6906 | 0.8547 | 0.8469 | 0.8588 | 0.8490 |
1.0258 | 4.0310 | 260 | 0.6617 | 0.8690 | 0.8634 | 0.8756 | 0.8641 |
0.9731 | 4.1860 | 270 | 0.6541 | 0.8632 | 0.8549 | 0.8669 | 0.8577 |
0.9641 | 4.3411 | 280 | 0.6383 | 0.8630 | 0.8556 | 0.8686 | 0.8580 |
0.9656 | 4.4961 | 290 | 0.6161 | 0.8661 | 0.8594 | 0.8719 | 0.8611 |
0.9798 | 4.6512 | 300 | 0.6060 | 0.8652 | 0.8609 | 0.8703 | 0.8609 |
0.935 | 4.8062 | 310 | 0.5934 | 0.8649 | 0.8599 | 0.8694 | 0.8603 |
0.9218 | 4.9612 | 320 | 0.5911 | 0.8659 | 0.8621 | 0.8698 | 0.8614 |
0.9105 | 5.1163 | 330 | 0.5750 | 0.8669 | 0.8622 | 0.8689 | 0.8624 |
0.8954 | 5.2713 | 340 | 0.5639 | 0.8690 | 0.8630 | 0.8720 | 0.8644 |
0.8363 | 5.4264 | 350 | 0.5637 | 0.8705 | 0.8651 | 0.8714 | 0.8663 |
0.8548 | 5.5814 | 360 | 0.5581 | 0.8654 | 0.8599 | 0.8701 | 0.8607 |
0.7945 | 5.7364 | 370 | 0.5430 | 0.8681 | 0.8620 | 0.8692 | 0.8634 |
0.8321 | 5.8915 | 380 | 0.5394 | 0.8698 | 0.8645 | 0.8723 | 0.8654 |
0.8032 | 6.0465 | 390 | 0.5291 | 0.8763 | 0.8705 | 0.8776 | 0.8720 |
0.8116 | 6.2016 | 400 | 0.5252 | 0.8688 | 0.8634 | 0.8697 | 0.8647 |
0.7665 | 6.3566 | 410 | 0.5244 | 0.8717 | 0.8671 | 0.8739 | 0.8675 |
0.7807 | 6.5116 | 420 | 0.5148 | 0.8734 | 0.8692 | 0.8745 | 0.8694 |
0.7796 | 6.6667 | 430 | 0.5035 | 0.8734 | 0.8693 | 0.8761 | 0.8691 |
0.7669 | 6.8217 | 440 | 0.5016 | 0.8756 | 0.8698 | 0.8764 | 0.8715 |
0.78 | 6.9767 | 450 | 0.5031 | 0.8739 | 0.8686 | 0.8790 | 0.8696 |
0.7408 | 7.1318 | 460 | 0.4984 | 0.8717 | 0.8666 | 0.8800 | 0.8681 |
0.73 | 7.2868 | 470 | 0.4917 | 0.8737 | 0.8687 | 0.8761 | 0.8701 |
0.7057 | 7.4419 | 480 | 0.4912 | 0.8766 | 0.8706 | 0.8795 | 0.8725 |
0.7325 | 7.5969 | 490 | 0.4839 | 0.8795 | 0.8753 | 0.8841 | 0.8756 |
0.6938 | 7.7519 | 500 | 0.4840 | 0.8788 | 0.8755 | 0.8834 | 0.8756 |
0.7084 | 7.9070 | 510 | 0.4817 | 0.8744 | 0.8705 | 0.8783 | 0.8708 |
0.7342 | 8.0620 | 520 | 0.4761 | 0.8771 | 0.8735 | 0.8798 | 0.8741 |
0.6689 | 8.2171 | 530 | 0.4767 | 0.8746 | 0.8701 | 0.8788 | 0.8709 |
0.6857 | 8.3721 | 540 | 0.4768 | 0.8741 | 0.8701 | 0.8774 | 0.8703 |
0.694 | 8.5271 | 550 | 0.4723 | 0.8729 | 0.8683 | 0.8760 | 0.8688 |
0.6821 | 8.6822 | 560 | 0.4671 | 0.8763 | 0.8727 | 0.8795 | 0.8731 |
0.6752 | 8.8372 | 570 | 0.4618 | 0.8771 | 0.8724 | 0.8785 | 0.8733 |
0.7315 | 8.9922 | 580 | 0.4632 | 0.8768 | 0.8721 | 0.8791 | 0.8730 |
0.6561 | 9.1473 | 590 | 0.4552 | 0.8807 | 0.8765 | 0.8843 | 0.8768 |
0.6302 | 9.3023 | 600 | 0.4560 | 0.8793 | 0.8751 | 0.8822 | 0.8758 |
0.6376 | 9.4574 | 610 | 0.4586 | 0.8800 | 0.8757 | 0.8817 | 0.8769 |
0.6397 | 9.6124 | 620 | 0.4586 | 0.8776 | 0.8730 | 0.8797 | 0.8740 |
0.6883 | 9.7674 | 630 | 0.4532 | 0.8785 | 0.8740 | 0.8805 | 0.8748 |
0.614 | 9.9225 | 640 | 0.4571 | 0.8763 | 0.8722 | 0.8797 | 0.8728 |
0.6666 | 10.0775 | 650 | 0.4572 | 0.8761 | 0.8728 | 0.8801 | 0.8733 |
0.6014 | 10.2326 | 660 | 0.4493 | 0.8812 | 0.8770 | 0.8847 | 0.8775 |
0.6254 | 10.3876 | 670 | 0.4516 | 0.8776 | 0.8733 | 0.8808 | 0.8741 |
0.6449 | 10.5426 | 680 | 0.4435 | 0.8810 | 0.8765 | 0.8829 | 0.8774 |
0.6585 | 10.6977 | 690 | 0.4434 | 0.8829 | 0.8786 | 0.8854 | 0.8792 |
0.6371 | 10.8527 | 700 | 0.4409 | 0.8812 | 0.8774 | 0.8838 | 0.8776 |
0.6408 | 11.0078 | 710 | 0.4397 | 0.8844 | 0.8810 | 0.8867 | 0.8812 |
0.6098 | 11.1628 | 720 | 0.4407 | 0.8824 | 0.8783 | 0.8850 | 0.8788 |
0.5738 | 11.3178 | 730 | 0.4404 | 0.8793 | 0.8747 | 0.8811 | 0.8757 |
0.591 | 11.4729 | 740 | 0.4399 | 0.8822 | 0.8782 | 0.8836 | 0.8788 |
0.631 | 11.6279 | 750 | 0.4368 | 0.8812 | 0.8777 | 0.8838 | 0.8780 |
0.5467 | 11.7829 | 760 | 0.4363 | 0.8827 | 0.8792 | 0.8852 | 0.8796 |
0.6188 | 11.9380 | 770 | 0.4372 | 0.8817 | 0.8782 | 0.8845 | 0.8786 |
0.6116 | 12.0930 | 780 | 0.4368 | 0.8810 | 0.8778 | 0.8836 | 0.8779 |
0.5964 | 12.2481 | 790 | 0.4365 | 0.8814 | 0.8776 | 0.8841 | 0.8779 |
0.547 | 12.4031 | 800 | 0.4352 | 0.8785 | 0.8742 | 0.8797 | 0.8750 |
0.6151 | 12.5581 | 810 | 0.4331 | 0.8814 | 0.8779 | 0.8841 | 0.8784 |
0.5889 | 12.7132 | 820 | 0.4317 | 0.8819 | 0.8786 | 0.8850 | 0.8786 |
0.5662 | 12.8682 | 830 | 0.4301 | 0.8841 | 0.8811 | 0.8879 | 0.8810 |
0.5806 | 13.0233 | 840 | 0.4315 | 0.8805 | 0.8768 | 0.8834 | 0.8770 |
0.5863 | 13.1783 | 850 | 0.4291 | 0.8819 | 0.8778 | 0.8837 | 0.8787 |
0.5704 | 13.3333 | 860 | 0.4295 | 0.8824 | 0.8786 | 0.8845 | 0.8791 |
0.5879 | 13.4884 | 870 | 0.4293 | 0.8831 | 0.8797 | 0.8860 | 0.8797 |
0.5824 | 13.6434 | 880 | 0.4286 | 0.8822 | 0.8784 | 0.8845 | 0.8786 |
0.5525 | 13.7984 | 890 | 0.4289 | 0.8817 | 0.8776 | 0.8842 | 0.8780 |
0.5781 | 13.9535 | 900 | 0.4286 | 0.8824 | 0.8783 | 0.8845 | 0.8788 |
0.5929 | 14.1085 | 910 | 0.4282 | 0.8814 | 0.8777 | 0.8840 | 0.8779 |
0.5374 | 14.2636 | 920 | 0.4283 | 0.8819 | 0.8779 | 0.8840 | 0.8783 |
0.5691 | 14.4186 | 930 | 0.4297 | 0.8810 | 0.8765 | 0.8823 | 0.8774 |
0.5406 | 14.5736 | 940 | 0.4280 | 0.8810 | 0.8767 | 0.8825 | 0.8774 |
0.5387 | 14.7287 | 950 | 0.4274 | 0.8812 | 0.8771 | 0.8831 | 0.8778 |
0.5501 | 14.8837 | 960 | 0.4278 | 0.8822 | 0.8780 | 0.8841 | 0.8787 |
0.5729 | 15.0388 | 970 | 0.4280 | 0.8827 | 0.8783 | 0.8844 | 0.8791 |
0.5373 | 15.1938 | 980 | 0.4280 | 0.8831 | 0.8789 | 0.8849 | 0.8795 |
0.537 | 15.3488 | 990 | 0.4279 | 0.8827 | 0.8784 | 0.8844 | 0.8791 |
0.5463 | 15.5039 | 1000 | 0.4279 | 0.8827 | 0.8784 | 0.8844 | 0.8791 |
Framework versions
- Transformers 4.40.2
- Pytorch 2.3.0
- Datasets 2.19.1
- Tokenizers 0.19.1
- Downloads last month
- 218
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for amaye15/google-vit-base-patch16-224-batch64-lr0.005-standford-dogs
Base model
google/vit-base-patch16-224Evaluation results
- Accuracy on stanford-dogsself-reported0.883
- F1 on stanford-dogsself-reported0.878
- Precision on stanford-dogsself-reported0.884
- Recall on stanford-dogsself-reported0.879