File size: 1,523 Bytes
9d1992b
 
 
 
 
 
 
c03cfbd
 
c13ef30
 
 
 
 
 
 
 
949199b
c13ef30
 
f57fda5
 
 
 
 
c13ef30
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
---
library_name: transformers
tags: []
---

# Model Card for Model ID

This model is a finetuned image classificator based on [vit-base-patch16-224](https://huggingface.co./google/vit-base-patch16-224)
This model takes as input a picture from google maps' street view representing a road and returns a walkability score from 0 (worst score) to 4 (best score)

# How to Use
Load the model with the following code:
```python
from transformers import AutoModelForImageClassification
model = AutoModelForImageClassification.from_pretrained("AEnigmista/Sardegna-ViT", num_labels=5, ignore_mismatched_sizes=True)
```

For more information on the code: please visit the [github repo](https://github.com/MatteoMocci/Most-Walkability-AI)
# Training Hyper-parameters
This version's hyper-parameters for training are:
- Fp16 = True
- batch size = 32
- 10 epochs
- learning rate = 1e-4
- optimizer = 'adamw_hf'

# Metrics
The metrics that are used for evaluation are accuracy, recall, precision, mse, confusion matrix and a custom metric called one_out. The one_out_accuracy uses
the confusion matrix to check how many predictions of the model are within 1 from the ground truth (so label 2 is considered correct if ground truth is 1 or 3, incorrect if 0 or 5).
Since each label is actually a walkability score, this metric is useful to see how many predictions of the model are correct or pretty close to the expected value, and, thus,
how many predictions are way off (for example a street with 0 walkability score is predicted as 4)