amyeroberts HF staff commited on
Commit
c14adcb
1 Parent(s): e74ad6b

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,35 +1,64 @@
1
  {
2
- "_name_or_path": "google/vit-base-patch16-224-in21k",
 
3
  "architectures": [
4
- "ViTForImageClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.0,
7
- "encoder_stride": 16,
8
  "finetuning_task": "image-classification",
9
- "hidden_act": "gelu",
10
- "hidden_dropout_prob": 0.0,
11
- "hidden_size": 768,
12
  "id2label": {
13
  "0": "angular_leaf_spot",
14
  "1": "bean_rust",
15
  "2": "healthy"
16
  },
17
- "image_size": 224,
18
  "initializer_range": 0.02,
19
- "intermediate_size": 3072,
20
  "label2id": {
21
  "angular_leaf_spot": "0",
22
  "bean_rust": "1",
23
  "healthy": "2"
24
  },
25
- "layer_norm_eps": 1e-12,
26
- "model_type": "vit",
27
- "num_attention_heads": 12,
28
- "num_channels": 3,
29
- "num_hidden_layers": 12,
30
- "patch_size": 16,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
  "problem_type": "single_label_classification",
32
- "qkv_bias": true,
33
  "torch_dtype": "float32",
34
- "transformers_version": "4.27.0.dev0"
35
  }
 
1
  {
2
+ "_name_or_path": "timm/resnet18.a1_in1k",
3
+ "architecture": "resnet18",
4
  "architectures": [
5
+ "TimmWrapperForImageClassification"
6
  ],
 
 
7
  "finetuning_task": "image-classification",
 
 
 
8
  "id2label": {
9
  "0": "angular_leaf_spot",
10
  "1": "bean_rust",
11
  "2": "healthy"
12
  },
 
13
  "initializer_range": 0.02,
 
14
  "label2id": {
15
  "angular_leaf_spot": "0",
16
  "bean_rust": "1",
17
  "healthy": "2"
18
  },
19
+ "model_name": "timm/resnet18.a1_in1k",
20
+ "model_type": "timm_wrapper",
21
+ "num_classes": 1000,
22
+ "num_features": 512,
23
+ "pretrained_cfg": {
24
+ "classifier": "fc",
25
+ "crop_mode": "center",
26
+ "crop_pct": 0.95,
27
+ "custom_load": false,
28
+ "first_conv": "conv1",
29
+ "fixed_input_size": false,
30
+ "input_size": [
31
+ 3,
32
+ 224,
33
+ 224
34
+ ],
35
+ "interpolation": "bicubic",
36
+ "mean": [
37
+ 0.485,
38
+ 0.456,
39
+ 0.406
40
+ ],
41
+ "num_classes": 1000,
42
+ "origin_url": "https://github.com/huggingface/pytorch-image-models",
43
+ "paper_ids": "arXiv:2110.00476",
44
+ "pool_size": [
45
+ 7,
46
+ 7
47
+ ],
48
+ "std": [
49
+ 0.229,
50
+ 0.224,
51
+ 0.225
52
+ ],
53
+ "tag": "a1_in1k",
54
+ "test_crop_pct": 1.0,
55
+ "test_input_size": [
56
+ 3,
57
+ 288,
58
+ 288
59
+ ]
60
+ },
61
  "problem_type": "single_label_classification",
 
62
  "torch_dtype": "float32",
63
+ "transformers_version": "4.46.0.dev0"
64
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:043ec38d6470689b0104eea2df8cf79e57a270b12439d48886b6778561b9a248
3
+ size 44762892
preprocessor_config.json CHANGED
@@ -1,22 +1,64 @@
1
  {
2
- "do_normalize": true,
3
- "do_rescale": true,
4
- "do_resize": true,
5
- "image_mean": [
6
- 0.5,
7
- 0.5,
8
- 0.5
9
- ],
10
- "image_processor_type": "ViTImageProcessor",
11
- "image_std": [
12
- 0.5,
13
- 0.5,
14
- 0.5
15
- ],
16
- "resample": 2,
17
- "rescale_factor": 0.00392156862745098,
18
- "size": {
19
- "height": 224,
20
- "width": 224
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  }
22
  }
 
1
  {
2
+ "architecture": "resnet18",
3
+ "data_config": {
4
+ "crop_mode": "center",
5
+ "crop_pct": 0.95,
6
+ "input_size": [
7
+ 3,
8
+ 224,
9
+ 224
10
+ ],
11
+ "interpolation": "bicubic",
12
+ "mean": [
13
+ 0.485,
14
+ 0.456,
15
+ 0.406
16
+ ],
17
+ "std": [
18
+ 0.229,
19
+ 0.224,
20
+ 0.225
21
+ ]
22
+ },
23
+ "image_processor_type": "TimmWrapperImageProcessor",
24
+ "num_classes": 1000,
25
+ "num_features": 512,
26
+ "pretrained_cfg": {
27
+ "classifier": "fc",
28
+ "crop_mode": "center",
29
+ "crop_pct": 0.95,
30
+ "custom_load": false,
31
+ "first_conv": "conv1",
32
+ "fixed_input_size": false,
33
+ "input_size": [
34
+ 3,
35
+ 224,
36
+ 224
37
+ ],
38
+ "interpolation": "bicubic",
39
+ "mean": [
40
+ 0.485,
41
+ 0.456,
42
+ 0.406
43
+ ],
44
+ "num_classes": 1000,
45
+ "origin_url": "https://github.com/huggingface/pytorch-image-models",
46
+ "paper_ids": "arXiv:2110.00476",
47
+ "pool_size": [
48
+ 7,
49
+ 7
50
+ ],
51
+ "std": [
52
+ 0.229,
53
+ 0.224,
54
+ 0.225
55
+ ],
56
+ "tag": "a1_in1k",
57
+ "test_crop_pct": 1.0,
58
+ "test_input_size": [
59
+ 3,
60
+ 288,
61
+ 288
62
+ ]
63
  }
64
  }
runs/Sep30_20-11-09_Amys-MBP-2.communityfibre.co.uk/events.out.tfevents.1727723476.Amys-MBP-2.communityfibre.co.uk.11627.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0607107311eb2e18b77a08fca49be6fdacd3f8149596bbc2881eaf508454cd5f
3
+ size 8548
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b447ebb7685228fc4112bec8c2c3da722a5a15e27defbbb737e96a5cad7c00a
3
- size 3579
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15c6975d22c1fbaef774b8d87f3a9ce454a4951a9ba9333093626322e9867bd0
3
+ size 5304