franibm commited on
Commit
136beea
1 Parent(s): b06716b

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - image-classification
6
+ base_model: Organika/sdxl-detector
7
+ widget:
8
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
9
+ example_title: Tiger
10
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
11
+ example_title: Teapot
12
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
13
+ example_title: Palace
14
+ ---
15
+
16
+ # Model Trained Using AutoTrain
17
+
18
+ - Problem type: Image Classification
19
+
20
+ ## Validation Metrics
21
+ loss: 0.2632964849472046
22
+
23
+ f1: 0.9655172413793104
24
+
25
+ precision: 1.0
26
+
27
+ recall: 0.9333333333333333
28
+
29
+ auc: 0.9866666666666667
30
+
31
+ accuracy: 0.9666666666666667
checkpoint-18/config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Organika/sdxl-detector",
3
+ "_num_labels": 2,
4
+ "architectures": [
5
+ "SwinForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "depths": [
9
+ 2,
10
+ 2,
11
+ 18,
12
+ 2
13
+ ],
14
+ "drop_path_rate": 0.1,
15
+ "embed_dim": 128,
16
+ "encoder_stride": 32,
17
+ "hidden_act": "gelu",
18
+ "hidden_dropout_prob": 0.0,
19
+ "hidden_size": 1024,
20
+ "id2label": {
21
+ "0": "manipulated-Images",
22
+ "1": "non-manipulated-Images"
23
+ },
24
+ "image_size": 224,
25
+ "initializer_range": 0.02,
26
+ "label2id": {
27
+ "manipulated-Images": 0,
28
+ "non-manipulated-Images": 1
29
+ },
30
+ "layer_norm_eps": 1e-05,
31
+ "max_length": 128,
32
+ "mlp_ratio": 4.0,
33
+ "model_type": "swin",
34
+ "num_channels": 3,
35
+ "num_heads": [
36
+ 4,
37
+ 8,
38
+ 16,
39
+ 32
40
+ ],
41
+ "num_layers": 4,
42
+ "out_features": [
43
+ "stage4"
44
+ ],
45
+ "out_indices": [
46
+ 4
47
+ ],
48
+ "padding": "max_length",
49
+ "patch_size": 4,
50
+ "path_norm": true,
51
+ "problem_type": "single_label_classification",
52
+ "qkv_bias": true,
53
+ "stage_names": [
54
+ "stem",
55
+ "stage1",
56
+ "stage2",
57
+ "stage3",
58
+ "stage4"
59
+ ],
60
+ "torch_dtype": "float32",
61
+ "transformers_version": "4.44.2",
62
+ "use_absolute_embeddings": false,
63
+ "window_size": 7
64
+ }
checkpoint-18/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2536e96ee9b0571645bbc1af6bf0cd64e2e1df4cec1fc346c9ac817daaef03ad
3
+ size 347498816
checkpoint-18/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42de37e3d94d4e69948d93bc5efe1587d654a81a433ee2bb36e18705440ac258
3
+ size 694305741
checkpoint-18/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d88128180060408f5c5227c971931a5461d1c62bf0d020aca0a58f675285db3
3
+ size 13990
checkpoint-18/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0629355acb3346c7ab6602b74b10b2b769270cc4a37bc58ea47a3612e4b6122
3
+ size 1064
checkpoint-18/trainer_state.json ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.2632964849472046,
3
+ "best_model_checkpoint": "autotrain-Chiara2/checkpoint-18",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 18,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.1111111111111111,
13
+ "grad_norm": 364.52386474609375,
14
+ "learning_rate": 0.00016666666666666666,
15
+ "loss": 6.9638,
16
+ "step": 1
17
+ },
18
+ {
19
+ "epoch": 0.2222222222222222,
20
+ "grad_norm": 246.92811584472656,
21
+ "learning_rate": 0.0003333333333333333,
22
+ "loss": 4.562,
23
+ "step": 2
24
+ },
25
+ {
26
+ "epoch": 0.3333333333333333,
27
+ "grad_norm": 59.629798889160156,
28
+ "learning_rate": 0.0005,
29
+ "loss": 2.5174,
30
+ "step": 3
31
+ },
32
+ {
33
+ "epoch": 0.4444444444444444,
34
+ "grad_norm": 20.26913070678711,
35
+ "learning_rate": 0.0004791666666666667,
36
+ "loss": 0.8195,
37
+ "step": 4
38
+ },
39
+ {
40
+ "epoch": 0.5555555555555556,
41
+ "grad_norm": 14.140233993530273,
42
+ "learning_rate": 0.0004583333333333333,
43
+ "loss": 1.2522,
44
+ "step": 5
45
+ },
46
+ {
47
+ "epoch": 0.6666666666666666,
48
+ "grad_norm": 14.12821102142334,
49
+ "learning_rate": 0.0004375,
50
+ "loss": 1.3973,
51
+ "step": 6
52
+ },
53
+ {
54
+ "epoch": 0.7777777777777778,
55
+ "grad_norm": 10.597712516784668,
56
+ "learning_rate": 0.0004166666666666667,
57
+ "loss": 1.1615,
58
+ "step": 7
59
+ },
60
+ {
61
+ "epoch": 0.8888888888888888,
62
+ "grad_norm": 1.9146018028259277,
63
+ "learning_rate": 0.0003958333333333333,
64
+ "loss": 0.6808,
65
+ "step": 8
66
+ },
67
+ {
68
+ "epoch": 1.0,
69
+ "grad_norm": 0.8343021273612976,
70
+ "learning_rate": 0.000375,
71
+ "loss": 0.6365,
72
+ "step": 9
73
+ },
74
+ {
75
+ "epoch": 1.0,
76
+ "eval_accuracy": 0.5,
77
+ "eval_auc": 0.9822222222222223,
78
+ "eval_f1": 0.6666666666666666,
79
+ "eval_loss": 0.7654130458831787,
80
+ "eval_precision": 0.5,
81
+ "eval_recall": 1.0,
82
+ "eval_runtime": 11.1692,
83
+ "eval_samples_per_second": 2.686,
84
+ "eval_steps_per_second": 0.179,
85
+ "step": 9
86
+ },
87
+ {
88
+ "epoch": 1.1111111111111112,
89
+ "grad_norm": 2.730844020843506,
90
+ "learning_rate": 0.0003541666666666667,
91
+ "loss": 0.7581,
92
+ "step": 10
93
+ },
94
+ {
95
+ "epoch": 1.2222222222222223,
96
+ "grad_norm": 3.2819716930389404,
97
+ "learning_rate": 0.0003333333333333333,
98
+ "loss": 0.6775,
99
+ "step": 11
100
+ },
101
+ {
102
+ "epoch": 1.3333333333333333,
103
+ "grad_norm": 1.4362558126449585,
104
+ "learning_rate": 0.0003125,
105
+ "loss": 0.65,
106
+ "step": 12
107
+ },
108
+ {
109
+ "epoch": 1.4444444444444444,
110
+ "grad_norm": 2.433623790740967,
111
+ "learning_rate": 0.0002916666666666667,
112
+ "loss": 0.6205,
113
+ "step": 13
114
+ },
115
+ {
116
+ "epoch": 1.5555555555555556,
117
+ "grad_norm": 2.65726637840271,
118
+ "learning_rate": 0.0002708333333333333,
119
+ "loss": 0.6386,
120
+ "step": 14
121
+ },
122
+ {
123
+ "epoch": 1.6666666666666665,
124
+ "grad_norm": 5.401534557342529,
125
+ "learning_rate": 0.00025,
126
+ "loss": 0.7259,
127
+ "step": 15
128
+ },
129
+ {
130
+ "epoch": 1.7777777777777777,
131
+ "grad_norm": 2.8221542835235596,
132
+ "learning_rate": 0.00022916666666666666,
133
+ "loss": 0.4751,
134
+ "step": 16
135
+ },
136
+ {
137
+ "epoch": 1.8888888888888888,
138
+ "grad_norm": 4.188050270080566,
139
+ "learning_rate": 0.00020833333333333335,
140
+ "loss": 0.7153,
141
+ "step": 17
142
+ },
143
+ {
144
+ "epoch": 2.0,
145
+ "grad_norm": 6.374919891357422,
146
+ "learning_rate": 0.0001875,
147
+ "loss": 0.5568,
148
+ "step": 18
149
+ },
150
+ {
151
+ "epoch": 2.0,
152
+ "eval_accuracy": 0.9666666666666667,
153
+ "eval_auc": 0.9866666666666667,
154
+ "eval_f1": 0.9655172413793104,
155
+ "eval_loss": 0.2632964849472046,
156
+ "eval_precision": 1.0,
157
+ "eval_recall": 0.9333333333333333,
158
+ "eval_runtime": 10.6144,
159
+ "eval_samples_per_second": 2.826,
160
+ "eval_steps_per_second": 0.188,
161
+ "step": 18
162
+ }
163
+ ],
164
+ "logging_steps": 1,
165
+ "max_steps": 27,
166
+ "num_input_tokens_seen": 0,
167
+ "num_train_epochs": 3,
168
+ "save_steps": 500,
169
+ "stateful_callbacks": {
170
+ "EarlyStoppingCallback": {
171
+ "args": {
172
+ "early_stopping_patience": 5,
173
+ "early_stopping_threshold": 0.01
174
+ },
175
+ "attributes": {
176
+ "early_stopping_patience_counter": 0
177
+ }
178
+ },
179
+ "TrainerControl": {
180
+ "args": {
181
+ "should_epoch_stop": false,
182
+ "should_evaluate": false,
183
+ "should_log": false,
184
+ "should_save": true,
185
+ "should_training_stop": false
186
+ },
187
+ "attributes": {}
188
+ }
189
+ },
190
+ "total_flos": 1.096837778792448e+16,
191
+ "train_batch_size": 8,
192
+ "trial_name": null,
193
+ "trial_params": null
194
+ }
checkpoint-18/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2be25b38b707148b6d62ff0db410629ff1b803e3081e058131bed06097508989
3
+ size 5240
config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Organika/sdxl-detector",
3
+ "_num_labels": 2,
4
+ "architectures": [
5
+ "SwinForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "depths": [
9
+ 2,
10
+ 2,
11
+ 18,
12
+ 2
13
+ ],
14
+ "drop_path_rate": 0.1,
15
+ "embed_dim": 128,
16
+ "encoder_stride": 32,
17
+ "hidden_act": "gelu",
18
+ "hidden_dropout_prob": 0.0,
19
+ "hidden_size": 1024,
20
+ "id2label": {
21
+ "0": "manipulated-Images",
22
+ "1": "non-manipulated-Images"
23
+ },
24
+ "image_size": 224,
25
+ "initializer_range": 0.02,
26
+ "label2id": {
27
+ "manipulated-Images": 0,
28
+ "non-manipulated-Images": 1
29
+ },
30
+ "layer_norm_eps": 1e-05,
31
+ "max_length": 128,
32
+ "mlp_ratio": 4.0,
33
+ "model_type": "swin",
34
+ "num_channels": 3,
35
+ "num_heads": [
36
+ 4,
37
+ 8,
38
+ 16,
39
+ 32
40
+ ],
41
+ "num_layers": 4,
42
+ "out_features": [
43
+ "stage4"
44
+ ],
45
+ "out_indices": [
46
+ 4
47
+ ],
48
+ "padding": "max_length",
49
+ "patch_size": 4,
50
+ "path_norm": true,
51
+ "problem_type": "single_label_classification",
52
+ "qkv_bias": true,
53
+ "stage_names": [
54
+ "stem",
55
+ "stage1",
56
+ "stage2",
57
+ "stage3",
58
+ "stage4"
59
+ ],
60
+ "torch_dtype": "float32",
61
+ "transformers_version": "4.44.2",
62
+ "use_absolute_embeddings": false,
63
+ "window_size": 7
64
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2536e96ee9b0571645bbc1af6bf0cd64e2e1df4cec1fc346c9ac817daaef03ad
3
+ size 347498816
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.485,
7
+ 0.456,
8
+ 0.406
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.229,
13
+ 0.224,
14
+ 0.225
15
+ ],
16
+ "resample": 3,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
runs/Sep09_07-29-40_r-franibm-fraud-detection-4vdek1u4-c8466-vymxo/events.out.tfevents.1725866984.r-franibm-fraud-detection-4vdek1u4-c8466-vymxo.146.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2319156f6ad48a7c06454e1210f46a346ae01305f504e819fc8f9c7c5c381323
3
- size 5467
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce934dde4a85176c03a06762fa87ddbc78f7f0b24335013a6bbc605801387a6c
3
+ size 12931
runs/Sep09_07-29-40_r-franibm-fraud-detection-4vdek1u4-c8466-vymxo/events.out.tfevents.1725867265.r-franibm-fraud-detection-4vdek1u4-c8466-vymxo.146.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc63d4c21db5165e4a0cefb3d9dc6cda5855fe0da885752119461aa1dad37fb2
3
+ size 597
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2be25b38b707148b6d62ff0db410629ff1b803e3081e058131bed06097508989
3
+ size 5240
training_params.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "autotrain-Chiara2/autotrain-data",
3
+ "model": "Organika/sdxl-detector",
4
+ "username": "franibm",
5
+ "lr": 0.0005,
6
+ "epochs": 3,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "autotrain-Chiara2",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": "fp16",
21
+ "save_total_limit": 1,
22
+ "push_to_hub": true,
23
+ "eval_strategy": "epoch",
24
+ "image_column": "autotrain_image",
25
+ "target_column": "autotrain_label",
26
+ "log": "tensorboard",
27
+ "early_stopping_patience": 5,
28
+ "early_stopping_threshold": 0.01
29
+ }