Rausda6 commited on
Commit
dd29b1d
·
verified ·
1 Parent(s): b99e782

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - transformers
6
+ - image-classification
7
+ base_model: google/vit-large-patch16-224
8
+ widget:
9
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
10
+ example_title: Tiger
11
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
12
+ example_title: Teapot
13
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
14
+ example_title: Palace
15
+ ---
16
+
17
+ # Model Trained Using AutoTrain
18
+
19
+ - Problem type: Image Classification
20
+
21
+ ## Validation Metrics
22
+ loss: 1.0500575304031372
23
+
24
+ f1_macro: 0.611111111111111
25
+
26
+ f1_micro: 0.7142857142857143
27
+
28
+ f1_weighted: 0.619047619047619
29
+
30
+ precision_macro: 0.5833333333333334
31
+
32
+ precision_micro: 0.7142857142857143
33
+
34
+ precision_weighted: 0.5714285714285714
35
+
36
+ recall_macro: 0.6666666666666666
37
+
38
+ recall_micro: 0.7142857142857143
39
+
40
+ recall_weighted: 0.7142857142857143
41
+
42
+ accuracy: 0.7142857142857143
checkpoint-9/config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-large-patch16-224",
3
+ "_num_labels": 6,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 1024,
12
+ "id2label": {
13
+ "0": "blur",
14
+ "1": "dots",
15
+ "2": "good",
16
+ "3": "illum",
17
+ "4": "noise",
18
+ "5": "radblur"
19
+ },
20
+ "image_size": 224,
21
+ "initializer_range": 0.02,
22
+ "intermediate_size": 4096,
23
+ "label2id": {
24
+ "blur": 0,
25
+ "dots": 1,
26
+ "good": 2,
27
+ "illum": 3,
28
+ "noise": 4,
29
+ "radblur": 5
30
+ },
31
+ "layer_norm_eps": 1e-12,
32
+ "model_type": "vit",
33
+ "num_attention_heads": 16,
34
+ "num_channels": 3,
35
+ "num_hidden_layers": 24,
36
+ "patch_size": 16,
37
+ "problem_type": "single_label_classification",
38
+ "qkv_bias": true,
39
+ "torch_dtype": "float32",
40
+ "transformers_version": "4.48.0"
41
+ }
checkpoint-9/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aacb1ee67afdbab32a1d3d303fb2b3c1c9752aeed15e5a5cb92565e78eee194
3
+ size 1213277672
checkpoint-9/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e603c7c93a949806e57f50ab6960f7d9e1266ac72815d485852f4eb4fbb3e66
3
+ size 2426790314
checkpoint-9/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95b6047bd8cc6f4cdf7c46dea47edb8e542435510070c6cd1e0a7d9ccf5fd7da
3
+ size 14244
checkpoint-9/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe28afbf6131c4aaf26ff58140d1779aa47b73f2084ea00758f7b4ca1fc0182a
3
+ size 1064
checkpoint-9/trainer_state.json ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.0500575304031372,
3
+ "best_model_checkpoint": "autotrain-uo2t1-gvgzu/checkpoint-9",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 9,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.3333333333333333,
13
+ "grad_norm": 13.932000160217285,
14
+ "learning_rate": 5e-05,
15
+ "loss": 2.0226,
16
+ "step": 1
17
+ },
18
+ {
19
+ "epoch": 0.6666666666666666,
20
+ "grad_norm": 14.935710906982422,
21
+ "learning_rate": 4.375e-05,
22
+ "loss": 2.1974,
23
+ "step": 2
24
+ },
25
+ {
26
+ "epoch": 1.0,
27
+ "grad_norm": 14.412321090698242,
28
+ "learning_rate": 3.7500000000000003e-05,
29
+ "loss": 1.7189,
30
+ "step": 3
31
+ },
32
+ {
33
+ "epoch": 1.0,
34
+ "eval_accuracy": 0.5714285714285714,
35
+ "eval_f1_macro": 0.5666666666666667,
36
+ "eval_f1_micro": 0.5714285714285714,
37
+ "eval_f1_weighted": 0.4857142857142857,
38
+ "eval_loss": 1.58984375,
39
+ "eval_precision_macro": 0.5416666666666666,
40
+ "eval_precision_micro": 0.5714285714285714,
41
+ "eval_precision_weighted": 0.4642857142857143,
42
+ "eval_recall_macro": 0.6666666666666666,
43
+ "eval_recall_micro": 0.5714285714285714,
44
+ "eval_recall_weighted": 0.5714285714285714,
45
+ "eval_runtime": 0.0967,
46
+ "eval_samples_per_second": 72.423,
47
+ "eval_steps_per_second": 10.346,
48
+ "step": 3
49
+ },
50
+ {
51
+ "epoch": 1.3333333333333333,
52
+ "grad_norm": 14.967069625854492,
53
+ "learning_rate": 3.125e-05,
54
+ "loss": 1.4587,
55
+ "step": 4
56
+ },
57
+ {
58
+ "epoch": 1.6666666666666665,
59
+ "grad_norm": 13.65169620513916,
60
+ "learning_rate": 2.5e-05,
61
+ "loss": 1.9072,
62
+ "step": 5
63
+ },
64
+ {
65
+ "epoch": 2.0,
66
+ "grad_norm": 14.333626747131348,
67
+ "learning_rate": 1.8750000000000002e-05,
68
+ "loss": 1.4889,
69
+ "step": 6
70
+ },
71
+ {
72
+ "epoch": 2.0,
73
+ "eval_accuracy": 0.5714285714285714,
74
+ "eval_f1_macro": 0.42857142857142855,
75
+ "eval_f1_micro": 0.5714285714285714,
76
+ "eval_f1_weighted": 0.4489795918367347,
77
+ "eval_loss": 1.3249949216842651,
78
+ "eval_precision_macro": 0.39999999999999997,
79
+ "eval_precision_micro": 0.5714285714285714,
80
+ "eval_precision_weighted": 0.39999999999999997,
81
+ "eval_recall_macro": 0.5,
82
+ "eval_recall_micro": 0.5714285714285714,
83
+ "eval_recall_weighted": 0.5714285714285714,
84
+ "eval_runtime": 0.0922,
85
+ "eval_samples_per_second": 75.902,
86
+ "eval_steps_per_second": 10.843,
87
+ "step": 6
88
+ },
89
+ {
90
+ "epoch": 2.3333333333333335,
91
+ "grad_norm": 11.617460250854492,
92
+ "learning_rate": 1.25e-05,
93
+ "loss": 1.4087,
94
+ "step": 7
95
+ },
96
+ {
97
+ "epoch": 2.6666666666666665,
98
+ "grad_norm": 10.838333129882812,
99
+ "learning_rate": 6.25e-06,
100
+ "loss": 1.616,
101
+ "step": 8
102
+ },
103
+ {
104
+ "epoch": 3.0,
105
+ "grad_norm": 11.89291763305664,
106
+ "learning_rate": 0.0,
107
+ "loss": 1.6632,
108
+ "step": 9
109
+ },
110
+ {
111
+ "epoch": 3.0,
112
+ "eval_accuracy": 0.7142857142857143,
113
+ "eval_f1_macro": 0.611111111111111,
114
+ "eval_f1_micro": 0.7142857142857143,
115
+ "eval_f1_weighted": 0.619047619047619,
116
+ "eval_loss": 1.0500575304031372,
117
+ "eval_precision_macro": 0.5833333333333334,
118
+ "eval_precision_micro": 0.7142857142857143,
119
+ "eval_precision_weighted": 0.5714285714285714,
120
+ "eval_recall_macro": 0.6666666666666666,
121
+ "eval_recall_micro": 0.7142857142857143,
122
+ "eval_recall_weighted": 0.7142857142857143,
123
+ "eval_runtime": 0.0917,
124
+ "eval_samples_per_second": 76.369,
125
+ "eval_steps_per_second": 10.91,
126
+ "step": 9
127
+ }
128
+ ],
129
+ "logging_steps": 1,
130
+ "max_steps": 9,
131
+ "num_input_tokens_seen": 0,
132
+ "num_train_epochs": 3,
133
+ "save_steps": 500,
134
+ "stateful_callbacks": {
135
+ "EarlyStoppingCallback": {
136
+ "args": {
137
+ "early_stopping_patience": 5,
138
+ "early_stopping_threshold": 0.01
139
+ },
140
+ "attributes": {
141
+ "early_stopping_patience_counter": 0
142
+ }
143
+ },
144
+ "TrainerControl": {
145
+ "args": {
146
+ "should_epoch_stop": false,
147
+ "should_evaluate": false,
148
+ "should_log": false,
149
+ "should_save": true,
150
+ "should_training_stop": true
151
+ },
152
+ "attributes": {}
153
+ }
154
+ },
155
+ "total_flos": 1.972352756544307e+16,
156
+ "train_batch_size": 8,
157
+ "trial_name": null,
158
+ "trial_params": null
159
+ }
checkpoint-9/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92f24a37f95f800f8303cb8953c661537e5997df39bf02ad5b2509f12c8c33f0
3
+ size 5368
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-large-patch16-224",
3
+ "_num_labels": 6,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 1024,
12
+ "id2label": {
13
+ "0": "blur",
14
+ "1": "dots",
15
+ "2": "good",
16
+ "3": "illum",
17
+ "4": "noise",
18
+ "5": "radblur"
19
+ },
20
+ "image_size": 224,
21
+ "initializer_range": 0.02,
22
+ "intermediate_size": 4096,
23
+ "label2id": {
24
+ "blur": 0,
25
+ "dots": 1,
26
+ "good": 2,
27
+ "illum": 3,
28
+ "noise": 4,
29
+ "radblur": 5
30
+ },
31
+ "layer_norm_eps": 1e-12,
32
+ "model_type": "vit",
33
+ "num_attention_heads": 16,
34
+ "num_channels": 3,
35
+ "num_hidden_layers": 24,
36
+ "patch_size": 16,
37
+ "problem_type": "single_label_classification",
38
+ "qkv_bias": true,
39
+ "torch_dtype": "float32",
40
+ "transformers_version": "4.48.0"
41
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aacb1ee67afdbab32a1d3d303fb2b3c1c9752aeed15e5a5cb92565e78eee194
3
+ size 1213277672
preprocessor_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "ViTImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "resample": 2,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "height": 224,
21
+ "width": 224
22
+ }
23
+ }
runs/Apr07_07-47-29_r-rausda6-autotrain-advanced-wfgr8ym7-992e8-hvd5g/events.out.tfevents.1744012051.r-rausda6-autotrain-advanced-wfgr8ym7-992e8-hvd5g.100.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87bd4c95b635f346cb97d083a8ffcbd9c68cb6c454101a688445455a5f518ac6
3
- size 5241
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f49c87a7525185d41e426c17d11d9ac161c3388ee37378c666ce3104e8db6241
3
+ size 9906
runs/Apr07_07-47-29_r-rausda6-autotrain-advanced-wfgr8ym7-992e8-hvd5g/events.out.tfevents.1744012135.r-rausda6-autotrain-advanced-wfgr8ym7-992e8-hvd5g.100.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2a2686d2a345df4441852b5922596bb8b3b4bba2c284495251fc5d7809886d3
3
+ size 906
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92f24a37f95f800f8303cb8953c661537e5997df39bf02ad5b2509f12c8c33f0
3
+ size 5368
training_params.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "autotrain-uo2t1-gvgzu/autotrain-data",
3
+ "model": "google/vit-large-patch16-224",
4
+ "username": "Rausda6",
5
+ "lr": 5e-05,
6
+ "epochs": 3,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "autotrain-uo2t1-gvgzu",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": "fp16",
21
+ "save_total_limit": 1,
22
+ "push_to_hub": true,
23
+ "eval_strategy": "epoch",
24
+ "image_column": "autotrain_image",
25
+ "target_column": "autotrain_label",
26
+ "log": "tensorboard",
27
+ "early_stopping_patience": 5,
28
+ "early_stopping_threshold": 0.01
29
+ }