GuntramG commited on
Commit
1f52f54
1 Parent(s): 00527f0

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - image-classification
6
+ base_model: facebook/convnextv2-tiny-1k-224
7
+ widget:
8
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
9
+ example_title: Tiger
10
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
11
+ example_title: Teapot
12
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
13
+ example_title: Palace
14
+ ---
15
+
16
+ # Model Trained Using AutoTrain
17
+
18
+ - Problem type: Image Classification
19
+
20
+ ## Validation Metrics
21
+ loss: 0.29826417565345764
22
+
23
+ f1_macro: 0.8819634453630242
24
+
25
+ f1_micro: 0.9117484831770546
26
+
27
+ f1_weighted: 0.9111571757544475
28
+
29
+ precision_macro: 0.9005385696343319
30
+
31
+ precision_micro: 0.9117484831770546
32
+
33
+ precision_weighted: 0.917903672857539
34
+
35
+ recall_macro: 0.872875172834113
36
+
37
+ recall_micro: 0.9117484831770546
38
+
39
+ recall_weighted: 0.9117484831770546
40
+
41
+ accuracy: 0.9117484831770546
checkpoint-1362/config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/convnextv2-tiny-1k-224",
3
+ "_num_labels": 6,
4
+ "architectures": [
5
+ "ConvNextV2ForImageClassification"
6
+ ],
7
+ "depths": [
8
+ 3,
9
+ 3,
10
+ 9,
11
+ 3
12
+ ],
13
+ "drop_path_rate": 0.0,
14
+ "hidden_act": "gelu",
15
+ "hidden_sizes": [
16
+ 96,
17
+ 192,
18
+ 384,
19
+ 768
20
+ ],
21
+ "id2label": {
22
+ "0": "food",
23
+ "1": "guru",
24
+ "2": "inside",
25
+ "3": "menu",
26
+ "4": "outside",
27
+ "5": "people"
28
+ },
29
+ "image_size": 224,
30
+ "initializer_range": 0.02,
31
+ "label2id": {
32
+ "food": 0,
33
+ "guru": 1,
34
+ "inside": 2,
35
+ "menu": 3,
36
+ "outside": 4,
37
+ "people": 5
38
+ },
39
+ "layer_norm_eps": 1e-12,
40
+ "model_type": "convnextv2",
41
+ "num_channels": 3,
42
+ "num_stages": 4,
43
+ "out_features": [
44
+ "stage4"
45
+ ],
46
+ "out_indices": [
47
+ 4
48
+ ],
49
+ "patch_size": 4,
50
+ "problem_type": "single_label_classification",
51
+ "stage_names": [
52
+ "stem",
53
+ "stage1",
54
+ "stage2",
55
+ "stage3",
56
+ "stage4"
57
+ ],
58
+ "torch_dtype": "float32",
59
+ "transformers_version": "4.44.2"
60
+ }
checkpoint-1362/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:950fbdef2ef7f2d0ecbbf2bfc309d5a9ae0097e41171579e89f8cbd87e6e7e19
3
+ size 111508128
checkpoint-1362/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:423602a4660af7f2aae1ab040474021497ccd5dbab299b77de4c75c42b814f28
3
+ size 223137210
checkpoint-1362/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce02224b2a82857db8f61afe7b20cb83aa24f31949d5cde44b26f427be1c0d01
3
+ size 14244
checkpoint-1362/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acf7c1213ec83c65808f9dc4a225d73d14cb2ec4df8eaeab3a4c93c4bb3a3918
3
+ size 1064
checkpoint-1362/trainer_state.json ADDED
@@ -0,0 +1,474 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.29826417565345764,
3
+ "best_model_checkpoint": "autotrain-nb4wh-9me5w/checkpoint-1362",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1362,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.05506607929515418,
13
+ "grad_norm": 17.699098587036133,
14
+ "learning_rate": 8.02919708029197e-06,
15
+ "loss": 1.789,
16
+ "step": 25
17
+ },
18
+ {
19
+ "epoch": 0.11013215859030837,
20
+ "grad_norm": 24.31856918334961,
21
+ "learning_rate": 1.715328467153285e-05,
22
+ "loss": 1.6122,
23
+ "step": 50
24
+ },
25
+ {
26
+ "epoch": 0.16519823788546256,
27
+ "grad_norm": 22.508787155151367,
28
+ "learning_rate": 2.6277372262773724e-05,
29
+ "loss": 1.3215,
30
+ "step": 75
31
+ },
32
+ {
33
+ "epoch": 0.22026431718061673,
34
+ "grad_norm": 20.886804580688477,
35
+ "learning_rate": 3.467153284671533e-05,
36
+ "loss": 1.0448,
37
+ "step": 100
38
+ },
39
+ {
40
+ "epoch": 0.2753303964757709,
41
+ "grad_norm": 38.399627685546875,
42
+ "learning_rate": 4.379562043795621e-05,
43
+ "loss": 0.7886,
44
+ "step": 125
45
+ },
46
+ {
47
+ "epoch": 0.3303964757709251,
48
+ "grad_norm": 26.71021842956543,
49
+ "learning_rate": 4.967346938775511e-05,
50
+ "loss": 0.5643,
51
+ "step": 150
52
+ },
53
+ {
54
+ "epoch": 0.3854625550660793,
55
+ "grad_norm": 17.157649993896484,
56
+ "learning_rate": 4.8653061224489796e-05,
57
+ "loss": 0.5184,
58
+ "step": 175
59
+ },
60
+ {
61
+ "epoch": 0.44052863436123346,
62
+ "grad_norm": 27.793960571289062,
63
+ "learning_rate": 4.763265306122449e-05,
64
+ "loss": 0.4071,
65
+ "step": 200
66
+ },
67
+ {
68
+ "epoch": 0.4955947136563877,
69
+ "grad_norm": 3.2827534675598145,
70
+ "learning_rate": 4.6612244897959185e-05,
71
+ "loss": 0.3605,
72
+ "step": 225
73
+ },
74
+ {
75
+ "epoch": 0.5506607929515418,
76
+ "grad_norm": 14.194032669067383,
77
+ "learning_rate": 4.559183673469388e-05,
78
+ "loss": 0.3437,
79
+ "step": 250
80
+ },
81
+ {
82
+ "epoch": 0.6057268722466961,
83
+ "grad_norm": 25.65274429321289,
84
+ "learning_rate": 4.4571428571428574e-05,
85
+ "loss": 0.5132,
86
+ "step": 275
87
+ },
88
+ {
89
+ "epoch": 0.6607929515418502,
90
+ "grad_norm": 13.928115844726562,
91
+ "learning_rate": 4.355102040816327e-05,
92
+ "loss": 0.2919,
93
+ "step": 300
94
+ },
95
+ {
96
+ "epoch": 0.7158590308370044,
97
+ "grad_norm": 14.263054847717285,
98
+ "learning_rate": 4.253061224489796e-05,
99
+ "loss": 0.3613,
100
+ "step": 325
101
+ },
102
+ {
103
+ "epoch": 0.7709251101321586,
104
+ "grad_norm": 63.43931198120117,
105
+ "learning_rate": 4.151020408163265e-05,
106
+ "loss": 0.1767,
107
+ "step": 350
108
+ },
109
+ {
110
+ "epoch": 0.8259911894273128,
111
+ "grad_norm": 13.402261734008789,
112
+ "learning_rate": 4.048979591836735e-05,
113
+ "loss": 0.4263,
114
+ "step": 375
115
+ },
116
+ {
117
+ "epoch": 0.8810572687224669,
118
+ "grad_norm": 80.55384063720703,
119
+ "learning_rate": 3.9469387755102045e-05,
120
+ "loss": 0.3597,
121
+ "step": 400
122
+ },
123
+ {
124
+ "epoch": 0.9361233480176211,
125
+ "grad_norm": 2.136183500289917,
126
+ "learning_rate": 3.844897959183674e-05,
127
+ "loss": 0.2739,
128
+ "step": 425
129
+ },
130
+ {
131
+ "epoch": 0.9911894273127754,
132
+ "grad_norm": 34.432132720947266,
133
+ "learning_rate": 3.742857142857143e-05,
134
+ "loss": 0.4018,
135
+ "step": 450
136
+ },
137
+ {
138
+ "epoch": 1.0,
139
+ "eval_accuracy": 0.8665195808052951,
140
+ "eval_f1_macro": 0.8283997321654719,
141
+ "eval_f1_micro": 0.8665195808052951,
142
+ "eval_f1_weighted": 0.8686614641516474,
143
+ "eval_loss": 0.3665294647216797,
144
+ "eval_precision_macro": 0.851578163810387,
145
+ "eval_precision_micro": 0.8665195808052951,
146
+ "eval_precision_weighted": 0.8862860589163636,
147
+ "eval_recall_macro": 0.8254227786088771,
148
+ "eval_recall_micro": 0.8665195808052951,
149
+ "eval_recall_weighted": 0.8665195808052951,
150
+ "eval_runtime": 24.5021,
151
+ "eval_samples_per_second": 147.987,
152
+ "eval_steps_per_second": 9.265,
153
+ "step": 454
154
+ },
155
+ {
156
+ "epoch": 1.0462555066079295,
157
+ "grad_norm": 32.53194046020508,
158
+ "learning_rate": 3.640816326530612e-05,
159
+ "loss": 0.3578,
160
+ "step": 475
161
+ },
162
+ {
163
+ "epoch": 1.1013215859030836,
164
+ "grad_norm": 12.507505416870117,
165
+ "learning_rate": 3.538775510204082e-05,
166
+ "loss": 0.283,
167
+ "step": 500
168
+ },
169
+ {
170
+ "epoch": 1.1563876651982379,
171
+ "grad_norm": 21.437583923339844,
172
+ "learning_rate": 3.436734693877551e-05,
173
+ "loss": 0.3169,
174
+ "step": 525
175
+ },
176
+ {
177
+ "epoch": 1.2114537444933922,
178
+ "grad_norm": 24.21172332763672,
179
+ "learning_rate": 3.3346938775510205e-05,
180
+ "loss": 0.2746,
181
+ "step": 550
182
+ },
183
+ {
184
+ "epoch": 1.2665198237885462,
185
+ "grad_norm": 16.129947662353516,
186
+ "learning_rate": 3.23265306122449e-05,
187
+ "loss": 0.224,
188
+ "step": 575
189
+ },
190
+ {
191
+ "epoch": 1.3215859030837005,
192
+ "grad_norm": 0.1952405422925949,
193
+ "learning_rate": 3.1306122448979594e-05,
194
+ "loss": 0.2148,
195
+ "step": 600
196
+ },
197
+ {
198
+ "epoch": 1.3766519823788546,
199
+ "grad_norm": 8.216381072998047,
200
+ "learning_rate": 3.0285714285714288e-05,
201
+ "loss": 0.33,
202
+ "step": 625
203
+ },
204
+ {
205
+ "epoch": 1.4317180616740088,
206
+ "grad_norm": 14.885648727416992,
207
+ "learning_rate": 2.9265306122448982e-05,
208
+ "loss": 0.2717,
209
+ "step": 650
210
+ },
211
+ {
212
+ "epoch": 1.4867841409691631,
213
+ "grad_norm": 40.765113830566406,
214
+ "learning_rate": 2.8244897959183673e-05,
215
+ "loss": 0.3455,
216
+ "step": 675
217
+ },
218
+ {
219
+ "epoch": 1.5418502202643172,
220
+ "grad_norm": 19.627338409423828,
221
+ "learning_rate": 2.7224489795918368e-05,
222
+ "loss": 0.262,
223
+ "step": 700
224
+ },
225
+ {
226
+ "epoch": 1.5969162995594712,
227
+ "grad_norm": 7.5291972160339355,
228
+ "learning_rate": 2.620408163265306e-05,
229
+ "loss": 0.4149,
230
+ "step": 725
231
+ },
232
+ {
233
+ "epoch": 1.6519823788546255,
234
+ "grad_norm": 18.426895141601562,
235
+ "learning_rate": 2.518367346938776e-05,
236
+ "loss": 0.2655,
237
+ "step": 750
238
+ },
239
+ {
240
+ "epoch": 1.7070484581497798,
241
+ "grad_norm": 24.276607513427734,
242
+ "learning_rate": 2.416326530612245e-05,
243
+ "loss": 0.401,
244
+ "step": 775
245
+ },
246
+ {
247
+ "epoch": 1.7621145374449338,
248
+ "grad_norm": 4.47360372543335,
249
+ "learning_rate": 2.3142857142857145e-05,
250
+ "loss": 0.2396,
251
+ "step": 800
252
+ },
253
+ {
254
+ "epoch": 1.8171806167400881,
255
+ "grad_norm": 89.49807739257812,
256
+ "learning_rate": 2.2122448979591836e-05,
257
+ "loss": 0.2951,
258
+ "step": 825
259
+ },
260
+ {
261
+ "epoch": 1.8722466960352424,
262
+ "grad_norm": 13.01455020904541,
263
+ "learning_rate": 2.110204081632653e-05,
264
+ "loss": 0.2654,
265
+ "step": 850
266
+ },
267
+ {
268
+ "epoch": 1.9273127753303965,
269
+ "grad_norm": 0.8529973030090332,
270
+ "learning_rate": 2.0081632653061225e-05,
271
+ "loss": 0.3684,
272
+ "step": 875
273
+ },
274
+ {
275
+ "epoch": 1.9823788546255505,
276
+ "grad_norm": 18.54584503173828,
277
+ "learning_rate": 1.906122448979592e-05,
278
+ "loss": 0.3434,
279
+ "step": 900
280
+ },
281
+ {
282
+ "epoch": 2.0,
283
+ "eval_accuracy": 0.8626585769442913,
284
+ "eval_f1_macro": 0.8149603266772006,
285
+ "eval_f1_micro": 0.8626585769442913,
286
+ "eval_f1_weighted": 0.862521320160242,
287
+ "eval_loss": 0.42735883593559265,
288
+ "eval_precision_macro": 0.864044899871371,
289
+ "eval_precision_micro": 0.8626585769442913,
290
+ "eval_precision_weighted": 0.8869813676514895,
291
+ "eval_recall_macro": 0.8017329444734634,
292
+ "eval_recall_micro": 0.8626585769442913,
293
+ "eval_recall_weighted": 0.8626585769442913,
294
+ "eval_runtime": 24.5304,
295
+ "eval_samples_per_second": 147.816,
296
+ "eval_steps_per_second": 9.254,
297
+ "step": 908
298
+ },
299
+ {
300
+ "epoch": 2.037444933920705,
301
+ "grad_norm": 32.09855651855469,
302
+ "learning_rate": 1.8040816326530614e-05,
303
+ "loss": 0.3931,
304
+ "step": 925
305
+ },
306
+ {
307
+ "epoch": 2.092511013215859,
308
+ "grad_norm": 12.878518104553223,
309
+ "learning_rate": 1.7020408163265305e-05,
310
+ "loss": 0.2092,
311
+ "step": 950
312
+ },
313
+ {
314
+ "epoch": 2.147577092511013,
315
+ "grad_norm": 6.512836933135986,
316
+ "learning_rate": 1.6000000000000003e-05,
317
+ "loss": 0.1755,
318
+ "step": 975
319
+ },
320
+ {
321
+ "epoch": 2.202643171806167,
322
+ "grad_norm": 13.462218284606934,
323
+ "learning_rate": 1.4979591836734694e-05,
324
+ "loss": 0.265,
325
+ "step": 1000
326
+ },
327
+ {
328
+ "epoch": 2.2577092511013217,
329
+ "grad_norm": 6.5825347900390625,
330
+ "learning_rate": 1.395918367346939e-05,
331
+ "loss": 0.3132,
332
+ "step": 1025
333
+ },
334
+ {
335
+ "epoch": 2.3127753303964758,
336
+ "grad_norm": 8.169792175292969,
337
+ "learning_rate": 1.2938775510204082e-05,
338
+ "loss": 0.2036,
339
+ "step": 1050
340
+ },
341
+ {
342
+ "epoch": 2.36784140969163,
343
+ "grad_norm": 0.07600528746843338,
344
+ "learning_rate": 1.1918367346938777e-05,
345
+ "loss": 0.2417,
346
+ "step": 1075
347
+ },
348
+ {
349
+ "epoch": 2.4229074889867843,
350
+ "grad_norm": 37.740264892578125,
351
+ "learning_rate": 1.089795918367347e-05,
352
+ "loss": 0.3434,
353
+ "step": 1100
354
+ },
355
+ {
356
+ "epoch": 2.4779735682819384,
357
+ "grad_norm": 0.7481518983840942,
358
+ "learning_rate": 9.877551020408164e-06,
359
+ "loss": 0.3904,
360
+ "step": 1125
361
+ },
362
+ {
363
+ "epoch": 2.5330396475770924,
364
+ "grad_norm": 19.178939819335938,
365
+ "learning_rate": 8.857142857142857e-06,
366
+ "loss": 0.2762,
367
+ "step": 1150
368
+ },
369
+ {
370
+ "epoch": 2.5881057268722465,
371
+ "grad_norm": 31.451202392578125,
372
+ "learning_rate": 7.836734693877551e-06,
373
+ "loss": 0.1749,
374
+ "step": 1175
375
+ },
376
+ {
377
+ "epoch": 2.643171806167401,
378
+ "grad_norm": 19.86487579345703,
379
+ "learning_rate": 6.857142857142858e-06,
380
+ "loss": 0.2828,
381
+ "step": 1200
382
+ },
383
+ {
384
+ "epoch": 2.698237885462555,
385
+ "grad_norm": 26.013954162597656,
386
+ "learning_rate": 5.8367346938775515e-06,
387
+ "loss": 0.2932,
388
+ "step": 1225
389
+ },
390
+ {
391
+ "epoch": 2.753303964757709,
392
+ "grad_norm": 9.863497734069824,
393
+ "learning_rate": 4.816326530612245e-06,
394
+ "loss": 0.1348,
395
+ "step": 1250
396
+ },
397
+ {
398
+ "epoch": 2.8083700440528636,
399
+ "grad_norm": 5.155836582183838,
400
+ "learning_rate": 3.7959183673469385e-06,
401
+ "loss": 0.1825,
402
+ "step": 1275
403
+ },
404
+ {
405
+ "epoch": 2.8634361233480177,
406
+ "grad_norm": 34.157859802246094,
407
+ "learning_rate": 2.775510204081633e-06,
408
+ "loss": 0.2959,
409
+ "step": 1300
410
+ },
411
+ {
412
+ "epoch": 2.9185022026431717,
413
+ "grad_norm": 18.83913803100586,
414
+ "learning_rate": 1.7551020408163264e-06,
415
+ "loss": 0.147,
416
+ "step": 1325
417
+ },
418
+ {
419
+ "epoch": 2.9735682819383262,
420
+ "grad_norm": 23.70765495300293,
421
+ "learning_rate": 7.346938775510204e-07,
422
+ "loss": 0.3109,
423
+ "step": 1350
424
+ },
425
+ {
426
+ "epoch": 3.0,
427
+ "eval_accuracy": 0.9117484831770546,
428
+ "eval_f1_macro": 0.8819634453630242,
429
+ "eval_f1_micro": 0.9117484831770546,
430
+ "eval_f1_weighted": 0.9111571757544475,
431
+ "eval_loss": 0.29826417565345764,
432
+ "eval_precision_macro": 0.9005385696343319,
433
+ "eval_precision_micro": 0.9117484831770546,
434
+ "eval_precision_weighted": 0.917903672857539,
435
+ "eval_recall_macro": 0.872875172834113,
436
+ "eval_recall_micro": 0.9117484831770546,
437
+ "eval_recall_weighted": 0.9117484831770546,
438
+ "eval_runtime": 24.4755,
439
+ "eval_samples_per_second": 148.148,
440
+ "eval_steps_per_second": 9.275,
441
+ "step": 1362
442
+ }
443
+ ],
444
+ "logging_steps": 25,
445
+ "max_steps": 1362,
446
+ "num_input_tokens_seen": 0,
447
+ "num_train_epochs": 3,
448
+ "save_steps": 500,
449
+ "stateful_callbacks": {
450
+ "EarlyStoppingCallback": {
451
+ "args": {
452
+ "early_stopping_patience": 5,
453
+ "early_stopping_threshold": 0.01
454
+ },
455
+ "attributes": {
456
+ "early_stopping_patience_counter": 0
457
+ }
458
+ },
459
+ "TrainerControl": {
460
+ "args": {
461
+ "should_epoch_stop": false,
462
+ "should_evaluate": false,
463
+ "should_log": false,
464
+ "should_save": true,
465
+ "should_training_stop": true
466
+ },
467
+ "attributes": {}
468
+ }
469
+ },
470
+ "total_flos": 2.7382422149074944e+17,
471
+ "train_batch_size": 8,
472
+ "trial_name": null,
473
+ "trial_params": null
474
+ }
checkpoint-1362/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69381925c578000fab5b86ca3b17154f6d67ee878e2c84197b673b90aff7fc37
3
+ size 5240
config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/convnextv2-tiny-1k-224",
3
+ "_num_labels": 6,
4
+ "architectures": [
5
+ "ConvNextV2ForImageClassification"
6
+ ],
7
+ "depths": [
8
+ 3,
9
+ 3,
10
+ 9,
11
+ 3
12
+ ],
13
+ "drop_path_rate": 0.0,
14
+ "hidden_act": "gelu",
15
+ "hidden_sizes": [
16
+ 96,
17
+ 192,
18
+ 384,
19
+ 768
20
+ ],
21
+ "id2label": {
22
+ "0": "food",
23
+ "1": "guru",
24
+ "2": "inside",
25
+ "3": "menu",
26
+ "4": "outside",
27
+ "5": "people"
28
+ },
29
+ "image_size": 224,
30
+ "initializer_range": 0.02,
31
+ "label2id": {
32
+ "food": 0,
33
+ "guru": 1,
34
+ "inside": 2,
35
+ "menu": 3,
36
+ "outside": 4,
37
+ "people": 5
38
+ },
39
+ "layer_norm_eps": 1e-12,
40
+ "model_type": "convnextv2",
41
+ "num_channels": 3,
42
+ "num_stages": 4,
43
+ "out_features": [
44
+ "stage4"
45
+ ],
46
+ "out_indices": [
47
+ 4
48
+ ],
49
+ "patch_size": 4,
50
+ "problem_type": "single_label_classification",
51
+ "stage_names": [
52
+ "stem",
53
+ "stage1",
54
+ "stage2",
55
+ "stage3",
56
+ "stage4"
57
+ ],
58
+ "torch_dtype": "float32",
59
+ "transformers_version": "4.44.2"
60
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:950fbdef2ef7f2d0ecbbf2bfc309d5a9ae0097e41171579e89f8cbd87e6e7e19
3
+ size 111508128
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
runs/Sep06_06-48-37_r-guntramg-autotrain-advanced-7pokmkts-78696-5v30l/events.out.tfevents.1725605319.r-guntramg-autotrain-advanced-7pokmkts-78696-5v30l.148.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a89ba7efb8340619e2accce077f2ca0d8a9536b04163e1cbff5c8208512f9d33
3
- size 5239
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13c1ba646c9654a1b23a4f14ab01146a1dfbf55379d3852bdfcb237bca81ea92
3
+ size 19466
runs/Sep06_06-48-37_r-guntramg-autotrain-advanced-7pokmkts-78696-5v30l/events.out.tfevents.1725605617.r-guntramg-autotrain-advanced-7pokmkts-78696-5v30l.148.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c42809226585c417b4b146287396652fef747f52c5ff7168bc8b372d3082408
3
+ size 921
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69381925c578000fab5b86ca3b17154f6d67ee878e2c84197b673b90aff7fc37
3
+ size 5240
training_params.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "autotrain-nb4wh-9me5w/autotrain-data",
3
+ "model": "facebook/convnextv2-tiny-1k-224",
4
+ "username": "GuntramG",
5
+ "lr": 5e-05,
6
+ "epochs": 3,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "autotrain-nb4wh-9me5w",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": "fp16",
21
+ "save_total_limit": 1,
22
+ "push_to_hub": true,
23
+ "eval_strategy": "epoch",
24
+ "image_column": "autotrain_image",
25
+ "target_column": "autotrain_label",
26
+ "log": "tensorboard",
27
+ "early_stopping_patience": 5,
28
+ "early_stopping_threshold": 0.01
29
+ }