kaijun123 commited on
Commit
aeba744
·
verified ·
1 Parent(s): b054178

Upload add-1-extra-layer weights part 3

Browse files
add-1-extra-layer-mixup-batch-128/epoch-1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faae425dd32d3084ba5f39137ab939c8ee7bfca6b809c390e22ff20ae0d72ebe
3
+ size 542075809
add-1-extra-layer-mixup-batch-128/results.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "checkpoint/add-1-extra-layer-mixup-batch-128/epoch-7.pt": {
3
+ "train_loss": [
4
+ 3.1375856399536133,
5
+ 2.5574276447296143,
6
+ 2.327446222305298,
7
+ 2.6292409896850586,
8
+ 2.0438895225524902,
9
+ 2.3215715885162354,
10
+ 2.319085121154785,
11
+ 2.208163261413574
12
+ ],
13
+ "train_acc": [
14
+ NaN,
15
+ NaN,
16
+ NaN,
17
+ NaN,
18
+ NaN,
19
+ NaN,
20
+ NaN,
21
+ NaN
22
+ ],
23
+ "val_loss": [
24
+ 2.5025672912597656,
25
+ 2.2167279720306396,
26
+ 2.224395990371704,
27
+ 1.9122543334960938,
28
+ 1.8773566484451294,
29
+ 2.318298578262329,
30
+ 2.186471939086914,
31
+ 2.23453426361084
32
+ ],
33
+ "val_acc": [
34
+ NaN,
35
+ NaN,
36
+ NaN,
37
+ NaN,
38
+ NaN,
39
+ NaN,
40
+ NaN,
41
+ NaN
42
+ ]
43
+ }
44
+ }
add-1-extra-layer-with-augmentation-batch-128/epoch-12.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0505be7d0a246de8cf6e6bd56d58d137aacc303b8ab183f7587afb26571305a
3
+ size 542076364
add-1-extra-layer-with-augmentation-batch-128/results.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "checkpoint/add-1-extra-layer-with-new-augmentation-batch-128/epoch-14.pt": {
3
+ "train_loss": [
4
+ 2.2322847843170166,
5
+ 0.7974888682365417,
6
+ 0.5749911665916443,
7
+ 0.4667174220085144,
8
+ 0.39886176586151123,
9
+ 0.3391243815422058,
10
+ 0.3057149648666382,
11
+ 0.2873183488845825,
12
+ 0.2537812292575836,
13
+ 0.2264219969511032,
14
+ 0.21699164807796478,
15
+ 0.18806761503219604,
16
+ 0.18919824063777924,
17
+ 0.18341554701328278,
18
+ 0.16843673586845398
19
+ ],
20
+ "train_acc": [
21
+ 0.5214342948717948,
22
+ 0.818359375,
23
+ 0.8731971153846154,
24
+ 0.889473157051282,
25
+ 0.904296875,
26
+ 0.9158153044871795,
27
+ 0.9221754807692307,
28
+ 0.9300881410256411,
29
+ 0.9399539262820512,
30
+ 0.9430588942307693,
31
+ 0.9459134615384616,
32
+ 0.9564302884615384,
33
+ 0.9548778044871795,
34
+ 0.9528245192307693,
35
+ 0.9569811698717949
36
+ ],
37
+ "val_loss": [
38
+ 1.050006628036499,
39
+ 0.7672672867774963,
40
+ 0.6427950859069824,
41
+ 0.5706731677055359,
42
+ 0.5588288903236389,
43
+ 0.4984721839427948,
44
+ 0.4937463104724884,
45
+ 0.4651919901371002,
46
+ 0.46598437428474426,
47
+ 0.47189193964004517,
48
+ 0.4518733024597168,
49
+ 0.4319031834602356,
50
+ 0.44288355112075806,
51
+ 0.4392522871494293,
52
+ 0.44642576575279236
53
+ ],
54
+ "val_acc": [
55
+ 0.7913839285714286,
56
+ 0.8460714285714286,
57
+ 0.8600892857142857,
58
+ 0.8836607142857142,
59
+ 0.8773214285714286,
60
+ 0.8933482142857143,
61
+ 0.8842857142857142,
62
+ 0.8829017857142858,
63
+ 0.8966964285714286,
64
+ 0.8773214285714286,
65
+ 0.8909821428571428,
66
+ 0.8987946428571428,
67
+ 0.8886160714285715,
68
+ 0.8925892857142858,
69
+ 0.8935714285714286
70
+ ]
71
+ }
72
+ }
add-1-extra-layer-with-augmentation-batch-32/epoch-14.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd8fd93e054020728d9b045a1a166b39b3896b53e8a9345456aa8fdf1c461c6e
3
+ size 542076364
add-1-extra-layer-with-augmentation-batch-32/results.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "checkpoint/add-1-extra-layer-with-new-augmentation-batch-32/epoch-14.pt": {
3
+ "train_loss": [
4
+ 1.6449568271636963,
5
+ 0.5982732772827148,
6
+ 0.4511142373085022,
7
+ 0.3671591281890869,
8
+ 0.29329684376716614,
9
+ 0.2738831341266632,
10
+ 0.24829310178756714,
11
+ 0.2246854454278946,
12
+ 0.20776110887527466,
13
+ 0.2026463747024536,
14
+ 0.19877706468105316,
15
+ 0.1738194227218628,
16
+ 0.1780676245689392,
17
+ 0.15821893513202667
18
+ ],
19
+ "train_acc": [
20
+ 0.620579268292683,
21
+ 0.8388719512195122,
22
+ 0.8737296747967479,
23
+ 0.8959349593495936,
24
+ 0.9127032520325205,
25
+ 0.9153963414634146,
26
+ 0.9245426829268293,
27
+ 0.9333333333333332,
28
+ 0.9340447154471546,
29
+ 0.9325203252032521,
30
+ 0.9381605691056911,
31
+ 0.9461890243902439,
32
+ 0.9397357723577234,
33
+ 0.949491869918699
34
+ ],
35
+ "val_loss": [
36
+ 0.7619845271110535,
37
+ 0.6232855916023254,
38
+ 0.5467333197593689,
39
+ 0.4960460066795349,
40
+ 0.48379069566726685,
41
+ 0.5179194211959839,
42
+ 0.47222471237182617,
43
+ 0.4744321405887604,
44
+ 0.4705529808998108,
45
+ 0.46792635321617126,
46
+ 0.4847172200679779,
47
+ 0.4650755524635315,
48
+ 0.4923202395439148,
49
+ 0.48628875613212585
50
+ ],
51
+ "val_acc": [
52
+ 0.8111645299145299,
53
+ 0.8454861111111112,
54
+ 0.8616452991452992,
55
+ 0.8728632478632478,
56
+ 0.8824786324786325,
57
+ 0.8664529914529915,
58
+ 0.8791399572649573,
59
+ 0.8799412393162392,
60
+ 0.8884882478632478,
61
+ 0.8863514957264957,
62
+ 0.8824786324786325,
63
+ 0.8815438034188035,
64
+ 0.8830128205128205,
65
+ 0.8896901709401709
66
+ ]
67
+ }
68
+ }
add-1-extra-layer-with-augmentation-batch-64-patience-5/epoch-25.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2328a7180e874bb31eeecc48481362da3d1d3e9d187c9a49406294339c8233c5
3
+ size 542070860
add-1-extra-layer-with-augmentation-batch-64-patience-5/results.json ADDED
@@ -0,0 +1,124 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "checkpoint/add-1-extra-layer-with-augmentation-batch-64-patience-5/epoch-27.pt": {
3
+ "train_loss": [
4
+ 2.4420857429504395,
5
+ 1.394033670425415,
6
+ 1.1942534446716309,
7
+ 1.0598857402801514,
8
+ 1.021833062171936,
9
+ 0.9981536865234375,
10
+ 0.9502733945846558,
11
+ 0.9285341501235962,
12
+ 0.9162173271179199,
13
+ 0.8687137961387634,
14
+ 0.8923584222793579,
15
+ 0.851234495639801,
16
+ 0.8546091914176941,
17
+ 0.842034101486206,
18
+ 0.8525927662849426,
19
+ 0.8396016955375671,
20
+ 0.8286609649658203,
21
+ 0.8080465793609619,
22
+ 0.7989516258239746,
23
+ 0.8347412943840027,
24
+ 0.8213275671005249,
25
+ 0.8117914199829102,
26
+ 0.8040000200271606,
27
+ 0.8193655610084534,
28
+ 0.7828842401504517,
29
+ 0.8075124025344849,
30
+ 0.8347918391227722,
31
+ 0.8170263171195984
32
+ ],
33
+ "train_acc": [
34
+ 0.4307241100323625,
35
+ 0.6312196601941747,
36
+ 0.6688410194174758,
37
+ 0.6947309870550162,
38
+ 0.7160699838187702,
39
+ 0.71875,
40
+ 0.7253741909385114,
41
+ 0.7401395631067961,
42
+ 0.7360942556634303,
43
+ 0.7441848705501618,
44
+ 0.7476739482200647,
45
+ 0.7515169902912622,
46
+ 0.7498483009708737,
47
+ 0.7551577669902912,
48
+ 0.7522754854368932,
49
+ 0.7549049352750808,
50
+ 0.7622876213592233,
51
+ 0.7667374595469255,
52
+ 0.7652204692556633,
53
+ 0.7571804207119741,
54
+ 0.7630461165048543,
55
+ 0.7595570388349514,
56
+ 0.7692152103559871,
57
+ 0.7590008090614886,
58
+ 0.7717940938511327,
59
+ 0.7675465210355987,
60
+ 0.7640068770226537,
61
+ 0.7680521844660194
62
+ ],
63
+ "val_loss": [
64
+ 1.5376410484313965,
65
+ 1.199355959892273,
66
+ 1.0828953981399536,
67
+ 0.9289830923080444,
68
+ 0.9297991991043091,
69
+ 0.838363766670227,
70
+ 0.8346579074859619,
71
+ 0.8298070430755615,
72
+ 0.8346342444419861,
73
+ 0.8001875281333923,
74
+ 0.7981680631637573,
75
+ 0.8030799627304077,
76
+ 0.7411221265792847,
77
+ 0.7895455956459045,
78
+ 0.7537394165992737,
79
+ 0.7421697378158569,
80
+ 0.7877919673919678,
81
+ 0.7363275289535522,
82
+ 0.7357286214828491,
83
+ 0.7229769825935364,
84
+ 0.718157172203064,
85
+ 0.7274062633514404,
86
+ 0.6983769536018372,
87
+ 0.7422466278076172,
88
+ 0.7319095134735107,
89
+ 0.7372193932533264,
90
+ 0.7254167795181274,
91
+ 0.7337458729743958
92
+ ],
93
+ "val_acc": [
94
+ 0.6275480769230769,
95
+ 0.7191826923076923,
96
+ 0.7185096153846153,
97
+ 0.77,
98
+ 0.7564423076923077,
99
+ 0.7760096153846153,
100
+ 0.7823557692307692,
101
+ 0.8000480769230769,
102
+ 0.7949038461538461,
103
+ 0.7892307692307692,
104
+ 0.7991826923076922,
105
+ 0.7914423076923077,
106
+ 0.8008653846153847,
107
+ 0.7842307692307693,
108
+ 0.8005288461538461,
109
+ 0.7993750000000001,
110
+ 0.7991826923076922,
111
+ 0.8053365384615384,
112
+ 0.806923076923077,
113
+ 0.8120192307692308,
114
+ 0.8104807692307693,
115
+ 0.8113461538461538,
116
+ 0.805673076923077,
117
+ 0.7976442307692307,
118
+ 0.8151442307692308,
119
+ 0.8065384615384615,
120
+ 0.7974519230769231,
121
+ 0.8065384615384615
122
+ ]
123
+ }
124
+ }
add-1-extra-layer-with-augmentation-batch-64/epoch-11.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f92c2e052af5a746b3042b7e2e7b06cf2cdbbffc154a7c6897b04439a9578ca
3
+ size 542076364
add-1-extra-layer-with-augmentation-batch-64/results.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "checkpoint/add-1-extra-layer-with-new-augmentation-batch-64/epoch-11.pt": {
3
+ "train_loss": [
4
+ 1.8546489477157593,
5
+ 0.6760998964309692,
6
+ 0.48841801285743713,
7
+ 0.40059995651245117,
8
+ 0.3372989296913147,
9
+ 0.3018609285354614,
10
+ 0.2741910219192505,
11
+ 0.2439292073249817,
12
+ 0.21093396842479706,
13
+ 0.20598289370536804,
14
+ 0.18278439342975616,
15
+ 0.16971677541732788
16
+ ],
17
+ "train_acc": [
18
+ 0.5833333333333334,
19
+ 0.83667071197411,
20
+ 0.8778822815533981,
21
+ 0.8962378640776699,
22
+ 0.9093851132686085,
23
+ 0.9154025080906149,
24
+ 0.9263248381877022,
25
+ 0.9307241100323624,
26
+ 0.9410396440129449,
27
+ 0.9414947411003236,
28
+ 0.946703074433657,
29
+ 0.9524676375404532
30
+ ],
31
+ "val_loss": [
32
+ 0.8411495685577393,
33
+ 0.6467218995094299,
34
+ 0.589273989200592,
35
+ 0.5351311564445496,
36
+ 0.5155400633811951,
37
+ 0.4804815948009491,
38
+ 0.47603434324264526,
39
+ 0.46770334243774414,
40
+ 0.433291494846344,
41
+ 0.442001610994339,
42
+ 0.43963420391082764,
43
+ 0.43448227643966675
44
+ ],
45
+ "val_acc": [
46
+ 0.8306250000000001,
47
+ 0.8539423076923077,
48
+ 0.8534134615384615,
49
+ 0.8628846153846154,
50
+ 0.8759134615384615,
51
+ 0.8786538461538461,
52
+ 0.877451923076923,
53
+ 0.8874038461538463,
54
+ 0.8901442307692308,
55
+ 0.891875,
56
+ 0.8951442307692308,
57
+ 0.885
58
+ ]
59
+ }
60
+ }