kaijun123 commited on
Commit
94d16ce
·
verified ·
1 Parent(s): 2257f3f

Add add-1-extra-layer-batch-32-patience-5 weights

Browse files
add-1-extra-layer-batch-32-patience-5/epoch-16.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd228587929305f989cd37ec81dc714288072dbaf1f823b1f3503c01f13b6000
3
+ size 542070860
add-1-extra-layer-batch-32-patience-5/results.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "checkpoint/add-1-extra-layer-batch-32-patience-5/epoch-16.pt": {
3
+ "train_loss": [
4
+ 1.6538326740264893,
5
+ 0.6112778782844543,
6
+ 0.4547349810600281,
7
+ 0.3801289498806,
8
+ 0.3178323805332184,
9
+ 0.2622621953487396,
10
+ 0.25763607025146484,
11
+ 0.22383306920528412,
12
+ 0.2025839388370514,
13
+ 0.19861480593681335,
14
+ 0.1775054931640625,
15
+ 0.18190531432628632,
16
+ 0.18431280553340912,
17
+ 0.16741526126861572,
18
+ 0.16761712729930878,
19
+ 0.15633295476436615,
20
+ 0.14906401932239532
21
+ ],
22
+ "train_acc": [
23
+ 0.6047256097560976,
24
+ 0.8355182926829269,
25
+ 0.874390243902439,
26
+ 0.8867886178861789,
27
+ 0.9119410569105691,
28
+ 0.9226117886178863,
29
+ 0.924288617886179,
30
+ 0.9303353658536585,
31
+ 0.9360772357723577,
32
+ 0.9341971544715448,
33
+ 0.945630081300813,
34
+ 0.942428861788618,
35
+ 0.9380589430894308,
36
+ 0.943140243902439,
37
+ 0.9451727642276424,
38
+ 0.949491869918699,
39
+ 0.9493394308943088
40
+ ],
41
+ "val_loss": [
42
+ 0.7604890465736389,
43
+ 0.5971790552139282,
44
+ 0.5464081168174744,
45
+ 0.5170701146125793,
46
+ 0.5036472678184509,
47
+ 0.5031166672706604,
48
+ 0.4747341275215149,
49
+ 0.4623942971229553,
50
+ 0.4764615595340729,
51
+ 0.4960789084434509,
52
+ 0.46395400166511536,
53
+ 0.4391948878765106,
54
+ 0.4876585304737091,
55
+ 0.460769921541214,
56
+ 0.4802149534225464,
57
+ 0.47685402631759644,
58
+ 0.48240378499031067
59
+ ],
60
+ "val_acc": [
61
+ 0.8250534188034189,
62
+ 0.8623130341880342,
63
+ 0.8625801282051282,
64
+ 0.8707264957264957,
65
+ 0.8782051282051282,
66
+ 0.8697916666666666,
67
+ 0.8788728632478632,
68
+ 0.8872863247863247,
69
+ 0.8875534188034189,
70
+ 0.8728632478632478,
71
+ 0.8890224358974359,
72
+ 0.8918269230769231,
73
+ 0.8818108974358974,
74
+ 0.8824786324786325,
75
+ 0.8851495726495726,
76
+ 0.8950320512820512,
77
+ 0.8791399572649573
78
+ ]
79
+ }
80
+ }