re-ran the tests

This commit is contained in:
2025-04-21 22:34:51 -04:00
parent 23899c703f
commit ebc64d766f
95 changed files with 2446 additions and 74 deletions
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.0048419479751587,0.26248,1.6749397346496582,0.3807
2,1.803226097946167,0.34018,1.5852585729598998,0.4277
3,1.7538686734390259,0.3595,1.5596578207015992,0.44
4,1.7227864822769166,0.36774,1.5332639671325683,0.4412
5,1.707428610610962,0.37698,1.5266790023803711,0.446
6,1.687287041053772,0.3853,1.5208749200820924,0.4534
7,1.6827863134002685,0.38752,1.4709485807418823,0.4678
8,1.6679418573379516,0.39552,1.4785521780014037,0.4569
9,1.6778243729400635,0.38932,1.4824026586532593,0.4535
10,1.661981136817932,0.39328,1.4527232948303224,0.4733
11,1.6635226394271851,0.39552,1.4467358991622925,0.4761
12,1.6477233642578124,0.39888,1.4165599590301514,0.4825
13,1.64288627204895,0.4028,1.4934470342636108,0.4554
14,1.64764131275177,0.40076,1.3920714778900147,0.4971
15,1.6471917071151734,0.39942,1.4170124416351317,0.4897
16,1.632711910057068,0.40306,1.437167017364502,0.4766
17,1.6346079250717163,0.40256,1.4856373008728028,0.4602
18,1.6376697371673583,0.40028,1.441266096687317,0.4795
19,1.630727212524414,0.40478,1.405259503364563,0.4869
20,1.6380860889434814,0.40302,1.500902590560913,0.4542
1 epoch train_loss train_acc test_loss test_acc
2 1 2.0048419479751587 0.26248 1.6749397346496582 0.3807
3 2 1.803226097946167 0.34018 1.5852585729598998 0.4277
4 3 1.7538686734390259 0.3595 1.5596578207015992 0.44
5 4 1.7227864822769166 0.36774 1.5332639671325683 0.4412
6 5 1.707428610610962 0.37698 1.5266790023803711 0.446
7 6 1.687287041053772 0.3853 1.5208749200820924 0.4534
8 7 1.6827863134002685 0.38752 1.4709485807418823 0.4678
9 8 1.6679418573379516 0.39552 1.4785521780014037 0.4569
10 9 1.6778243729400635 0.38932 1.4824026586532593 0.4535
11 10 1.661981136817932 0.39328 1.4527232948303224 0.4733
12 11 1.6635226394271851 0.39552 1.4467358991622925 0.4761
13 12 1.6477233642578124 0.39888 1.4165599590301514 0.4825
14 13 1.64288627204895 0.4028 1.4934470342636108 0.4554
15 14 1.64764131275177 0.40076 1.3920714778900147 0.4971
16 15 1.6471917071151734 0.39942 1.4170124416351317 0.4897
17 16 1.632711910057068 0.40306 1.437167017364502 0.4766
18 17 1.6346079250717163 0.40256 1.4856373008728028 0.4602
19 18 1.6376697371673583 0.40028 1.441266096687317 0.4795
20 19 1.630727212524414 0.40478 1.405259503364563 0.4869
21 20 1.6380860889434814 0.40302 1.500902590560913 0.4542
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,1.9924071334075928,0.2727,1.6629206789016724,0.3942
2,1.782769969177246,0.34778,1.6269580562591552,0.4111
3,1.7246631295776367,0.36938,1.512624101638794,0.4465
4,1.699350281906128,0.37616,1.561422179031372,0.4365
5,1.6983959941864013,0.37842,1.5700433086395265,0.4367
6,1.6805000713348388,0.38506,1.4903933980941773,0.4524
7,1.6756603227996827,0.39022,1.589101205253601,0.4264
8,1.6851841028594972,0.38372,1.53794698677063,0.4287
9,1.667366398010254,0.39178,1.489094132232666,0.4576
10,1.6613886919021605,0.39288,1.5648905296325684,0.4195
11,1.6709651499176026,0.38908,1.4680940185546876,0.4564
12,1.6781271075439452,0.38556,1.6323107133865356,0.3965
13,1.6682115633773804,0.38786,1.67539608707428,0.394
14,1.6545760455322265,0.39296,1.484992802810669,0.4598
15,1.664471877670288,0.39126,1.4973457614898682,0.4527
16,1.6631774948501588,0.39178,1.4977839450836181,0.4523
17,1.6620827933502198,0.39418,1.4580686128616334,0.4648
18,1.6693129856109619,0.3915,1.4670265132904052,0.4499
19,1.6571670418548583,0.39328,1.5775521347045898,0.4209
20,1.6555169297027588,0.39302,1.4923007045745849,0.45
1 epoch train_loss train_acc test_loss test_acc
2 1 1.9924071334075928 0.2727 1.6629206789016724 0.3942
3 2 1.782769969177246 0.34778 1.6269580562591552 0.4111
4 3 1.7246631295776367 0.36938 1.512624101638794 0.4465
5 4 1.699350281906128 0.37616 1.561422179031372 0.4365
6 5 1.6983959941864013 0.37842 1.5700433086395265 0.4367
7 6 1.6805000713348388 0.38506 1.4903933980941773 0.4524
8 7 1.6756603227996827 0.39022 1.589101205253601 0.4264
9 8 1.6851841028594972 0.38372 1.53794698677063 0.4287
10 9 1.667366398010254 0.39178 1.489094132232666 0.4576
11 10 1.6613886919021605 0.39288 1.5648905296325684 0.4195
12 11 1.6709651499176026 0.38908 1.4680940185546876 0.4564
13 12 1.6781271075439452 0.38556 1.6323107133865356 0.3965
14 13 1.6682115633773804 0.38786 1.67539608707428 0.394
15 14 1.6545760455322265 0.39296 1.484992802810669 0.4598
16 15 1.664471877670288 0.39126 1.4973457614898682 0.4527
17 16 1.6631774948501588 0.39178 1.4977839450836181 0.4523
18 17 1.6620827933502198 0.39418 1.4580686128616334 0.4648
19 18 1.6693129856109619 0.3915 1.4670265132904052 0.4499
20 19 1.6571670418548583 0.39328 1.5775521347045898 0.4209
21 20 1.6555169297027588 0.39302 1.4923007045745849 0.45
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.0259005422592162,0.26528,1.7835535663604736,0.3501
2,1.867787606163025,0.313,1.7164635316848755,0.3679
3,1.8262985235214233,0.33078,1.6985314193725587,0.3729
4,1.8109176992416383,0.33478,1.6716279788970947,0.3847
5,1.7937457250976563,0.34112,1.6525777326583861,0.3855
6,1.7862395941925049,0.3438,1.6767313753128053,0.3771
7,1.7756537421417236,0.34192,1.649860231399536,0.3899
8,1.7808906607055663,0.34136,1.642415731048584,0.3915
9,1.7794113501358033,0.33866,1.6533834308624267,0.3913
10,1.7733372591781615,0.34298,1.679184874534607,0.3706
11,1.7687813821411134,0.34438,1.6319498205184937,0.3919
12,1.7672993996810913,0.34524,1.5924089086532593,0.4055
13,1.7586094732284545,0.34774,1.5894474613189697,0.4183
14,1.7682015967178344,0.34308,1.6354400550842285,0.3904
15,1.7509896782684327,0.35068,1.6326687992095947,0.3796
16,1.7584374965667724,0.34796,1.633549732208252,0.3948
17,1.7580876070785523,0.35102,1.5947255432128906,0.4044
18,1.750145943069458,0.35058,1.5891325847625732,0.418
19,1.7573828769683837,0.3461,1.5960699188232421,0.4032
20,1.7578185536956787,0.3469,1.5921159523010253,0.4039
1 epoch train_loss train_acc test_loss test_acc
2 1 2.0259005422592162 0.26528 1.7835535663604736 0.3501
3 2 1.867787606163025 0.313 1.7164635316848755 0.3679
4 3 1.8262985235214233 0.33078 1.6985314193725587 0.3729
5 4 1.8109176992416383 0.33478 1.6716279788970947 0.3847
6 5 1.7937457250976563 0.34112 1.6525777326583861 0.3855
7 6 1.7862395941925049 0.3438 1.6767313753128053 0.3771
8 7 1.7756537421417236 0.34192 1.649860231399536 0.3899
9 8 1.7808906607055663 0.34136 1.642415731048584 0.3915
10 9 1.7794113501358033 0.33866 1.6533834308624267 0.3913
11 10 1.7733372591781615 0.34298 1.679184874534607 0.3706
12 11 1.7687813821411134 0.34438 1.6319498205184937 0.3919
13 12 1.7672993996810913 0.34524 1.5924089086532593 0.4055
14 13 1.7586094732284545 0.34774 1.5894474613189697 0.4183
15 14 1.7682015967178344 0.34308 1.6354400550842285 0.3904
16 15 1.7509896782684327 0.35068 1.6326687992095947 0.3796
17 16 1.7584374965667724 0.34796 1.633549732208252 0.3948
18 17 1.7580876070785523 0.35102 1.5947255432128906 0.4044
19 18 1.750145943069458 0.35058 1.5891325847625732 0.418
20 19 1.7573828769683837 0.3461 1.5960699188232421 0.4032
21 20 1.7578185536956787 0.3469 1.5921159523010253 0.4039
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,1.8579250992584229,0.32392,1.5558856401443482,0.4313
2,1.6377606032562255,0.4044,1.5286170751571655,0.4396
3,1.5881630603408814,0.41738,1.5185808382034303,0.4484
4,1.5531442667388915,0.43602,1.4579217964172364,0.4586
5,1.5474866276931762,0.43182,1.45462444896698,0.4727
6,1.5150492757415772,0.44956,1.4247750133514405,0.475
7,1.5141647922515868,0.44442,1.4379277662277221,0.4691
8,1.5064603635406495,0.44778,1.4320051357269288,0.4814
9,1.4982417543029785,0.45322,1.4486036586761475,0.4788
10,1.495431781539917,0.45466,1.4063307929992677,0.4955
11,1.477474406814575,0.46004,1.3848083011627197,0.5074
12,1.478807368774414,0.46202,1.4433747045516967,0.4806
13,1.481698868713379,0.45994,1.4207568691253663,0.4875
14,1.4778631386184693,0.46108,1.3489213497161865,0.5107
15,1.4712807468032838,0.46604,1.3919997245788573,0.4893
16,1.483530824584961,0.46112,1.4437157917022705,0.4853
17,1.4729563687133789,0.4616,1.3912005447387696,0.496
18,1.4708208094787598,0.46484,1.4403230434417724,0.4778
19,1.469524460105896,0.46858,1.3713474117279052,0.5112
20,1.4658287371444703,0.46598,1.4386186960220337,0.4857
1 epoch train_loss train_acc test_loss test_acc
2 1 1.8579250992584229 0.32392 1.5558856401443482 0.4313
3 2 1.6377606032562255 0.4044 1.5286170751571655 0.4396
4 3 1.5881630603408814 0.41738 1.5185808382034303 0.4484
5 4 1.5531442667388915 0.43602 1.4579217964172364 0.4586
6 5 1.5474866276931762 0.43182 1.45462444896698 0.4727
7 6 1.5150492757415772 0.44956 1.4247750133514405 0.475
8 7 1.5141647922515868 0.44442 1.4379277662277221 0.4691
9 8 1.5064603635406495 0.44778 1.4320051357269288 0.4814
10 9 1.4982417543029785 0.45322 1.4486036586761475 0.4788
11 10 1.495431781539917 0.45466 1.4063307929992677 0.4955
12 11 1.477474406814575 0.46004 1.3848083011627197 0.5074
13 12 1.478807368774414 0.46202 1.4433747045516967 0.4806
14 13 1.481698868713379 0.45994 1.4207568691253663 0.4875
15 14 1.4778631386184693 0.46108 1.3489213497161865 0.5107
16 15 1.4712807468032838 0.46604 1.3919997245788573 0.4893
17 16 1.483530824584961 0.46112 1.4437157917022705 0.4853
18 17 1.4729563687133789 0.4616 1.3912005447387696 0.496
19 18 1.4708208094787598 0.46484 1.4403230434417724 0.4778
20 19 1.469524460105896 0.46858 1.3713474117279052 0.5112
21 20 1.4658287371444703 0.46598 1.4386186960220337 0.4857
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,1.7651092416763305,0.34864,1.4801026742935182,0.4452
2,1.4773385808563233,0.4608,1.4107468299865722,0.4928
3,1.4245695472717286,0.48826,1.3724896883010864,0.5095
4,1.3900219127273559,0.49736,1.2948982368469237,0.5366
5,1.3694689385223389,0.50896,1.2803543266296387,0.5367
6,1.3547902798843383,0.51456,1.2770936399459838,0.5403
7,1.3447893646621705,0.51488,1.2862837057113647,0.5363
8,1.3417145071411132,0.519,1.2915318742752075,0.5294
9,1.338515231552124,0.52032,1.247458812904358,0.5552
10,1.3313588520050048,0.52276,1.23068475151062,0.5559
11,1.3147309407806396,0.52892,1.2065896244049072,0.5765
12,1.3206113652801514,0.52802,1.2014692079544067,0.5784
13,1.3099163648223877,0.52914,1.219766689491272,0.5604
14,1.3167620611953736,0.52754,1.1866582777023316,0.5766
15,1.3133597641754151,0.53076,1.2739620433807373,0.5413
16,1.3288053150939942,0.5251,1.280159574699402,0.5474
17,1.3213993626785279,0.5305,1.3356346948623656,0.5339
18,1.3234788956451415,0.5274,1.2488118535995483,0.5621
19,1.3162768488693237,0.53068,1.1890131019592285,0.5786
20,1.3078575289154053,0.53372,1.2257861446380616,0.5658
1 epoch train_loss train_acc test_loss test_acc
2 1 1.7651092416763305 0.34864 1.4801026742935182 0.4452
3 2 1.4773385808563233 0.4608 1.4107468299865722 0.4928
4 3 1.4245695472717286 0.48826 1.3724896883010864 0.5095
5 4 1.3900219127273559 0.49736 1.2948982368469237 0.5366
6 5 1.3694689385223389 0.50896 1.2803543266296387 0.5367
7 6 1.3547902798843383 0.51456 1.2770936399459838 0.5403
8 7 1.3447893646621705 0.51488 1.2862837057113647 0.5363
9 8 1.3417145071411132 0.519 1.2915318742752075 0.5294
10 9 1.338515231552124 0.52032 1.247458812904358 0.5552
11 10 1.3313588520050048 0.52276 1.23068475151062 0.5559
12 11 1.3147309407806396 0.52892 1.2065896244049072 0.5765
13 12 1.3206113652801514 0.52802 1.2014692079544067 0.5784
14 13 1.3099163648223877 0.52914 1.219766689491272 0.5604
15 14 1.3167620611953736 0.52754 1.1866582777023316 0.5766
16 15 1.3133597641754151 0.53076 1.2739620433807373 0.5413
17 16 1.3288053150939942 0.5251 1.280159574699402 0.5474
18 17 1.3213993626785279 0.5305 1.3356346948623656 0.5339
19 18 1.3234788956451415 0.5274 1.2488118535995483 0.5621
20 19 1.3162768488693237 0.53068 1.1890131019592285 0.5786
21 20 1.3078575289154053 0.53372 1.2257861446380616 0.5658
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,1.8653128273773194,0.32796,1.611533346939087,0.3995
2,1.612140337791443,0.41292,1.623063315963745,0.4098
3,1.5402255586242677,0.44038,1.4548933393478394,0.4697
4,1.5275923446655273,0.44896,1.440786859703064,0.4802
5,1.4979061844635009,0.45434,1.417161144065857,0.4818
6,1.4915729135131837,0.46008,1.3997222789764405,0.4988
7,1.4728927141571044,0.46468,1.387175965309143,0.4972
8,1.4769946060943604,0.46384,1.466834031867981,0.4687
9,1.4653232444000244,0.46896,1.4313162572860718,0.4816
10,1.4634774406433106,0.47178,1.3978804470062256,0.4935
11,1.464304472732544,0.47028,1.4218554107666015,0.4782
12,1.4602648656463624,0.47172,1.4064278043746947,0.4873
13,1.4479452781295776,0.47628,1.4901286466598511,0.4706
14,1.4542893864059447,0.47666,1.3517237024307251,0.5077
15,1.4511225904846192,0.4727,1.3578523971557617,0.5074
16,1.4690848288726808,0.47104,1.3588831977844238,0.5053
17,1.4494867483520508,0.47718,1.3755928480148316,0.5045
18,1.460067735671997,0.4717,1.3654316268920899,0.5149
19,1.4619887506103515,0.47406,1.3814095928192138,0.5031
20,1.4534882698059082,0.47412,1.34085482711792,0.5189
1 epoch train_loss train_acc test_loss test_acc
2 1 1.8653128273773194 0.32796 1.611533346939087 0.3995
3 2 1.612140337791443 0.41292 1.623063315963745 0.4098
4 3 1.5402255586242677 0.44038 1.4548933393478394 0.4697
5 4 1.5275923446655273 0.44896 1.440786859703064 0.4802
6 5 1.4979061844635009 0.45434 1.417161144065857 0.4818
7 6 1.4915729135131837 0.46008 1.3997222789764405 0.4988
8 7 1.4728927141571044 0.46468 1.387175965309143 0.4972
9 8 1.4769946060943604 0.46384 1.466834031867981 0.4687
10 9 1.4653232444000244 0.46896 1.4313162572860718 0.4816
11 10 1.4634774406433106 0.47178 1.3978804470062256 0.4935
12 11 1.464304472732544 0.47028 1.4218554107666015 0.4782
13 12 1.4602648656463624 0.47172 1.4064278043746947 0.4873
14 13 1.4479452781295776 0.47628 1.4901286466598511 0.4706
15 14 1.4542893864059447 0.47666 1.3517237024307251 0.5077
16 15 1.4511225904846192 0.4727 1.3578523971557617 0.5074
17 16 1.4690848288726808 0.47104 1.3588831977844238 0.5053
18 17 1.4494867483520508 0.47718 1.3755928480148316 0.5045
19 18 1.460067735671997 0.4717 1.3654316268920899 0.5149
20 19 1.4619887506103515 0.47406 1.3814095928192138 0.5031
21 20 1.4534882698059082 0.47412 1.34085482711792 0.5189
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.0278418547058106,0.24944,1.8780883220672608,0.3251
2,1.8143817639541626,0.331,1.6300555017471314,0.4045
3,1.7264616962432862,0.3622,1.597921866607666,0.4111
4,1.6720347478866577,0.38138,1.608277184677124,0.41
5,1.6579552675628662,0.3885,1.5234451871871948,0.435
6,1.6298340225982666,0.39974,1.5039015756607055,0.4597
7,1.6189628750610352,0.40322,1.5221681720733642,0.4422
8,1.6198538191986085,0.40798,1.4737193891525269,0.4565
9,1.6145962512969971,0.4075,1.466329651069641,0.4629
10,1.621643498878479,0.40454,1.6072639640808106,0.4081
11,1.6055096750259399,0.40826,1.527749220275879,0.4358
12,1.6175734860229491,0.40692,1.4493362300872803,0.4722
13,1.599906416091919,0.4124,1.4468452342987062,0.4715
14,1.5933954864883424,0.41398,1.4764380693435668,0.4626
15,1.5933787399291992,0.41372,1.533870811843872,0.4468
16,1.5964966577911377,0.41302,1.4542985107421875,0.467
17,1.6135705094146728,0.40712,1.4721424545288087,0.4527
18,1.5964763764190675,0.41558,1.4527687072753905,0.4593
19,1.591530082244873,0.41652,1.4388617012023925,0.4642
20,1.589130754928589,0.41602,1.4646002044677735,0.4536
1 epoch train_loss train_acc test_loss test_acc
2 1 2.0278418547058106 0.24944 1.8780883220672608 0.3251
3 2 1.8143817639541626 0.331 1.6300555017471314 0.4045
4 3 1.7264616962432862 0.3622 1.597921866607666 0.4111
5 4 1.6720347478866577 0.38138 1.608277184677124 0.41
6 5 1.6579552675628662 0.3885 1.5234451871871948 0.435
7 6 1.6298340225982666 0.39974 1.5039015756607055 0.4597
8 7 1.6189628750610352 0.40322 1.5221681720733642 0.4422
9 8 1.6198538191986085 0.40798 1.4737193891525269 0.4565
10 9 1.6145962512969971 0.4075 1.466329651069641 0.4629
11 10 1.621643498878479 0.40454 1.6072639640808106 0.4081
12 11 1.6055096750259399 0.40826 1.527749220275879 0.4358
13 12 1.6175734860229491 0.40692 1.4493362300872803 0.4722
14 13 1.599906416091919 0.4124 1.4468452342987062 0.4715
15 14 1.5933954864883424 0.41398 1.4764380693435668 0.4626
16 15 1.5933787399291992 0.41372 1.533870811843872 0.4468
17 16 1.5964966577911377 0.41302 1.4542985107421875 0.467
18 17 1.6135705094146728 0.40712 1.4721424545288087 0.4527
19 18 1.5964763764190675 0.41558 1.4527687072753905 0.4593
20 19 1.591530082244873 0.41652 1.4388617012023925 0.4642
21 20 1.589130754928589 0.41602 1.4646002044677735 0.4536
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.0309293317031862,0.24762,1.7998561183929442,0.3421
2,1.7957917514801025,0.33852,1.6648591684341432,0.3857
3,1.7348672610855103,0.35496,1.654585442352295,0.4019
4,1.7112164305877686,0.36548,1.5846896692276,0.4141
5,1.6914103266525269,0.37268,1.5373584619522094,0.4395
6,1.676686876564026,0.37698,1.637582375717163,0.3837
7,1.6700877303695678,0.38384,1.5936841482162476,0.4161
8,1.6569111339569091,0.38968,1.5129856206893921,0.4351
9,1.6577720043182373,0.3908,1.499471364212036,0.4555
10,1.6537026256942748,0.39124,1.537830891418457,0.4293
11,1.6379817530059815,0.401,1.5753344255447388,0.4145
12,1.64174579536438,0.39604,1.6890381210327148,0.3761
13,1.6474020536422729,0.39648,1.5377606107711792,0.4443
14,1.631554557762146,0.40042,1.5029741109848023,0.4553
15,1.6246128929138184,0.40476,1.4533207773208618,0.4794
16,1.6243007053375245,0.40454,1.4570706197738648,0.4646
17,1.6160261626815795,0.40666,1.4699448907852173,0.4644
18,1.6250457259750366,0.40384,1.43235032081604,0.4734
19,1.6166238045501709,0.40848,1.6371612733840943,0.3933
20,1.6181047994995117,0.40592,1.5512634952545166,0.4394
1 epoch train_loss train_acc test_loss test_acc
2 1 2.0309293317031862 0.24762 1.7998561183929442 0.3421
3 2 1.7957917514801025 0.33852 1.6648591684341432 0.3857
4 3 1.7348672610855103 0.35496 1.654585442352295 0.4019
5 4 1.7112164305877686 0.36548 1.5846896692276 0.4141
6 5 1.6914103266525269 0.37268 1.5373584619522094 0.4395
7 6 1.676686876564026 0.37698 1.637582375717163 0.3837
8 7 1.6700877303695678 0.38384 1.5936841482162476 0.4161
9 8 1.6569111339569091 0.38968 1.5129856206893921 0.4351
10 9 1.6577720043182373 0.3908 1.499471364212036 0.4555
11 10 1.6537026256942748 0.39124 1.537830891418457 0.4293
12 11 1.6379817530059815 0.401 1.5753344255447388 0.4145
13 12 1.64174579536438 0.39604 1.6890381210327148 0.3761
14 13 1.6474020536422729 0.39648 1.5377606107711792 0.4443
15 14 1.631554557762146 0.40042 1.5029741109848023 0.4553
16 15 1.6246128929138184 0.40476 1.4533207773208618 0.4794
17 16 1.6243007053375245 0.40454 1.4570706197738648 0.4646
18 17 1.6160261626815795 0.40666 1.4699448907852173 0.4644
19 18 1.6250457259750366 0.40384 1.43235032081604 0.4734
20 19 1.6166238045501709 0.40848 1.6371612733840943 0.3933
21 20 1.6181047994995117 0.40592 1.5512634952545166 0.4394
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,1.9290429125213624,0.29754,1.6215127836227416,0.4071
2,1.7239283940124512,0.36334,1.738866775894165,0.3631
3,1.6748943075942992,0.38214,1.5513272310256958,0.4274
4,1.6455059506988525,0.39562,1.508594535446167,0.4514
5,1.624495430870056,0.40286,1.5517113952636719,0.4313
6,1.6052836887359618,0.41002,1.50381274394989,0.4484
7,1.6106374563980101,0.40804,1.5364446432113648,0.4302
8,1.5953116985702516,0.41482,1.4611921619415282,0.4639
9,1.5824666970062256,0.4168,1.4630669622421264,0.4662
10,1.5835675071334838,0.4205,1.4603066038131713,0.4595
11,1.5734461785125733,0.42412,1.4167298784255982,0.4862
12,1.5816114051818848,0.42086,1.4909738641738892,0.4603
13,1.5812189110946655,0.42152,1.4088649713516235,0.4889
14,1.5847568724822998,0.42022,1.4178233968734741,0.4823
15,1.576828946876526,0.42272,1.4326917642593384,0.4809
16,1.576308758392334,0.42412,1.4116268701553345,0.4911
17,1.5560429149246215,0.43062,1.4361749185562134,0.4727
18,1.5523672427749633,0.43476,1.4626990962982178,0.4581
19,1.5581282339859008,0.42906,1.4140942724227905,0.4822
20,1.5584481320190429,0.42816,1.4160622207641602,0.4934
1 epoch train_loss train_acc test_loss test_acc
2 1 1.9290429125213624 0.29754 1.6215127836227416 0.4071
3 2 1.7239283940124512 0.36334 1.738866775894165 0.3631
4 3 1.6748943075942992 0.38214 1.5513272310256958 0.4274
5 4 1.6455059506988525 0.39562 1.508594535446167 0.4514
6 5 1.624495430870056 0.40286 1.5517113952636719 0.4313
7 6 1.6052836887359618 0.41002 1.50381274394989 0.4484
8 7 1.6106374563980101 0.40804 1.5364446432113648 0.4302
9 8 1.5953116985702516 0.41482 1.4611921619415282 0.4639
10 9 1.5824666970062256 0.4168 1.4630669622421264 0.4662
11 10 1.5835675071334838 0.4205 1.4603066038131713 0.4595
12 11 1.5734461785125733 0.42412 1.4167298784255982 0.4862
13 12 1.5816114051818848 0.42086 1.4909738641738892 0.4603
14 13 1.5812189110946655 0.42152 1.4088649713516235 0.4889
15 14 1.5847568724822998 0.42022 1.4178233968734741 0.4823
16 15 1.576828946876526 0.42272 1.4326917642593384 0.4809
17 16 1.576308758392334 0.42412 1.4116268701553345 0.4911
18 17 1.5560429149246215 0.43062 1.4361749185562134 0.4727
19 18 1.5523672427749633 0.43476 1.4626990962982178 0.4581
20 19 1.5581282339859008 0.42906 1.4140942724227905 0.4822
21 20 1.5584481320190429 0.42816 1.4160622207641602 0.4934
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.0803506112289427,0.23,1.8052742357254028,0.3649
2,1.795086921234131,0.35172,1.513138659286499,0.4536
3,1.6374745911407471,0.4064,1.4209930673599243,0.4924
4,1.5572372722625731,0.43516,1.3496723909378052,0.5132
5,1.51327647895813,0.45306,1.2944782655715943,0.5399
6,1.4705483396148682,0.47014,1.2843341438293456,0.5457
7,1.419877719154358,0.48872,1.2024114828109742,0.5757
8,1.3825100279998779,0.50334,1.1727403156280518,0.5862
9,1.349350403137207,0.51694,1.1227089519500733,0.6043
10,1.3265677324295044,0.5264,1.1120436277389527,0.6032
11,1.2989790354156494,0.53608,1.09653370552063,0.6181
12,1.271951921310425,0.5476,1.0683507353782653,0.6288
13,1.252325573387146,0.55396,1.0399812242507935,0.6344
14,1.2363645384979247,0.56054,1.005698392677307,0.6526
15,1.2197590786361694,0.56746,0.9838214920043945,0.6531
16,1.2122340144348145,0.5668,0.9973555490493774,0.6472
17,1.1860296492004394,0.58086,0.9641439270019532,0.659
18,1.1860070882797242,0.5799,0.9424298759460449,0.672
19,1.1792696018218993,0.57868,0.9344734144210816,0.6784
20,1.1561780276107787,0.5907,0.9342237223625183,0.6766
1 epoch train_loss train_acc test_loss test_acc
2 1 2.0803506112289427 0.23 1.8052742357254028 0.3649
3 2 1.795086921234131 0.35172 1.513138659286499 0.4536
4 3 1.6374745911407471 0.4064 1.4209930673599243 0.4924
5 4 1.5572372722625731 0.43516 1.3496723909378052 0.5132
6 5 1.51327647895813 0.45306 1.2944782655715943 0.5399
7 6 1.4705483396148682 0.47014 1.2843341438293456 0.5457
8 7 1.419877719154358 0.48872 1.2024114828109742 0.5757
9 8 1.3825100279998779 0.50334 1.1727403156280518 0.5862
10 9 1.349350403137207 0.51694 1.1227089519500733 0.6043
11 10 1.3265677324295044 0.5264 1.1120436277389527 0.6032
12 11 1.2989790354156494 0.53608 1.09653370552063 0.6181
13 12 1.271951921310425 0.5476 1.0683507353782653 0.6288
14 13 1.252325573387146 0.55396 1.0399812242507935 0.6344
15 14 1.2363645384979247 0.56054 1.005698392677307 0.6526
16 15 1.2197590786361694 0.56746 0.9838214920043945 0.6531
17 16 1.2122340144348145 0.5668 0.9973555490493774 0.6472
18 17 1.1860296492004394 0.58086 0.9641439270019532 0.659
19 18 1.1860070882797242 0.5799 0.9424298759460449 0.672
20 19 1.1792696018218993 0.57868 0.9344734144210816 0.6784
21 20 1.1561780276107787 0.5907 0.9342237223625183 0.6766
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.0498613494491575,0.24344,1.827182295036316,0.3315
2,1.7751250897216797,0.35796,1.5328051206588744,0.4525
3,1.6295417392730713,0.40922,1.4538747985839844,0.4708
4,1.556714581451416,0.4379,1.3427102214813234,0.5272
5,1.5068591191864014,0.45632,1.3105488367080689,0.5268
6,1.4472922039413452,0.48044,1.2710780916213988,0.5427
7,1.416664746017456,0.49122,1.184156767177582,0.5828
8,1.3932555706787109,0.50056,1.1800921089172363,0.5861
9,1.354030286064148,0.51452,1.11541519241333,0.6093
10,1.3268484888839722,0.52786,1.0645180792808533,0.6262
11,1.2946206281661987,0.53744,1.0727122479438782,0.621
12,1.2794264878082275,0.54414,1.044085719871521,0.6339
13,1.2605588024902343,0.55172,1.0073520627975463,0.6379
14,1.2451800212478639,0.55708,0.989365707397461,0.6558
15,1.2263247829437256,0.56498,0.9746960914611816,0.6631
16,1.2118944458770753,0.56814,0.9805354339599609,0.6508
17,1.2089972533798217,0.57022,0.9678337673187256,0.6643
18,1.1906279234695434,0.57802,0.9852053064346313,0.6507
19,1.1747162343597413,0.58308,0.9369933985710144,0.6728
20,1.160263595199585,0.58824,0.9429312159538269,0.6703
1 epoch train_loss train_acc test_loss test_acc
2 1 2.0498613494491575 0.24344 1.827182295036316 0.3315
3 2 1.7751250897216797 0.35796 1.5328051206588744 0.4525
4 3 1.6295417392730713 0.40922 1.4538747985839844 0.4708
5 4 1.556714581451416 0.4379 1.3427102214813234 0.5272
6 5 1.5068591191864014 0.45632 1.3105488367080689 0.5268
7 6 1.4472922039413452 0.48044 1.2710780916213988 0.5427
8 7 1.416664746017456 0.49122 1.184156767177582 0.5828
9 8 1.3932555706787109 0.50056 1.1800921089172363 0.5861
10 9 1.354030286064148 0.51452 1.11541519241333 0.6093
11 10 1.3268484888839722 0.52786 1.0645180792808533 0.6262
12 11 1.2946206281661987 0.53744 1.0727122479438782 0.621
13 12 1.2794264878082275 0.54414 1.044085719871521 0.6339
14 13 1.2605588024902343 0.55172 1.0073520627975463 0.6379
15 14 1.2451800212478639 0.55708 0.989365707397461 0.6558
16 15 1.2263247829437256 0.56498 0.9746960914611816 0.6631
17 16 1.2118944458770753 0.56814 0.9805354339599609 0.6508
18 17 1.2089972533798217 0.57022 0.9678337673187256 0.6643
19 18 1.1906279234695434 0.57802 0.9852053064346313 0.6507
20 19 1.1747162343597413 0.58308 0.9369933985710144 0.6728
21 20 1.160263595199585 0.58824 0.9429312159538269 0.6703
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.0627308975601197,0.24156,1.80689540309906,0.3478
2,1.7932613165283202,0.35134,1.6185862417221069,0.4216
3,1.658083698501587,0.39922,1.4435414199829102,0.4786
4,1.563901499633789,0.43574,1.3859216415405273,0.5008
5,1.5054525188827514,0.45402,1.2911276792526245,0.5423
6,1.450156503715515,0.47896,1.2477545137405395,0.5579
7,1.4208231299591065,0.49338,1.2044965368270875,0.5692
8,1.3701529790115357,0.50918,1.1438802488327027,0.5956
9,1.3449303218460082,0.51834,1.1281687593460084,0.6007
10,1.3100552983856202,0.5339,1.0984893451690674,0.617
11,1.2957517767715454,0.53988,1.0663194850921631,0.6201
12,1.2759582391357422,0.54506,1.0382993364334105,0.6332
13,1.256755221672058,0.55376,1.0107640748977662,0.638
14,1.240665273475647,0.55926,1.0230664975166321,0.6393
15,1.232104567527771,0.55946,1.0093883613586425,0.645
16,1.2164045375442505,0.56734,0.9853534309387207,0.6538
17,1.202742511291504,0.57248,0.9765661770820617,0.654
18,1.1869954927825928,0.57676,0.9522229454994202,0.6696
19,1.1841716981315613,0.5796,0.982388532447815,0.6555
20,1.1647253083992004,0.58448,0.9557765811920166,0.663
1 epoch train_loss train_acc test_loss test_acc
2 1 2.0627308975601197 0.24156 1.80689540309906 0.3478
3 2 1.7932613165283202 0.35134 1.6185862417221069 0.4216
4 3 1.658083698501587 0.39922 1.4435414199829102 0.4786
5 4 1.563901499633789 0.43574 1.3859216415405273 0.5008
6 5 1.5054525188827514 0.45402 1.2911276792526245 0.5423
7 6 1.450156503715515 0.47896 1.2477545137405395 0.5579
8 7 1.4208231299591065 0.49338 1.2044965368270875 0.5692
9 8 1.3701529790115357 0.50918 1.1438802488327027 0.5956
10 9 1.3449303218460082 0.51834 1.1281687593460084 0.6007
11 10 1.3100552983856202 0.5339 1.0984893451690674 0.617
12 11 1.2957517767715454 0.53988 1.0663194850921631 0.6201
13 12 1.2759582391357422 0.54506 1.0382993364334105 0.6332
14 13 1.256755221672058 0.55376 1.0107640748977662 0.638
15 14 1.240665273475647 0.55926 1.0230664975166321 0.6393
16 15 1.232104567527771 0.55946 1.0093883613586425 0.645
17 16 1.2164045375442505 0.56734 0.9853534309387207 0.6538
18 17 1.202742511291504 0.57248 0.9765661770820617 0.654
19 18 1.1869954927825928 0.57676 0.9522229454994202 0.6696
20 19 1.1841716981315613 0.5796 0.982388532447815 0.6555
21 20 1.1647253083992004 0.58448 0.9557765811920166 0.663
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,1.933640638999939,0.29732,1.6238486095428466,0.419
2,1.5058992380523681,0.4574,1.416843348312378,0.4962
3,1.3495764307403564,0.51736,1.2509430753707886,0.5525
4,1.2513321809768676,0.55462,1.1895510234832765,0.5781
5,1.1647147640228273,0.58834,1.0852176672935485,0.6106
6,1.090117392807007,0.61272,1.0372767268180847,0.6397
7,1.0261711755180358,0.6359,1.0134801021575928,0.6398
8,0.9832253789138794,0.65152,0.958988918209076,0.6642
9,0.9298850146102905,0.67178,0.9321723978996277,0.6715
10,0.8931379676055908,0.68462,0.9411900254249572,0.6734
11,0.8506363994216919,0.69908,0.9130176817893982,0.6856
12,0.8162826980781556,0.71196,0.9254231359481812,0.6771
13,0.7824999229240418,0.72322,0.8887297651290894,0.69
14,0.7469561821365357,0.73758,0.8926811159133912,0.6945
15,0.7194658122444153,0.7479,0.8924074723243713,0.6951
16,0.6972272609138489,0.75446,0.8889803544044494,0.6973
17,0.6697472100067139,0.76272,0.9309483677864074,0.6851
18,0.64141212266922,0.77758,0.9145923362731934,0.6949
19,0.6034097666549683,0.78874,0.9149182550430298,0.6954
20,0.5872987679862977,0.7941,0.8907693174362182,0.7002
1 epoch train_loss train_acc test_loss test_acc
2 1 1.933640638999939 0.29732 1.6238486095428466 0.419
3 2 1.5058992380523681 0.4574 1.416843348312378 0.4962
4 3 1.3495764307403564 0.51736 1.2509430753707886 0.5525
5 4 1.2513321809768676 0.55462 1.1895510234832765 0.5781
6 5 1.1647147640228273 0.58834 1.0852176672935485 0.6106
7 6 1.090117392807007 0.61272 1.0372767268180847 0.6397
8 7 1.0261711755180358 0.6359 1.0134801021575928 0.6398
9 8 0.9832253789138794 0.65152 0.958988918209076 0.6642
10 9 0.9298850146102905 0.67178 0.9321723978996277 0.6715
11 10 0.8931379676055908 0.68462 0.9411900254249572 0.6734
12 11 0.8506363994216919 0.69908 0.9130176817893982 0.6856
13 12 0.8162826980781556 0.71196 0.9254231359481812 0.6771
14 13 0.7824999229240418 0.72322 0.8887297651290894 0.69
15 14 0.7469561821365357 0.73758 0.8926811159133912 0.6945
16 15 0.7194658122444153 0.7479 0.8924074723243713 0.6951
17 16 0.6972272609138489 0.75446 0.8889803544044494 0.6973
18 17 0.6697472100067139 0.76272 0.9309483677864074 0.6851
19 18 0.64141212266922 0.77758 0.9145923362731934 0.6949
20 19 0.6034097666549683 0.78874 0.9149182550430298 0.6954
21 20 0.5872987679862977 0.7941 0.8907693174362182 0.7002
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.0053396127319334,0.26608,1.7093295513153077,0.3923
2,1.5551787852859498,0.44096,1.4028151916503906,0.4985
3,1.3658983665466309,0.51162,1.303923954963684,0.5425
4,1.2634439575958252,0.54832,1.163003251361847,0.5884
5,1.1752150338745118,0.58098,1.1534721800804137,0.5867
6,1.0963337196731568,0.6116,1.0367407133102418,0.6368
7,1.0309874046707153,0.6359,1.01473597574234,0.6424
8,0.9726004527282714,0.65736,0.9714125958442688,0.6589
9,0.9323004844665528,0.67234,0.9494790088653564,0.6647
10,0.8843243282318115,0.68848,0.9313362251281738,0.6754
11,0.8454029673957825,0.70276,0.9059185691833496,0.6839
12,0.8084154852294921,0.71458,0.9137783187866211,0.6844
13,0.7736339636802674,0.72754,0.8896457827568054,0.6905
14,0.7404502736854554,0.7391,0.9187201057434082,0.6847
15,0.7178288528823853,0.74732,0.8655570640563964,0.7029
16,0.6816242917251587,0.7612,0.902590276145935,0.6944
17,0.6591360913085937,0.7693,0.9042986038208007,0.6946
18,0.6323869649124145,0.77696,0.8995812829971314,0.7046
19,0.6078444783401489,0.78464,0.9112963752746582,0.7003
20,0.5815039796447754,0.79348,0.8971220227241516,0.7013
1 epoch train_loss train_acc test_loss test_acc
2 1 2.0053396127319334 0.26608 1.7093295513153077 0.3923
3 2 1.5551787852859498 0.44096 1.4028151916503906 0.4985
4 3 1.3658983665466309 0.51162 1.303923954963684 0.5425
5 4 1.2634439575958252 0.54832 1.163003251361847 0.5884
6 5 1.1752150338745118 0.58098 1.1534721800804137 0.5867
7 6 1.0963337196731568 0.6116 1.0367407133102418 0.6368
8 7 1.0309874046707153 0.6359 1.01473597574234 0.6424
9 8 0.9726004527282714 0.65736 0.9714125958442688 0.6589
10 9 0.9323004844665528 0.67234 0.9494790088653564 0.6647
11 10 0.8843243282318115 0.68848 0.9313362251281738 0.6754
12 11 0.8454029673957825 0.70276 0.9059185691833496 0.6839
13 12 0.8084154852294921 0.71458 0.9137783187866211 0.6844
14 13 0.7736339636802674 0.72754 0.8896457827568054 0.6905
15 14 0.7404502736854554 0.7391 0.9187201057434082 0.6847
16 15 0.7178288528823853 0.74732 0.8655570640563964 0.7029
17 16 0.6816242917251587 0.7612 0.902590276145935 0.6944
18 17 0.6591360913085937 0.7693 0.9042986038208007 0.6946
19 18 0.6323869649124145 0.77696 0.8995812829971314 0.7046
20 19 0.6078444783401489 0.78464 0.9112963752746582 0.7003
21 20 0.5815039796447754 0.79348 0.8971220227241516 0.7013
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,1.9614395635604858,0.28388,1.643307664680481,0.4046
2,1.5448385535430909,0.44616,1.395233334159851,0.4981
3,1.384000219039917,0.50478,1.2697911640167237,0.5481
4,1.27566444480896,0.5418,1.2024863681793212,0.5712
5,1.1857435577011108,0.57704,1.110471813774109,0.5998
6,1.10453219291687,0.60846,1.063853458595276,0.6177
7,1.0465589532089234,0.63092,1.056383687210083,0.6266
8,0.9939296613693237,0.65138,1.0162584516525268,0.6429
9,0.952554736328125,0.66554,0.9766125122070313,0.6542
10,0.9012007156181335,0.68404,0.9544676315307618,0.6643
11,0.8649646133804322,0.69644,0.9271421590805053,0.676
12,0.8223739467048645,0.70954,0.9386976690292358,0.6767
13,0.7874418495559692,0.72296,0.9196794429779053,0.679
14,0.7571423944664002,0.73368,0.9027493181228637,0.6858
15,0.7220863299465179,0.74638,0.902780525970459,0.6937
16,0.6927142573738098,0.7568,0.8872590173721313,0.6978
17,0.6688128486442566,0.7631,0.8854528800964355,0.7009
18,0.6569424513626099,0.769,0.8944921180725097,0.6999
19,0.6169235229301453,0.78028,0.886359746170044,0.6992
20,0.5988713124084473,0.78786,0.9267883558273315,0.6961
1 epoch train_loss train_acc test_loss test_acc
2 1 1.9614395635604858 0.28388 1.643307664680481 0.4046
3 2 1.5448385535430909 0.44616 1.395233334159851 0.4981
4 3 1.384000219039917 0.50478 1.2697911640167237 0.5481
5 4 1.27566444480896 0.5418 1.2024863681793212 0.5712
6 5 1.1857435577011108 0.57704 1.110471813774109 0.5998
7 6 1.10453219291687 0.60846 1.063853458595276 0.6177
8 7 1.0465589532089234 0.63092 1.056383687210083 0.6266
9 8 0.9939296613693237 0.65138 1.0162584516525268 0.6429
10 9 0.952554736328125 0.66554 0.9766125122070313 0.6542
11 10 0.9012007156181335 0.68404 0.9544676315307618 0.6643
12 11 0.8649646133804322 0.69644 0.9271421590805053 0.676
13 12 0.8223739467048645 0.70954 0.9386976690292358 0.6767
14 13 0.7874418495559692 0.72296 0.9196794429779053 0.679
15 14 0.7571423944664002 0.73368 0.9027493181228637 0.6858
16 15 0.7220863299465179 0.74638 0.902780525970459 0.6937
17 16 0.6927142573738098 0.7568 0.8872590173721313 0.6978
18 17 0.6688128486442566 0.7631 0.8854528800964355 0.7009
19 18 0.6569424513626099 0.769 0.8944921180725097 0.6999
20 19 0.6169235229301453 0.78028 0.886359746170044 0.6992
21 20 0.5988713124084473 0.78786 0.9267883558273315 0.6961
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.054507377700806,0.24836,1.7629061944961548,0.3766
2,1.7023405939102172,0.38108,1.4944005809783936,0.4511
3,1.5391481832122802,0.44006,1.3729140111923217,0.5054
4,1.48024245677948,0.46148,1.3319352336883545,0.5222
5,1.4120918664550781,0.48858,1.2639450727462769,0.5524
6,1.3601713945770264,0.51062,1.1844155906677245,0.5765
7,1.3003378784942627,0.53042,1.153577501964569,0.5892
8,1.2586501470565796,0.54992,1.0892585767745973,0.6125
9,1.2181877610397338,0.56466,1.0839576932907105,0.621
10,1.1902454167175294,0.57382,1.0581496836662292,0.6269
11,1.1582570567703248,0.58646,1.0145035231590271,0.6385
12,1.1427707640457154,0.5913,0.9702991489410401,0.6493
13,1.1167496506881713,0.60266,0.9598955163002014,0.6567
14,1.1066592629623413,0.60824,0.9935148837089539,0.652
15,1.0919563820648193,0.61256,0.9485135807991028,0.6679
16,1.07323411693573,0.61778,0.9448822968482972,0.6692
17,1.0547150045776368,0.62656,0.9020288042068482,0.6852
18,1.0447941234588622,0.62932,0.8850298257827759,0.6878
19,1.0325043148803712,0.63322,0.8606620573997498,0.7026
20,1.0181197985839843,0.64104,0.8649613096237183,0.6951
1 epoch train_loss train_acc test_loss test_acc
2 1 2.054507377700806 0.24836 1.7629061944961548 0.3766
3 2 1.7023405939102172 0.38108 1.4944005809783936 0.4511
4 3 1.5391481832122802 0.44006 1.3729140111923217 0.5054
5 4 1.48024245677948 0.46148 1.3319352336883545 0.5222
6 5 1.4120918664550781 0.48858 1.2639450727462769 0.5524
7 6 1.3601713945770264 0.51062 1.1844155906677245 0.5765
8 7 1.3003378784942627 0.53042 1.153577501964569 0.5892
9 8 1.2586501470565796 0.54992 1.0892585767745973 0.6125
10 9 1.2181877610397338 0.56466 1.0839576932907105 0.621
11 10 1.1902454167175294 0.57382 1.0581496836662292 0.6269
12 11 1.1582570567703248 0.58646 1.0145035231590271 0.6385
13 12 1.1427707640457154 0.5913 0.9702991489410401 0.6493
14 13 1.1167496506881713 0.60266 0.9598955163002014 0.6567
15 14 1.1066592629623413 0.60824 0.9935148837089539 0.652
16 15 1.0919563820648193 0.61256 0.9485135807991028 0.6679
17 16 1.07323411693573 0.61778 0.9448822968482972 0.6692
18 17 1.0547150045776368 0.62656 0.9020288042068482 0.6852
19 18 1.0447941234588622 0.62932 0.8850298257827759 0.6878
20 19 1.0325043148803712 0.63322 0.8606620573997498 0.7026
21 20 1.0181197985839843 0.64104 0.8649613096237183 0.6951
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.035904112930298,0.2552,1.7296793979644776,0.3902
2,1.708837389717102,0.3825,1.4909840774536134,0.4608
3,1.539429332046509,0.4392,1.3871454195022583,0.5077
4,1.4547308851242065,0.47234,1.304142402458191,0.5314
5,1.38470291431427,0.50112,1.2006133085250854,0.5736
6,1.3219089529800414,0.52174,1.1560321960449218,0.5886
7,1.2655390045928956,0.54448,1.0789848009109497,0.62
8,1.2301358563995362,0.56088,1.0741067754745484,0.615
9,1.1941656197357178,0.57536,1.0340390792846679,0.6317
10,1.1679337029647827,0.58268,0.9948765556335449,0.6496
11,1.1529586157226563,0.58736,0.9822504173278809,0.6523
12,1.120876918144226,0.59952,0.9725488736152649,0.6542
13,1.100623334312439,0.6071,0.9678766529083251,0.6636
14,1.0869316192626952,0.61298,0.9187479448318482,0.6743
15,1.066665187110901,0.6217,0.9006691802978516,0.6889
16,1.0647497671890258,0.62334,0.8880218561172485,0.6859
17,1.0438396575164794,0.62966,0.8780597747802734,0.6906
18,1.0385048112487794,0.63038,0.8763846193313599,0.6939
19,1.0159992222976684,0.64046,0.8701905465126037,0.6895
20,1.0131669343566894,0.64074,0.8963505390167237,0.6791
1 epoch train_loss train_acc test_loss test_acc
2 1 2.035904112930298 0.2552 1.7296793979644776 0.3902
3 2 1.708837389717102 0.3825 1.4909840774536134 0.4608
4 3 1.539429332046509 0.4392 1.3871454195022583 0.5077
5 4 1.4547308851242065 0.47234 1.304142402458191 0.5314
6 5 1.38470291431427 0.50112 1.2006133085250854 0.5736
7 6 1.3219089529800414 0.52174 1.1560321960449218 0.5886
8 7 1.2655390045928956 0.54448 1.0789848009109497 0.62
9 8 1.2301358563995362 0.56088 1.0741067754745484 0.615
10 9 1.1941656197357178 0.57536 1.0340390792846679 0.6317
11 10 1.1679337029647827 0.58268 0.9948765556335449 0.6496
12 11 1.1529586157226563 0.58736 0.9822504173278809 0.6523
13 12 1.120876918144226 0.59952 0.9725488736152649 0.6542
14 13 1.100623334312439 0.6071 0.9678766529083251 0.6636
15 14 1.0869316192626952 0.61298 0.9187479448318482 0.6743
16 15 1.066665187110901 0.6217 0.9006691802978516 0.6889
17 16 1.0647497671890258 0.62334 0.8880218561172485 0.6859
18 17 1.0438396575164794 0.62966 0.8780597747802734 0.6906
19 18 1.0385048112487794 0.63038 0.8763846193313599 0.6939
20 19 1.0159992222976684 0.64046 0.8701905465126037 0.6895
21 20 1.0131669343566894 0.64074 0.8963505390167237 0.6791
@@ -0,0 +1,21 @@
epoch,train_loss,train_acc,test_loss,test_acc
1,2.031018338546753,0.25506,1.7647437294006347,0.3568
2,1.7171751900482177,0.37778,1.510964904975891,0.4514
3,1.5605473971176147,0.43146,1.4323553239822389,0.4787
4,1.4770329965972901,0.45964,1.3029861038208008,0.5318
5,1.4138119129180908,0.48766,1.25504387550354,0.5538
6,1.3561226502227783,0.51288,1.1646997854232788,0.5827
7,1.3055281218719483,0.52908,1.1090957656860352,0.6042
8,1.256667589073181,0.55272,1.106753503894806,0.6148
9,1.2256172871017457,0.56104,1.0723959255218505,0.6215
10,1.1950459407806397,0.57334,1.0152456411361694,0.6431
11,1.1654781759643555,0.58468,0.9916627891540527,0.6515
12,1.1399137686538696,0.59356,1.0249745522499085,0.6378
13,1.1239594958877563,0.60214,0.944364684677124,0.6656
14,1.1052272381401063,0.60858,0.9644020311355591,0.6585
15,1.0886156544876098,0.61204,0.92658356590271,0.6714
16,1.0744145578384399,0.61968,0.9169919846534729,0.6851
17,1.0608087825012207,0.62458,0.906572629737854,0.6838
18,1.0402798016357422,0.62994,0.8794814926147461,0.6894
19,1.039906374988556,0.62976,0.8636875254631042,0.6966
20,1.0233247353744508,0.63816,0.8742800078392029,0.6896
1 epoch train_loss train_acc test_loss test_acc
2 1 2.031018338546753 0.25506 1.7647437294006347 0.3568
3 2 1.7171751900482177 0.37778 1.510964904975891 0.4514
4 3 1.5605473971176147 0.43146 1.4323553239822389 0.4787
5 4 1.4770329965972901 0.45964 1.3029861038208008 0.5318
6 5 1.4138119129180908 0.48766 1.25504387550354 0.5538
7 6 1.3561226502227783 0.51288 1.1646997854232788 0.5827
8 7 1.3055281218719483 0.52908 1.1090957656860352 0.6042
9 8 1.256667589073181 0.55272 1.106753503894806 0.6148
10 9 1.2256172871017457 0.56104 1.0723959255218505 0.6215
11 10 1.1950459407806397 0.57334 1.0152456411361694 0.6431
12 11 1.1654781759643555 0.58468 0.9916627891540527 0.6515
13 12 1.1399137686538696 0.59356 1.0249745522499085 0.6378
14 13 1.1239594958877563 0.60214 0.944364684677124 0.6656
15 14 1.1052272381401063 0.60858 0.9644020311355591 0.6585
16 15 1.0886156544876098 0.61204 0.92658356590271 0.6714
17 16 1.0744145578384399 0.61968 0.9169919846534729 0.6851
18 17 1.0608087825012207 0.62458 0.906572629737854 0.6838
19 18 1.0402798016357422 0.62994 0.8794814926147461 0.6894
20 19 1.039906374988556 0.62976 0.8636875254631042 0.6966
21 20 1.0233247353744508 0.63816 0.8742800078392029 0.6896