epoch: 0 training_loss 0.20512352526187896 test_loss: 0.1502478837966919
epoch: 1 training_loss 0.13533398628234863 test_loss: 0.12685002088546754
epoch: 2 training_loss 0.12733291141688824 test_loss: 0.15344160795211792
epoch: 3 training_loss 0.13637833528220653 test_loss: 0.11854150295257568
epoch: 4 training_loss 0.11971571303904056 test_loss: 0.12529221773147584
epoch: 5 training_loss 0.1306749427318573 test_loss: 0.13917789459228516
epoch: 6 training_loss 0.13354827731847763 test_loss: 0.12901957035064698
epoch: 7 training_loss 0.11712410863488913 test_loss: 0.12023504972457885
epoch: 8 training_loss 0.1325275420024991 test_loss: 0.13398342132568358
epoch: 9 training_loss 0.12064668513834477 test_loss: 0.11568043231964112
epoch: 10 training_loss 0.12976188957691193 test_loss: 0.10675914287567138
epoch: 11 training_loss 0.12108262483030557 test_loss: 0.107962965965271
epoch: 12 training_loss 0.12168973140418529 test_loss: 0.15484267473220825
epoch: 13 training_loss 0.1227904373779893 test_loss: 0.12214380502700806
epoch: 14 training_loss 0.12365275278687476 test_loss: 0.12090052366256714
epoch: 15 training_loss 0.12112020336091518 test_loss: 0.1299978256225586
epoch: 16 training_loss 0.11768414355814456 test_loss: 0.1262538194656372
epoch: 17 training_loss 0.1211352650076151 test_loss: 0.10836278200149536
epoch: 18 training_loss 0.11764661055058241 test_loss: 0.10724151134490967
epoch: 19 training_loss 0.11348951928317547 test_loss: 0.11612029075622558
epoch: 20 training_loss 0.12053389664739371 test_loss: 0.1467921257019043
epoch: 21 training_loss 0.12210227973759175 test_loss: 0.13739088773727418
epoch: 22 training_loss 0.11939424455165863 test_loss: 0.11939741373062134
epoch: 23 training_loss 0.12139536142349243 test_loss: 0.11360291242599488
epoch: 24 training_loss 0.1153323569893837 test_loss: 0.1124030590057373
epoch: 25 training_loss 0.11596849445253611 test_loss: 0.11642088890075683
epoch: 26 training_loss 0.11876296870410442 test_loss: 0.1029886245727539
epoch: 27 training_loss 0.11404175374656916 test_loss: 0.10524525642395019
epoch: 28 training_loss 0.11734450846910477 test_loss: 0.12066969871520997
epoch: 29 training_loss 0.11342820566147566 test_loss: 0.10312731266021728
epoch: 30 training_loss 0.1217008725553751 test_loss: 0.1252826929092407
epoch: 31 training_loss 0.11085608791559935 test_loss: 0.11034984588623047
epoch: 32 training_loss 0.11044720660895109 test_loss: 0.14181480407714844
epoch: 33 training_loss 0.11158621244132519 test_loss: 0.10024091005325317
epoch: 34 training_loss 0.11220039535313844 test_loss: 0.1068395733833313
epoch: 35 training_loss 0.10667034443467856 test_loss: 0.10828264951705932
epoch: 36 training_loss 0.11818572528660297 test_loss: 0.10905570983886718
epoch: 37 training_loss 0.11276911471039057 test_loss: 0.12264074087142944
epoch: 38 training_loss 0.11790960520505905 test_loss: 0.10243414640426636
epoch: 39 training_loss 0.11988300856202841 test_loss: 0.07993549704551697
epoch: 40 training_loss 0.11638030990958213 test_loss: 0.1139346718788147
epoch: 41 training_loss 0.11162076260894536 test_loss: 0.12651089429855347
epoch: 42 training_loss 0.11014925979077816 test_loss: 0.11369597911834717
epoch: 43 training_loss 0.11824331734329462 test_loss: 0.0967624008655548
epoch: 44 training_loss 0.11754941578954459 test_loss: 0.10988242626190185
epoch: 45 training_loss 0.11153876550495624 test_loss: 0.11283475160598755
epoch: 46 training_loss 0.11561407301574946 test_loss: 0.12018373012542724
epoch: 47 training_loss 0.11312195837497711 test_loss: 0.10279287099838257
epoch: 48 training_loss 0.11048832807689905 test_loss: 0.11430805921554565
epoch: 49 training_loss 0.12162161372601986 test_loss: 0.12255902290344238
epoch: 50 training_loss 0.12087772965431214 test_loss: 0.10267426967620849
epoch: 51 training_loss 0.11232612647116184 test_loss: 0.1400610089302063
epoch: 52 training_loss 0.1180714912340045 test_loss: 0.10666842460632324
epoch: 53 training_loss 0.12305308934301137 test_loss: 0.11989656686782837
epoch: 54 training_loss 0.11400994032621384 test_loss: 0.11137471199035645
epoch: 55 training_loss 0.11633414518088102 test_loss: 0.11543241739273072
epoch: 56 training_loss 0.11658500511199237 test_loss: 0.13154990673065187
epoch: 57 training_loss 0.11455311290919781 test_loss: 0.11428438425064087
epoch: 58 training_loss 0.11276597559452056 test_loss: 0.09742043614387512
epoch: 59 training_loss 0.1116810967028141 test_loss: 0.11508196592330933
epoch: 60 training_loss 0.11730686470866203 test_loss: 0.11643419265747071
epoch: 61 training_loss 0.11735550470650197 test_loss: 0.10167733430862427
epoch: 62 training_loss 0.11106583137065172 test_loss: 0.11245167255401611
epoch: 63 training_loss 0.11324246857315302 test_loss: 0.12697443962097169
epoch: 64 training_loss 0.12332496393471956 test_loss: 0.13073086738586426
epoch: 65 training_loss 0.11345750711858273 test_loss: 0.1190083384513855
epoch: 66 training_loss 0.11637738347053528 test_loss: 0.15065590143203736
epoch: 67 training_loss 0.11196409739553928 test_loss: 0.09724677801132202
epoch: 68 training_loss 0.11049443576484919 test_loss: 0.11223167181015015
epoch: 69 training_loss 0.11533140402287245 test_loss: 0.1273547649383545
epoch: 70 training_loss 0.11584050707519054 test_loss: 0.12246919870376587
epoch: 71 training_loss 0.11205051504075528 test_loss: 0.10967077016830444
epoch: 72 training_loss 0.1176248823106289 test_loss: 0.10316472053527832
epoch: 73 training_loss 0.11353109415620566 test_loss: 0.1219712734222412
epoch: 74 training_loss 0.11146369751542806 test_loss: 0.13097023963928223
epoch: 75 training_loss 0.11667269095778465 test_loss: 0.13019125461578368
epoch: 76 training_loss 0.11626988288015128 test_loss: 0.11900352239608765
epoch: 77 training_loss 0.11496006201952697 test_loss: 0.11252480745315552
epoch: 78 training_loss 0.11572079248726368 test_loss: 0.12281737327575684
epoch: 79 training_loss 0.11802765984088183 test_loss: 0.110194993019104
epoch: 80 training_loss 0.12071499295532703 test_loss: 0.11677283048629761
epoch: 81 training_loss 0.11477162022143603 test_loss: 0.12288655042648315
epoch: 82 training_loss 0.11477930061519145 test_loss: 0.09916423559188843
epoch: 83 training_loss 0.12155274976044893 test_loss: 0.11351150274276733
epoch: 84 training_loss 0.116127187050879 test_loss: 0.11187376976013183
epoch: 85 training_loss 0.11447627600282431 test_loss: 0.11356980800628662
epoch: 86 training_loss 0.11144329942762851 test_loss: 0.1312330722808838
epoch: 87 training_loss 0.11333112880587577 test_loss: 0.10905390977859497
epoch: 88 training_loss 0.11827943131327628 test_loss: 0.11482948064804077
epoch: 89 training_loss 0.11725325133651494 test_loss: 0.12670500278472902
epoch: 90 training_loss 0.11966450016945601 test_loss: 0.11573266983032227
epoch: 91 training_loss 0.11840264815837145 test_loss: 0.09846985340118408
epoch: 92 training_loss 0.10909337613731623 test_loss: 0.12918264865875245
epoch: 93 training_loss 0.11277208287268876 test_loss: 0.12697012424468995
epoch: 94 training_loss 0.11031651049852371 test_loss: 0.11179383993148803
epoch: 95 training_loss 0.11214132662862539 test_loss: 0.10003495216369629
epoch: 96 training_loss 0.11007829591631889 test_loss: 0.11746692657470703
epoch: 97 training_loss 0.11328576751053333 test_loss: 0.1164049744606018
epoch: 98 training_loss 0.10985464185476303 test_loss: 0.1139423131942749
epoch: 99 training_loss 0.11716186545789242 test_loss: 0.13697924613952636
epoch: 100 training_loss 0.11152001932263374 test_loss: 0.10282888412475585
epoch: 101 training_loss 0.11637570209801197 test_loss: 0.12125955820083618
epoch: 102 training_loss 0.10736185081303119 test_loss: 0.12311302423477173
epoch: 103 training_loss 0.1155161502584815 test_loss: 0.10445091724395753
epoch: 104 training_loss 0.11355633929371833 test_loss: 0.11672236919403076
epoch: 105 training_loss 0.11566997848451138 test_loss: 0.10234028100967407
epoch: 106 training_loss 0.11590654369443655 test_loss: 0.10588644742965699
epoch: 107 training_loss 0.11301120828837156 test_loss: 0.11642276048660279
epoch: 108 training_loss 0.11181147884577512 test_loss: 0.09765737652778625
epoch: 109 training_loss 0.1080929021537304 test_loss: 0.10576404333114624
epoch: 110 training_loss 0.11234679594635963 test_loss: 0.11727083921432495
epoch: 111 training_loss 0.10967713870108127 test_loss: 0.10933833122253418
epoch: 112 training_loss 0.11346535120159387 test_loss: 0.11651817560195923
epoch: 113 training_loss 0.11680526081472635 test_loss: 0.10567742586135864
epoch: 114 training_loss 0.11164254389703274 test_loss: 0.14643317461013794
epoch: 115 training_loss 0.11799527738243341 test_loss: 0.11018768548965455
epoch: 116 training_loss 0.1123328273370862 test_loss: 0.11691049337387086
epoch: 117 training_loss 0.11596338603645563 test_loss: 0.11124585866928101
epoch: 118 training_loss 0.11134001839905977 test_loss: 0.13465863466262817
epoch: 119 training_loss 0.10836497072130441 test_loss: 0.12454842329025269
epoch: 120 training_loss 0.1059259955585003 test_loss: 0.10803905725479127
epoch: 121 training_loss 0.1129729438200593 test_loss: 0.11908682584762573
epoch: 122 training_loss 0.11581955544650555 test_loss: 0.10528895854949952
epoch: 123 training_loss 0.11573912214487792 test_loss: 0.10143550634384155
epoch: 124 training_loss 0.10859044410288334 test_loss: 0.11530616283416747
epoch: 125 training_loss 0.12164287131279707 test_loss: 0.1383710980415344
epoch: 126 training_loss 0.11377838611602784 test_loss: 0.11226680278778076
epoch: 127 training_loss 0.10963735934346915 test_loss: 0.10847997665405273
epoch: 128 training_loss 0.10293643701821566 test_loss: 0.11241117715835572
epoch: 129 training_loss 0.11470914054661989 test_loss: 0.10313831567764283
epoch: 130 training_loss 0.11239326484501362 test_loss: 0.13148891925811768
epoch: 131 training_loss 0.11829243492335081 test_loss: 0.11640145778656005
epoch: 132 training_loss 0.1135164614021778 test_loss: 0.1052773118019104
epoch: 133 training_loss 0.1137196921929717 test_loss: 0.11673933267593384
epoch: 134 training_loss 0.11619590446352959 test_loss: 0.14669045209884643
epoch: 135 training_loss 0.1064752110466361 test_loss: 0.12202458381652832
epoch: 136 training_loss 0.11334576543420553 test_loss: 0.1305975317955017
epoch: 137 training_loss 0.11279871642589569 test_loss: 0.11621116399765015
epoch: 138 training_loss 0.11086263749748468 test_loss: 0.11186004877090454
epoch: 139 training_loss 0.11806750731542706 test_loss: 0.11132359504699707
epoch: 140 training_loss 0.10774043671786786 test_loss: 0.10687161684036255
epoch: 141 training_loss 0.12000003159046173 test_loss: 0.12035788297653198
epoch: 142 training_loss 0.11238252457231283 test_loss: 0.11819846630096435
epoch: 143 training_loss 0.11573359124362469 test_loss: 0.12187104225158692
epoch: 144 training_loss 0.11052491914480925 test_loss: 0.11869204044342041
epoch: 145 training_loss 0.10708265896886587 test_loss: 0.10500273704528809
epoch: 146 training_loss 0.11257677834481 test_loss: 0.13811018466949462
epoch: 147 training_loss 0.11489514697343112 test_loss: 0.13346794843673707
epoch: 148 training_loss 0.11488017238676548 test_loss: 0.10986872911453247
epoch: 149 training_loss 0.11470595609396696 test_loss: 0.10071921348571777
epoch: 0 training_loss 21.573029346466065 test_loss: 16.636624145507813
epoch: 1 training_loss 13.631988964080811 test_loss: 10.543934631347657
epoch: 2 training_loss 9.180405368804932 test_loss: 8.078938293457032
epoch: 3 training_loss 7.667288732528687 test_loss: 7.394482421875
epoch: 4 training_loss 6.920551309585571 test_loss: 6.703471374511719
epoch: 5 training_loss 6.396016459465027 test_loss: 5.991786956787109
epoch: 6 training_loss 5.8402870702743535 test_loss: 5.607751083374024
epoch: 7 training_loss 5.38578631401062 test_loss: 4.974869918823242
epoch: 8 training_loss 5.024591629505157 test_loss: 5.038148880004883
epoch: 9 training_loss 4.667191405296325 test_loss: 4.644622802734375
epoch: 10 training_loss 4.512156252861023 test_loss: 4.391397476196289
epoch: 11 training_loss 4.354286816120148 test_loss: 4.271553802490234
epoch: 12 training_loss 4.278302457332611 test_loss: 4.176342391967774
epoch: 13 training_loss 4.077052991390229 test_loss: 4.010981750488281
epoch: 14 training_loss 4.020831568241119 test_loss: 4.199015045166016
epoch: 15 training_loss 3.6899440336227416 test_loss: 3.694986343383789
epoch: 16 training_loss 3.7000263118743897 test_loss: 4.334593963623047
epoch: 17 training_loss 3.7298887872695925 test_loss: 3.637584686279297
epoch: 18 training_loss 3.55797073841095 test_loss: 3.783072662353516
epoch: 19 training_loss 3.585324487686157 test_loss: 3.364807891845703
epoch: 20 training_loss 3.449791657924652 test_loss: 3.496717071533203
epoch: 21 training_loss 3.3777775502204896 test_loss: 3.3789295196533202
epoch: 22 training_loss 3.3752532458305358 test_loss: 3.3016780853271483
epoch: 23 training_loss 3.2176043581962586 test_loss: 3.2589908599853517
epoch: 24 training_loss 3.2226833939552306 test_loss: 3.270094299316406
epoch: 25 training_loss 3.2075881218910216 test_loss: 3.3855873107910157
epoch: 26 training_loss 3.340355157852173 test_loss: 3.208770751953125
epoch: 27 training_loss 3.3137051177024843 test_loss: 3.22091064453125
epoch: 28 training_loss 3.194711468219757 test_loss: 3.163705062866211
epoch: 29 training_loss 3.12110821723938 test_loss: 3.358011245727539
epoch: 30 training_loss 3.1276717233657836 test_loss: 3.102565383911133
epoch: 31 training_loss 3.0738562393188475 test_loss: 3.239360046386719
epoch: 32 training_loss 3.1154960799217224 test_loss: 2.966333770751953
epoch: 33 training_loss 2.9654207181930543 test_loss: 3.017695999145508
epoch: 34 training_loss 2.914234025478363 test_loss: 2.773025131225586
epoch: 35 training_loss 3.009013969898224 test_loss: 2.940969467163086
epoch: 36 training_loss 2.986836392879486 test_loss: 3.1353816986083984
epoch: 37 training_loss 2.916114320755005 test_loss: 2.908725357055664
epoch: 38 training_loss 2.829265470504761 test_loss: 2.886844825744629
epoch: 39 training_loss 2.8418248200416567 test_loss: 2.7989839553833007
epoch: 40 training_loss 2.8087170195579527 test_loss: 2.9860713958740233
epoch: 41 training_loss 2.929391710758209 test_loss: 2.7019638061523437
epoch: 42 training_loss 2.8364457309246065 test_loss: 2.8835683822631837
epoch: 43 training_loss 2.7779732012748717 test_loss: 2.9585393905639648
epoch: 44 training_loss 2.735625584125519 test_loss: 2.6769153594970705
epoch: 45 training_loss 2.7108316814899442 test_loss: 2.8835453033447265
epoch: 46 training_loss 2.717259986400604 test_loss: 2.780647850036621
epoch: 47 training_loss 2.701658775806427 test_loss: 2.8814842224121096
epoch: 48 training_loss 2.758029032945633 test_loss: 2.7973608016967773
epoch: 49 training_loss 2.714447786808014 test_loss: 2.6335630416870117
epoch: 50 training_loss 2.7444849956035613 test_loss: 2.5593473434448244
epoch: 51 training_loss 2.559681625366211 test_loss: 2.535341262817383
epoch: 52 training_loss 2.6089124727249144 test_loss: 2.8608585357666017
epoch: 53 training_loss 2.6487641954421997 test_loss: 2.5710018157958983
epoch: 54 training_loss 2.631725332736969 test_loss: 2.4514944076538088
epoch: 55 training_loss 2.691622896194458 test_loss: 2.607416343688965
epoch: 56 training_loss 2.5754676020145415 test_loss: 2.4901466369628906
epoch: 57 training_loss 2.538533592224121 test_loss: 2.6705133438110353
epoch: 58 training_loss 2.543350911140442 test_loss: 2.541670799255371
epoch: 59 training_loss 2.594285100698471 test_loss: 2.75119686126709
epoch: 60 training_loss 2.6082566010951997 test_loss: 2.5103586196899412
epoch: 61 training_loss 2.463497505187988 test_loss: 2.5418764114379884
epoch: 62 training_loss 2.5198745775222777 test_loss: 2.664948654174805
epoch: 63 training_loss 2.6061455607414246 test_loss: 2.3838788986206056
epoch: 64 training_loss 2.5446089136600496 test_loss: 2.5658796310424803
epoch: 65 training_loss 2.4992623496055604 test_loss: 2.467142105102539
epoch: 66 training_loss 2.460998463630676 test_loss: 2.578066825866699
epoch: 67 training_loss 2.5129178166389465 test_loss: 2.3729698181152346
epoch: 68 training_loss 2.417237820625305 test_loss: 2.347146415710449
epoch: 69 training_loss 2.4153116047382355 test_loss: 2.528362846374512
epoch: 70 training_loss 2.472103008031845 test_loss: 2.5804195404052734
epoch: 71 training_loss 2.428881471157074 test_loss: 2.583315849304199
epoch: 72 training_loss 2.444689158201218 test_loss: 2.5729393005371093
epoch: 73 training_loss 2.4584236109256743 test_loss: 2.3882028579711916
epoch: 74 training_loss 2.404282627105713 test_loss: 2.4086427688598633
epoch: 75 training_loss 2.52040874838829 test_loss: 2.531126594543457
epoch: 76 training_loss 2.3591092360019683 test_loss: 2.492465019226074
epoch: 77 training_loss 2.4633518481254577 test_loss: 2.568226623535156
epoch: 78 training_loss 2.3849116337299345 test_loss: 2.6450700759887695
epoch: 79 training_loss 2.4078734874725343 test_loss: 2.6313055038452147
epoch: 80 training_loss 2.4131437230110167 test_loss: 2.2949766159057616
epoch: 81 training_loss 2.334482203722 test_loss: 2.319297027587891
epoch: 82 training_loss 2.3111549746990203 test_loss: 2.3955972671508787
epoch: 83 training_loss 2.38420423746109 test_loss: 2.409368133544922
epoch: 84 training_loss 2.370595352649689 test_loss: 2.334493064880371
epoch: 85 training_loss 2.3509212565422057 test_loss: 2.464259719848633
epoch: 86 training_loss 2.3902075302600863 test_loss: 2.299953651428223
epoch: 87 training_loss 2.257431888580322 test_loss: 2.644358825683594
epoch: 88 training_loss 2.3651016497612 test_loss: 2.5032075881958007
epoch: 89 training_loss 2.295545768737793 test_loss: 2.203098487854004
epoch: 90 training_loss 2.255348044633865 test_loss: 2.3004133224487306
epoch: 91 training_loss 2.318042677640915 test_loss: 2.3404054641723633
epoch: 92 training_loss 2.4001905941963195 test_loss: 2.473400688171387
epoch: 93 training_loss 2.2847881615161896 test_loss: 2.768084907531738
epoch: 94 training_loss 2.4172480618953704 test_loss: 2.248846435546875
epoch: 95 training_loss 2.251271365880966 test_loss: 2.4652307510375975
epoch: 96 training_loss 2.2689791536331176 test_loss: 2.259823799133301
epoch: 97 training_loss 2.412476271390915 test_loss: 2.227724075317383
epoch: 98 training_loss 2.3271822261810304 test_loss: 2.3627208709716796
epoch: 99 training_loss 2.2015037298202516 test_loss: 2.300973892211914
epoch: 100 training_loss 2.2216091454029083 test_loss: 2.6368101119995115
epoch: 101 training_loss 2.2128307914733885 test_loss: 2.2838001251220703
epoch: 102 training_loss 2.2074019432067873 test_loss: 2.204355812072754
epoch: 103 training_loss 2.337263959646225 test_loss: 2.388104057312012
epoch: 104 training_loss 2.2478219878673555 test_loss: 2.2041019439697265
epoch: 105 training_loss 2.3224806928634645 test_loss: 2.2302955627441405
epoch: 106 training_loss 2.248555415868759 test_loss: 2.1641590118408205
epoch: 107 training_loss 2.28831032037735 test_loss: 2.7551321029663085
epoch: 108 training_loss 2.1684975814819336 test_loss: 2.4088768005371093
epoch: 109 training_loss 2.1891833221912385 test_loss: 2.0454824447631834
epoch: 110 training_loss 2.1467277109622955 test_loss: 2.370128631591797
epoch: 111 training_loss 2.1876045620441436 test_loss: 2.1798301696777345
epoch: 112 training_loss 2.223087774515152 test_loss: 2.2042531967163086
epoch: 113 training_loss 2.1227455019950865 test_loss: 2.0865877151489256
epoch: 114 training_loss 2.1551481223106386 test_loss: 2.008028030395508
epoch: 115 training_loss 2.2059657061100006 test_loss: 2.460549545288086
epoch: 116 training_loss 2.2401337695121764 test_loss: 2.2385591506958007
epoch: 117 training_loss 2.143348568677902 test_loss: 2.248321533203125
epoch: 118 training_loss 2.1378835356235504 test_loss: 2.2958791732788084
epoch: 119 training_loss 2.1618668067455293 test_loss: 2.34678955078125
epoch: 120 training_loss 2.182463662624359 test_loss: 2.184596824645996
epoch: 121 training_loss 2.185797575712204 test_loss: 2.1275590896606444
epoch: 122 training_loss 2.1188914155960084 test_loss: 2.131983757019043
epoch: 123 training_loss 2.156797387599945 test_loss: 2.1615095138549805
epoch: 124 training_loss 2.1839335119724272 test_loss: 2.123635482788086
epoch: 125 training_loss 2.0971682298183443 test_loss: 2.1211729049682617
epoch: 126 training_loss 2.0783621287345886 test_loss: 2.0263906478881837
epoch: 127 training_loss 2.1673564219474795 test_loss: 2.152370834350586
epoch: 128 training_loss 2.0853256928920745 test_loss: 2.2363853454589844
epoch: 129 training_loss 2.1110061371326445 test_loss: 2.0347305297851563
epoch: 130 training_loss 2.183183118104935 test_loss: 2.209701919555664
epoch: 131 training_loss 2.1166810262203217 test_loss: 2.0660905838012695
epoch: 132 training_loss 2.121368024349213 test_loss: 2.095881462097168
epoch: 133 training_loss 2.147753781080246 test_loss: 1.9107576370239259
epoch: 134 training_loss 2.080243376493454 test_loss: 1.990595245361328
epoch: 135 training_loss 2.0573231089115143 test_loss: 2.0197830200195312
epoch: 136 training_loss 2.0351369512081146 test_loss: 1.9752490997314454
epoch: 137 training_loss 1.9996338486671448 test_loss: 2.0599571228027345
epoch: 138 training_loss 1.9903526830673217 test_loss: 1.9256326675415039
epoch: 139 training_loss 2.013538875579834 test_loss: 1.954897689819336
epoch: 140 training_loss 2.048191992044449 test_loss: 1.978067398071289
epoch: 141 training_loss 2.004863295555115 test_loss: 1.9062204360961914
epoch: 142 training_loss 1.928358612060547 test_loss: 1.9248777389526368
epoch: 143 training_loss 1.9396076142787932 test_loss: 2.0618715286254883
epoch: 144 training_loss 1.9672032582759857 test_loss: 2.125514793395996
epoch: 145 training_loss 2.0225631332397462 test_loss: 2.0212705612182615
epoch: 146 training_loss 1.9752636528015137 test_loss: 2.1066722869873047
epoch: 147 training_loss 1.9961149299144745 test_loss: 1.9287656784057616
epoch: 148 training_loss 1.9873255932331084 test_loss: 1.8772632598876953
epoch: 149 training_loss 1.9264598596096039 test_loss: 1.934238815307617
134.02200541302295
episode: 0 training return: tensor(183.8496, device='cuda:0')
episode: 1 training return: tensor(202.3858, device='cuda:0')
episode: 2 training return: tensor(199.7733, device='cuda:0')
episode: 3 training return: tensor(200.3551, device='cuda:0')
epoch: 1 test_true_pfm: 135.47572033160145 sim_pfm: 196.62817116390215
episode: 4 training return: tensor(196.2247, device='cuda:0')
episode: 5 training return: tensor(198.7367, device='cuda:0')
episode: 6 training return: tensor(194.3855, device='cuda:0')
episode: 7 training return: tensor(202.2873, device='cuda:0')
epoch: 2 test_true_pfm: 133.27601336083495 sim_pfm: 201.05258194258204
episode: 8 training return: tensor(190.0256, device='cuda:0')
episode: 9 training return: tensor(190.4975, device='cuda:0')
episode: 10 training return: tensor(181.8666, device='cuda:0')
episode: 11 training return: tensor(194.3716, device='cuda:0')
epoch: 3 test_true_pfm: 133.97272460636194 sim_pfm: 200.845760968572
episode: 12 training return: tensor(192.3785, device='cuda:0')
episode: 13 training return: tensor(217.2488, device='cuda:0')
episode: 14 training return: tensor(196.7170, device='cuda:0')
episode: 15 training return: tensor(189.9425, device='cuda:0')
epoch: 4 test_true_pfm: 133.34575604999418 sim_pfm: 200.571312569984
episode: 16 training return: tensor(214.9957, device='cuda:0')
episode: 17 training return: tensor(213.0181, device='cuda:0')
episode: 18 training return: tensor(199.8522, device='cuda:0')
episode: 19 training return: tensor(195.0894, device='cuda:0')
epoch: 5 test_true_pfm: 130.3530152036695 sim_pfm: 198.42334819515236
episode: 20 training return: tensor(202.1256, device='cuda:0')
episode: 21 training return: tensor(201.3310, device='cuda:0')
episode: 22 training return: tensor(197.2664, device='cuda:0')
episode: 23 training return: tensor(188.0228, device='cuda:0')
epoch: 6 test_true_pfm: 132.20585269500504 sim_pfm: 200.30103759155026
episode: 24 training return: tensor(188.2405, device='cuda:0')
episode: 25 training return: tensor(194.4871, device='cuda:0')
episode: 26 training return: tensor(200.4477, device='cuda:0')
episode: 27 training return: tensor(197.1451, device='cuda:0')
epoch: 7 test_true_pfm: 133.59663861941624 sim_pfm: 204.83179811072768
episode: 28 training return: tensor(203.9750, device='cuda:0')
episode: 29 training return: tensor(188.5202, device='cuda:0')
episode: 30 training return: tensor(222.8669, device='cuda:0')
episode: 31 training return: tensor(202.7320, device='cuda:0')
epoch: 8 test_true_pfm: 131.174786884998 sim_pfm: 199.30554879379923
episode: 32 training return: tensor(192.0011, device='cuda:0')
episode: 33 training return: tensor(196.4520, device='cuda:0')
episode: 34 training return: tensor(204.9351, device='cuda:0')
episode: 35 training return: tensor(204.5766, device='cuda:0')
epoch: 9 test_true_pfm: 134.12384888055286 sim_pfm: 197.22066736154957
episode: 36 training return: tensor(187.3781, device='cuda:0')
episode: 37 training return: tensor(198.0790, device='cuda:0')
episode: 38 training return: tensor(207.8201, device='cuda:0')
episode: 39 training return: tensor(205.7694, device='cuda:0')
epoch: 10 test_true_pfm: 131.52238993304462 sim_pfm: 198.40909560644067
episode: 40 training return: tensor(200.9856, device='cuda:0')
episode: 41 training return: tensor(214.0660, device='cuda:0')
episode: 42 training return: tensor(213.8047, device='cuda:0')
episode: 43 training return: tensor(200.0671, device='cuda:0')
epoch: 11 test_true_pfm: 130.4192554718132 sim_pfm: 205.3969060128089
episode: 44 training return: tensor(198.3538, device='cuda:0')
episode: 45 training return: tensor(196.9913, device='cuda:0')
episode: 46 training return: tensor(209.5625, device='cuda:0')
episode: 47 training return: tensor(193.3424, device='cuda:0')
epoch: 12 test_true_pfm: 132.14301611724514 sim_pfm: 205.32183105261066
episode: 48 training return: tensor(205.7638, device='cuda:0')
episode: 49 training return: tensor(205.3596, device='cuda:0')
episode: 50 training return: tensor(208.5628, device='cuda:0')
episode: 51 training return: tensor(199.7554, device='cuda:0')
epoch: 13 test_true_pfm: 131.21752124802822 sim_pfm: 194.25139867949184
episode: 52 training return: tensor(225.3083, device='cuda:0')
episode: 53 training return: tensor(202.1114, device='cuda:0')
episode: 54 training return: tensor(202.4562, device='cuda:0')
episode: 55 training return: tensor(217.8452, device='cuda:0')
epoch: 14 test_true_pfm: 134.4635399274237 sim_pfm: 202.58185009367298
episode: 56 training return: tensor(202.8356, device='cuda:0')
episode: 57 training return: tensor(191.5358, device='cuda:0')
episode: 58 training return: tensor(212.4695, device='cuda:0')
episode: 59 training return: tensor(196.8281, device='cuda:0')
epoch: 15 test_true_pfm: 130.11449581028222 sim_pfm: 201.94027010827557
episode: 60 training return: tensor(209.3899, device='cuda:0')
episode: 61 training return: tensor(212.8251, device='cuda:0')
episode: 62 training return: tensor(212.3771, device='cuda:0')
episode: 63 training return: tensor(197.9348, device='cuda:0')
epoch: 16 test_true_pfm: 130.1935951118631 sim_pfm: 194.90834238487295
episode: 64 training return: tensor(208.9903, device='cuda:0')
episode: 65 training return: tensor(218.5938, device='cuda:0')
episode: 66 training return: tensor(216.2376, device='cuda:0')
episode: 67 training return: tensor(203.5377, device='cuda:0')
epoch: 17 test_true_pfm: 131.70868169879716 sim_pfm: 202.01417554444635
episode: 68 training return: tensor(204.0751, device='cuda:0')
episode: 69 training return: tensor(204.3832, device='cuda:0')
episode: 70 training return: tensor(193.2763, device='cuda:0')
episode: 71 training return: tensor(196.3363, device='cuda:0')
epoch: 18 test_true_pfm: 132.23527776500427 sim_pfm: 204.3442852949491
episode: 72 training return: tensor(206.9882, device='cuda:0')
episode: 73 training return: tensor(192.1489, device='cuda:0')
episode: 74 training return: tensor(207.1553, device='cuda:0')
episode: 75 training return: tensor(208.4590, device='cuda:0')
epoch: 19 test_true_pfm: 131.7104610609458 sim_pfm: 201.25919056570274
episode: 76 training return: tensor(204.5057, device='cuda:0')
episode: 77 training return: tensor(200.7806, device='cuda:0')
episode: 78 training return: tensor(204.8297, device='cuda:0')
episode: 79 training return: tensor(194.6098, device='cuda:0')
epoch: 20 test_true_pfm: 131.94190886903866 sim_pfm: 198.70823068592
episode: 80 training return: tensor(190.2458, device='cuda:0')
episode: 81 training return: tensor(209.6750, device='cuda:0')
episode: 82 training return: tensor(183.6454, device='cuda:0')
episode: 83 training return: tensor(194.5255, device='cuda:0')
epoch: 21 test_true_pfm: 134.070842929337 sim_pfm: 205.08385920484434
episode: 84 training return: tensor(208.2897, device='cuda:0')
episode: 85 training return: tensor(199.8347, device='cuda:0')
episode: 86 training return: tensor(209.0398, device='cuda:0')
episode: 87 training return: tensor(206.5356, device='cuda:0')
epoch: 22 test_true_pfm: 131.0827723135718 sim_pfm: 204.71214764656617
episode: 88 training return: tensor(215.5016, device='cuda:0')
episode: 89 training return: tensor(205.1506, device='cuda:0')
episode: 90 training return: tensor(191.6516, device='cuda:0')
episode: 91 training return: tensor(201.9619, device='cuda:0')
epoch: 23 test_true_pfm: 131.52437359651185 sim_pfm: 202.1755506109621
episode: 92 training return: tensor(200.7282, device='cuda:0')
episode: 93 training return: tensor(210.2711, device='cuda:0')
episode: 94 training return: tensor(191.5421, device='cuda:0')
episode: 95 training return: tensor(214.3203, device='cuda:0')
epoch: 24 test_true_pfm: 133.54344670537503 sim_pfm: 205.4813659542706
episode: 96 training return: tensor(199.2846, device='cuda:0')
episode: 97 training return: tensor(197.0630, device='cuda:0')
episode: 98 training return: tensor(197.3077, device='cuda:0')
episode: 99 training return: tensor(204.2638, device='cuda:0')
epoch: 25 test_true_pfm: 133.69226006719893 sim_pfm: 201.51153460405766
episode: 100 training return: tensor(203.3342, device='cuda:0')
episode: 101 training return: tensor(203.8913, device='cuda:0')
episode: 102 training return: tensor(203.2430, device='cuda:0')
episode: 103 training return: tensor(206.6611, device='cuda:0')
epoch: 26 test_true_pfm: 132.9699356390742 sim_pfm: 195.89990800808883
episode: 104 training return: tensor(195.8702, device='cuda:0')
episode: 105 training return: tensor(204.8474, device='cuda:0')
episode: 106 training return: tensor(201.0067, device='cuda:0')
episode: 107 training return: tensor(207.6003, device='cuda:0')
epoch: 27 test_true_pfm: 133.70272996207478 sim_pfm: 201.4010338196531
episode: 108 training return: tensor(197.5099, device='cuda:0')
episode: 109 training return: tensor(204.9125, device='cuda:0')
episode: 110 training return: tensor(212.1743, device='cuda:0')
episode: 111 training return: tensor(198.4196, device='cuda:0')
epoch: 28 test_true_pfm: 133.45140105813076 sim_pfm: 201.98481027127127
episode: 112 training return: tensor(206.6999, device='cuda:0')
episode: 113 training return: tensor(207.4590, device='cuda:0')
episode: 114 training return: tensor(206.2982, device='cuda:0')
episode: 115 training return: tensor(201.8513, device='cuda:0')
epoch: 29 test_true_pfm: 129.774086207359 sim_pfm: 210.90243649830808
episode: 116 training return: tensor(195.6995, device='cuda:0')
episode: 117 training return: tensor(194.2142, device='cuda:0')
episode: 118 training return: tensor(208.1585, device='cuda:0')
episode: 119 training return: tensor(208.2074, device='cuda:0')
epoch: 30 test_true_pfm: 131.91120192871935 sim_pfm: 202.2076707775006
episode: 120 training return: tensor(201.7211, device='cuda:0')
episode: 121 training return: tensor(202.2839, device='cuda:0')
episode: 122 training return: tensor(195.9510, device='cuda:0')
episode: 123 training return: tensor(204.4671, device='cuda:0')
epoch: 31 test_true_pfm: 132.72515448436158 sim_pfm: 203.360440160695
episode: 124 training return: tensor(202.8023, device='cuda:0')
episode: 125 training return: tensor(200.7649, device='cuda:0')
episode: 126 training return: tensor(224.9503, device='cuda:0')
episode: 127 training return: tensor(202.1278, device='cuda:0')
epoch: 32 test_true_pfm: 132.141421053371 sim_pfm: 202.67887820680625
episode: 128 training return: tensor(195.0558, device='cuda:0')
episode: 129 training return: tensor(205.4637, device='cuda:0')
episode: 130 training return: tensor(217.9359, device='cuda:0')
episode: 131 training return: tensor(210.1452, device='cuda:0')
epoch: 33 test_true_pfm: 133.86958226746938 sim_pfm: 205.1334510830813
episode: 132 training return: tensor(206.0554, device='cuda:0')
episode: 133 training return: tensor(203.1953, device='cuda:0')
episode: 134 training return: tensor(205.4389, device='cuda:0')
episode: 135 training return: tensor(200.8858, device='cuda:0')
epoch: 34 test_true_pfm: 131.70673229019062 sim_pfm: 207.93208732315105
episode: 136 training return: tensor(196.8439, device='cuda:0')
episode: 137 training return: tensor(203.1251, device='cuda:0')
episode: 138 training return: tensor(219.6143, device='cuda:0')
episode: 139 training return: tensor(192.5166, device='cuda:0')
epoch: 35 test_true_pfm: 128.67059485035657 sim_pfm: 201.74074705179083
episode: 140 training return: tensor(193.2437, device='cuda:0')
episode: 141 training return: tensor(210.5028, device='cuda:0')
episode: 142 training return: tensor(214.4365, device='cuda:0')
episode: 143 training return: tensor(225.6475, device='cuda:0')
epoch: 36 test_true_pfm: 132.37541577411045 sim_pfm: 202.0583036406897
episode: 144 training return: tensor(200.2948, device='cuda:0')
episode: 145 training return: tensor(204.3956, device='cuda:0')
episode: 146 training return: tensor(204.0555, device='cuda:0')
episode: 147 training return: tensor(204.0988, device='cuda:0')
epoch: 37 test_true_pfm: 128.9653111979088 sim_pfm: 208.55612529780484
episode: 148 training return: tensor(191.0357, device='cuda:0')
episode: 149 training return: tensor(211.3804, device='cuda:0')
episode: 150 training return: tensor(201.7403, device='cuda:0')
episode: 151 training return: tensor(208.0396, device='cuda:0')
epoch: 38 test_true_pfm: 133.47874351293427 sim_pfm: 202.7260555687535
episode: 152 training return: tensor(199.5798, device='cuda:0')
episode: 153 training return: tensor(189.1030, device='cuda:0')
episode: 154 training return: tensor(201.9967, device='cuda:0')
episode: 155 training return: tensor(209.3683, device='cuda:0')
epoch: 39 test_true_pfm: 132.67622678304897 sim_pfm: 208.30965130081168
episode: 156 training return: tensor(205.4051, device='cuda:0')
episode: 157 training return: tensor(197.7125, device='cuda:0')
episode: 158 training return: tensor(201.3085, device='cuda:0')
episode: 159 training return: tensor(209.0545, device='cuda:0')
epoch: 40 test_true_pfm: 133.0415746611671 sim_pfm: 201.63897539907484
episode: 160 training return: tensor(206.5434, device='cuda:0')
episode: 161 training return: tensor(187.2848, device='cuda:0')
episode: 162 training return: tensor(207.8055, device='cuda:0')
episode: 163 training return: tensor(205.2324, device='cuda:0')
epoch: 41 test_true_pfm: 133.80155114880768 sim_pfm: 203.9192322164541
episode: 164 training return: tensor(201.9914, device='cuda:0')
episode: 165 training return: tensor(206.8931, device='cuda:0')
episode: 166 training return: tensor(200.7605, device='cuda:0')
episode: 167 training return: tensor(199.4457, device='cuda:0')
epoch: 42 test_true_pfm: 129.21381632018907 sim_pfm: 200.52152441858198
episode: 168 training return: tensor(194.2650, device='cuda:0')
episode: 169 training return: tensor(194.2904, device='cuda:0')
episode: 170 training return: tensor(196.9227, device='cuda:0')
episode: 171 training return: tensor(212.6402, device='cuda:0')
epoch: 43 test_true_pfm: 133.25105210874622 sim_pfm: 200.78357270105045
episode: 172 training return: tensor(203.8576, device='cuda:0')
episode: 173 training return: tensor(192.6623, device='cuda:0')
episode: 174 training return: tensor(185.6261, device='cuda:0')
episode: 175 training return: tensor(199.3466, device='cuda:0')
epoch: 44 test_true_pfm: 132.09549644117095 sim_pfm: 204.47153452227357
episode: 176 training return: tensor(204.3959, device='cuda:0')
episode: 177 training return: tensor(209.6620, device='cuda:0')
episode: 178 training return: tensor(212.8295, device='cuda:0')
episode: 179 training return: tensor(199.9319, device='cuda:0')
epoch: 45 test_true_pfm: 132.19312667591802 sim_pfm: 206.24440333461388
episode: 180 training return: tensor(193.5167, device='cuda:0')
episode: 181 training return: tensor(196.8618, device='cuda:0')
episode: 182 training return: tensor(214.3304, device='cuda:0')
episode: 183 training return: tensor(199.3215, device='cuda:0')
epoch: 46 test_true_pfm: 132.4895894372126 sim_pfm: 207.0064636815514
episode: 184 training return: tensor(205.0826, device='cuda:0')
episode: 185 training return: tensor(194.8839, device='cuda:0')
episode: 186 training return: tensor(206.8226, device='cuda:0')
episode: 187 training return: tensor(202.5862, device='cuda:0')
epoch: 47 test_true_pfm: 131.21279029685348 sim_pfm: 204.85540295627433
episode: 188 training return: tensor(202.2237, device='cuda:0')
episode: 189 training return: tensor(212.8995, device='cuda:0')
episode: 190 training return: tensor(214.2545, device='cuda:0')
episode: 191 training return: tensor(197.7864, device='cuda:0')
epoch: 48 test_true_pfm: 130.93447728148936 sim_pfm: 205.22794417145195
episode: 192 training return: tensor(211.9930, device='cuda:0')
episode: 193 training return: tensor(208.3807, device='cuda:0')
episode: 194 training return: tensor(208.2966, device='cuda:0')
episode: 195 training return: tensor(203.8325, device='cuda:0')
epoch: 49 test_true_pfm: 131.11742000152884 sim_pfm: 200.55719573765526
episode: 196 training return: tensor(203.3018, device='cuda:0')
episode: 197 training return: tensor(202.1881, device='cuda:0')
episode: 198 training return: tensor(205.6220, device='cuda:0')
episode: 199 training return: tensor(198.2418, device='cuda:0')
epoch: 50 test_true_pfm: 130.67102243986545 sim_pfm: 205.99315837088506
episode: 200 training return: tensor(216.8046, device='cuda:0')
episode: 201 training return: tensor(207.9481, device='cuda:0')
episode: 202 training return: tensor(210.1443, device='cuda:0')
episode: 203 training return: tensor(188.5242, device='cuda:0')
epoch: 51 test_true_pfm: 131.21172707091313 sim_pfm: 204.619309070264
episode: 204 training return: tensor(211.9102, device='cuda:0')
episode: 205 training return: tensor(203.9512, device='cuda:0')
episode: 206 training return: tensor(212.8149, device='cuda:0')
episode: 207 training return: tensor(207.2771, device='cuda:0')
epoch: 52 test_true_pfm: 132.09964552973423 sim_pfm: 202.5763304719585
episode: 208 training return: tensor(198.5143, device='cuda:0')
episode: 209 training return: tensor(215.1185, device='cuda:0')
episode: 210 training return: tensor(198.5946, device='cuda:0')
episode: 211 training return: tensor(202.0555, device='cuda:0')
epoch: 53 test_true_pfm: 131.39849095093908 sim_pfm: 197.64432387322887
episode: 212 training return: tensor(195.2679, device='cuda:0')
episode: 213 training return: tensor(206.3486, device='cuda:0')
episode: 214 training return: tensor(222.9507, device='cuda:0')
episode: 215 training return: tensor(202.4759, device='cuda:0')
epoch: 54 test_true_pfm: 131.50495526034794 sim_pfm: 206.64543946115882
episode: 216 training return: tensor(195.7923, device='cuda:0')
episode: 217 training return: tensor(225.8762, device='cuda:0')
episode: 218 training return: tensor(204.2999, device='cuda:0')
episode: 219 training return: tensor(206.4103, device='cuda:0')
epoch: 55 test_true_pfm: 130.81109947522052 sim_pfm: 206.69154060302535
episode: 220 training return: tensor(208.7832, device='cuda:0')
episode: 221 training return: tensor(203.6977, device='cuda:0')
episode: 222 training return: tensor(196.9815, device='cuda:0')
episode: 223 training return: tensor(203.3742, device='cuda:0')
epoch: 56 test_true_pfm: 131.62878950456425 sim_pfm: 198.5921149334754
episode: 224 training return: tensor(204.0696, device='cuda:0')
episode: 225 training return: tensor(201.1969, device='cuda:0')
episode: 226 training return: tensor(211.5711, device='cuda:0')
episode: 227 training return: tensor(200.8612, device='cuda:0')
epoch: 57 test_true_pfm: 131.45051614469136 sim_pfm: 202.55720917435247
episode: 228 training return: tensor(206.5570, device='cuda:0')
episode: 229 training return: tensor(206.4603, device='cuda:0')
episode: 230 training return: tensor(203.9321, device='cuda:0')
episode: 231 training return: tensor(213.7417, device='cuda:0')
epoch: 58 test_true_pfm: 131.20652132995005 sim_pfm: 205.56789290923626
episode: 232 training return: tensor(216.8958, device='cuda:0')
episode: 233 training return: tensor(209.9420, device='cuda:0')
episode: 234 training return: tensor(210.1249, device='cuda:0')
episode: 235 training return: tensor(204.0370, device='cuda:0')
epoch: 59 test_true_pfm: 131.20549540804967 sim_pfm: 205.59743111666177
episode: 236 training return: tensor(205.2557, device='cuda:0')
episode: 237 training return: tensor(205.4894, device='cuda:0')
episode: 238 training return: tensor(193.8160, device='cuda:0')
episode: 239 training return: tensor(213.8626, device='cuda:0')
epoch: 60 test_true_pfm: 130.23384829111296 sim_pfm: 202.98583564039436
episode: 240 training return: tensor(215.0332, device='cuda:0')
episode: 241 training return: tensor(194.0728, device='cuda:0')
episode: 242 training return: tensor(199.2920, device='cuda:0')
episode: 243 training return: tensor(202.4922, device='cuda:0')
epoch: 61 test_true_pfm: 130.9506939308398 sim_pfm: 205.07476712099742
