Zethearc's picture
End of training
d7002f1 verified
{
"best_metric": 0.9605053856292324,
"best_model_checkpoint": "swin-tiny-patch4-window7-224-finetuned-cancer/checkpoint-1547",
"epoch": 0.9998384230085636,
"eval_steps": 500,
"global_step": 1547,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006463079657456778,
"grad_norm": 7.331174850463867,
"learning_rate": 3.225806451612903e-06,
"loss": 0.7162,
"step": 10
},
{
"epoch": 0.012926159314913557,
"grad_norm": 4.575641632080078,
"learning_rate": 6.451612903225806e-06,
"loss": 0.6444,
"step": 20
},
{
"epoch": 0.019389238972370333,
"grad_norm": 3.2741119861602783,
"learning_rate": 9.67741935483871e-06,
"loss": 0.5618,
"step": 30
},
{
"epoch": 0.025852318629827113,
"grad_norm": 4.564846515655518,
"learning_rate": 1.2903225806451613e-05,
"loss": 0.4959,
"step": 40
},
{
"epoch": 0.03231539828728389,
"grad_norm": 4.165364742279053,
"learning_rate": 1.6129032258064517e-05,
"loss": 0.447,
"step": 50
},
{
"epoch": 0.038778477944740666,
"grad_norm": 4.854116439819336,
"learning_rate": 1.935483870967742e-05,
"loss": 0.409,
"step": 60
},
{
"epoch": 0.045241557602197446,
"grad_norm": 4.5222907066345215,
"learning_rate": 2.258064516129032e-05,
"loss": 0.3992,
"step": 70
},
{
"epoch": 0.051704637259654226,
"grad_norm": 5.930300235748291,
"learning_rate": 2.5806451612903226e-05,
"loss": 0.3832,
"step": 80
},
{
"epoch": 0.058167716917111006,
"grad_norm": 6.286398410797119,
"learning_rate": 2.9032258064516133e-05,
"loss": 0.399,
"step": 90
},
{
"epoch": 0.06463079657456779,
"grad_norm": 8.192097663879395,
"learning_rate": 3.2258064516129034e-05,
"loss": 0.3631,
"step": 100
},
{
"epoch": 0.07109387623202457,
"grad_norm": 7.482578754425049,
"learning_rate": 3.548387096774194e-05,
"loss": 0.3702,
"step": 110
},
{
"epoch": 0.07755695588948133,
"grad_norm": 5.62371301651001,
"learning_rate": 3.870967741935484e-05,
"loss": 0.352,
"step": 120
},
{
"epoch": 0.08402003554693811,
"grad_norm": 5.444067001342773,
"learning_rate": 4.1935483870967746e-05,
"loss": 0.3338,
"step": 130
},
{
"epoch": 0.09048311520439489,
"grad_norm": 4.892695426940918,
"learning_rate": 4.516129032258064e-05,
"loss": 0.3463,
"step": 140
},
{
"epoch": 0.09694619486185167,
"grad_norm": 5.656785488128662,
"learning_rate": 4.8387096774193554e-05,
"loss": 0.3718,
"step": 150
},
{
"epoch": 0.10340927451930845,
"grad_norm": 7.789802074432373,
"learning_rate": 4.982040229885058e-05,
"loss": 0.3412,
"step": 160
},
{
"epoch": 0.10987235417676523,
"grad_norm": 7.198510646820068,
"learning_rate": 4.946120689655172e-05,
"loss": 0.3577,
"step": 170
},
{
"epoch": 0.11633543383422201,
"grad_norm": 7.838411808013916,
"learning_rate": 4.9102011494252875e-05,
"loss": 0.3522,
"step": 180
},
{
"epoch": 0.12279851349167878,
"grad_norm": 7.711742401123047,
"learning_rate": 4.8742816091954026e-05,
"loss": 0.3381,
"step": 190
},
{
"epoch": 0.12926159314913557,
"grad_norm": 5.089247703552246,
"learning_rate": 4.838362068965517e-05,
"loss": 0.3507,
"step": 200
},
{
"epoch": 0.13572467280659234,
"grad_norm": 4.4015913009643555,
"learning_rate": 4.802442528735632e-05,
"loss": 0.3162,
"step": 210
},
{
"epoch": 0.14218775246404913,
"grad_norm": 4.59859561920166,
"learning_rate": 4.7665229885057474e-05,
"loss": 0.274,
"step": 220
},
{
"epoch": 0.1486508321215059,
"grad_norm": 3.7790114879608154,
"learning_rate": 4.730603448275862e-05,
"loss": 0.3161,
"step": 230
},
{
"epoch": 0.15511391177896267,
"grad_norm": 5.165287017822266,
"learning_rate": 4.694683908045977e-05,
"loss": 0.2878,
"step": 240
},
{
"epoch": 0.16157699143641946,
"grad_norm": 6.651523590087891,
"learning_rate": 4.658764367816092e-05,
"loss": 0.3133,
"step": 250
},
{
"epoch": 0.16804007109387623,
"grad_norm": 6.430370807647705,
"learning_rate": 4.622844827586207e-05,
"loss": 0.2744,
"step": 260
},
{
"epoch": 0.17450315075133302,
"grad_norm": 3.7495827674865723,
"learning_rate": 4.5869252873563225e-05,
"loss": 0.2638,
"step": 270
},
{
"epoch": 0.18096623040878979,
"grad_norm": 4.071202278137207,
"learning_rate": 4.551005747126437e-05,
"loss": 0.2698,
"step": 280
},
{
"epoch": 0.18742931006624658,
"grad_norm": 5.824757099151611,
"learning_rate": 4.515086206896552e-05,
"loss": 0.2916,
"step": 290
},
{
"epoch": 0.19389238972370335,
"grad_norm": 4.58206033706665,
"learning_rate": 4.4791666666666673e-05,
"loss": 0.2674,
"step": 300
},
{
"epoch": 0.2003554693811601,
"grad_norm": 7.343913555145264,
"learning_rate": 4.443247126436782e-05,
"loss": 0.3069,
"step": 310
},
{
"epoch": 0.2068185490386169,
"grad_norm": 4.5890302658081055,
"learning_rate": 4.407327586206897e-05,
"loss": 0.2889,
"step": 320
},
{
"epoch": 0.21328162869607367,
"grad_norm": 6.601052761077881,
"learning_rate": 4.371408045977012e-05,
"loss": 0.2509,
"step": 330
},
{
"epoch": 0.21974470835353047,
"grad_norm": 2.973667621612549,
"learning_rate": 4.3354885057471266e-05,
"loss": 0.2667,
"step": 340
},
{
"epoch": 0.22620778801098723,
"grad_norm": 7.966582298278809,
"learning_rate": 4.299568965517242e-05,
"loss": 0.2888,
"step": 350
},
{
"epoch": 0.23267086766844403,
"grad_norm": 2.4454495906829834,
"learning_rate": 4.263649425287356e-05,
"loss": 0.2638,
"step": 360
},
{
"epoch": 0.2391339473259008,
"grad_norm": 4.780755996704102,
"learning_rate": 4.2277298850574714e-05,
"loss": 0.2771,
"step": 370
},
{
"epoch": 0.24559702698335756,
"grad_norm": 7.205840587615967,
"learning_rate": 4.1918103448275866e-05,
"loss": 0.289,
"step": 380
},
{
"epoch": 0.25206010664081435,
"grad_norm": 6.309414863586426,
"learning_rate": 4.155890804597701e-05,
"loss": 0.2471,
"step": 390
},
{
"epoch": 0.25852318629827115,
"grad_norm": 4.736650466918945,
"learning_rate": 4.119971264367816e-05,
"loss": 0.2204,
"step": 400
},
{
"epoch": 0.2649862659557279,
"grad_norm": 4.276492595672607,
"learning_rate": 4.0840517241379314e-05,
"loss": 0.2546,
"step": 410
},
{
"epoch": 0.2714493456131847,
"grad_norm": 4.888199329376221,
"learning_rate": 4.048132183908046e-05,
"loss": 0.2343,
"step": 420
},
{
"epoch": 0.27791242527064147,
"grad_norm": 3.0027356147766113,
"learning_rate": 4.012212643678161e-05,
"loss": 0.2644,
"step": 430
},
{
"epoch": 0.28437550492809827,
"grad_norm": 3.5034611225128174,
"learning_rate": 3.976293103448276e-05,
"loss": 0.2318,
"step": 440
},
{
"epoch": 0.290838584585555,
"grad_norm": 4.029373645782471,
"learning_rate": 3.940373563218391e-05,
"loss": 0.2393,
"step": 450
},
{
"epoch": 0.2973016642430118,
"grad_norm": 2.631014585494995,
"learning_rate": 3.904454022988506e-05,
"loss": 0.2497,
"step": 460
},
{
"epoch": 0.3037647439004686,
"grad_norm": 3.7511560916900635,
"learning_rate": 3.86853448275862e-05,
"loss": 0.2776,
"step": 470
},
{
"epoch": 0.31022782355792533,
"grad_norm": 4.721391677856445,
"learning_rate": 3.8326149425287355e-05,
"loss": 0.2353,
"step": 480
},
{
"epoch": 0.3166909032153821,
"grad_norm": 3.97963285446167,
"learning_rate": 3.796695402298851e-05,
"loss": 0.2549,
"step": 490
},
{
"epoch": 0.3231539828728389,
"grad_norm": 4.646782398223877,
"learning_rate": 3.760775862068966e-05,
"loss": 0.2473,
"step": 500
},
{
"epoch": 0.3296170625302957,
"grad_norm": 4.661077976226807,
"learning_rate": 3.724856321839081e-05,
"loss": 0.2581,
"step": 510
},
{
"epoch": 0.33608014218775245,
"grad_norm": 7.082584381103516,
"learning_rate": 3.6889367816091954e-05,
"loss": 0.2427,
"step": 520
},
{
"epoch": 0.34254322184520924,
"grad_norm": 4.724052906036377,
"learning_rate": 3.6530172413793106e-05,
"loss": 0.2683,
"step": 530
},
{
"epoch": 0.34900630150266604,
"grad_norm": 3.6246917247772217,
"learning_rate": 3.617097701149426e-05,
"loss": 0.2636,
"step": 540
},
{
"epoch": 0.3554693811601228,
"grad_norm": 9.237018585205078,
"learning_rate": 3.58117816091954e-05,
"loss": 0.2502,
"step": 550
},
{
"epoch": 0.36193246081757957,
"grad_norm": 3.987290382385254,
"learning_rate": 3.5452586206896554e-05,
"loss": 0.2359,
"step": 560
},
{
"epoch": 0.36839554047503636,
"grad_norm": 6.946052074432373,
"learning_rate": 3.5093390804597706e-05,
"loss": 0.2445,
"step": 570
},
{
"epoch": 0.37485862013249316,
"grad_norm": 4.0834736824035645,
"learning_rate": 3.473419540229885e-05,
"loss": 0.2515,
"step": 580
},
{
"epoch": 0.3813216997899499,
"grad_norm": 3.3558614253997803,
"learning_rate": 3.4375e-05,
"loss": 0.2511,
"step": 590
},
{
"epoch": 0.3877847794474067,
"grad_norm": 4.314471244812012,
"learning_rate": 3.4015804597701154e-05,
"loss": 0.232,
"step": 600
},
{
"epoch": 0.3942478591048635,
"grad_norm": 3.44812273979187,
"learning_rate": 3.36566091954023e-05,
"loss": 0.1993,
"step": 610
},
{
"epoch": 0.4007109387623202,
"grad_norm": 4.297704219818115,
"learning_rate": 3.329741379310345e-05,
"loss": 0.2556,
"step": 620
},
{
"epoch": 0.407174018419777,
"grad_norm": 2.1731884479522705,
"learning_rate": 3.2938218390804595e-05,
"loss": 0.2168,
"step": 630
},
{
"epoch": 0.4136370980772338,
"grad_norm": 3.119593620300293,
"learning_rate": 3.2579022988505747e-05,
"loss": 0.2089,
"step": 640
},
{
"epoch": 0.4201001777346906,
"grad_norm": 3.971031665802002,
"learning_rate": 3.22198275862069e-05,
"loss": 0.2373,
"step": 650
},
{
"epoch": 0.42656325739214734,
"grad_norm": 3.3221435546875,
"learning_rate": 3.186063218390804e-05,
"loss": 0.243,
"step": 660
},
{
"epoch": 0.43302633704960414,
"grad_norm": 3.1322009563446045,
"learning_rate": 3.1501436781609195e-05,
"loss": 0.2239,
"step": 670
},
{
"epoch": 0.43948941670706093,
"grad_norm": 3.9616572856903076,
"learning_rate": 3.1142241379310346e-05,
"loss": 0.2316,
"step": 680
},
{
"epoch": 0.44595249636451767,
"grad_norm": 2.667262077331543,
"learning_rate": 3.078304597701149e-05,
"loss": 0.204,
"step": 690
},
{
"epoch": 0.45241557602197446,
"grad_norm": 5.354103088378906,
"learning_rate": 3.042385057471265e-05,
"loss": 0.2035,
"step": 700
},
{
"epoch": 0.45887865567943126,
"grad_norm": 3.422032356262207,
"learning_rate": 3.0064655172413798e-05,
"loss": 0.2518,
"step": 710
},
{
"epoch": 0.46534173533688805,
"grad_norm": 6.190507411956787,
"learning_rate": 2.9705459770114946e-05,
"loss": 0.2148,
"step": 720
},
{
"epoch": 0.4718048149943448,
"grad_norm": 4.419785976409912,
"learning_rate": 2.9346264367816094e-05,
"loss": 0.2399,
"step": 730
},
{
"epoch": 0.4782678946518016,
"grad_norm": 6.475925922393799,
"learning_rate": 2.8987068965517246e-05,
"loss": 0.2308,
"step": 740
},
{
"epoch": 0.4847309743092584,
"grad_norm": 2.907485246658325,
"learning_rate": 2.8627873563218394e-05,
"loss": 0.2071,
"step": 750
},
{
"epoch": 0.4911940539667151,
"grad_norm": 3.476051092147827,
"learning_rate": 2.8268678160919542e-05,
"loss": 0.2374,
"step": 760
},
{
"epoch": 0.4976571336241719,
"grad_norm": 6.298080921173096,
"learning_rate": 2.7909482758620694e-05,
"loss": 0.2277,
"step": 770
},
{
"epoch": 0.5041202132816287,
"grad_norm": 4.822377681732178,
"learning_rate": 2.7550287356321842e-05,
"loss": 0.2259,
"step": 780
},
{
"epoch": 0.5105832929390854,
"grad_norm": 3.966501474380493,
"learning_rate": 2.719109195402299e-05,
"loss": 0.2212,
"step": 790
},
{
"epoch": 0.5170463725965423,
"grad_norm": 2.599501848220825,
"learning_rate": 2.6831896551724138e-05,
"loss": 0.2202,
"step": 800
},
{
"epoch": 0.523509452253999,
"grad_norm": 4.360517501831055,
"learning_rate": 2.647270114942529e-05,
"loss": 0.2482,
"step": 810
},
{
"epoch": 0.5299725319114558,
"grad_norm": 4.807380199432373,
"learning_rate": 2.6113505747126438e-05,
"loss": 0.2222,
"step": 820
},
{
"epoch": 0.5364356115689126,
"grad_norm": 3.3132829666137695,
"learning_rate": 2.5754310344827586e-05,
"loss": 0.2421,
"step": 830
},
{
"epoch": 0.5428986912263694,
"grad_norm": 6.560469150543213,
"learning_rate": 2.5395114942528734e-05,
"loss": 0.196,
"step": 840
},
{
"epoch": 0.5493617708838261,
"grad_norm": 7.168336391448975,
"learning_rate": 2.5035919540229886e-05,
"loss": 0.244,
"step": 850
},
{
"epoch": 0.5558248505412829,
"grad_norm": 3.6738462448120117,
"learning_rate": 2.4676724137931034e-05,
"loss": 0.2361,
"step": 860
},
{
"epoch": 0.5622879301987397,
"grad_norm": 4.0579833984375,
"learning_rate": 2.4317528735632186e-05,
"loss": 0.2195,
"step": 870
},
{
"epoch": 0.5687510098561965,
"grad_norm": 5.091606616973877,
"learning_rate": 2.3958333333333334e-05,
"loss": 0.2128,
"step": 880
},
{
"epoch": 0.5752140895136533,
"grad_norm": 7.0507354736328125,
"learning_rate": 2.3599137931034486e-05,
"loss": 0.2259,
"step": 890
},
{
"epoch": 0.58167716917111,
"grad_norm": 4.719695568084717,
"learning_rate": 2.3239942528735634e-05,
"loss": 0.2119,
"step": 900
},
{
"epoch": 0.5881402488285669,
"grad_norm": 4.4290289878845215,
"learning_rate": 2.2880747126436782e-05,
"loss": 0.2138,
"step": 910
},
{
"epoch": 0.5946033284860236,
"grad_norm": 4.132259368896484,
"learning_rate": 2.2521551724137934e-05,
"loss": 0.2111,
"step": 920
},
{
"epoch": 0.6010664081434803,
"grad_norm": 6.010085105895996,
"learning_rate": 2.2162356321839082e-05,
"loss": 0.2106,
"step": 930
},
{
"epoch": 0.6075294878009372,
"grad_norm": 3.847933769226074,
"learning_rate": 2.180316091954023e-05,
"loss": 0.2146,
"step": 940
},
{
"epoch": 0.6139925674583939,
"grad_norm": 4.39457893371582,
"learning_rate": 2.144396551724138e-05,
"loss": 0.2279,
"step": 950
},
{
"epoch": 0.6204556471158507,
"grad_norm": 3.269622564315796,
"learning_rate": 2.108477011494253e-05,
"loss": 0.2259,
"step": 960
},
{
"epoch": 0.6269187267733075,
"grad_norm": 3.5994322299957275,
"learning_rate": 2.0725574712643678e-05,
"loss": 0.1996,
"step": 970
},
{
"epoch": 0.6333818064307642,
"grad_norm": 4.982938289642334,
"learning_rate": 2.036637931034483e-05,
"loss": 0.2311,
"step": 980
},
{
"epoch": 0.639844886088221,
"grad_norm": 5.611133098602295,
"learning_rate": 2.0007183908045978e-05,
"loss": 0.2073,
"step": 990
},
{
"epoch": 0.6463079657456778,
"grad_norm": 4.818994522094727,
"learning_rate": 1.964798850574713e-05,
"loss": 0.2233,
"step": 1000
},
{
"epoch": 0.6527710454031346,
"grad_norm": 2.8192803859710693,
"learning_rate": 1.9288793103448278e-05,
"loss": 0.1841,
"step": 1010
},
{
"epoch": 0.6592341250605914,
"grad_norm": 3.74814510345459,
"learning_rate": 1.8929597701149426e-05,
"loss": 0.1908,
"step": 1020
},
{
"epoch": 0.6656972047180482,
"grad_norm": 5.619290351867676,
"learning_rate": 1.8570402298850574e-05,
"loss": 0.2135,
"step": 1030
},
{
"epoch": 0.6721602843755049,
"grad_norm": 3.431131362915039,
"learning_rate": 1.8211206896551726e-05,
"loss": 0.2036,
"step": 1040
},
{
"epoch": 0.6786233640329618,
"grad_norm": 4.3107781410217285,
"learning_rate": 1.7852011494252874e-05,
"loss": 0.2013,
"step": 1050
},
{
"epoch": 0.6850864436904185,
"grad_norm": 2.6932215690612793,
"learning_rate": 1.7492816091954022e-05,
"loss": 0.1654,
"step": 1060
},
{
"epoch": 0.6915495233478752,
"grad_norm": 5.07429313659668,
"learning_rate": 1.7133620689655174e-05,
"loss": 0.2128,
"step": 1070
},
{
"epoch": 0.6980126030053321,
"grad_norm": 6.4850873947143555,
"learning_rate": 1.6774425287356325e-05,
"loss": 0.1926,
"step": 1080
},
{
"epoch": 0.7044756826627888,
"grad_norm": 4.696744441986084,
"learning_rate": 1.6415229885057474e-05,
"loss": 0.2329,
"step": 1090
},
{
"epoch": 0.7109387623202456,
"grad_norm": 7.928685188293457,
"learning_rate": 1.6056034482758622e-05,
"loss": 0.2494,
"step": 1100
},
{
"epoch": 0.7174018419777024,
"grad_norm": 6.6810455322265625,
"learning_rate": 1.569683908045977e-05,
"loss": 0.2082,
"step": 1110
},
{
"epoch": 0.7238649216351591,
"grad_norm": 6.868622303009033,
"learning_rate": 1.533764367816092e-05,
"loss": 0.2153,
"step": 1120
},
{
"epoch": 0.7303280012926159,
"grad_norm": 4.763729095458984,
"learning_rate": 1.497844827586207e-05,
"loss": 0.1972,
"step": 1130
},
{
"epoch": 0.7367910809500727,
"grad_norm": 3.208186149597168,
"learning_rate": 1.461925287356322e-05,
"loss": 0.201,
"step": 1140
},
{
"epoch": 0.7432541606075295,
"grad_norm": 4.111704349517822,
"learning_rate": 1.4260057471264368e-05,
"loss": 0.2154,
"step": 1150
},
{
"epoch": 0.7497172402649863,
"grad_norm": 6.802502632141113,
"learning_rate": 1.3900862068965518e-05,
"loss": 0.1819,
"step": 1160
},
{
"epoch": 0.7561803199224431,
"grad_norm": 4.15958833694458,
"learning_rate": 1.3541666666666666e-05,
"loss": 0.1823,
"step": 1170
},
{
"epoch": 0.7626433995798998,
"grad_norm": 2.806739568710327,
"learning_rate": 1.3182471264367816e-05,
"loss": 0.1963,
"step": 1180
},
{
"epoch": 0.7691064792373566,
"grad_norm": 5.593130111694336,
"learning_rate": 1.2823275862068968e-05,
"loss": 0.1948,
"step": 1190
},
{
"epoch": 0.7755695588948134,
"grad_norm": 7.160956859588623,
"learning_rate": 1.2464080459770116e-05,
"loss": 0.2249,
"step": 1200
},
{
"epoch": 0.7820326385522701,
"grad_norm": 5.844871520996094,
"learning_rate": 1.2104885057471264e-05,
"loss": 0.2109,
"step": 1210
},
{
"epoch": 0.788495718209727,
"grad_norm": 4.273647785186768,
"learning_rate": 1.1745689655172416e-05,
"loss": 0.1976,
"step": 1220
},
{
"epoch": 0.7949587978671837,
"grad_norm": 5.624157428741455,
"learning_rate": 1.1386494252873564e-05,
"loss": 0.2017,
"step": 1230
},
{
"epoch": 0.8014218775246404,
"grad_norm": 5.627654552459717,
"learning_rate": 1.1027298850574714e-05,
"loss": 0.2153,
"step": 1240
},
{
"epoch": 0.8078849571820973,
"grad_norm": 4.900112628936768,
"learning_rate": 1.0668103448275862e-05,
"loss": 0.2436,
"step": 1250
},
{
"epoch": 0.814348036839554,
"grad_norm": 4.697551727294922,
"learning_rate": 1.0308908045977012e-05,
"loss": 0.1912,
"step": 1260
},
{
"epoch": 0.8208111164970108,
"grad_norm": 7.477961540222168,
"learning_rate": 9.949712643678162e-06,
"loss": 0.2109,
"step": 1270
},
{
"epoch": 0.8272741961544676,
"grad_norm": 4.095799922943115,
"learning_rate": 9.590517241379312e-06,
"loss": 0.2129,
"step": 1280
},
{
"epoch": 0.8337372758119244,
"grad_norm": 6.4529924392700195,
"learning_rate": 9.23132183908046e-06,
"loss": 0.2272,
"step": 1290
},
{
"epoch": 0.8402003554693812,
"grad_norm": 5.603827476501465,
"learning_rate": 8.87212643678161e-06,
"loss": 0.1771,
"step": 1300
},
{
"epoch": 0.846663435126838,
"grad_norm": 9.214598655700684,
"learning_rate": 8.512931034482758e-06,
"loss": 0.1775,
"step": 1310
},
{
"epoch": 0.8531265147842947,
"grad_norm": 3.415295124053955,
"learning_rate": 8.153735632183908e-06,
"loss": 0.1973,
"step": 1320
},
{
"epoch": 0.8595895944417515,
"grad_norm": 3.975978136062622,
"learning_rate": 7.794540229885058e-06,
"loss": 0.187,
"step": 1330
},
{
"epoch": 0.8660526740992083,
"grad_norm": 2.804140090942383,
"learning_rate": 7.435344827586208e-06,
"loss": 0.2051,
"step": 1340
},
{
"epoch": 0.872515753756665,
"grad_norm": 6.0080647468566895,
"learning_rate": 7.076149425287357e-06,
"loss": 0.216,
"step": 1350
},
{
"epoch": 0.8789788334141219,
"grad_norm": 4.182724475860596,
"learning_rate": 6.716954022988506e-06,
"loss": 0.1982,
"step": 1360
},
{
"epoch": 0.8854419130715786,
"grad_norm": 7.1342854499816895,
"learning_rate": 6.357758620689655e-06,
"loss": 0.1881,
"step": 1370
},
{
"epoch": 0.8919049927290353,
"grad_norm": 4.865418434143066,
"learning_rate": 5.998563218390805e-06,
"loss": 0.2099,
"step": 1380
},
{
"epoch": 0.8983680723864922,
"grad_norm": 4.020890235900879,
"learning_rate": 5.639367816091954e-06,
"loss": 0.1819,
"step": 1390
},
{
"epoch": 0.9048311520439489,
"grad_norm": 3.826892852783203,
"learning_rate": 5.280172413793104e-06,
"loss": 0.1846,
"step": 1400
},
{
"epoch": 0.9112942317014057,
"grad_norm": 5.284408092498779,
"learning_rate": 4.920977011494253e-06,
"loss": 0.1888,
"step": 1410
},
{
"epoch": 0.9177573113588625,
"grad_norm": 4.285012722015381,
"learning_rate": 4.561781609195403e-06,
"loss": 0.1886,
"step": 1420
},
{
"epoch": 0.9242203910163193,
"grad_norm": 3.0137135982513428,
"learning_rate": 4.202586206896552e-06,
"loss": 0.1976,
"step": 1430
},
{
"epoch": 0.9306834706737761,
"grad_norm": 7.759015083312988,
"learning_rate": 3.843390804597701e-06,
"loss": 0.1891,
"step": 1440
},
{
"epoch": 0.9371465503312328,
"grad_norm": 4.454775810241699,
"learning_rate": 3.484195402298851e-06,
"loss": 0.2085,
"step": 1450
},
{
"epoch": 0.9436096299886896,
"grad_norm": 4.0864949226379395,
"learning_rate": 3.125e-06,
"loss": 0.1859,
"step": 1460
},
{
"epoch": 0.9500727096461464,
"grad_norm": 6.020027160644531,
"learning_rate": 2.7658045977011496e-06,
"loss": 0.1994,
"step": 1470
},
{
"epoch": 0.9565357893036032,
"grad_norm": 5.507518768310547,
"learning_rate": 2.4066091954022987e-06,
"loss": 0.1908,
"step": 1480
},
{
"epoch": 0.9629988689610599,
"grad_norm": 4.609635829925537,
"learning_rate": 2.047413793103448e-06,
"loss": 0.2239,
"step": 1490
},
{
"epoch": 0.9694619486185168,
"grad_norm": 5.80496072769165,
"learning_rate": 1.6882183908045979e-06,
"loss": 0.1824,
"step": 1500
},
{
"epoch": 0.9759250282759735,
"grad_norm": 4.583664417266846,
"learning_rate": 1.3290229885057471e-06,
"loss": 0.1866,
"step": 1510
},
{
"epoch": 0.9823881079334302,
"grad_norm": 5.140854358673096,
"learning_rate": 9.698275862068966e-07,
"loss": 0.1693,
"step": 1520
},
{
"epoch": 0.9888511875908871,
"grad_norm": 4.870802402496338,
"learning_rate": 6.106321839080461e-07,
"loss": 0.1786,
"step": 1530
},
{
"epoch": 0.9953142672483438,
"grad_norm": 4.085971832275391,
"learning_rate": 2.514367816091954e-07,
"loss": 0.1819,
"step": 1540
},
{
"epoch": 0.9998384230085636,
"eval_accuracy": 0.9605053856292324,
"eval_loss": 0.11349575221538544,
"eval_runtime": 169.4025,
"eval_samples_per_second": 129.886,
"eval_steps_per_second": 4.061,
"step": 1547
},
{
"epoch": 0.9998384230085636,
"step": 1547,
"total_flos": 4.921883515023262e+18,
"train_loss": 0.25020766805354133,
"train_runtime": 3368.4805,
"train_samples_per_second": 58.787,
"train_steps_per_second": 0.459
}
],
"logging_steps": 10,
"max_steps": 1547,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"total_flos": 4.921883515023262e+18,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}