|
{
|
|
"best_metric": 0.5242346938775511,
|
|
"best_model_checkpoint": "swin-tiny-patch4-window7-224-finetuned-ginger\\checkpoint-551",
|
|
"epoch": 2.9918404351767904,
|
|
"eval_steps": 500,
|
|
"global_step": 825,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.03626473254759746,
|
|
"grad_norm": 66.77265930175781,
|
|
"learning_rate": 6.024096385542169e-06,
|
|
"loss": 2.0316,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.07252946509519492,
|
|
"grad_norm": 98.40676879882812,
|
|
"learning_rate": 1.2048192771084338e-05,
|
|
"loss": 1.8051,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.10879419764279238,
|
|
"grad_norm": 75.0258560180664,
|
|
"learning_rate": 1.8072289156626505e-05,
|
|
"loss": 1.2977,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.14505893019038985,
|
|
"grad_norm": 193.25062561035156,
|
|
"learning_rate": 2.4096385542168677e-05,
|
|
"loss": 0.7561,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 0.1813236627379873,
|
|
"grad_norm": 448.8358459472656,
|
|
"learning_rate": 3.012048192771085e-05,
|
|
"loss": 0.5479,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.21758839528558477,
|
|
"grad_norm": 450.2969970703125,
|
|
"learning_rate": 3.614457831325301e-05,
|
|
"loss": 0.5176,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 0.25385312783318226,
|
|
"grad_norm": 179.40008544921875,
|
|
"learning_rate": 4.2168674698795186e-05,
|
|
"loss": 0.4498,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 0.2901178603807797,
|
|
"grad_norm": 431.70550537109375,
|
|
"learning_rate": 4.8192771084337354e-05,
|
|
"loss": 0.3653,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 0.3263825929283772,
|
|
"grad_norm": 448.7202453613281,
|
|
"learning_rate": 4.952830188679246e-05,
|
|
"loss": 0.3826,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 0.3626473254759746,
|
|
"grad_norm": 146.0322723388672,
|
|
"learning_rate": 4.88544474393531e-05,
|
|
"loss": 0.3937,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.3989120580235721,
|
|
"grad_norm": 230.89141845703125,
|
|
"learning_rate": 4.818059299191375e-05,
|
|
"loss": 0.3891,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 0.43517679057116954,
|
|
"grad_norm": 291.6907043457031,
|
|
"learning_rate": 4.750673854447439e-05,
|
|
"loss": 0.4013,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 0.471441523118767,
|
|
"grad_norm": 299.92999267578125,
|
|
"learning_rate": 4.683288409703504e-05,
|
|
"loss": 0.3297,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 0.5077062556663645,
|
|
"grad_norm": 158.54531860351562,
|
|
"learning_rate": 4.615902964959569e-05,
|
|
"loss": 0.317,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 0.543970988213962,
|
|
"grad_norm": 171.34927368164062,
|
|
"learning_rate": 4.548517520215634e-05,
|
|
"loss": 0.3041,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.5802357207615594,
|
|
"grad_norm": 359.9927978515625,
|
|
"learning_rate": 4.4811320754716985e-05,
|
|
"loss": 0.3252,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 0.6165004533091568,
|
|
"grad_norm": 191.6513671875,
|
|
"learning_rate": 4.413746630727763e-05,
|
|
"loss": 0.3096,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 0.6527651858567544,
|
|
"grad_norm": 212.29483032226562,
|
|
"learning_rate": 4.3463611859838275e-05,
|
|
"loss": 0.2636,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 0.6890299184043518,
|
|
"grad_norm": 141.14498901367188,
|
|
"learning_rate": 4.2789757412398926e-05,
|
|
"loss": 0.2549,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 0.7252946509519492,
|
|
"grad_norm": 146.6569061279297,
|
|
"learning_rate": 4.211590296495957e-05,
|
|
"loss": 0.2518,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.7615593834995467,
|
|
"grad_norm": 126.00825500488281,
|
|
"learning_rate": 4.1442048517520216e-05,
|
|
"loss": 0.2204,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 0.7978241160471442,
|
|
"grad_norm": 181.69412231445312,
|
|
"learning_rate": 4.076819407008086e-05,
|
|
"loss": 0.2087,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 0.8340888485947416,
|
|
"grad_norm": 555.3983154296875,
|
|
"learning_rate": 4.009433962264151e-05,
|
|
"loss": 0.2254,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 0.8703535811423391,
|
|
"grad_norm": 289.5895690917969,
|
|
"learning_rate": 3.942048517520216e-05,
|
|
"loss": 0.2591,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 0.9066183136899365,
|
|
"grad_norm": 171.5518798828125,
|
|
"learning_rate": 3.874663072776281e-05,
|
|
"loss": 0.2474,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 0.942883046237534,
|
|
"grad_norm": 245.42327880859375,
|
|
"learning_rate": 3.807277628032345e-05,
|
|
"loss": 0.249,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 0.9791477787851315,
|
|
"grad_norm": 481.2547302246094,
|
|
"learning_rate": 3.73989218328841e-05,
|
|
"loss": 0.241,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 0.9972801450589301,
|
|
"eval_accuracy": 0.510969387755102,
|
|
"eval_loss": 1.5232751369476318,
|
|
"eval_runtime": 14.3392,
|
|
"eval_samples_per_second": 273.376,
|
|
"eval_steps_per_second": 8.578,
|
|
"step": 275
|
|
},
|
|
{
|
|
"epoch": 1.015412511332729,
|
|
"grad_norm": 127.02649688720703,
|
|
"learning_rate": 3.672506738544474e-05,
|
|
"loss": 0.2281,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 1.0516772438803264,
|
|
"grad_norm": 288.0721130371094,
|
|
"learning_rate": 3.605121293800539e-05,
|
|
"loss": 0.2058,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 1.087941976427924,
|
|
"grad_norm": 349.9566345214844,
|
|
"learning_rate": 3.537735849056604e-05,
|
|
"loss": 0.2228,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 1.1242067089755212,
|
|
"grad_norm": 268.7991027832031,
|
|
"learning_rate": 3.470350404312669e-05,
|
|
"loss": 0.2157,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 1.1604714415231188,
|
|
"grad_norm": 119.46910858154297,
|
|
"learning_rate": 3.4029649595687336e-05,
|
|
"loss": 0.1892,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 1.1967361740707163,
|
|
"grad_norm": 151.6316680908203,
|
|
"learning_rate": 3.335579514824798e-05,
|
|
"loss": 0.2094,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 1.2330009066183136,
|
|
"grad_norm": 179.56629943847656,
|
|
"learning_rate": 3.2681940700808625e-05,
|
|
"loss": 0.2061,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 1.2692656391659112,
|
|
"grad_norm": 111.70588684082031,
|
|
"learning_rate": 3.200808625336928e-05,
|
|
"loss": 0.2413,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 1.3055303717135085,
|
|
"grad_norm": 349.2808837890625,
|
|
"learning_rate": 3.133423180592992e-05,
|
|
"loss": 0.1837,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 1.341795104261106,
|
|
"grad_norm": 188.50990295410156,
|
|
"learning_rate": 3.0660377358490567e-05,
|
|
"loss": 0.2154,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 1.3780598368087036,
|
|
"grad_norm": 190.62208557128906,
|
|
"learning_rate": 2.998652291105121e-05,
|
|
"loss": 0.1856,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 1.414324569356301,
|
|
"grad_norm": 493.2279968261719,
|
|
"learning_rate": 2.931266846361186e-05,
|
|
"loss": 0.1999,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 1.4505893019038985,
|
|
"grad_norm": 244.08197021484375,
|
|
"learning_rate": 2.863881401617251e-05,
|
|
"loss": 0.1687,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 1.486854034451496,
|
|
"grad_norm": 85.91020965576172,
|
|
"learning_rate": 2.7964959568733156e-05,
|
|
"loss": 0.1607,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 1.5231187669990933,
|
|
"grad_norm": 80.52980041503906,
|
|
"learning_rate": 2.7291105121293804e-05,
|
|
"loss": 0.1967,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 1.5593834995466909,
|
|
"grad_norm": 623.0601806640625,
|
|
"learning_rate": 2.661725067385445e-05,
|
|
"loss": 0.1779,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 1.5956482320942884,
|
|
"grad_norm": 77.6041488647461,
|
|
"learning_rate": 2.5943396226415094e-05,
|
|
"loss": 0.1588,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 1.6319129646418857,
|
|
"grad_norm": 126.40304565429688,
|
|
"learning_rate": 2.5269541778975742e-05,
|
|
"loss": 0.1516,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 1.6681776971894833,
|
|
"grad_norm": 219.29595947265625,
|
|
"learning_rate": 2.459568733153639e-05,
|
|
"loss": 0.1477,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 1.7044424297370808,
|
|
"grad_norm": 144.26341247558594,
|
|
"learning_rate": 2.3921832884097038e-05,
|
|
"loss": 0.1714,
|
|
"step": 470
|
|
},
|
|
{
|
|
"epoch": 1.7407071622846781,
|
|
"grad_norm": 177.91326904296875,
|
|
"learning_rate": 2.3247978436657683e-05,
|
|
"loss": 0.1446,
|
|
"step": 480
|
|
},
|
|
{
|
|
"epoch": 1.7769718948322755,
|
|
"grad_norm": 160.88905334472656,
|
|
"learning_rate": 2.2574123989218328e-05,
|
|
"loss": 0.1681,
|
|
"step": 490
|
|
},
|
|
{
|
|
"epoch": 1.8132366273798732,
|
|
"grad_norm": 118.04940032958984,
|
|
"learning_rate": 2.1900269541778976e-05,
|
|
"loss": 0.153,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 1.8495013599274706,
|
|
"grad_norm": 42.61888122558594,
|
|
"learning_rate": 2.1226415094339624e-05,
|
|
"loss": 0.1268,
|
|
"step": 510
|
|
},
|
|
{
|
|
"epoch": 1.8857660924750679,
|
|
"grad_norm": 383.0787353515625,
|
|
"learning_rate": 2.055256064690027e-05,
|
|
"loss": 0.1489,
|
|
"step": 520
|
|
},
|
|
{
|
|
"epoch": 1.9220308250226654,
|
|
"grad_norm": 281.9739685058594,
|
|
"learning_rate": 1.9878706199460917e-05,
|
|
"loss": 0.1966,
|
|
"step": 530
|
|
},
|
|
{
|
|
"epoch": 1.958295557570263,
|
|
"grad_norm": 175.12257385253906,
|
|
"learning_rate": 1.9204851752021562e-05,
|
|
"loss": 0.1479,
|
|
"step": 540
|
|
},
|
|
{
|
|
"epoch": 1.9945602901178603,
|
|
"grad_norm": 250.56533813476562,
|
|
"learning_rate": 1.8530997304582214e-05,
|
|
"loss": 0.1661,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 1.9981867633726202,
|
|
"eval_accuracy": 0.5242346938775511,
|
|
"eval_loss": 1.3872867822647095,
|
|
"eval_runtime": 14.3219,
|
|
"eval_samples_per_second": 273.707,
|
|
"eval_steps_per_second": 8.588,
|
|
"step": 551
|
|
},
|
|
{
|
|
"epoch": 2.030825022665458,
|
|
"grad_norm": 128.275146484375,
|
|
"learning_rate": 1.785714285714286e-05,
|
|
"loss": 0.1605,
|
|
"step": 560
|
|
},
|
|
{
|
|
"epoch": 2.0670897552130554,
|
|
"grad_norm": 267.3365478515625,
|
|
"learning_rate": 1.7183288409703503e-05,
|
|
"loss": 0.1295,
|
|
"step": 570
|
|
},
|
|
{
|
|
"epoch": 2.1033544877606527,
|
|
"grad_norm": 62.166568756103516,
|
|
"learning_rate": 1.650943396226415e-05,
|
|
"loss": 0.1209,
|
|
"step": 580
|
|
},
|
|
{
|
|
"epoch": 2.13961922030825,
|
|
"grad_norm": 398.4904479980469,
|
|
"learning_rate": 1.58355795148248e-05,
|
|
"loss": 0.1541,
|
|
"step": 590
|
|
},
|
|
{
|
|
"epoch": 2.175883952855848,
|
|
"grad_norm": 89.24071502685547,
|
|
"learning_rate": 1.5161725067385446e-05,
|
|
"loss": 0.1427,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 2.212148685403445,
|
|
"grad_norm": 227.34228515625,
|
|
"learning_rate": 1.4487870619946093e-05,
|
|
"loss": 0.1557,
|
|
"step": 610
|
|
},
|
|
{
|
|
"epoch": 2.2484134179510424,
|
|
"grad_norm": 333.01739501953125,
|
|
"learning_rate": 1.381401617250674e-05,
|
|
"loss": 0.1194,
|
|
"step": 620
|
|
},
|
|
{
|
|
"epoch": 2.28467815049864,
|
|
"grad_norm": 250.48745727539062,
|
|
"learning_rate": 1.3140161725067384e-05,
|
|
"loss": 0.1585,
|
|
"step": 630
|
|
},
|
|
{
|
|
"epoch": 2.3209428830462375,
|
|
"grad_norm": 48.325416564941406,
|
|
"learning_rate": 1.2466307277628032e-05,
|
|
"loss": 0.1384,
|
|
"step": 640
|
|
},
|
|
{
|
|
"epoch": 2.357207615593835,
|
|
"grad_norm": 111.5351791381836,
|
|
"learning_rate": 1.179245283018868e-05,
|
|
"loss": 0.1096,
|
|
"step": 650
|
|
},
|
|
{
|
|
"epoch": 2.3934723481414326,
|
|
"grad_norm": 149.7464141845703,
|
|
"learning_rate": 1.1118598382749327e-05,
|
|
"loss": 0.1327,
|
|
"step": 660
|
|
},
|
|
{
|
|
"epoch": 2.42973708068903,
|
|
"grad_norm": 145.99148559570312,
|
|
"learning_rate": 1.0444743935309973e-05,
|
|
"loss": 0.099,
|
|
"step": 670
|
|
},
|
|
{
|
|
"epoch": 2.4660018132366273,
|
|
"grad_norm": 67.97899627685547,
|
|
"learning_rate": 9.77088948787062e-06,
|
|
"loss": 0.1545,
|
|
"step": 680
|
|
},
|
|
{
|
|
"epoch": 2.5022665457842246,
|
|
"grad_norm": 288.68731689453125,
|
|
"learning_rate": 9.097035040431268e-06,
|
|
"loss": 0.1155,
|
|
"step": 690
|
|
},
|
|
{
|
|
"epoch": 2.5385312783318223,
|
|
"grad_norm": 70.22994232177734,
|
|
"learning_rate": 8.423180592991915e-06,
|
|
"loss": 0.1244,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 2.5747960108794197,
|
|
"grad_norm": 98.32154083251953,
|
|
"learning_rate": 7.749326145552561e-06,
|
|
"loss": 0.1182,
|
|
"step": 710
|
|
},
|
|
{
|
|
"epoch": 2.611060743427017,
|
|
"grad_norm": 107.6678466796875,
|
|
"learning_rate": 7.0754716981132075e-06,
|
|
"loss": 0.1247,
|
|
"step": 720
|
|
},
|
|
{
|
|
"epoch": 2.6473254759746148,
|
|
"grad_norm": 61.7956657409668,
|
|
"learning_rate": 6.401617250673856e-06,
|
|
"loss": 0.1066,
|
|
"step": 730
|
|
},
|
|
{
|
|
"epoch": 2.683590208522212,
|
|
"grad_norm": 105.1893539428711,
|
|
"learning_rate": 5.727762803234501e-06,
|
|
"loss": 0.1216,
|
|
"step": 740
|
|
},
|
|
{
|
|
"epoch": 2.7198549410698094,
|
|
"grad_norm": 170.1961212158203,
|
|
"learning_rate": 5.053908355795149e-06,
|
|
"loss": 0.1394,
|
|
"step": 750
|
|
},
|
|
{
|
|
"epoch": 2.756119673617407,
|
|
"grad_norm": 202.67616271972656,
|
|
"learning_rate": 4.380053908355795e-06,
|
|
"loss": 0.0923,
|
|
"step": 760
|
|
},
|
|
{
|
|
"epoch": 2.7923844061650045,
|
|
"grad_norm": 73.72498321533203,
|
|
"learning_rate": 3.706199460916442e-06,
|
|
"loss": 0.112,
|
|
"step": 770
|
|
},
|
|
{
|
|
"epoch": 2.828649138712602,
|
|
"grad_norm": 50.71122741699219,
|
|
"learning_rate": 3.032345013477089e-06,
|
|
"loss": 0.1025,
|
|
"step": 780
|
|
},
|
|
{
|
|
"epoch": 2.8649138712601996,
|
|
"grad_norm": 65.91510009765625,
|
|
"learning_rate": 2.358490566037736e-06,
|
|
"loss": 0.0911,
|
|
"step": 790
|
|
},
|
|
{
|
|
"epoch": 2.901178603807797,
|
|
"grad_norm": 72.48229217529297,
|
|
"learning_rate": 1.6846361185983827e-06,
|
|
"loss": 0.1068,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 2.9374433363553942,
|
|
"grad_norm": 156.2762451171875,
|
|
"learning_rate": 1.0107816711590296e-06,
|
|
"loss": 0.1022,
|
|
"step": 810
|
|
},
|
|
{
|
|
"epoch": 2.973708068902992,
|
|
"grad_norm": 270.1289367675781,
|
|
"learning_rate": 3.369272237196766e-07,
|
|
"loss": 0.1177,
|
|
"step": 820
|
|
},
|
|
{
|
|
"epoch": 2.9918404351767904,
|
|
"eval_accuracy": 0.5104591836734694,
|
|
"eval_loss": 1.3372142314910889,
|
|
"eval_runtime": 14.3839,
|
|
"eval_samples_per_second": 272.527,
|
|
"eval_steps_per_second": 8.551,
|
|
"step": 825
|
|
},
|
|
{
|
|
"epoch": 2.9918404351767904,
|
|
"step": 825,
|
|
"total_flos": 2.6244369700391485e+18,
|
|
"train_loss": 0.26469580238515683,
|
|
"train_runtime": 846.2486,
|
|
"train_samples_per_second": 125.07,
|
|
"train_steps_per_second": 0.975
|
|
}
|
|
],
|
|
"logging_steps": 10,
|
|
"max_steps": 825,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 3,
|
|
"save_steps": 500,
|
|
"stateful_callbacks": {
|
|
"TrainerControl": {
|
|
"args": {
|
|
"should_epoch_stop": false,
|
|
"should_evaluate": false,
|
|
"should_log": false,
|
|
"should_save": true,
|
|
"should_training_stop": true
|
|
},
|
|
"attributes": {}
|
|
}
|
|
},
|
|
"total_flos": 2.6244369700391485e+18,
|
|
"train_batch_size": 32,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|
|
|