{ "best_metric": 0.5242346938775511, "best_model_checkpoint": "swin-tiny-patch4-window7-224-finetuned-ginger\\checkpoint-551", "epoch": 2.9918404351767904, "eval_steps": 500, "global_step": 825, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03626473254759746, "grad_norm": 66.77265930175781, "learning_rate": 6.024096385542169e-06, "loss": 2.0316, "step": 10 }, { "epoch": 0.07252946509519492, "grad_norm": 98.40676879882812, "learning_rate": 1.2048192771084338e-05, "loss": 1.8051, "step": 20 }, { "epoch": 0.10879419764279238, "grad_norm": 75.0258560180664, "learning_rate": 1.8072289156626505e-05, "loss": 1.2977, "step": 30 }, { "epoch": 0.14505893019038985, "grad_norm": 193.25062561035156, "learning_rate": 2.4096385542168677e-05, "loss": 0.7561, "step": 40 }, { "epoch": 0.1813236627379873, "grad_norm": 448.8358459472656, "learning_rate": 3.012048192771085e-05, "loss": 0.5479, "step": 50 }, { "epoch": 0.21758839528558477, "grad_norm": 450.2969970703125, "learning_rate": 3.614457831325301e-05, "loss": 0.5176, "step": 60 }, { "epoch": 0.25385312783318226, "grad_norm": 179.40008544921875, "learning_rate": 4.2168674698795186e-05, "loss": 0.4498, "step": 70 }, { "epoch": 0.2901178603807797, "grad_norm": 431.70550537109375, "learning_rate": 4.8192771084337354e-05, "loss": 0.3653, "step": 80 }, { "epoch": 0.3263825929283772, "grad_norm": 448.7202453613281, "learning_rate": 4.952830188679246e-05, "loss": 0.3826, "step": 90 }, { "epoch": 0.3626473254759746, "grad_norm": 146.0322723388672, "learning_rate": 4.88544474393531e-05, "loss": 0.3937, "step": 100 }, { "epoch": 0.3989120580235721, "grad_norm": 230.89141845703125, "learning_rate": 4.818059299191375e-05, "loss": 0.3891, "step": 110 }, { "epoch": 0.43517679057116954, "grad_norm": 291.6907043457031, "learning_rate": 4.750673854447439e-05, "loss": 0.4013, "step": 120 }, { "epoch": 0.471441523118767, "grad_norm": 299.92999267578125, "learning_rate": 4.683288409703504e-05, "loss": 0.3297, "step": 130 }, { "epoch": 0.5077062556663645, "grad_norm": 158.54531860351562, "learning_rate": 4.615902964959569e-05, "loss": 0.317, "step": 140 }, { "epoch": 0.543970988213962, "grad_norm": 171.34927368164062, "learning_rate": 4.548517520215634e-05, "loss": 0.3041, "step": 150 }, { "epoch": 0.5802357207615594, "grad_norm": 359.9927978515625, "learning_rate": 4.4811320754716985e-05, "loss": 0.3252, "step": 160 }, { "epoch": 0.6165004533091568, "grad_norm": 191.6513671875, "learning_rate": 4.413746630727763e-05, "loss": 0.3096, "step": 170 }, { "epoch": 0.6527651858567544, "grad_norm": 212.29483032226562, "learning_rate": 4.3463611859838275e-05, "loss": 0.2636, "step": 180 }, { "epoch": 0.6890299184043518, "grad_norm": 141.14498901367188, "learning_rate": 4.2789757412398926e-05, "loss": 0.2549, "step": 190 }, { "epoch": 0.7252946509519492, "grad_norm": 146.6569061279297, "learning_rate": 4.211590296495957e-05, "loss": 0.2518, "step": 200 }, { "epoch": 0.7615593834995467, "grad_norm": 126.00825500488281, "learning_rate": 4.1442048517520216e-05, "loss": 0.2204, "step": 210 }, { "epoch": 0.7978241160471442, "grad_norm": 181.69412231445312, "learning_rate": 4.076819407008086e-05, "loss": 0.2087, "step": 220 }, { "epoch": 0.8340888485947416, "grad_norm": 555.3983154296875, "learning_rate": 4.009433962264151e-05, "loss": 0.2254, "step": 230 }, { "epoch": 0.8703535811423391, "grad_norm": 289.5895690917969, "learning_rate": 3.942048517520216e-05, "loss": 0.2591, "step": 240 }, { "epoch": 0.9066183136899365, "grad_norm": 171.5518798828125, "learning_rate": 3.874663072776281e-05, "loss": 0.2474, "step": 250 }, { "epoch": 0.942883046237534, "grad_norm": 245.42327880859375, "learning_rate": 3.807277628032345e-05, "loss": 0.249, "step": 260 }, { "epoch": 0.9791477787851315, "grad_norm": 481.2547302246094, "learning_rate": 3.73989218328841e-05, "loss": 0.241, "step": 270 }, { "epoch": 0.9972801450589301, "eval_accuracy": 0.510969387755102, "eval_loss": 1.5232751369476318, "eval_runtime": 14.3392, "eval_samples_per_second": 273.376, "eval_steps_per_second": 8.578, "step": 275 }, { "epoch": 1.015412511332729, "grad_norm": 127.02649688720703, "learning_rate": 3.672506738544474e-05, "loss": 0.2281, "step": 280 }, { "epoch": 1.0516772438803264, "grad_norm": 288.0721130371094, "learning_rate": 3.605121293800539e-05, "loss": 0.2058, "step": 290 }, { "epoch": 1.087941976427924, "grad_norm": 349.9566345214844, "learning_rate": 3.537735849056604e-05, "loss": 0.2228, "step": 300 }, { "epoch": 1.1242067089755212, "grad_norm": 268.7991027832031, "learning_rate": 3.470350404312669e-05, "loss": 0.2157, "step": 310 }, { "epoch": 1.1604714415231188, "grad_norm": 119.46910858154297, "learning_rate": 3.4029649595687336e-05, "loss": 0.1892, "step": 320 }, { "epoch": 1.1967361740707163, "grad_norm": 151.6316680908203, "learning_rate": 3.335579514824798e-05, "loss": 0.2094, "step": 330 }, { "epoch": 1.2330009066183136, "grad_norm": 179.56629943847656, "learning_rate": 3.2681940700808625e-05, "loss": 0.2061, "step": 340 }, { "epoch": 1.2692656391659112, "grad_norm": 111.70588684082031, "learning_rate": 3.200808625336928e-05, "loss": 0.2413, "step": 350 }, { "epoch": 1.3055303717135085, "grad_norm": 349.2808837890625, "learning_rate": 3.133423180592992e-05, "loss": 0.1837, "step": 360 }, { "epoch": 1.341795104261106, "grad_norm": 188.50990295410156, "learning_rate": 3.0660377358490567e-05, "loss": 0.2154, "step": 370 }, { "epoch": 1.3780598368087036, "grad_norm": 190.62208557128906, "learning_rate": 2.998652291105121e-05, "loss": 0.1856, "step": 380 }, { "epoch": 1.414324569356301, "grad_norm": 493.2279968261719, "learning_rate": 2.931266846361186e-05, "loss": 0.1999, "step": 390 }, { "epoch": 1.4505893019038985, "grad_norm": 244.08197021484375, "learning_rate": 2.863881401617251e-05, "loss": 0.1687, "step": 400 }, { "epoch": 1.486854034451496, "grad_norm": 85.91020965576172, "learning_rate": 2.7964959568733156e-05, "loss": 0.1607, "step": 410 }, { "epoch": 1.5231187669990933, "grad_norm": 80.52980041503906, "learning_rate": 2.7291105121293804e-05, "loss": 0.1967, "step": 420 }, { "epoch": 1.5593834995466909, "grad_norm": 623.0601806640625, "learning_rate": 2.661725067385445e-05, "loss": 0.1779, "step": 430 }, { "epoch": 1.5956482320942884, "grad_norm": 77.6041488647461, "learning_rate": 2.5943396226415094e-05, "loss": 0.1588, "step": 440 }, { "epoch": 1.6319129646418857, "grad_norm": 126.40304565429688, "learning_rate": 2.5269541778975742e-05, "loss": 0.1516, "step": 450 }, { "epoch": 1.6681776971894833, "grad_norm": 219.29595947265625, "learning_rate": 2.459568733153639e-05, "loss": 0.1477, "step": 460 }, { "epoch": 1.7044424297370808, "grad_norm": 144.26341247558594, "learning_rate": 2.3921832884097038e-05, "loss": 0.1714, "step": 470 }, { "epoch": 1.7407071622846781, "grad_norm": 177.91326904296875, "learning_rate": 2.3247978436657683e-05, "loss": 0.1446, "step": 480 }, { "epoch": 1.7769718948322755, "grad_norm": 160.88905334472656, "learning_rate": 2.2574123989218328e-05, "loss": 0.1681, "step": 490 }, { "epoch": 1.8132366273798732, "grad_norm": 118.04940032958984, "learning_rate": 2.1900269541778976e-05, "loss": 0.153, "step": 500 }, { "epoch": 1.8495013599274706, "grad_norm": 42.61888122558594, "learning_rate": 2.1226415094339624e-05, "loss": 0.1268, "step": 510 }, { "epoch": 1.8857660924750679, "grad_norm": 383.0787353515625, "learning_rate": 2.055256064690027e-05, "loss": 0.1489, "step": 520 }, { "epoch": 1.9220308250226654, "grad_norm": 281.9739685058594, "learning_rate": 1.9878706199460917e-05, "loss": 0.1966, "step": 530 }, { "epoch": 1.958295557570263, "grad_norm": 175.12257385253906, "learning_rate": 1.9204851752021562e-05, "loss": 0.1479, "step": 540 }, { "epoch": 1.9945602901178603, "grad_norm": 250.56533813476562, "learning_rate": 1.8530997304582214e-05, "loss": 0.1661, "step": 550 }, { "epoch": 1.9981867633726202, "eval_accuracy": 0.5242346938775511, "eval_loss": 1.3872867822647095, "eval_runtime": 14.3219, "eval_samples_per_second": 273.707, "eval_steps_per_second": 8.588, "step": 551 }, { "epoch": 2.030825022665458, "grad_norm": 128.275146484375, "learning_rate": 1.785714285714286e-05, "loss": 0.1605, "step": 560 }, { "epoch": 2.0670897552130554, "grad_norm": 267.3365478515625, "learning_rate": 1.7183288409703503e-05, "loss": 0.1295, "step": 570 }, { "epoch": 2.1033544877606527, "grad_norm": 62.166568756103516, "learning_rate": 1.650943396226415e-05, "loss": 0.1209, "step": 580 }, { "epoch": 2.13961922030825, "grad_norm": 398.4904479980469, "learning_rate": 1.58355795148248e-05, "loss": 0.1541, "step": 590 }, { "epoch": 2.175883952855848, "grad_norm": 89.24071502685547, "learning_rate": 1.5161725067385446e-05, "loss": 0.1427, "step": 600 }, { "epoch": 2.212148685403445, "grad_norm": 227.34228515625, "learning_rate": 1.4487870619946093e-05, "loss": 0.1557, "step": 610 }, { "epoch": 2.2484134179510424, "grad_norm": 333.01739501953125, "learning_rate": 1.381401617250674e-05, "loss": 0.1194, "step": 620 }, { "epoch": 2.28467815049864, "grad_norm": 250.48745727539062, "learning_rate": 1.3140161725067384e-05, "loss": 0.1585, "step": 630 }, { "epoch": 2.3209428830462375, "grad_norm": 48.325416564941406, "learning_rate": 1.2466307277628032e-05, "loss": 0.1384, "step": 640 }, { "epoch": 2.357207615593835, "grad_norm": 111.5351791381836, "learning_rate": 1.179245283018868e-05, "loss": 0.1096, "step": 650 }, { "epoch": 2.3934723481414326, "grad_norm": 149.7464141845703, "learning_rate": 1.1118598382749327e-05, "loss": 0.1327, "step": 660 }, { "epoch": 2.42973708068903, "grad_norm": 145.99148559570312, "learning_rate": 1.0444743935309973e-05, "loss": 0.099, "step": 670 }, { "epoch": 2.4660018132366273, "grad_norm": 67.97899627685547, "learning_rate": 9.77088948787062e-06, "loss": 0.1545, "step": 680 }, { "epoch": 2.5022665457842246, "grad_norm": 288.68731689453125, "learning_rate": 9.097035040431268e-06, "loss": 0.1155, "step": 690 }, { "epoch": 2.5385312783318223, "grad_norm": 70.22994232177734, "learning_rate": 8.423180592991915e-06, "loss": 0.1244, "step": 700 }, { "epoch": 2.5747960108794197, "grad_norm": 98.32154083251953, "learning_rate": 7.749326145552561e-06, "loss": 0.1182, "step": 710 }, { "epoch": 2.611060743427017, "grad_norm": 107.6678466796875, "learning_rate": 7.0754716981132075e-06, "loss": 0.1247, "step": 720 }, { "epoch": 2.6473254759746148, "grad_norm": 61.7956657409668, "learning_rate": 6.401617250673856e-06, "loss": 0.1066, "step": 730 }, { "epoch": 2.683590208522212, "grad_norm": 105.1893539428711, "learning_rate": 5.727762803234501e-06, "loss": 0.1216, "step": 740 }, { "epoch": 2.7198549410698094, "grad_norm": 170.1961212158203, "learning_rate": 5.053908355795149e-06, "loss": 0.1394, "step": 750 }, { "epoch": 2.756119673617407, "grad_norm": 202.67616271972656, "learning_rate": 4.380053908355795e-06, "loss": 0.0923, "step": 760 }, { "epoch": 2.7923844061650045, "grad_norm": 73.72498321533203, "learning_rate": 3.706199460916442e-06, "loss": 0.112, "step": 770 }, { "epoch": 2.828649138712602, "grad_norm": 50.71122741699219, "learning_rate": 3.032345013477089e-06, "loss": 0.1025, "step": 780 }, { "epoch": 2.8649138712601996, "grad_norm": 65.91510009765625, "learning_rate": 2.358490566037736e-06, "loss": 0.0911, "step": 790 }, { "epoch": 2.901178603807797, "grad_norm": 72.48229217529297, "learning_rate": 1.6846361185983827e-06, "loss": 0.1068, "step": 800 }, { "epoch": 2.9374433363553942, "grad_norm": 156.2762451171875, "learning_rate": 1.0107816711590296e-06, "loss": 0.1022, "step": 810 }, { "epoch": 2.973708068902992, "grad_norm": 270.1289367675781, "learning_rate": 3.369272237196766e-07, "loss": 0.1177, "step": 820 }, { "epoch": 2.9918404351767904, "eval_accuracy": 0.5104591836734694, "eval_loss": 1.3372142314910889, "eval_runtime": 14.3839, "eval_samples_per_second": 272.527, "eval_steps_per_second": 8.551, "step": 825 }, { "epoch": 2.9918404351767904, "step": 825, "total_flos": 2.6244369700391485e+18, "train_loss": 0.26469580238515683, "train_runtime": 846.2486, "train_samples_per_second": 125.07, "train_steps_per_second": 0.975 } ], "logging_steps": 10, "max_steps": 825, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.6244369700391485e+18, "train_batch_size": 32, "trial_name": null, "trial_params": null }