k-code's picture
add check point
47e3f8f
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 2391,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.012547051442910916,
"grad_norm": 0.716044008731842,
"learning_rate": 1.9924717691342537e-05,
"loss": 1.0972,
"step": 10
},
{
"epoch": 0.025094102885821833,
"grad_norm": 1.4613193273544312,
"learning_rate": 1.9841070681723128e-05,
"loss": 1.088,
"step": 20
},
{
"epoch": 0.037641154328732745,
"grad_norm": 0.627024233341217,
"learning_rate": 1.9757423672103725e-05,
"loss": 1.0716,
"step": 30
},
{
"epoch": 0.050188205771643665,
"grad_norm": 1.224279522895813,
"learning_rate": 1.9673776662484318e-05,
"loss": 1.0661,
"step": 40
},
{
"epoch": 0.06273525721455459,
"grad_norm": 1.368181586265564,
"learning_rate": 1.959012965286491e-05,
"loss": 1.035,
"step": 50
},
{
"epoch": 0.07528230865746549,
"grad_norm": 1.0884851217269897,
"learning_rate": 1.9506482643245505e-05,
"loss": 1.0288,
"step": 60
},
{
"epoch": 0.08782936010037641,
"grad_norm": 1.9844820499420166,
"learning_rate": 1.9422835633626102e-05,
"loss": 0.9855,
"step": 70
},
{
"epoch": 0.10037641154328733,
"grad_norm": 1.6572856903076172,
"learning_rate": 1.9339188624006692e-05,
"loss": 0.991,
"step": 80
},
{
"epoch": 0.11292346298619825,
"grad_norm": 2.2602856159210205,
"learning_rate": 1.925554161438729e-05,
"loss": 1.0092,
"step": 90
},
{
"epoch": 0.12547051442910917,
"grad_norm": 4.111606597900391,
"learning_rate": 1.9171894604767882e-05,
"loss": 0.9062,
"step": 100
},
{
"epoch": 0.13801756587202008,
"grad_norm": 3.383659601211548,
"learning_rate": 1.9088247595148476e-05,
"loss": 0.9749,
"step": 110
},
{
"epoch": 0.15056461731493098,
"grad_norm": 1.9777194261550903,
"learning_rate": 1.900460058552907e-05,
"loss": 0.9193,
"step": 120
},
{
"epoch": 0.16311166875784192,
"grad_norm": 6.869212627410889,
"learning_rate": 1.8920953575909663e-05,
"loss": 0.9695,
"step": 130
},
{
"epoch": 0.17565872020075282,
"grad_norm": 2.800828695297241,
"learning_rate": 1.8837306566290257e-05,
"loss": 0.9375,
"step": 140
},
{
"epoch": 0.18820577164366373,
"grad_norm": 3.783500909805298,
"learning_rate": 1.875365955667085e-05,
"loss": 0.9294,
"step": 150
},
{
"epoch": 0.20075282308657466,
"grad_norm": 2.961916208267212,
"learning_rate": 1.8670012547051444e-05,
"loss": 0.9651,
"step": 160
},
{
"epoch": 0.21329987452948557,
"grad_norm": 4.599372386932373,
"learning_rate": 1.8586365537432037e-05,
"loss": 0.9706,
"step": 170
},
{
"epoch": 0.2258469259723965,
"grad_norm": 3.3036141395568848,
"learning_rate": 1.8502718527812634e-05,
"loss": 0.9289,
"step": 180
},
{
"epoch": 0.2383939774153074,
"grad_norm": 6.022705078125,
"learning_rate": 1.8419071518193224e-05,
"loss": 0.9015,
"step": 190
},
{
"epoch": 0.25094102885821834,
"grad_norm": 13.482539176940918,
"learning_rate": 1.833542450857382e-05,
"loss": 0.9033,
"step": 200
},
{
"epoch": 0.26348808030112925,
"grad_norm": 4.047461032867432,
"learning_rate": 1.8251777498954414e-05,
"loss": 0.8447,
"step": 210
},
{
"epoch": 0.27603513174404015,
"grad_norm": 2.8806040287017822,
"learning_rate": 1.8168130489335008e-05,
"loss": 0.8333,
"step": 220
},
{
"epoch": 0.28858218318695106,
"grad_norm": 5.517472743988037,
"learning_rate": 1.80844834797156e-05,
"loss": 0.8697,
"step": 230
},
{
"epoch": 0.30112923462986196,
"grad_norm": 7.646421432495117,
"learning_rate": 1.8000836470096195e-05,
"loss": 0.9159,
"step": 240
},
{
"epoch": 0.3136762860727729,
"grad_norm": 4.211366176605225,
"learning_rate": 1.791718946047679e-05,
"loss": 0.8848,
"step": 250
},
{
"epoch": 0.32622333751568383,
"grad_norm": 7.600694179534912,
"learning_rate": 1.7833542450857385e-05,
"loss": 0.8085,
"step": 260
},
{
"epoch": 0.33877038895859474,
"grad_norm": 4.872680187225342,
"learning_rate": 1.7749895441237976e-05,
"loss": 0.8301,
"step": 270
},
{
"epoch": 0.35131744040150564,
"grad_norm": 6.380251884460449,
"learning_rate": 1.7666248431618572e-05,
"loss": 0.9484,
"step": 280
},
{
"epoch": 0.36386449184441655,
"grad_norm": 4.365920066833496,
"learning_rate": 1.7582601421999166e-05,
"loss": 0.9457,
"step": 290
},
{
"epoch": 0.37641154328732745,
"grad_norm": 3.710908889770508,
"learning_rate": 1.749895441237976e-05,
"loss": 0.8773,
"step": 300
},
{
"epoch": 0.3889585947302384,
"grad_norm": 4.279157638549805,
"learning_rate": 1.7415307402760353e-05,
"loss": 0.8892,
"step": 310
},
{
"epoch": 0.4015056461731493,
"grad_norm": 17.142894744873047,
"learning_rate": 1.7331660393140947e-05,
"loss": 0.8723,
"step": 320
},
{
"epoch": 0.41405269761606023,
"grad_norm": 3.349973440170288,
"learning_rate": 1.724801338352154e-05,
"loss": 0.8092,
"step": 330
},
{
"epoch": 0.42659974905897113,
"grad_norm": 7.343029022216797,
"learning_rate": 1.7164366373902134e-05,
"loss": 0.9051,
"step": 340
},
{
"epoch": 0.43914680050188204,
"grad_norm": 3.02512264251709,
"learning_rate": 1.7080719364282727e-05,
"loss": 0.8959,
"step": 350
},
{
"epoch": 0.451693851944793,
"grad_norm": 13.072802543640137,
"learning_rate": 1.699707235466332e-05,
"loss": 0.9339,
"step": 360
},
{
"epoch": 0.4642409033877039,
"grad_norm": 3.5314481258392334,
"learning_rate": 1.6913425345043917e-05,
"loss": 0.8883,
"step": 370
},
{
"epoch": 0.4767879548306148,
"grad_norm": 3.5592713356018066,
"learning_rate": 1.6829778335424508e-05,
"loss": 0.9539,
"step": 380
},
{
"epoch": 0.4893350062735257,
"grad_norm": 2.2764039039611816,
"learning_rate": 1.6746131325805104e-05,
"loss": 0.9251,
"step": 390
},
{
"epoch": 0.5018820577164367,
"grad_norm": 2.953474998474121,
"learning_rate": 1.6662484316185698e-05,
"loss": 0.9561,
"step": 400
},
{
"epoch": 0.5144291091593476,
"grad_norm": 2.250779628753662,
"learning_rate": 1.657883730656629e-05,
"loss": 0.9405,
"step": 410
},
{
"epoch": 0.5269761606022585,
"grad_norm": 5.86421537399292,
"learning_rate": 1.6495190296946885e-05,
"loss": 0.8936,
"step": 420
},
{
"epoch": 0.5395232120451694,
"grad_norm": 4.82138729095459,
"learning_rate": 1.6411543287327482e-05,
"loss": 0.8553,
"step": 430
},
{
"epoch": 0.5520702634880803,
"grad_norm": 3.090364456176758,
"learning_rate": 1.6327896277708072e-05,
"loss": 0.9059,
"step": 440
},
{
"epoch": 0.5646173149309912,
"grad_norm": 5.363993167877197,
"learning_rate": 1.624424926808867e-05,
"loss": 0.9631,
"step": 450
},
{
"epoch": 0.5771643663739021,
"grad_norm": 3.4371540546417236,
"learning_rate": 1.616060225846926e-05,
"loss": 0.794,
"step": 460
},
{
"epoch": 0.589711417816813,
"grad_norm": 4.080657482147217,
"learning_rate": 1.6076955248849856e-05,
"loss": 0.9348,
"step": 470
},
{
"epoch": 0.6022584692597239,
"grad_norm": 5.351539134979248,
"learning_rate": 1.599330823923045e-05,
"loss": 0.9286,
"step": 480
},
{
"epoch": 0.6148055207026348,
"grad_norm": 3.6495888233184814,
"learning_rate": 1.5909661229611043e-05,
"loss": 0.7525,
"step": 490
},
{
"epoch": 0.6273525721455459,
"grad_norm": 4.849205017089844,
"learning_rate": 1.5826014219991636e-05,
"loss": 0.9534,
"step": 500
},
{
"epoch": 0.6398996235884568,
"grad_norm": 4.923251152038574,
"learning_rate": 1.574236721037223e-05,
"loss": 0.9681,
"step": 510
},
{
"epoch": 0.6524466750313677,
"grad_norm": 4.8032073974609375,
"learning_rate": 1.5658720200752824e-05,
"loss": 0.8674,
"step": 520
},
{
"epoch": 0.6649937264742786,
"grad_norm": 6.583972454071045,
"learning_rate": 1.5575073191133417e-05,
"loss": 0.8463,
"step": 530
},
{
"epoch": 0.6775407779171895,
"grad_norm": 6.276474475860596,
"learning_rate": 1.5491426181514014e-05,
"loss": 0.8481,
"step": 540
},
{
"epoch": 0.6900878293601004,
"grad_norm": 6.980823040008545,
"learning_rate": 1.5407779171894604e-05,
"loss": 0.8441,
"step": 550
},
{
"epoch": 0.7026348808030113,
"grad_norm": 5.919472694396973,
"learning_rate": 1.53241321622752e-05,
"loss": 0.7705,
"step": 560
},
{
"epoch": 0.7151819322459222,
"grad_norm": 14.738715171813965,
"learning_rate": 1.5240485152655793e-05,
"loss": 0.8883,
"step": 570
},
{
"epoch": 0.7277289836888331,
"grad_norm": 9.944400787353516,
"learning_rate": 1.5156838143036388e-05,
"loss": 0.843,
"step": 580
},
{
"epoch": 0.740276035131744,
"grad_norm": 5.619649410247803,
"learning_rate": 1.5073191133416981e-05,
"loss": 0.9996,
"step": 590
},
{
"epoch": 0.7528230865746549,
"grad_norm": 7.122910499572754,
"learning_rate": 1.4989544123797575e-05,
"loss": 0.9,
"step": 600
},
{
"epoch": 0.7653701380175659,
"grad_norm": 10.766443252563477,
"learning_rate": 1.490589711417817e-05,
"loss": 0.8326,
"step": 610
},
{
"epoch": 0.7779171894604768,
"grad_norm": 4.103264808654785,
"learning_rate": 1.4822250104558764e-05,
"loss": 0.8552,
"step": 620
},
{
"epoch": 0.7904642409033877,
"grad_norm": 11.64675521850586,
"learning_rate": 1.4738603094939357e-05,
"loss": 0.774,
"step": 630
},
{
"epoch": 0.8030112923462986,
"grad_norm": 6.035724639892578,
"learning_rate": 1.465495608531995e-05,
"loss": 0.9782,
"step": 640
},
{
"epoch": 0.8155583437892095,
"grad_norm": 6.475778102874756,
"learning_rate": 1.4571309075700546e-05,
"loss": 0.7997,
"step": 650
},
{
"epoch": 0.8281053952321205,
"grad_norm": 5.5077314376831055,
"learning_rate": 1.4487662066081138e-05,
"loss": 0.7985,
"step": 660
},
{
"epoch": 0.8406524466750314,
"grad_norm": 9.788359642028809,
"learning_rate": 1.4404015056461733e-05,
"loss": 0.9237,
"step": 670
},
{
"epoch": 0.8531994981179423,
"grad_norm": 5.642941951751709,
"learning_rate": 1.4320368046842325e-05,
"loss": 0.7795,
"step": 680
},
{
"epoch": 0.8657465495608532,
"grad_norm": 7.9442315101623535,
"learning_rate": 1.423672103722292e-05,
"loss": 0.891,
"step": 690
},
{
"epoch": 0.8782936010037641,
"grad_norm": 4.023198127746582,
"learning_rate": 1.4153074027603515e-05,
"loss": 0.8391,
"step": 700
},
{
"epoch": 0.890840652446675,
"grad_norm": 11.08212661743164,
"learning_rate": 1.4069427017984107e-05,
"loss": 0.9049,
"step": 710
},
{
"epoch": 0.903387703889586,
"grad_norm": 7.001065254211426,
"learning_rate": 1.3985780008364702e-05,
"loss": 0.9636,
"step": 720
},
{
"epoch": 0.9159347553324969,
"grad_norm": 7.488030433654785,
"learning_rate": 1.3902132998745297e-05,
"loss": 0.8491,
"step": 730
},
{
"epoch": 0.9284818067754078,
"grad_norm": 5.319072246551514,
"learning_rate": 1.381848598912589e-05,
"loss": 0.921,
"step": 740
},
{
"epoch": 0.9410288582183187,
"grad_norm": 7.3197126388549805,
"learning_rate": 1.3734838979506484e-05,
"loss": 0.9866,
"step": 750
},
{
"epoch": 0.9535759096612296,
"grad_norm": 5.280470848083496,
"learning_rate": 1.3651191969887078e-05,
"loss": 0.8146,
"step": 760
},
{
"epoch": 0.9661229611041405,
"grad_norm": 3.304518938064575,
"learning_rate": 1.3567544960267671e-05,
"loss": 0.8724,
"step": 770
},
{
"epoch": 0.9786700125470514,
"grad_norm": 3.306471109390259,
"learning_rate": 1.3483897950648267e-05,
"loss": 0.8796,
"step": 780
},
{
"epoch": 0.9912170639899623,
"grad_norm": 5.108331203460693,
"learning_rate": 1.3400250941028858e-05,
"loss": 0.8931,
"step": 790
},
{
"epoch": 1.0037641154328734,
"grad_norm": 9.46453857421875,
"learning_rate": 1.3316603931409454e-05,
"loss": 0.8834,
"step": 800
},
{
"epoch": 1.0163111668757843,
"grad_norm": 8.54848861694336,
"learning_rate": 1.3232956921790047e-05,
"loss": 0.7073,
"step": 810
},
{
"epoch": 1.0288582183186952,
"grad_norm": 8.689090728759766,
"learning_rate": 1.314930991217064e-05,
"loss": 0.8242,
"step": 820
},
{
"epoch": 1.041405269761606,
"grad_norm": 8.000664710998535,
"learning_rate": 1.3065662902551234e-05,
"loss": 0.8314,
"step": 830
},
{
"epoch": 1.053952321204517,
"grad_norm": 11.315167427062988,
"learning_rate": 1.298201589293183e-05,
"loss": 0.6794,
"step": 840
},
{
"epoch": 1.066499372647428,
"grad_norm": 6.037744522094727,
"learning_rate": 1.2898368883312423e-05,
"loss": 0.7796,
"step": 850
},
{
"epoch": 1.0790464240903388,
"grad_norm": 9.521272659301758,
"learning_rate": 1.2814721873693016e-05,
"loss": 0.92,
"step": 860
},
{
"epoch": 1.0915934755332497,
"grad_norm": 18.728450775146484,
"learning_rate": 1.2731074864073612e-05,
"loss": 0.8371,
"step": 870
},
{
"epoch": 1.1041405269761606,
"grad_norm": 7.143416404724121,
"learning_rate": 1.2647427854454203e-05,
"loss": 0.8534,
"step": 880
},
{
"epoch": 1.1166875784190715,
"grad_norm": 7.581761837005615,
"learning_rate": 1.2563780844834799e-05,
"loss": 0.894,
"step": 890
},
{
"epoch": 1.1292346298619824,
"grad_norm": 3.4014008045196533,
"learning_rate": 1.248013383521539e-05,
"loss": 0.8155,
"step": 900
},
{
"epoch": 1.1417816813048933,
"grad_norm": 13.978635787963867,
"learning_rate": 1.2396486825595986e-05,
"loss": 0.754,
"step": 910
},
{
"epoch": 1.1543287327478042,
"grad_norm": 5.18642520904541,
"learning_rate": 1.2312839815976581e-05,
"loss": 0.8493,
"step": 920
},
{
"epoch": 1.1668757841907151,
"grad_norm": 4.153924942016602,
"learning_rate": 1.2229192806357173e-05,
"loss": 0.813,
"step": 930
},
{
"epoch": 1.179422835633626,
"grad_norm": 9.386292457580566,
"learning_rate": 1.2145545796737768e-05,
"loss": 0.7526,
"step": 940
},
{
"epoch": 1.191969887076537,
"grad_norm": 5.997525215148926,
"learning_rate": 1.2061898787118363e-05,
"loss": 0.8602,
"step": 950
},
{
"epoch": 1.2045169385194479,
"grad_norm": 7.475282669067383,
"learning_rate": 1.1978251777498955e-05,
"loss": 0.6822,
"step": 960
},
{
"epoch": 1.2170639899623588,
"grad_norm": 11.12500286102295,
"learning_rate": 1.189460476787955e-05,
"loss": 0.7505,
"step": 970
},
{
"epoch": 1.2296110414052697,
"grad_norm": 10.952397346496582,
"learning_rate": 1.1810957758260144e-05,
"loss": 0.8621,
"step": 980
},
{
"epoch": 1.2421580928481806,
"grad_norm": 16.098909378051758,
"learning_rate": 1.1727310748640737e-05,
"loss": 0.8263,
"step": 990
},
{
"epoch": 1.2547051442910915,
"grad_norm": 7.630945682525635,
"learning_rate": 1.164366373902133e-05,
"loss": 0.7283,
"step": 1000
},
{
"epoch": 1.2672521957340024,
"grad_norm": 5.213379859924316,
"learning_rate": 1.1560016729401924e-05,
"loss": 0.7849,
"step": 1010
},
{
"epoch": 1.2797992471769133,
"grad_norm": 9.576950073242188,
"learning_rate": 1.147636971978252e-05,
"loss": 0.7845,
"step": 1020
},
{
"epoch": 1.2923462986198244,
"grad_norm": 7.526945114135742,
"learning_rate": 1.1392722710163113e-05,
"loss": 0.8226,
"step": 1030
},
{
"epoch": 1.3048933500627353,
"grad_norm": 8.773786544799805,
"learning_rate": 1.1309075700543706e-05,
"loss": 0.7912,
"step": 1040
},
{
"epoch": 1.3174404015056462,
"grad_norm": 5.114284038543701,
"learning_rate": 1.12254286909243e-05,
"loss": 0.7414,
"step": 1050
},
{
"epoch": 1.3299874529485571,
"grad_norm": 6.488368988037109,
"learning_rate": 1.1141781681304895e-05,
"loss": 0.8588,
"step": 1060
},
{
"epoch": 1.342534504391468,
"grad_norm": 6.658433437347412,
"learning_rate": 1.1058134671685487e-05,
"loss": 0.9214,
"step": 1070
},
{
"epoch": 1.355081555834379,
"grad_norm": 10.264144897460938,
"learning_rate": 1.0974487662066082e-05,
"loss": 0.7492,
"step": 1080
},
{
"epoch": 1.3676286072772899,
"grad_norm": 7.86182165145874,
"learning_rate": 1.0890840652446677e-05,
"loss": 0.7919,
"step": 1090
},
{
"epoch": 1.3801756587202008,
"grad_norm": 7.3965535163879395,
"learning_rate": 1.080719364282727e-05,
"loss": 0.8695,
"step": 1100
},
{
"epoch": 1.3927227101631117,
"grad_norm": 9.31979751586914,
"learning_rate": 1.0723546633207864e-05,
"loss": 0.7495,
"step": 1110
},
{
"epoch": 1.4052697616060226,
"grad_norm": 9.471373558044434,
"learning_rate": 1.0639899623588456e-05,
"loss": 0.8729,
"step": 1120
},
{
"epoch": 1.4178168130489335,
"grad_norm": 13.156733512878418,
"learning_rate": 1.0556252613969051e-05,
"loss": 0.8688,
"step": 1130
},
{
"epoch": 1.4303638644918444,
"grad_norm": 13.249898910522461,
"learning_rate": 1.0472605604349647e-05,
"loss": 0.8097,
"step": 1140
},
{
"epoch": 1.4429109159347553,
"grad_norm": 6.98345947265625,
"learning_rate": 1.0388958594730238e-05,
"loss": 0.7635,
"step": 1150
},
{
"epoch": 1.4554579673776662,
"grad_norm": 18.78076171875,
"learning_rate": 1.0305311585110834e-05,
"loss": 0.957,
"step": 1160
},
{
"epoch": 1.468005018820577,
"grad_norm": 25.165569305419922,
"learning_rate": 1.0221664575491427e-05,
"loss": 0.7616,
"step": 1170
},
{
"epoch": 1.480552070263488,
"grad_norm": 8.38501262664795,
"learning_rate": 1.013801756587202e-05,
"loss": 0.7768,
"step": 1180
},
{
"epoch": 1.4930991217063991,
"grad_norm": 10.194948196411133,
"learning_rate": 1.0054370556252616e-05,
"loss": 0.8924,
"step": 1190
},
{
"epoch": 1.50564617314931,
"grad_norm": 10.704676628112793,
"learning_rate": 9.97072354663321e-06,
"loss": 0.9066,
"step": 1200
},
{
"epoch": 1.518193224592221,
"grad_norm": 21.507707595825195,
"learning_rate": 9.887076537013803e-06,
"loss": 0.9013,
"step": 1210
},
{
"epoch": 1.5307402760351319,
"grad_norm": 8.45915699005127,
"learning_rate": 9.803429527394396e-06,
"loss": 1.0474,
"step": 1220
},
{
"epoch": 1.5432873274780428,
"grad_norm": 5.320367336273193,
"learning_rate": 9.71978251777499e-06,
"loss": 0.8313,
"step": 1230
},
{
"epoch": 1.5558343789209537,
"grad_norm": 10.026762962341309,
"learning_rate": 9.636135508155583e-06,
"loss": 0.8881,
"step": 1240
},
{
"epoch": 1.5683814303638646,
"grad_norm": 9.318647384643555,
"learning_rate": 9.552488498536177e-06,
"loss": 0.8274,
"step": 1250
},
{
"epoch": 1.5809284818067755,
"grad_norm": 9.438924789428711,
"learning_rate": 9.468841488916772e-06,
"loss": 0.9072,
"step": 1260
},
{
"epoch": 1.5934755332496864,
"grad_norm": 11.8627290725708,
"learning_rate": 9.385194479297366e-06,
"loss": 0.746,
"step": 1270
},
{
"epoch": 1.6060225846925973,
"grad_norm": 5.647990703582764,
"learning_rate": 9.30154746967796e-06,
"loss": 0.8526,
"step": 1280
},
{
"epoch": 1.6185696361355082,
"grad_norm": 17.41633415222168,
"learning_rate": 9.217900460058554e-06,
"loss": 0.7216,
"step": 1290
},
{
"epoch": 1.631116687578419,
"grad_norm": 10.73770523071289,
"learning_rate": 9.134253450439148e-06,
"loss": 0.7405,
"step": 1300
},
{
"epoch": 1.64366373902133,
"grad_norm": 9.316720962524414,
"learning_rate": 9.050606440819741e-06,
"loss": 0.727,
"step": 1310
},
{
"epoch": 1.656210790464241,
"grad_norm": 8.067049980163574,
"learning_rate": 8.966959431200335e-06,
"loss": 0.7992,
"step": 1320
},
{
"epoch": 1.6687578419071518,
"grad_norm": 23.564279556274414,
"learning_rate": 8.88331242158093e-06,
"loss": 0.7617,
"step": 1330
},
{
"epoch": 1.6813048933500627,
"grad_norm": 10.805792808532715,
"learning_rate": 8.799665411961524e-06,
"loss": 0.9413,
"step": 1340
},
{
"epoch": 1.6938519447929736,
"grad_norm": 8.173852920532227,
"learning_rate": 8.716018402342117e-06,
"loss": 0.8204,
"step": 1350
},
{
"epoch": 1.7063989962358845,
"grad_norm": 16.444934844970703,
"learning_rate": 8.63237139272271e-06,
"loss": 0.7124,
"step": 1360
},
{
"epoch": 1.7189460476787954,
"grad_norm": 19.09043312072754,
"learning_rate": 8.548724383103306e-06,
"loss": 0.8573,
"step": 1370
},
{
"epoch": 1.7314930991217063,
"grad_norm": 15.454293251037598,
"learning_rate": 8.4650773734839e-06,
"loss": 0.7964,
"step": 1380
},
{
"epoch": 1.7440401505646173,
"grad_norm": 9.219696044921875,
"learning_rate": 8.381430363864493e-06,
"loss": 0.7494,
"step": 1390
},
{
"epoch": 1.7565872020075282,
"grad_norm": 12.434812545776367,
"learning_rate": 8.297783354245086e-06,
"loss": 0.7066,
"step": 1400
},
{
"epoch": 1.769134253450439,
"grad_norm": 8.64736270904541,
"learning_rate": 8.21413634462568e-06,
"loss": 0.8049,
"step": 1410
},
{
"epoch": 1.78168130489335,
"grad_norm": 8.343708038330078,
"learning_rate": 8.130489335006273e-06,
"loss": 0.7935,
"step": 1420
},
{
"epoch": 1.7942283563362609,
"grad_norm": 6.48344612121582,
"learning_rate": 8.046842325386867e-06,
"loss": 0.7431,
"step": 1430
},
{
"epoch": 1.8067754077791718,
"grad_norm": 7.362094402313232,
"learning_rate": 7.963195315767462e-06,
"loss": 0.6778,
"step": 1440
},
{
"epoch": 1.8193224592220827,
"grad_norm": 10.41412353515625,
"learning_rate": 7.879548306148056e-06,
"loss": 0.8869,
"step": 1450
},
{
"epoch": 1.8318695106649936,
"grad_norm": 11.075960159301758,
"learning_rate": 7.79590129652865e-06,
"loss": 0.8913,
"step": 1460
},
{
"epoch": 1.8444165621079045,
"grad_norm": 3.257545232772827,
"learning_rate": 7.712254286909243e-06,
"loss": 0.8214,
"step": 1470
},
{
"epoch": 1.8569636135508154,
"grad_norm": 2.363823175430298,
"learning_rate": 7.628607277289838e-06,
"loss": 0.8268,
"step": 1480
},
{
"epoch": 1.8695106649937263,
"grad_norm": 16.772573471069336,
"learning_rate": 7.544960267670431e-06,
"loss": 0.9585,
"step": 1490
},
{
"epoch": 1.8820577164366374,
"grad_norm": 23.401941299438477,
"learning_rate": 7.461313258051025e-06,
"loss": 0.8636,
"step": 1500
},
{
"epoch": 1.8946047678795483,
"grad_norm": 8.009302139282227,
"learning_rate": 7.37766624843162e-06,
"loss": 0.6728,
"step": 1510
},
{
"epoch": 1.9071518193224593,
"grad_norm": 19.989112854003906,
"learning_rate": 7.294019238812214e-06,
"loss": 0.8526,
"step": 1520
},
{
"epoch": 1.9196988707653702,
"grad_norm": 11.849247932434082,
"learning_rate": 7.210372229192807e-06,
"loss": 0.8825,
"step": 1530
},
{
"epoch": 1.932245922208281,
"grad_norm": 8.389686584472656,
"learning_rate": 7.126725219573401e-06,
"loss": 0.8261,
"step": 1540
},
{
"epoch": 1.944792973651192,
"grad_norm": 11.961833000183105,
"learning_rate": 7.043078209953995e-06,
"loss": 0.6564,
"step": 1550
},
{
"epoch": 1.9573400250941029,
"grad_norm": 15.514620780944824,
"learning_rate": 6.9594312003345885e-06,
"loss": 0.7943,
"step": 1560
},
{
"epoch": 1.9698870765370138,
"grad_norm": 10.584662437438965,
"learning_rate": 6.875784190715182e-06,
"loss": 0.7593,
"step": 1570
},
{
"epoch": 1.9824341279799247,
"grad_norm": 15.600502014160156,
"learning_rate": 6.7921371810957755e-06,
"loss": 0.8643,
"step": 1580
},
{
"epoch": 1.9949811794228356,
"grad_norm": 11.883516311645508,
"learning_rate": 6.708490171476371e-06,
"loss": 0.8381,
"step": 1590
},
{
"epoch": 2.0075282308657467,
"grad_norm": 6.613313674926758,
"learning_rate": 6.624843161856964e-06,
"loss": 0.8074,
"step": 1600
},
{
"epoch": 2.0200752823086576,
"grad_norm": 10.316112518310547,
"learning_rate": 6.541196152237558e-06,
"loss": 0.7161,
"step": 1610
},
{
"epoch": 2.0326223337515685,
"grad_norm": 34.29435348510742,
"learning_rate": 6.457549142618152e-06,
"loss": 0.7553,
"step": 1620
},
{
"epoch": 2.0451693851944794,
"grad_norm": 13.848133087158203,
"learning_rate": 6.373902132998746e-06,
"loss": 0.7486,
"step": 1630
},
{
"epoch": 2.0577164366373903,
"grad_norm": 6.787230014801025,
"learning_rate": 6.29025512337934e-06,
"loss": 0.7867,
"step": 1640
},
{
"epoch": 2.0702634880803013,
"grad_norm": 12.09325122833252,
"learning_rate": 6.2066081137599335e-06,
"loss": 0.7999,
"step": 1650
},
{
"epoch": 2.082810539523212,
"grad_norm": 9.453817367553711,
"learning_rate": 6.122961104140528e-06,
"loss": 0.7668,
"step": 1660
},
{
"epoch": 2.095357590966123,
"grad_norm": 10.155051231384277,
"learning_rate": 6.039314094521121e-06,
"loss": 0.7595,
"step": 1670
},
{
"epoch": 2.107904642409034,
"grad_norm": 26.19032859802246,
"learning_rate": 5.955667084901715e-06,
"loss": 0.9065,
"step": 1680
},
{
"epoch": 2.120451693851945,
"grad_norm": 9.538314819335938,
"learning_rate": 5.872020075282308e-06,
"loss": 0.7759,
"step": 1690
},
{
"epoch": 2.132998745294856,
"grad_norm": 6.306014060974121,
"learning_rate": 5.788373065662904e-06,
"loss": 0.7431,
"step": 1700
},
{
"epoch": 2.1455457967377667,
"grad_norm": 8.156487464904785,
"learning_rate": 5.704726056043497e-06,
"loss": 0.7029,
"step": 1710
},
{
"epoch": 2.1580928481806776,
"grad_norm": 14.053692817687988,
"learning_rate": 5.621079046424091e-06,
"loss": 0.8357,
"step": 1720
},
{
"epoch": 2.1706398996235885,
"grad_norm": 10.932027816772461,
"learning_rate": 5.537432036804685e-06,
"loss": 0.723,
"step": 1730
},
{
"epoch": 2.1831869510664994,
"grad_norm": 23.12853240966797,
"learning_rate": 5.4537850271852785e-06,
"loss": 0.819,
"step": 1740
},
{
"epoch": 2.1957340025094103,
"grad_norm": 10.468832015991211,
"learning_rate": 5.370138017565872e-06,
"loss": 0.7771,
"step": 1750
},
{
"epoch": 2.208281053952321,
"grad_norm": 16.84822654724121,
"learning_rate": 5.286491007946466e-06,
"loss": 0.7895,
"step": 1760
},
{
"epoch": 2.220828105395232,
"grad_norm": 17.586076736450195,
"learning_rate": 5.202843998327061e-06,
"loss": 0.8417,
"step": 1770
},
{
"epoch": 2.233375156838143,
"grad_norm": 5.2970662117004395,
"learning_rate": 5.119196988707654e-06,
"loss": 0.7451,
"step": 1780
},
{
"epoch": 2.245922208281054,
"grad_norm": 11.92138385772705,
"learning_rate": 5.035549979088248e-06,
"loss": 0.8368,
"step": 1790
},
{
"epoch": 2.258469259723965,
"grad_norm": 22.60080909729004,
"learning_rate": 4.951902969468842e-06,
"loss": 0.7133,
"step": 1800
},
{
"epoch": 2.2710163111668757,
"grad_norm": 12.277490615844727,
"learning_rate": 4.868255959849436e-06,
"loss": 0.8418,
"step": 1810
},
{
"epoch": 2.2835633626097867,
"grad_norm": 6.862631320953369,
"learning_rate": 4.78460895023003e-06,
"loss": 0.7564,
"step": 1820
},
{
"epoch": 2.2961104140526976,
"grad_norm": 7.9091267585754395,
"learning_rate": 4.7009619406106235e-06,
"loss": 0.7582,
"step": 1830
},
{
"epoch": 2.3086574654956085,
"grad_norm": 8.295835494995117,
"learning_rate": 4.617314930991217e-06,
"loss": 0.8465,
"step": 1840
},
{
"epoch": 2.3212045169385194,
"grad_norm": 11.34477710723877,
"learning_rate": 4.533667921371811e-06,
"loss": 0.7846,
"step": 1850
},
{
"epoch": 2.3337515683814303,
"grad_norm": 32.95912170410156,
"learning_rate": 4.450020911752405e-06,
"loss": 0.6656,
"step": 1860
},
{
"epoch": 2.346298619824341,
"grad_norm": 17.355579376220703,
"learning_rate": 4.366373902132999e-06,
"loss": 0.8187,
"step": 1870
},
{
"epoch": 2.358845671267252,
"grad_norm": 8.741580963134766,
"learning_rate": 4.282726892513593e-06,
"loss": 0.744,
"step": 1880
},
{
"epoch": 2.371392722710163,
"grad_norm": 12.800196647644043,
"learning_rate": 4.199079882894187e-06,
"loss": 0.6891,
"step": 1890
},
{
"epoch": 2.383939774153074,
"grad_norm": 6.387624740600586,
"learning_rate": 4.115432873274781e-06,
"loss": 0.6968,
"step": 1900
},
{
"epoch": 2.396486825595985,
"grad_norm": 9.609091758728027,
"learning_rate": 4.031785863655375e-06,
"loss": 0.7412,
"step": 1910
},
{
"epoch": 2.4090338770388957,
"grad_norm": 16.369497299194336,
"learning_rate": 3.9481388540359685e-06,
"loss": 0.7136,
"step": 1920
},
{
"epoch": 2.4215809284818066,
"grad_norm": 7.245341777801514,
"learning_rate": 3.864491844416563e-06,
"loss": 0.6953,
"step": 1930
},
{
"epoch": 2.4341279799247175,
"grad_norm": 21.487659454345703,
"learning_rate": 3.7808448347971568e-06,
"loss": 0.776,
"step": 1940
},
{
"epoch": 2.4466750313676284,
"grad_norm": 13.91763973236084,
"learning_rate": 3.6971978251777503e-06,
"loss": 0.9644,
"step": 1950
},
{
"epoch": 2.4592220828105393,
"grad_norm": 16.034120559692383,
"learning_rate": 3.6135508155583442e-06,
"loss": 0.7451,
"step": 1960
},
{
"epoch": 2.4717691342534502,
"grad_norm": 15.478806495666504,
"learning_rate": 3.5299038059389377e-06,
"loss": 0.6669,
"step": 1970
},
{
"epoch": 2.484316185696361,
"grad_norm": 14.39084529876709,
"learning_rate": 3.446256796319532e-06,
"loss": 0.7545,
"step": 1980
},
{
"epoch": 2.496863237139272,
"grad_norm": 27.011598587036133,
"learning_rate": 3.3626097867001256e-06,
"loss": 0.8599,
"step": 1990
},
{
"epoch": 2.509410288582183,
"grad_norm": 15.559931755065918,
"learning_rate": 3.27896277708072e-06,
"loss": 0.8266,
"step": 2000
},
{
"epoch": 2.521957340025094,
"grad_norm": 8.093729019165039,
"learning_rate": 3.1953157674613135e-06,
"loss": 0.7978,
"step": 2010
},
{
"epoch": 2.5345043914680048,
"grad_norm": 13.65913200378418,
"learning_rate": 3.1116687578419074e-06,
"loss": 0.8145,
"step": 2020
},
{
"epoch": 2.5470514429109157,
"grad_norm": 10.85305404663086,
"learning_rate": 3.028021748222501e-06,
"loss": 0.7976,
"step": 2030
},
{
"epoch": 2.5595984943538266,
"grad_norm": 21.572189331054688,
"learning_rate": 2.9443747386030953e-06,
"loss": 0.7435,
"step": 2040
},
{
"epoch": 2.572145545796738,
"grad_norm": 14.962433815002441,
"learning_rate": 2.860727728983689e-06,
"loss": 0.7856,
"step": 2050
},
{
"epoch": 2.584692597239649,
"grad_norm": 8.81922435760498,
"learning_rate": 2.7770807193642827e-06,
"loss": 0.6729,
"step": 2060
},
{
"epoch": 2.5972396486825597,
"grad_norm": 36.92323684692383,
"learning_rate": 2.693433709744877e-06,
"loss": 0.7374,
"step": 2070
},
{
"epoch": 2.6097867001254706,
"grad_norm": 19.693674087524414,
"learning_rate": 2.6097867001254706e-06,
"loss": 0.6577,
"step": 2080
},
{
"epoch": 2.6223337515683816,
"grad_norm": 6.363992691040039,
"learning_rate": 2.526139690506065e-06,
"loss": 0.7902,
"step": 2090
},
{
"epoch": 2.6348808030112925,
"grad_norm": 18.565412521362305,
"learning_rate": 2.4424926808866585e-06,
"loss": 0.638,
"step": 2100
},
{
"epoch": 2.6474278544542034,
"grad_norm": 23.991640090942383,
"learning_rate": 2.3588456712672524e-06,
"loss": 0.8057,
"step": 2110
},
{
"epoch": 2.6599749058971143,
"grad_norm": 11.251315116882324,
"learning_rate": 2.2751986616478463e-06,
"loss": 0.7766,
"step": 2120
},
{
"epoch": 2.672521957340025,
"grad_norm": 13.534173011779785,
"learning_rate": 2.1915516520284403e-06,
"loss": 0.7792,
"step": 2130
},
{
"epoch": 2.685069008782936,
"grad_norm": 28.147571563720703,
"learning_rate": 2.107904642409034e-06,
"loss": 0.7725,
"step": 2140
},
{
"epoch": 2.697616060225847,
"grad_norm": 25.395660400390625,
"learning_rate": 2.024257632789628e-06,
"loss": 0.7114,
"step": 2150
},
{
"epoch": 2.710163111668758,
"grad_norm": 11.658987998962402,
"learning_rate": 1.9406106231702216e-06,
"loss": 0.7515,
"step": 2160
},
{
"epoch": 2.722710163111669,
"grad_norm": 7.770451068878174,
"learning_rate": 1.8569636135508156e-06,
"loss": 0.7042,
"step": 2170
},
{
"epoch": 2.7352572145545797,
"grad_norm": 15.549407005310059,
"learning_rate": 1.7733166039314095e-06,
"loss": 0.8029,
"step": 2180
},
{
"epoch": 2.7478042659974906,
"grad_norm": 9.49413013458252,
"learning_rate": 1.6896695943120034e-06,
"loss": 0.642,
"step": 2190
},
{
"epoch": 2.7603513174404015,
"grad_norm": 4.831557750701904,
"learning_rate": 1.6060225846925972e-06,
"loss": 0.6564,
"step": 2200
},
{
"epoch": 2.7728983688833124,
"grad_norm": 12.240572929382324,
"learning_rate": 1.5223755750731913e-06,
"loss": 0.7423,
"step": 2210
},
{
"epoch": 2.7854454203262233,
"grad_norm": 5.629863739013672,
"learning_rate": 1.4387285654537853e-06,
"loss": 0.8664,
"step": 2220
},
{
"epoch": 2.7979924717691342,
"grad_norm": 14.666424751281738,
"learning_rate": 1.3550815558343792e-06,
"loss": 0.8401,
"step": 2230
},
{
"epoch": 2.810539523212045,
"grad_norm": 21.368179321289062,
"learning_rate": 1.271434546214973e-06,
"loss": 0.8129,
"step": 2240
},
{
"epoch": 2.823086574654956,
"grad_norm": 16.279918670654297,
"learning_rate": 1.1877875365955668e-06,
"loss": 0.676,
"step": 2250
},
{
"epoch": 2.835633626097867,
"grad_norm": 9.378284454345703,
"learning_rate": 1.1041405269761608e-06,
"loss": 0.6352,
"step": 2260
},
{
"epoch": 2.848180677540778,
"grad_norm": 21.204431533813477,
"learning_rate": 1.0204935173567545e-06,
"loss": 0.7627,
"step": 2270
},
{
"epoch": 2.8607277289836888,
"grad_norm": 23.616268157958984,
"learning_rate": 9.368465077373484e-07,
"loss": 0.6356,
"step": 2280
},
{
"epoch": 2.8732747804265997,
"grad_norm": 25.798891067504883,
"learning_rate": 8.531994981179424e-07,
"loss": 0.7385,
"step": 2290
},
{
"epoch": 2.8858218318695106,
"grad_norm": 11.757165908813477,
"learning_rate": 7.695524884985363e-07,
"loss": 0.6858,
"step": 2300
},
{
"epoch": 2.8983688833124215,
"grad_norm": 9.811064720153809,
"learning_rate": 6.859054788791301e-07,
"loss": 0.6791,
"step": 2310
},
{
"epoch": 2.9109159347553324,
"grad_norm": 14.021916389465332,
"learning_rate": 6.02258469259724e-07,
"loss": 0.8106,
"step": 2320
},
{
"epoch": 2.9234629861982433,
"grad_norm": 16.58710479736328,
"learning_rate": 5.186114596403179e-07,
"loss": 0.8049,
"step": 2330
},
{
"epoch": 2.936010037641154,
"grad_norm": 8.84688663482666,
"learning_rate": 4.349644500209118e-07,
"loss": 0.6731,
"step": 2340
},
{
"epoch": 2.948557089084065,
"grad_norm": 9.54117488861084,
"learning_rate": 3.5131744040150566e-07,
"loss": 0.7293,
"step": 2350
},
{
"epoch": 2.961104140526976,
"grad_norm": 17.779142379760742,
"learning_rate": 2.6767043078209955e-07,
"loss": 0.7803,
"step": 2360
},
{
"epoch": 2.973651191969887,
"grad_norm": 7.330904006958008,
"learning_rate": 1.8402342116269345e-07,
"loss": 0.7281,
"step": 2370
},
{
"epoch": 2.9861982434127983,
"grad_norm": 31.703744888305664,
"learning_rate": 1.0037641154328732e-07,
"loss": 0.8839,
"step": 2380
},
{
"epoch": 2.998745294855709,
"grad_norm": 13.579190254211426,
"learning_rate": 1.672940192388122e-08,
"loss": 0.8382,
"step": 2390
}
],
"logging_steps": 10,
"max_steps": 2391,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 274122632540160.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}