diff --git "a/infer/30000/trainer_state.json" "b/infer/30000/trainer_state.json" new file mode 100644--- /dev/null +++ "b/infer/30000/trainer_state.json" @@ -0,0 +1,21514 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.1693659981679627, + "eval_steps": 500, + "global_step": 30000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003897951626420316, + "grad_norm": 64.1451416015625, + "learning_rate": 1.2992074834351045e-09, + "loss": 3.3225, + "step": 10 + }, + { + "epoch": 0.0007795903252840632, + "grad_norm": 62.536888122558594, + "learning_rate": 2.598414966870209e-09, + "loss": 3.3585, + "step": 20 + }, + { + "epoch": 0.001169385487926095, + "grad_norm": 61.82063293457031, + "learning_rate": 3.897622450305313e-09, + "loss": 3.4335, + "step": 30 + }, + { + "epoch": 0.0015591806505681264, + "grad_norm": 49.641357421875, + "learning_rate": 5.196829933740418e-09, + "loss": 3.4522, + "step": 40 + }, + { + "epoch": 0.001948975813210158, + "grad_norm": 56.13090515136719, + "learning_rate": 6.496037417175522e-09, + "loss": 3.3142, + "step": 50 + }, + { + "epoch": 0.00233877097585219, + "grad_norm": 60.67042541503906, + "learning_rate": 7.795244900610626e-09, + "loss": 3.5183, + "step": 60 + }, + { + "epoch": 0.0027285661384942213, + "grad_norm": 63.279563903808594, + "learning_rate": 9.094452384045731e-09, + "loss": 3.393, + "step": 70 + }, + { + "epoch": 0.0031183613011362527, + "grad_norm": 62.238983154296875, + "learning_rate": 1.0393659867480836e-08, + "loss": 3.4131, + "step": 80 + }, + { + "epoch": 0.0035081564637782846, + "grad_norm": 52.55228805541992, + "learning_rate": 1.1692867350915941e-08, + "loss": 3.2909, + "step": 90 + }, + { + "epoch": 0.003897951626420316, + "grad_norm": 76.7668228149414, + "learning_rate": 1.2992074834351044e-08, + "loss": 3.2567, + "step": 100 + }, + { + "epoch": 0.004287746789062348, + "grad_norm": 61.65962600708008, + "learning_rate": 1.429128231778615e-08, + "loss": 3.2486, + "step": 110 + }, + { + "epoch": 0.00467754195170438, + "grad_norm": 51.81744384765625, + "learning_rate": 1.5590489801221253e-08, + "loss": 3.2648, + "step": 120 + }, + { + "epoch": 0.005067337114346411, + "grad_norm": 62.810977935791016, + "learning_rate": 1.6889697284656358e-08, + "loss": 3.3637, + "step": 130 + }, + { + "epoch": 0.0054571322769884426, + "grad_norm": 55.76907730102539, + "learning_rate": 1.8188904768091463e-08, + "loss": 3.2763, + "step": 140 + }, + { + "epoch": 0.005846927439630474, + "grad_norm": 68.08555603027344, + "learning_rate": 1.9488112251526568e-08, + "loss": 3.3257, + "step": 150 + }, + { + "epoch": 0.006236722602272505, + "grad_norm": 59.73173141479492, + "learning_rate": 2.0787319734961672e-08, + "loss": 3.3037, + "step": 160 + }, + { + "epoch": 0.006626517764914537, + "grad_norm": 59.092342376708984, + "learning_rate": 2.2086527218396777e-08, + "loss": 3.3186, + "step": 170 + }, + { + "epoch": 0.007016312927556569, + "grad_norm": 61.11860275268555, + "learning_rate": 2.3385734701831882e-08, + "loss": 3.2191, + "step": 180 + }, + { + "epoch": 0.007406108090198601, + "grad_norm": 54.53586196899414, + "learning_rate": 2.4684942185266987e-08, + "loss": 3.2395, + "step": 190 + }, + { + "epoch": 0.007795903252840632, + "grad_norm": 56.4966926574707, + "learning_rate": 2.598414966870209e-08, + "loss": 3.1696, + "step": 200 + }, + { + "epoch": 0.008185698415482664, + "grad_norm": 49.62760543823242, + "learning_rate": 2.7283357152137194e-08, + "loss": 3.0164, + "step": 210 + }, + { + "epoch": 0.008575493578124696, + "grad_norm": 54.922706604003906, + "learning_rate": 2.85825646355723e-08, + "loss": 3.1971, + "step": 220 + }, + { + "epoch": 0.008965288740766728, + "grad_norm": 54.08602523803711, + "learning_rate": 2.9881772119007404e-08, + "loss": 3.1508, + "step": 230 + }, + { + "epoch": 0.00935508390340876, + "grad_norm": 57.31306457519531, + "learning_rate": 3.1180979602442505e-08, + "loss": 3.1948, + "step": 240 + }, + { + "epoch": 0.00974487906605079, + "grad_norm": 67.46495819091797, + "learning_rate": 3.248018708587761e-08, + "loss": 3.1149, + "step": 250 + }, + { + "epoch": 0.010134674228692821, + "grad_norm": 54.64365768432617, + "learning_rate": 3.3779394569312715e-08, + "loss": 2.89, + "step": 260 + }, + { + "epoch": 0.010524469391334853, + "grad_norm": 52.321868896484375, + "learning_rate": 3.507860205274782e-08, + "loss": 2.8456, + "step": 270 + }, + { + "epoch": 0.010914264553976885, + "grad_norm": 52.1832160949707, + "learning_rate": 3.6377809536182925e-08, + "loss": 2.8917, + "step": 280 + }, + { + "epoch": 0.011304059716618917, + "grad_norm": 52.72313690185547, + "learning_rate": 3.767701701961803e-08, + "loss": 2.8842, + "step": 290 + }, + { + "epoch": 0.011693854879260949, + "grad_norm": 51.243751525878906, + "learning_rate": 3.8976224503053135e-08, + "loss": 2.8475, + "step": 300 + }, + { + "epoch": 0.01208365004190298, + "grad_norm": 54.82512283325195, + "learning_rate": 4.027543198648824e-08, + "loss": 2.8628, + "step": 310 + }, + { + "epoch": 0.01247344520454501, + "grad_norm": 44.448665618896484, + "learning_rate": 4.1574639469923345e-08, + "loss": 2.6795, + "step": 320 + }, + { + "epoch": 0.012863240367187043, + "grad_norm": 59.023441314697266, + "learning_rate": 4.2873846953358447e-08, + "loss": 2.628, + "step": 330 + }, + { + "epoch": 0.013253035529829075, + "grad_norm": 42.821685791015625, + "learning_rate": 4.4173054436793555e-08, + "loss": 2.4313, + "step": 340 + }, + { + "epoch": 0.013642830692471106, + "grad_norm": 45.76029968261719, + "learning_rate": 4.5472261920228656e-08, + "loss": 2.2929, + "step": 350 + }, + { + "epoch": 0.014032625855113138, + "grad_norm": 44.448028564453125, + "learning_rate": 4.6771469403663765e-08, + "loss": 2.3314, + "step": 360 + }, + { + "epoch": 0.01442242101775517, + "grad_norm": 34.22445297241211, + "learning_rate": 4.8070676887098866e-08, + "loss": 2.1472, + "step": 370 + }, + { + "epoch": 0.014812216180397202, + "grad_norm": 29.940370559692383, + "learning_rate": 4.9369884370533975e-08, + "loss": 2.0653, + "step": 380 + }, + { + "epoch": 0.015202011343039232, + "grad_norm": 22.266698837280273, + "learning_rate": 5.0669091853969076e-08, + "loss": 1.8852, + "step": 390 + }, + { + "epoch": 0.015591806505681264, + "grad_norm": 28.12609100341797, + "learning_rate": 5.196829933740418e-08, + "loss": 2.1143, + "step": 400 + }, + { + "epoch": 0.015981601668323298, + "grad_norm": 23.73090171813965, + "learning_rate": 5.326750682083928e-08, + "loss": 1.9354, + "step": 410 + }, + { + "epoch": 0.016371396830965328, + "grad_norm": 20.142333984375, + "learning_rate": 5.456671430427439e-08, + "loss": 1.9727, + "step": 420 + }, + { + "epoch": 0.016761191993607358, + "grad_norm": 21.074663162231445, + "learning_rate": 5.586592178770949e-08, + "loss": 1.9947, + "step": 430 + }, + { + "epoch": 0.01715098715624939, + "grad_norm": 22.76948356628418, + "learning_rate": 5.71651292711446e-08, + "loss": 1.9143, + "step": 440 + }, + { + "epoch": 0.01754078231889142, + "grad_norm": 23.192638397216797, + "learning_rate": 5.84643367545797e-08, + "loss": 1.9317, + "step": 450 + }, + { + "epoch": 0.017930577481533455, + "grad_norm": 23.29318618774414, + "learning_rate": 5.976354423801481e-08, + "loss": 1.8379, + "step": 460 + }, + { + "epoch": 0.018320372644175485, + "grad_norm": 21.225448608398438, + "learning_rate": 6.106275172144992e-08, + "loss": 1.7689, + "step": 470 + }, + { + "epoch": 0.01871016780681752, + "grad_norm": 19.7229061126709, + "learning_rate": 6.236195920488501e-08, + "loss": 1.795, + "step": 480 + }, + { + "epoch": 0.01909996296945955, + "grad_norm": 23.65806007385254, + "learning_rate": 6.366116668832012e-08, + "loss": 1.8787, + "step": 490 + }, + { + "epoch": 0.01948975813210158, + "grad_norm": 22.190011978149414, + "learning_rate": 6.496037417175521e-08, + "loss": 1.7578, + "step": 500 + }, + { + "epoch": 0.01948975813210158, + "eval_loss": 1.766793966293335, + "eval_runtime": 82.4603, + "eval_samples_per_second": 50.291, + "eval_steps_per_second": 6.294, + "step": 500 + }, + { + "epoch": 0.019879553294743613, + "grad_norm": 21.708675384521484, + "learning_rate": 6.625958165519032e-08, + "loss": 1.6902, + "step": 510 + }, + { + "epoch": 0.020269348457385643, + "grad_norm": 18.43559455871582, + "learning_rate": 6.755878913862543e-08, + "loss": 1.7274, + "step": 520 + }, + { + "epoch": 0.020659143620027676, + "grad_norm": 21.447885513305664, + "learning_rate": 6.885799662206054e-08, + "loss": 1.7845, + "step": 530 + }, + { + "epoch": 0.021048938782669706, + "grad_norm": 22.138957977294922, + "learning_rate": 7.015720410549563e-08, + "loss": 1.6478, + "step": 540 + }, + { + "epoch": 0.02143873394531174, + "grad_norm": 19.850088119506836, + "learning_rate": 7.145641158893074e-08, + "loss": 1.683, + "step": 550 + }, + { + "epoch": 0.02182852910795377, + "grad_norm": 20.672582626342773, + "learning_rate": 7.275561907236585e-08, + "loss": 1.7562, + "step": 560 + }, + { + "epoch": 0.0222183242705958, + "grad_norm": 24.011667251586914, + "learning_rate": 7.405482655580096e-08, + "loss": 1.7358, + "step": 570 + }, + { + "epoch": 0.022608119433237834, + "grad_norm": 19.025850296020508, + "learning_rate": 7.535403403923605e-08, + "loss": 1.5996, + "step": 580 + }, + { + "epoch": 0.022997914595879864, + "grad_norm": 19.10131072998047, + "learning_rate": 7.665324152267116e-08, + "loss": 1.5468, + "step": 590 + }, + { + "epoch": 0.023387709758521898, + "grad_norm": 19.839317321777344, + "learning_rate": 7.795244900610627e-08, + "loss": 1.6533, + "step": 600 + }, + { + "epoch": 0.023777504921163928, + "grad_norm": 26.39604949951172, + "learning_rate": 7.925165648954138e-08, + "loss": 1.7154, + "step": 610 + }, + { + "epoch": 0.02416730008380596, + "grad_norm": 21.983118057250977, + "learning_rate": 8.055086397297647e-08, + "loss": 1.5979, + "step": 620 + }, + { + "epoch": 0.02455709524644799, + "grad_norm": 18.320154190063477, + "learning_rate": 8.185007145641158e-08, + "loss": 1.6694, + "step": 630 + }, + { + "epoch": 0.02494689040909002, + "grad_norm": 23.123022079467773, + "learning_rate": 8.314927893984669e-08, + "loss": 1.6903, + "step": 640 + }, + { + "epoch": 0.025336685571732055, + "grad_norm": 18.185483932495117, + "learning_rate": 8.44484864232818e-08, + "loss": 1.5708, + "step": 650 + }, + { + "epoch": 0.025726480734374085, + "grad_norm": 20.46993064880371, + "learning_rate": 8.574769390671689e-08, + "loss": 1.6113, + "step": 660 + }, + { + "epoch": 0.02611627589701612, + "grad_norm": 24.227840423583984, + "learning_rate": 8.7046901390152e-08, + "loss": 1.5922, + "step": 670 + }, + { + "epoch": 0.02650607105965815, + "grad_norm": 21.02799415588379, + "learning_rate": 8.834610887358711e-08, + "loss": 1.6674, + "step": 680 + }, + { + "epoch": 0.026895866222300183, + "grad_norm": 21.945404052734375, + "learning_rate": 8.964531635702222e-08, + "loss": 1.5487, + "step": 690 + }, + { + "epoch": 0.027285661384942213, + "grad_norm": 19.90485382080078, + "learning_rate": 9.094452384045731e-08, + "loss": 1.4443, + "step": 700 + }, + { + "epoch": 0.027675456547584243, + "grad_norm": 28.931657791137695, + "learning_rate": 9.224373132389242e-08, + "loss": 1.4925, + "step": 710 + }, + { + "epoch": 0.028065251710226276, + "grad_norm": 20.0445613861084, + "learning_rate": 9.354293880732753e-08, + "loss": 1.4638, + "step": 720 + }, + { + "epoch": 0.028455046872868307, + "grad_norm": 26.842493057250977, + "learning_rate": 9.484214629076264e-08, + "loss": 1.5618, + "step": 730 + }, + { + "epoch": 0.02884484203551034, + "grad_norm": 20.507217407226562, + "learning_rate": 9.614135377419773e-08, + "loss": 1.516, + "step": 740 + }, + { + "epoch": 0.02923463719815237, + "grad_norm": 18.522409439086914, + "learning_rate": 9.744056125763284e-08, + "loss": 1.4682, + "step": 750 + }, + { + "epoch": 0.029624432360794404, + "grad_norm": 27.31763458251953, + "learning_rate": 9.873976874106795e-08, + "loss": 1.473, + "step": 760 + }, + { + "epoch": 0.030014227523436434, + "grad_norm": 20.41707420349121, + "learning_rate": 1.0003897622450306e-07, + "loss": 1.3935, + "step": 770 + }, + { + "epoch": 0.030404022686078464, + "grad_norm": 21.07742691040039, + "learning_rate": 1.0133818370793815e-07, + "loss": 1.432, + "step": 780 + }, + { + "epoch": 0.030793817848720498, + "grad_norm": 27.64505958557129, + "learning_rate": 1.0263739119137326e-07, + "loss": 1.5643, + "step": 790 + }, + { + "epoch": 0.031183613011362528, + "grad_norm": 21.09249496459961, + "learning_rate": 1.0393659867480836e-07, + "loss": 1.3534, + "step": 800 + }, + { + "epoch": 0.03157340817400456, + "grad_norm": 26.328601837158203, + "learning_rate": 1.0523580615824346e-07, + "loss": 1.4714, + "step": 810 + }, + { + "epoch": 0.031963203336646595, + "grad_norm": 22.782474517822266, + "learning_rate": 1.0653501364167856e-07, + "loss": 1.4297, + "step": 820 + }, + { + "epoch": 0.032352998499288625, + "grad_norm": 21.920915603637695, + "learning_rate": 1.0783422112511367e-07, + "loss": 1.4412, + "step": 830 + }, + { + "epoch": 0.032742793661930655, + "grad_norm": 19.492868423461914, + "learning_rate": 1.0913342860854878e-07, + "loss": 1.3788, + "step": 840 + }, + { + "epoch": 0.033132588824572685, + "grad_norm": 22.896665573120117, + "learning_rate": 1.1043263609198388e-07, + "loss": 1.3062, + "step": 850 + }, + { + "epoch": 0.033522383987214716, + "grad_norm": 23.19571304321289, + "learning_rate": 1.1173184357541898e-07, + "loss": 1.4901, + "step": 860 + }, + { + "epoch": 0.03391217914985675, + "grad_norm": 24.01108169555664, + "learning_rate": 1.1303105105885409e-07, + "loss": 1.3784, + "step": 870 + }, + { + "epoch": 0.03430197431249878, + "grad_norm": 20.344667434692383, + "learning_rate": 1.143302585422892e-07, + "loss": 1.4096, + "step": 880 + }, + { + "epoch": 0.03469176947514081, + "grad_norm": 26.707822799682617, + "learning_rate": 1.156294660257243e-07, + "loss": 1.4687, + "step": 890 + }, + { + "epoch": 0.03508156463778284, + "grad_norm": 23.037599563598633, + "learning_rate": 1.169286735091594e-07, + "loss": 1.3896, + "step": 900 + }, + { + "epoch": 0.03547135980042488, + "grad_norm": 19.993160247802734, + "learning_rate": 1.1822788099259451e-07, + "loss": 1.3911, + "step": 910 + }, + { + "epoch": 0.03586115496306691, + "grad_norm": 31.847312927246094, + "learning_rate": 1.1952708847602962e-07, + "loss": 1.402, + "step": 920 + }, + { + "epoch": 0.03625095012570894, + "grad_norm": 23.46971893310547, + "learning_rate": 1.2082629595946472e-07, + "loss": 1.3151, + "step": 930 + }, + { + "epoch": 0.03664074528835097, + "grad_norm": 20.893728256225586, + "learning_rate": 1.2212550344289983e-07, + "loss": 1.3821, + "step": 940 + }, + { + "epoch": 0.037030540450993, + "grad_norm": 24.60841941833496, + "learning_rate": 1.2342471092633494e-07, + "loss": 1.3349, + "step": 950 + }, + { + "epoch": 0.03742033561363504, + "grad_norm": 39.57422637939453, + "learning_rate": 1.2472391840977002e-07, + "loss": 1.3614, + "step": 960 + }, + { + "epoch": 0.03781013077627707, + "grad_norm": 23.709590911865234, + "learning_rate": 1.2602312589320513e-07, + "loss": 1.3791, + "step": 970 + }, + { + "epoch": 0.0381999259389191, + "grad_norm": 25.98457908630371, + "learning_rate": 1.2732233337664024e-07, + "loss": 1.4767, + "step": 980 + }, + { + "epoch": 0.03858972110156113, + "grad_norm": 22.917917251586914, + "learning_rate": 1.2862154086007535e-07, + "loss": 1.3156, + "step": 990 + }, + { + "epoch": 0.03897951626420316, + "grad_norm": 20.277448654174805, + "learning_rate": 1.2992074834351043e-07, + "loss": 1.3303, + "step": 1000 + }, + { + "epoch": 0.03897951626420316, + "eval_loss": 1.3697881698608398, + "eval_runtime": 83.8319, + "eval_samples_per_second": 49.468, + "eval_steps_per_second": 6.191, + "step": 1000 + }, + { + "epoch": 0.039369311426845195, + "grad_norm": 25.3528995513916, + "learning_rate": 1.3121995582694556e-07, + "loss": 1.3373, + "step": 1010 + }, + { + "epoch": 0.039759106589487225, + "grad_norm": 38.638309478759766, + "learning_rate": 1.3251916331038064e-07, + "loss": 1.3174, + "step": 1020 + }, + { + "epoch": 0.040148901752129255, + "grad_norm": 22.78728675842285, + "learning_rate": 1.3381837079381578e-07, + "loss": 1.2947, + "step": 1030 + }, + { + "epoch": 0.040538696914771286, + "grad_norm": 32.03472137451172, + "learning_rate": 1.3511757827725086e-07, + "loss": 1.3397, + "step": 1040 + }, + { + "epoch": 0.04092849207741332, + "grad_norm": 25.601564407348633, + "learning_rate": 1.3641678576068597e-07, + "loss": 1.3274, + "step": 1050 + }, + { + "epoch": 0.04131828724005535, + "grad_norm": 21.4962100982666, + "learning_rate": 1.3771599324412108e-07, + "loss": 1.3597, + "step": 1060 + }, + { + "epoch": 0.04170808240269738, + "grad_norm": 23.354440689086914, + "learning_rate": 1.3901520072755619e-07, + "loss": 1.3443, + "step": 1070 + }, + { + "epoch": 0.04209787756533941, + "grad_norm": 29.751922607421875, + "learning_rate": 1.4031440821099127e-07, + "loss": 1.3801, + "step": 1080 + }, + { + "epoch": 0.04248767272798144, + "grad_norm": 22.104690551757812, + "learning_rate": 1.416136156944264e-07, + "loss": 1.2889, + "step": 1090 + }, + { + "epoch": 0.04287746789062348, + "grad_norm": 24.092578887939453, + "learning_rate": 1.4291282317786148e-07, + "loss": 1.3059, + "step": 1100 + }, + { + "epoch": 0.04326726305326551, + "grad_norm": 28.587032318115234, + "learning_rate": 1.4421203066129662e-07, + "loss": 1.3693, + "step": 1110 + }, + { + "epoch": 0.04365705821590754, + "grad_norm": 39.19141387939453, + "learning_rate": 1.455112381447317e-07, + "loss": 1.2924, + "step": 1120 + }, + { + "epoch": 0.04404685337854957, + "grad_norm": 31.168867111206055, + "learning_rate": 1.468104456281668e-07, + "loss": 1.1804, + "step": 1130 + }, + { + "epoch": 0.0444366485411916, + "grad_norm": 21.848552703857422, + "learning_rate": 1.4810965311160192e-07, + "loss": 1.3234, + "step": 1140 + }, + { + "epoch": 0.04482644370383364, + "grad_norm": 27.242977142333984, + "learning_rate": 1.4940886059503703e-07, + "loss": 1.4041, + "step": 1150 + }, + { + "epoch": 0.04521623886647567, + "grad_norm": 31.537796020507812, + "learning_rate": 1.507080680784721e-07, + "loss": 1.2986, + "step": 1160 + }, + { + "epoch": 0.0456060340291177, + "grad_norm": 32.458580017089844, + "learning_rate": 1.5200727556190724e-07, + "loss": 1.3143, + "step": 1170 + }, + { + "epoch": 0.04599582919175973, + "grad_norm": 25.82364845275879, + "learning_rate": 1.5330648304534232e-07, + "loss": 1.2583, + "step": 1180 + }, + { + "epoch": 0.046385624354401765, + "grad_norm": 27.868179321289062, + "learning_rate": 1.5460569052877746e-07, + "loss": 1.3071, + "step": 1190 + }, + { + "epoch": 0.046775419517043795, + "grad_norm": 28.524356842041016, + "learning_rate": 1.5590489801221254e-07, + "loss": 1.239, + "step": 1200 + }, + { + "epoch": 0.047165214679685825, + "grad_norm": 26.122072219848633, + "learning_rate": 1.5720410549564765e-07, + "loss": 1.2884, + "step": 1210 + }, + { + "epoch": 0.047555009842327856, + "grad_norm": 33.55978775024414, + "learning_rate": 1.5850331297908276e-07, + "loss": 1.2802, + "step": 1220 + }, + { + "epoch": 0.047944805004969886, + "grad_norm": 22.23946762084961, + "learning_rate": 1.5980252046251786e-07, + "loss": 1.2382, + "step": 1230 + }, + { + "epoch": 0.04833460016761192, + "grad_norm": 21.544361114501953, + "learning_rate": 1.6110172794595295e-07, + "loss": 1.2546, + "step": 1240 + }, + { + "epoch": 0.04872439533025395, + "grad_norm": 27.580764770507812, + "learning_rate": 1.6240093542938808e-07, + "loss": 1.304, + "step": 1250 + }, + { + "epoch": 0.04911419049289598, + "grad_norm": 19.470840454101562, + "learning_rate": 1.6370014291282316e-07, + "loss": 1.2626, + "step": 1260 + }, + { + "epoch": 0.04950398565553801, + "grad_norm": 32.23086929321289, + "learning_rate": 1.649993503962583e-07, + "loss": 1.2577, + "step": 1270 + }, + { + "epoch": 0.04989378081818004, + "grad_norm": 27.84480857849121, + "learning_rate": 1.6629855787969338e-07, + "loss": 1.259, + "step": 1280 + }, + { + "epoch": 0.05028357598082208, + "grad_norm": 34.83106994628906, + "learning_rate": 1.6759776536312846e-07, + "loss": 1.2759, + "step": 1290 + }, + { + "epoch": 0.05067337114346411, + "grad_norm": 35.22993469238281, + "learning_rate": 1.688969728465636e-07, + "loss": 1.209, + "step": 1300 + }, + { + "epoch": 0.05106316630610614, + "grad_norm": 24.666120529174805, + "learning_rate": 1.7019618032999868e-07, + "loss": 1.1823, + "step": 1310 + }, + { + "epoch": 0.05145296146874817, + "grad_norm": 20.746835708618164, + "learning_rate": 1.7149538781343379e-07, + "loss": 1.2481, + "step": 1320 + }, + { + "epoch": 0.05184275663139021, + "grad_norm": 28.15362548828125, + "learning_rate": 1.727945952968689e-07, + "loss": 1.2107, + "step": 1330 + }, + { + "epoch": 0.05223255179403224, + "grad_norm": 22.3405818939209, + "learning_rate": 1.74093802780304e-07, + "loss": 1.2569, + "step": 1340 + }, + { + "epoch": 0.05262234695667427, + "grad_norm": 25.35797691345215, + "learning_rate": 1.753930102637391e-07, + "loss": 1.2763, + "step": 1350 + }, + { + "epoch": 0.0530121421193163, + "grad_norm": 25.768844604492188, + "learning_rate": 1.7669221774717422e-07, + "loss": 1.2778, + "step": 1360 + }, + { + "epoch": 0.05340193728195833, + "grad_norm": 24.1910343170166, + "learning_rate": 1.779914252306093e-07, + "loss": 1.235, + "step": 1370 + }, + { + "epoch": 0.053791732444600365, + "grad_norm": 24.091421127319336, + "learning_rate": 1.7929063271404444e-07, + "loss": 1.282, + "step": 1380 + }, + { + "epoch": 0.054181527607242395, + "grad_norm": 20.77372169494629, + "learning_rate": 1.8058984019747952e-07, + "loss": 1.2546, + "step": 1390 + }, + { + "epoch": 0.054571322769884426, + "grad_norm": 17.36382484436035, + "learning_rate": 1.8188904768091463e-07, + "loss": 1.1963, + "step": 1400 + }, + { + "epoch": 0.054961117932526456, + "grad_norm": 17.44790267944336, + "learning_rate": 1.8318825516434973e-07, + "loss": 1.3036, + "step": 1410 + }, + { + "epoch": 0.055350913095168486, + "grad_norm": 20.17705535888672, + "learning_rate": 1.8448746264778484e-07, + "loss": 1.2578, + "step": 1420 + }, + { + "epoch": 0.05574070825781052, + "grad_norm": 29.456621170043945, + "learning_rate": 1.8578667013121995e-07, + "loss": 1.1798, + "step": 1430 + }, + { + "epoch": 0.05613050342045255, + "grad_norm": 18.45269203186035, + "learning_rate": 1.8708587761465506e-07, + "loss": 1.2811, + "step": 1440 + }, + { + "epoch": 0.05652029858309458, + "grad_norm": 32.55225372314453, + "learning_rate": 1.8838508509809014e-07, + "loss": 1.1683, + "step": 1450 + }, + { + "epoch": 0.05691009374573661, + "grad_norm": 24.9616641998291, + "learning_rate": 1.8968429258152528e-07, + "loss": 1.2833, + "step": 1460 + }, + { + "epoch": 0.05729988890837865, + "grad_norm": 19.256669998168945, + "learning_rate": 1.9098350006496036e-07, + "loss": 1.2059, + "step": 1470 + }, + { + "epoch": 0.05768968407102068, + "grad_norm": 18.240915298461914, + "learning_rate": 1.9228270754839547e-07, + "loss": 1.2216, + "step": 1480 + }, + { + "epoch": 0.05807947923366271, + "grad_norm": 18.207624435424805, + "learning_rate": 1.9358191503183057e-07, + "loss": 1.2127, + "step": 1490 + }, + { + "epoch": 0.05846927439630474, + "grad_norm": 21.78383445739746, + "learning_rate": 1.9488112251526568e-07, + "loss": 1.1926, + "step": 1500 + }, + { + "epoch": 0.05846927439630474, + "eval_loss": 1.2311373949050903, + "eval_runtime": 82.9565, + "eval_samples_per_second": 49.99, + "eval_steps_per_second": 6.256, + "step": 1500 + }, + { + "epoch": 0.05885906955894677, + "grad_norm": 21.91478729248047, + "learning_rate": 1.961803299987008e-07, + "loss": 1.2209, + "step": 1510 + }, + { + "epoch": 0.05924886472158881, + "grad_norm": 33.909385681152344, + "learning_rate": 1.974795374821359e-07, + "loss": 1.2217, + "step": 1520 + }, + { + "epoch": 0.05963865988423084, + "grad_norm": 19.926761627197266, + "learning_rate": 1.9877874496557098e-07, + "loss": 1.1819, + "step": 1530 + }, + { + "epoch": 0.06002845504687287, + "grad_norm": 20.056581497192383, + "learning_rate": 2.0007795244900611e-07, + "loss": 1.1496, + "step": 1540 + }, + { + "epoch": 0.0604182502095149, + "grad_norm": 33.30684280395508, + "learning_rate": 2.013771599324412e-07, + "loss": 1.2136, + "step": 1550 + }, + { + "epoch": 0.06080804537215693, + "grad_norm": 20.16193199157715, + "learning_rate": 2.026763674158763e-07, + "loss": 1.2147, + "step": 1560 + }, + { + "epoch": 0.061197840534798965, + "grad_norm": 27.292997360229492, + "learning_rate": 2.039755748993114e-07, + "loss": 1.1873, + "step": 1570 + }, + { + "epoch": 0.061587635697440996, + "grad_norm": 24.648168563842773, + "learning_rate": 2.0527478238274652e-07, + "loss": 1.2968, + "step": 1580 + }, + { + "epoch": 0.061977430860083026, + "grad_norm": 25.62425994873047, + "learning_rate": 2.0657398986618163e-07, + "loss": 1.1553, + "step": 1590 + }, + { + "epoch": 0.062367226022725056, + "grad_norm": 22.85311508178711, + "learning_rate": 2.078731973496167e-07, + "loss": 1.187, + "step": 1600 + }, + { + "epoch": 0.06275702118536709, + "grad_norm": 18.97858428955078, + "learning_rate": 2.0917240483305182e-07, + "loss": 1.1702, + "step": 1610 + }, + { + "epoch": 0.06314681634800912, + "grad_norm": 25.430551528930664, + "learning_rate": 2.1047161231648693e-07, + "loss": 1.2637, + "step": 1620 + }, + { + "epoch": 0.06353661151065115, + "grad_norm": 31.944725036621094, + "learning_rate": 2.1177081979992204e-07, + "loss": 1.1873, + "step": 1630 + }, + { + "epoch": 0.06392640667329319, + "grad_norm": 29.94383430480957, + "learning_rate": 2.1307002728335712e-07, + "loss": 1.1946, + "step": 1640 + }, + { + "epoch": 0.06431620183593521, + "grad_norm": 20.71239471435547, + "learning_rate": 2.1436923476679225e-07, + "loss": 1.2405, + "step": 1650 + }, + { + "epoch": 0.06470599699857725, + "grad_norm": 21.18262481689453, + "learning_rate": 2.1566844225022733e-07, + "loss": 1.2228, + "step": 1660 + }, + { + "epoch": 0.06509579216121927, + "grad_norm": 24.436254501342773, + "learning_rate": 2.1696764973366247e-07, + "loss": 1.2376, + "step": 1670 + }, + { + "epoch": 0.06548558732386131, + "grad_norm": 17.257238388061523, + "learning_rate": 2.1826685721709755e-07, + "loss": 1.1877, + "step": 1680 + }, + { + "epoch": 0.06587538248650335, + "grad_norm": 17.540054321289062, + "learning_rate": 2.1956606470053266e-07, + "loss": 1.2167, + "step": 1690 + }, + { + "epoch": 0.06626517764914537, + "grad_norm": 17.75524139404297, + "learning_rate": 2.2086527218396777e-07, + "loss": 1.18, + "step": 1700 + }, + { + "epoch": 0.06665497281178741, + "grad_norm": 20.71023178100586, + "learning_rate": 2.2216447966740288e-07, + "loss": 1.1599, + "step": 1710 + }, + { + "epoch": 0.06704476797442943, + "grad_norm": 21.568775177001953, + "learning_rate": 2.2346368715083796e-07, + "loss": 1.2124, + "step": 1720 + }, + { + "epoch": 0.06743456313707147, + "grad_norm": 19.144420623779297, + "learning_rate": 2.247628946342731e-07, + "loss": 1.254, + "step": 1730 + }, + { + "epoch": 0.0678243582997135, + "grad_norm": 18.634767532348633, + "learning_rate": 2.2606210211770817e-07, + "loss": 1.1915, + "step": 1740 + }, + { + "epoch": 0.06821415346235553, + "grad_norm": 18.5625057220459, + "learning_rate": 2.273613096011433e-07, + "loss": 1.1779, + "step": 1750 + }, + { + "epoch": 0.06860394862499757, + "grad_norm": 18.945903778076172, + "learning_rate": 2.286605170845784e-07, + "loss": 1.1449, + "step": 1760 + }, + { + "epoch": 0.06899374378763959, + "grad_norm": 27.690719604492188, + "learning_rate": 2.299597245680135e-07, + "loss": 1.1159, + "step": 1770 + }, + { + "epoch": 0.06938353895028163, + "grad_norm": 35.688194274902344, + "learning_rate": 2.312589320514486e-07, + "loss": 1.221, + "step": 1780 + }, + { + "epoch": 0.06977333411292366, + "grad_norm": 19.111961364746094, + "learning_rate": 2.3255813953488372e-07, + "loss": 1.267, + "step": 1790 + }, + { + "epoch": 0.07016312927556569, + "grad_norm": 24.548988342285156, + "learning_rate": 2.338573470183188e-07, + "loss": 1.1901, + "step": 1800 + }, + { + "epoch": 0.07055292443820772, + "grad_norm": 16.778841018676758, + "learning_rate": 2.3515655450175393e-07, + "loss": 1.1129, + "step": 1810 + }, + { + "epoch": 0.07094271960084976, + "grad_norm": 19.7405948638916, + "learning_rate": 2.3645576198518901e-07, + "loss": 1.123, + "step": 1820 + }, + { + "epoch": 0.07133251476349178, + "grad_norm": 22.7630615234375, + "learning_rate": 2.3775496946862415e-07, + "loss": 1.1761, + "step": 1830 + }, + { + "epoch": 0.07172230992613382, + "grad_norm": 21.28675651550293, + "learning_rate": 2.3905417695205923e-07, + "loss": 1.1757, + "step": 1840 + }, + { + "epoch": 0.07211210508877584, + "grad_norm": 22.512479782104492, + "learning_rate": 2.403533844354943e-07, + "loss": 1.1985, + "step": 1850 + }, + { + "epoch": 0.07250190025141788, + "grad_norm": 21.829023361206055, + "learning_rate": 2.4165259191892945e-07, + "loss": 1.2018, + "step": 1860 + }, + { + "epoch": 0.07289169541405992, + "grad_norm": 29.038089752197266, + "learning_rate": 2.4295179940236453e-07, + "loss": 1.1605, + "step": 1870 + }, + { + "epoch": 0.07328149057670194, + "grad_norm": 17.541589736938477, + "learning_rate": 2.4425100688579966e-07, + "loss": 1.1629, + "step": 1880 + }, + { + "epoch": 0.07367128573934398, + "grad_norm": 22.817169189453125, + "learning_rate": 2.4555021436923474e-07, + "loss": 1.1564, + "step": 1890 + }, + { + "epoch": 0.074061080901986, + "grad_norm": 25.73829460144043, + "learning_rate": 2.468494218526699e-07, + "loss": 1.1876, + "step": 1900 + }, + { + "epoch": 0.07445087606462804, + "grad_norm": 19.251325607299805, + "learning_rate": 2.4814862933610496e-07, + "loss": 1.0866, + "step": 1910 + }, + { + "epoch": 0.07484067122727008, + "grad_norm": 20.26015853881836, + "learning_rate": 2.4944783681954004e-07, + "loss": 1.1632, + "step": 1920 + }, + { + "epoch": 0.0752304663899121, + "grad_norm": 18.77247428894043, + "learning_rate": 2.507470443029752e-07, + "loss": 1.2252, + "step": 1930 + }, + { + "epoch": 0.07562026155255414, + "grad_norm": 18.1500301361084, + "learning_rate": 2.5204625178641026e-07, + "loss": 1.1154, + "step": 1940 + }, + { + "epoch": 0.07601005671519616, + "grad_norm": 17.62322425842285, + "learning_rate": 2.533454592698454e-07, + "loss": 1.1755, + "step": 1950 + }, + { + "epoch": 0.0763998518778382, + "grad_norm": 20.191869735717773, + "learning_rate": 2.546446667532805e-07, + "loss": 1.2767, + "step": 1960 + }, + { + "epoch": 0.07678964704048023, + "grad_norm": 19.961292266845703, + "learning_rate": 2.559438742367156e-07, + "loss": 1.1227, + "step": 1970 + }, + { + "epoch": 0.07717944220312226, + "grad_norm": 29.976518630981445, + "learning_rate": 2.572430817201507e-07, + "loss": 1.1429, + "step": 1980 + }, + { + "epoch": 0.0775692373657643, + "grad_norm": 23.046350479125977, + "learning_rate": 2.585422892035858e-07, + "loss": 1.1368, + "step": 1990 + }, + { + "epoch": 0.07795903252840632, + "grad_norm": 22.612689971923828, + "learning_rate": 2.5984149668702086e-07, + "loss": 1.1254, + "step": 2000 + }, + { + "epoch": 0.07795903252840632, + "eval_loss": 1.1641405820846558, + "eval_runtime": 83.0994, + "eval_samples_per_second": 49.904, + "eval_steps_per_second": 6.246, + "step": 2000 + }, + { + "epoch": 0.07834882769104835, + "grad_norm": 20.917192459106445, + "learning_rate": 2.6114070417045604e-07, + "loss": 1.0889, + "step": 2010 + }, + { + "epoch": 0.07873862285369039, + "grad_norm": 21.289188385009766, + "learning_rate": 2.624399116538911e-07, + "loss": 1.223, + "step": 2020 + }, + { + "epoch": 0.07912841801633241, + "grad_norm": 18.19302749633789, + "learning_rate": 2.637391191373262e-07, + "loss": 1.114, + "step": 2030 + }, + { + "epoch": 0.07951821317897445, + "grad_norm": 20.206937789916992, + "learning_rate": 2.650383266207613e-07, + "loss": 1.1474, + "step": 2040 + }, + { + "epoch": 0.07990800834161647, + "grad_norm": 40.2934455871582, + "learning_rate": 2.663375341041964e-07, + "loss": 1.1634, + "step": 2050 + }, + { + "epoch": 0.08029780350425851, + "grad_norm": 19.17534065246582, + "learning_rate": 2.6763674158763156e-07, + "loss": 1.1258, + "step": 2060 + }, + { + "epoch": 0.08068759866690055, + "grad_norm": 24.358631134033203, + "learning_rate": 2.6893594907106664e-07, + "loss": 1.207, + "step": 2070 + }, + { + "epoch": 0.08107739382954257, + "grad_norm": 16.418169021606445, + "learning_rate": 2.702351565545017e-07, + "loss": 1.1264, + "step": 2080 + }, + { + "epoch": 0.08146718899218461, + "grad_norm": 19.96869468688965, + "learning_rate": 2.7153436403793686e-07, + "loss": 1.2131, + "step": 2090 + }, + { + "epoch": 0.08185698415482665, + "grad_norm": 17.45491600036621, + "learning_rate": 2.7283357152137194e-07, + "loss": 1.0997, + "step": 2100 + }, + { + "epoch": 0.08224677931746867, + "grad_norm": 15.956841468811035, + "learning_rate": 2.7413277900480707e-07, + "loss": 1.127, + "step": 2110 + }, + { + "epoch": 0.0826365744801107, + "grad_norm": 24.947649002075195, + "learning_rate": 2.7543198648824216e-07, + "loss": 1.1553, + "step": 2120 + }, + { + "epoch": 0.08302636964275273, + "grad_norm": 30.256031036376953, + "learning_rate": 2.767311939716773e-07, + "loss": 1.1617, + "step": 2130 + }, + { + "epoch": 0.08341616480539477, + "grad_norm": 17.281461715698242, + "learning_rate": 2.7803040145511237e-07, + "loss": 1.1628, + "step": 2140 + }, + { + "epoch": 0.0838059599680368, + "grad_norm": 21.757123947143555, + "learning_rate": 2.7932960893854745e-07, + "loss": 1.1684, + "step": 2150 + }, + { + "epoch": 0.08419575513067883, + "grad_norm": 16.572677612304688, + "learning_rate": 2.8062881642198254e-07, + "loss": 1.1092, + "step": 2160 + }, + { + "epoch": 0.08458555029332086, + "grad_norm": 31.1545352935791, + "learning_rate": 2.819280239054177e-07, + "loss": 1.0977, + "step": 2170 + }, + { + "epoch": 0.08497534545596289, + "grad_norm": 22.96057891845703, + "learning_rate": 2.832272313888528e-07, + "loss": 1.0909, + "step": 2180 + }, + { + "epoch": 0.08536514061860492, + "grad_norm": 18.663755416870117, + "learning_rate": 2.845264388722879e-07, + "loss": 1.1981, + "step": 2190 + }, + { + "epoch": 0.08575493578124696, + "grad_norm": 15.627538681030273, + "learning_rate": 2.8582564635572297e-07, + "loss": 1.1849, + "step": 2200 + }, + { + "epoch": 0.08614473094388898, + "grad_norm": 22.33970832824707, + "learning_rate": 2.8712485383915805e-07, + "loss": 1.1101, + "step": 2210 + }, + { + "epoch": 0.08653452610653102, + "grad_norm": 20.601247787475586, + "learning_rate": 2.8842406132259324e-07, + "loss": 1.1344, + "step": 2220 + }, + { + "epoch": 0.08692432126917304, + "grad_norm": 24.503992080688477, + "learning_rate": 2.897232688060283e-07, + "loss": 1.1651, + "step": 2230 + }, + { + "epoch": 0.08731411643181508, + "grad_norm": 20.19804573059082, + "learning_rate": 2.910224762894634e-07, + "loss": 1.1342, + "step": 2240 + }, + { + "epoch": 0.08770391159445712, + "grad_norm": 28.511154174804688, + "learning_rate": 2.923216837728985e-07, + "loss": 1.1095, + "step": 2250 + }, + { + "epoch": 0.08809370675709914, + "grad_norm": 19.773250579833984, + "learning_rate": 2.936208912563336e-07, + "loss": 1.1991, + "step": 2260 + }, + { + "epoch": 0.08848350191974118, + "grad_norm": 20.530174255371094, + "learning_rate": 2.9492009873976875e-07, + "loss": 1.1637, + "step": 2270 + }, + { + "epoch": 0.0888732970823832, + "grad_norm": 19.899505615234375, + "learning_rate": 2.9621930622320383e-07, + "loss": 1.1901, + "step": 2280 + }, + { + "epoch": 0.08926309224502524, + "grad_norm": 23.331287384033203, + "learning_rate": 2.975185137066389e-07, + "loss": 1.096, + "step": 2290 + }, + { + "epoch": 0.08965288740766728, + "grad_norm": 22.39967155456543, + "learning_rate": 2.9881772119007405e-07, + "loss": 1.1002, + "step": 2300 + }, + { + "epoch": 0.0900426825703093, + "grad_norm": 17.148853302001953, + "learning_rate": 3.0011692867350913e-07, + "loss": 1.1317, + "step": 2310 + }, + { + "epoch": 0.09043247773295134, + "grad_norm": 17.86640167236328, + "learning_rate": 3.014161361569442e-07, + "loss": 1.1685, + "step": 2320 + }, + { + "epoch": 0.09082227289559336, + "grad_norm": 24.707834243774414, + "learning_rate": 3.0271534364037935e-07, + "loss": 1.1374, + "step": 2330 + }, + { + "epoch": 0.0912120680582354, + "grad_norm": 16.77250099182129, + "learning_rate": 3.040145511238145e-07, + "loss": 1.198, + "step": 2340 + }, + { + "epoch": 0.09160186322087743, + "grad_norm": 15.07976245880127, + "learning_rate": 3.0531375860724957e-07, + "loss": 1.1641, + "step": 2350 + }, + { + "epoch": 0.09199165838351946, + "grad_norm": 19.27408790588379, + "learning_rate": 3.0661296609068465e-07, + "loss": 1.1246, + "step": 2360 + }, + { + "epoch": 0.0923814535461615, + "grad_norm": 30.60783576965332, + "learning_rate": 3.0791217357411973e-07, + "loss": 1.1515, + "step": 2370 + }, + { + "epoch": 0.09277124870880353, + "grad_norm": 20.21698570251465, + "learning_rate": 3.092113810575549e-07, + "loss": 1.0558, + "step": 2380 + }, + { + "epoch": 0.09316104387144555, + "grad_norm": 19.32050132751465, + "learning_rate": 3.1051058854099e-07, + "loss": 1.0905, + "step": 2390 + }, + { + "epoch": 0.09355083903408759, + "grad_norm": 16.702831268310547, + "learning_rate": 3.118097960244251e-07, + "loss": 1.1467, + "step": 2400 + }, + { + "epoch": 0.09394063419672961, + "grad_norm": 19.29012680053711, + "learning_rate": 3.1310900350786016e-07, + "loss": 1.166, + "step": 2410 + }, + { + "epoch": 0.09433042935937165, + "grad_norm": 17.370508193969727, + "learning_rate": 3.144082109912953e-07, + "loss": 1.1463, + "step": 2420 + }, + { + "epoch": 0.09472022452201369, + "grad_norm": 20.080320358276367, + "learning_rate": 3.1570741847473043e-07, + "loss": 1.1051, + "step": 2430 + }, + { + "epoch": 0.09511001968465571, + "grad_norm": 16.047531127929688, + "learning_rate": 3.170066259581655e-07, + "loss": 1.1243, + "step": 2440 + }, + { + "epoch": 0.09549981484729775, + "grad_norm": 22.197399139404297, + "learning_rate": 3.183058334416006e-07, + "loss": 1.096, + "step": 2450 + }, + { + "epoch": 0.09588961000993977, + "grad_norm": 22.106971740722656, + "learning_rate": 3.1960504092503573e-07, + "loss": 1.1407, + "step": 2460 + }, + { + "epoch": 0.09627940517258181, + "grad_norm": 20.432201385498047, + "learning_rate": 3.209042484084708e-07, + "loss": 1.1017, + "step": 2470 + }, + { + "epoch": 0.09666920033522385, + "grad_norm": 21.421112060546875, + "learning_rate": 3.222034558919059e-07, + "loss": 1.0986, + "step": 2480 + }, + { + "epoch": 0.09705899549786587, + "grad_norm": 17.677053451538086, + "learning_rate": 3.2350266337534103e-07, + "loss": 1.0745, + "step": 2490 + }, + { + "epoch": 0.0974487906605079, + "grad_norm": 19.21147346496582, + "learning_rate": 3.2480187085877616e-07, + "loss": 1.0642, + "step": 2500 + }, + { + "epoch": 0.0974487906605079, + "eval_loss": 1.1195721626281738, + "eval_runtime": 85.2919, + "eval_samples_per_second": 48.621, + "eval_steps_per_second": 6.085, + "step": 2500 + }, + { + "epoch": 0.09783858582314993, + "grad_norm": 17.982376098632812, + "learning_rate": 3.2610107834221124e-07, + "loss": 1.1618, + "step": 2510 + }, + { + "epoch": 0.09822838098579197, + "grad_norm": 25.378965377807617, + "learning_rate": 3.2740028582564633e-07, + "loss": 1.1165, + "step": 2520 + }, + { + "epoch": 0.098618176148434, + "grad_norm": 26.69707489013672, + "learning_rate": 3.286994933090814e-07, + "loss": 1.165, + "step": 2530 + }, + { + "epoch": 0.09900797131107603, + "grad_norm": 25.33272361755371, + "learning_rate": 3.299987007925166e-07, + "loss": 1.1625, + "step": 2540 + }, + { + "epoch": 0.09939776647371806, + "grad_norm": 18.306480407714844, + "learning_rate": 3.312979082759517e-07, + "loss": 1.1538, + "step": 2550 + }, + { + "epoch": 0.09978756163636009, + "grad_norm": 17.6303653717041, + "learning_rate": 3.3259711575938676e-07, + "loss": 1.0943, + "step": 2560 + }, + { + "epoch": 0.10017735679900212, + "grad_norm": 23.567230224609375, + "learning_rate": 3.3389632324282184e-07, + "loss": 1.0164, + "step": 2570 + }, + { + "epoch": 0.10056715196164416, + "grad_norm": 23.226415634155273, + "learning_rate": 3.351955307262569e-07, + "loss": 1.0877, + "step": 2580 + }, + { + "epoch": 0.10095694712428618, + "grad_norm": 18.79073143005371, + "learning_rate": 3.364947382096921e-07, + "loss": 1.1188, + "step": 2590 + }, + { + "epoch": 0.10134674228692822, + "grad_norm": 20.62694549560547, + "learning_rate": 3.377939456931272e-07, + "loss": 1.0816, + "step": 2600 + }, + { + "epoch": 0.10173653744957024, + "grad_norm": 17.1717529296875, + "learning_rate": 3.390931531765623e-07, + "loss": 1.1468, + "step": 2610 + }, + { + "epoch": 0.10212633261221228, + "grad_norm": 21.884056091308594, + "learning_rate": 3.4039236065999736e-07, + "loss": 1.0441, + "step": 2620 + }, + { + "epoch": 0.10251612777485432, + "grad_norm": 15.515151023864746, + "learning_rate": 3.416915681434325e-07, + "loss": 1.1097, + "step": 2630 + }, + { + "epoch": 0.10290592293749634, + "grad_norm": 20.293100357055664, + "learning_rate": 3.4299077562686757e-07, + "loss": 1.1488, + "step": 2640 + }, + { + "epoch": 0.10329571810013838, + "grad_norm": 23.01701545715332, + "learning_rate": 3.442899831103027e-07, + "loss": 1.1199, + "step": 2650 + }, + { + "epoch": 0.10368551326278042, + "grad_norm": 25.233861923217773, + "learning_rate": 3.455891905937378e-07, + "loss": 1.1112, + "step": 2660 + }, + { + "epoch": 0.10407530842542244, + "grad_norm": 16.357101440429688, + "learning_rate": 3.468883980771729e-07, + "loss": 1.1227, + "step": 2670 + }, + { + "epoch": 0.10446510358806448, + "grad_norm": 17.625179290771484, + "learning_rate": 3.48187605560608e-07, + "loss": 1.1665, + "step": 2680 + }, + { + "epoch": 0.1048548987507065, + "grad_norm": 16.159347534179688, + "learning_rate": 3.494868130440431e-07, + "loss": 1.088, + "step": 2690 + }, + { + "epoch": 0.10524469391334854, + "grad_norm": 17.906139373779297, + "learning_rate": 3.507860205274782e-07, + "loss": 1.0644, + "step": 2700 + }, + { + "epoch": 0.10563448907599057, + "grad_norm": 16.22184944152832, + "learning_rate": 3.5208522801091336e-07, + "loss": 1.0982, + "step": 2710 + }, + { + "epoch": 0.1060242842386326, + "grad_norm": 20.092205047607422, + "learning_rate": 3.5338443549434844e-07, + "loss": 1.0855, + "step": 2720 + }, + { + "epoch": 0.10641407940127463, + "grad_norm": 39.1601676940918, + "learning_rate": 3.546836429777835e-07, + "loss": 1.1113, + "step": 2730 + }, + { + "epoch": 0.10680387456391666, + "grad_norm": 16.13973617553711, + "learning_rate": 3.559828504612186e-07, + "loss": 1.1487, + "step": 2740 + }, + { + "epoch": 0.1071936697265587, + "grad_norm": 18.15250587463379, + "learning_rate": 3.572820579446538e-07, + "loss": 1.1117, + "step": 2750 + }, + { + "epoch": 0.10758346488920073, + "grad_norm": 18.15755271911621, + "learning_rate": 3.5858126542808887e-07, + "loss": 1.0663, + "step": 2760 + }, + { + "epoch": 0.10797326005184275, + "grad_norm": 16.700483322143555, + "learning_rate": 3.5988047291152395e-07, + "loss": 1.0897, + "step": 2770 + }, + { + "epoch": 0.10836305521448479, + "grad_norm": 16.612733840942383, + "learning_rate": 3.6117968039495904e-07, + "loss": 1.0843, + "step": 2780 + }, + { + "epoch": 0.10875285037712681, + "grad_norm": 19.8333740234375, + "learning_rate": 3.6247888787839417e-07, + "loss": 1.1553, + "step": 2790 + }, + { + "epoch": 0.10914264553976885, + "grad_norm": 26.184743881225586, + "learning_rate": 3.6377809536182925e-07, + "loss": 1.0934, + "step": 2800 + }, + { + "epoch": 0.10953244070241089, + "grad_norm": 27.946151733398438, + "learning_rate": 3.650773028452644e-07, + "loss": 1.1118, + "step": 2810 + }, + { + "epoch": 0.10992223586505291, + "grad_norm": 23.988351821899414, + "learning_rate": 3.6637651032869947e-07, + "loss": 1.131, + "step": 2820 + }, + { + "epoch": 0.11031203102769495, + "grad_norm": 17.490507125854492, + "learning_rate": 3.676757178121346e-07, + "loss": 1.1974, + "step": 2830 + }, + { + "epoch": 0.11070182619033697, + "grad_norm": 15.807207107543945, + "learning_rate": 3.689749252955697e-07, + "loss": 1.0591, + "step": 2840 + }, + { + "epoch": 0.11109162135297901, + "grad_norm": 22.07007598876953, + "learning_rate": 3.7027413277900477e-07, + "loss": 1.1617, + "step": 2850 + }, + { + "epoch": 0.11148141651562105, + "grad_norm": 15.649239540100098, + "learning_rate": 3.715733402624399e-07, + "loss": 1.0867, + "step": 2860 + }, + { + "epoch": 0.11187121167826307, + "grad_norm": 17.1900577545166, + "learning_rate": 3.7287254774587504e-07, + "loss": 1.0747, + "step": 2870 + }, + { + "epoch": 0.1122610068409051, + "grad_norm": 16.546329498291016, + "learning_rate": 3.741717552293101e-07, + "loss": 1.1006, + "step": 2880 + }, + { + "epoch": 0.11265080200354713, + "grad_norm": 17.517803192138672, + "learning_rate": 3.754709627127452e-07, + "loss": 1.0932, + "step": 2890 + }, + { + "epoch": 0.11304059716618917, + "grad_norm": 17.336828231811523, + "learning_rate": 3.767701701961803e-07, + "loss": 1.094, + "step": 2900 + }, + { + "epoch": 0.1134303923288312, + "grad_norm": 18.63520050048828, + "learning_rate": 3.780693776796154e-07, + "loss": 1.0823, + "step": 2910 + }, + { + "epoch": 0.11382018749147323, + "grad_norm": 16.782115936279297, + "learning_rate": 3.7936858516305055e-07, + "loss": 1.1587, + "step": 2920 + }, + { + "epoch": 0.11420998265411526, + "grad_norm": 17.314008712768555, + "learning_rate": 3.8066779264648563e-07, + "loss": 1.0523, + "step": 2930 + }, + { + "epoch": 0.1145997778167573, + "grad_norm": 18.21299934387207, + "learning_rate": 3.819670001299207e-07, + "loss": 1.0673, + "step": 2940 + }, + { + "epoch": 0.11498957297939932, + "grad_norm": 18.328964233398438, + "learning_rate": 3.832662076133558e-07, + "loss": 1.0794, + "step": 2950 + }, + { + "epoch": 0.11537936814204136, + "grad_norm": 20.03628158569336, + "learning_rate": 3.8456541509679093e-07, + "loss": 1.1128, + "step": 2960 + }, + { + "epoch": 0.11576916330468338, + "grad_norm": 22.702531814575195, + "learning_rate": 3.8586462258022607e-07, + "loss": 1.1711, + "step": 2970 + }, + { + "epoch": 0.11615895846732542, + "grad_norm": 16.880537033081055, + "learning_rate": 3.8716383006366115e-07, + "loss": 1.1018, + "step": 2980 + }, + { + "epoch": 0.11654875362996746, + "grad_norm": 13.726103782653809, + "learning_rate": 3.8846303754709623e-07, + "loss": 0.9971, + "step": 2990 + }, + { + "epoch": 0.11693854879260948, + "grad_norm": 18.31511116027832, + "learning_rate": 3.8976224503053136e-07, + "loss": 1.1126, + "step": 3000 + }, + { + "epoch": 0.11693854879260948, + "eval_loss": 1.094496488571167, + "eval_runtime": 82.9835, + "eval_samples_per_second": 49.974, + "eval_steps_per_second": 6.254, + "step": 3000 + }, + { + "epoch": 0.11732834395525152, + "grad_norm": 15.176326751708984, + "learning_rate": 3.9106145251396645e-07, + "loss": 1.0562, + "step": 3010 + }, + { + "epoch": 0.11771813911789354, + "grad_norm": 17.718870162963867, + "learning_rate": 3.923606599974016e-07, + "loss": 1.0935, + "step": 3020 + }, + { + "epoch": 0.11810793428053558, + "grad_norm": 20.324752807617188, + "learning_rate": 3.9365986748083666e-07, + "loss": 1.1081, + "step": 3030 + }, + { + "epoch": 0.11849772944317762, + "grad_norm": 19.459096908569336, + "learning_rate": 3.949590749642718e-07, + "loss": 1.096, + "step": 3040 + }, + { + "epoch": 0.11888752460581964, + "grad_norm": 17.35738182067871, + "learning_rate": 3.962582824477069e-07, + "loss": 1.1063, + "step": 3050 + }, + { + "epoch": 0.11927731976846168, + "grad_norm": 20.396133422851562, + "learning_rate": 3.9755748993114196e-07, + "loss": 1.0465, + "step": 3060 + }, + { + "epoch": 0.1196671149311037, + "grad_norm": 17.108139038085938, + "learning_rate": 3.988566974145771e-07, + "loss": 1.0022, + "step": 3070 + }, + { + "epoch": 0.12005691009374574, + "grad_norm": 19.336660385131836, + "learning_rate": 4.0015590489801223e-07, + "loss": 1.0201, + "step": 3080 + }, + { + "epoch": 0.12044670525638777, + "grad_norm": 19.237716674804688, + "learning_rate": 4.014551123814473e-07, + "loss": 1.1178, + "step": 3090 + }, + { + "epoch": 0.1208365004190298, + "grad_norm": 19.837730407714844, + "learning_rate": 4.027543198648824e-07, + "loss": 0.9963, + "step": 3100 + }, + { + "epoch": 0.12122629558167183, + "grad_norm": 16.92448616027832, + "learning_rate": 4.040535273483175e-07, + "loss": 1.0775, + "step": 3110 + }, + { + "epoch": 0.12161609074431386, + "grad_norm": 17.480093002319336, + "learning_rate": 4.053527348317526e-07, + "loss": 1.1085, + "step": 3120 + }, + { + "epoch": 0.1220058859069559, + "grad_norm": 15.263054847717285, + "learning_rate": 4.0665194231518774e-07, + "loss": 1.0234, + "step": 3130 + }, + { + "epoch": 0.12239568106959793, + "grad_norm": 15.767720222473145, + "learning_rate": 4.079511497986228e-07, + "loss": 1.059, + "step": 3140 + }, + { + "epoch": 0.12278547623223995, + "grad_norm": 15.416570663452148, + "learning_rate": 4.092503572820579e-07, + "loss": 1.0649, + "step": 3150 + }, + { + "epoch": 0.12317527139488199, + "grad_norm": 19.639291763305664, + "learning_rate": 4.1054956476549304e-07, + "loss": 1.0827, + "step": 3160 + }, + { + "epoch": 0.12356506655752401, + "grad_norm": 19.81853485107422, + "learning_rate": 4.118487722489281e-07, + "loss": 1.0101, + "step": 3170 + }, + { + "epoch": 0.12395486172016605, + "grad_norm": 20.000398635864258, + "learning_rate": 4.1314797973236326e-07, + "loss": 1.0565, + "step": 3180 + }, + { + "epoch": 0.12434465688280809, + "grad_norm": 20.127769470214844, + "learning_rate": 4.1444718721579834e-07, + "loss": 1.1242, + "step": 3190 + }, + { + "epoch": 0.12473445204545011, + "grad_norm": 18.09425163269043, + "learning_rate": 4.157463946992334e-07, + "loss": 1.0892, + "step": 3200 + }, + { + "epoch": 0.12512424720809215, + "grad_norm": 18.4867000579834, + "learning_rate": 4.1704560218266856e-07, + "loss": 1.1157, + "step": 3210 + }, + { + "epoch": 0.12551404237073419, + "grad_norm": 19.374168395996094, + "learning_rate": 4.1834480966610364e-07, + "loss": 1.0751, + "step": 3220 + }, + { + "epoch": 0.12590383753337622, + "grad_norm": 18.61876678466797, + "learning_rate": 4.196440171495388e-07, + "loss": 1.0818, + "step": 3230 + }, + { + "epoch": 0.12629363269601823, + "grad_norm": 20.80162811279297, + "learning_rate": 4.2094322463297386e-07, + "loss": 1.0969, + "step": 3240 + }, + { + "epoch": 0.12668342785866027, + "grad_norm": 16.84150505065918, + "learning_rate": 4.22242432116409e-07, + "loss": 1.0474, + "step": 3250 + }, + { + "epoch": 0.1270732230213023, + "grad_norm": 18.334505081176758, + "learning_rate": 4.2354163959984407e-07, + "loss": 1.0308, + "step": 3260 + }, + { + "epoch": 0.12746301818394434, + "grad_norm": 19.96571922302246, + "learning_rate": 4.2484084708327915e-07, + "loss": 1.0634, + "step": 3270 + }, + { + "epoch": 0.12785281334658638, + "grad_norm": 17.380096435546875, + "learning_rate": 4.2614005456671424e-07, + "loss": 1.1257, + "step": 3280 + }, + { + "epoch": 0.1282426085092284, + "grad_norm": 20.934755325317383, + "learning_rate": 4.274392620501494e-07, + "loss": 1.1326, + "step": 3290 + }, + { + "epoch": 0.12863240367187043, + "grad_norm": 15.561760902404785, + "learning_rate": 4.287384695335845e-07, + "loss": 1.1235, + "step": 3300 + }, + { + "epoch": 0.12902219883451246, + "grad_norm": 24.218372344970703, + "learning_rate": 4.300376770170196e-07, + "loss": 1.0702, + "step": 3310 + }, + { + "epoch": 0.1294119939971545, + "grad_norm": 16.878768920898438, + "learning_rate": 4.3133688450045467e-07, + "loss": 1.0434, + "step": 3320 + }, + { + "epoch": 0.12980178915979654, + "grad_norm": 17.845272064208984, + "learning_rate": 4.326360919838898e-07, + "loss": 1.0436, + "step": 3330 + }, + { + "epoch": 0.13019158432243855, + "grad_norm": 15.354514122009277, + "learning_rate": 4.3393529946732494e-07, + "loss": 1.047, + "step": 3340 + }, + { + "epoch": 0.13058137948508058, + "grad_norm": 16.830114364624023, + "learning_rate": 4.3523450695076e-07, + "loss": 1.0343, + "step": 3350 + }, + { + "epoch": 0.13097117464772262, + "grad_norm": 16.88136100769043, + "learning_rate": 4.365337144341951e-07, + "loss": 1.0506, + "step": 3360 + }, + { + "epoch": 0.13136096981036466, + "grad_norm": 22.108646392822266, + "learning_rate": 4.3783292191763024e-07, + "loss": 1.0246, + "step": 3370 + }, + { + "epoch": 0.1317507649730067, + "grad_norm": 18.705514907836914, + "learning_rate": 4.391321294010653e-07, + "loss": 1.0613, + "step": 3380 + }, + { + "epoch": 0.1321405601356487, + "grad_norm": 17.11370849609375, + "learning_rate": 4.4043133688450045e-07, + "loss": 1.0662, + "step": 3390 + }, + { + "epoch": 0.13253035529829074, + "grad_norm": 18.835689544677734, + "learning_rate": 4.4173054436793554e-07, + "loss": 1.1196, + "step": 3400 + }, + { + "epoch": 0.13292015046093278, + "grad_norm": 17.418352127075195, + "learning_rate": 4.4302975185137067e-07, + "loss": 1.0669, + "step": 3410 + }, + { + "epoch": 0.13330994562357482, + "grad_norm": 15.820262908935547, + "learning_rate": 4.4432895933480575e-07, + "loss": 1.1085, + "step": 3420 + }, + { + "epoch": 0.13369974078621685, + "grad_norm": 15.507389068603516, + "learning_rate": 4.4562816681824083e-07, + "loss": 1.0401, + "step": 3430 + }, + { + "epoch": 0.13408953594885886, + "grad_norm": 16.742481231689453, + "learning_rate": 4.469273743016759e-07, + "loss": 1.0101, + "step": 3440 + }, + { + "epoch": 0.1344793311115009, + "grad_norm": 16.379539489746094, + "learning_rate": 4.482265817851111e-07, + "loss": 1.0597, + "step": 3450 + }, + { + "epoch": 0.13486912627414294, + "grad_norm": 18.920534133911133, + "learning_rate": 4.495257892685462e-07, + "loss": 1.0763, + "step": 3460 + }, + { + "epoch": 0.13525892143678497, + "grad_norm": 15.016353607177734, + "learning_rate": 4.5082499675198127e-07, + "loss": 1.0855, + "step": 3470 + }, + { + "epoch": 0.135648716599427, + "grad_norm": 16.04503631591797, + "learning_rate": 4.5212420423541635e-07, + "loss": 1.0547, + "step": 3480 + }, + { + "epoch": 0.13603851176206902, + "grad_norm": 17.34245491027832, + "learning_rate": 4.5342341171885143e-07, + "loss": 1.0304, + "step": 3490 + }, + { + "epoch": 0.13642830692471106, + "grad_norm": 14.309751510620117, + "learning_rate": 4.547226192022866e-07, + "loss": 1.0601, + "step": 3500 + }, + { + "epoch": 0.13642830692471106, + "eval_loss": 1.0717768669128418, + "eval_runtime": 83.0859, + "eval_samples_per_second": 49.912, + "eval_steps_per_second": 6.247, + "step": 3500 + }, + { + "epoch": 0.1368181020873531, + "grad_norm": 22.865070343017578, + "learning_rate": 4.560218266857217e-07, + "loss": 1.0725, + "step": 3510 + }, + { + "epoch": 0.13720789724999513, + "grad_norm": 17.023181915283203, + "learning_rate": 4.573210341691568e-07, + "loss": 1.0647, + "step": 3520 + }, + { + "epoch": 0.13759769241263717, + "grad_norm": 17.726045608520508, + "learning_rate": 4.5862024165259186e-07, + "loss": 1.0262, + "step": 3530 + }, + { + "epoch": 0.13798748757527918, + "grad_norm": 17.833837509155273, + "learning_rate": 4.59919449136027e-07, + "loss": 1.0871, + "step": 3540 + }, + { + "epoch": 0.13837728273792121, + "grad_norm": 20.43224334716797, + "learning_rate": 4.6121865661946213e-07, + "loss": 1.0729, + "step": 3550 + }, + { + "epoch": 0.13876707790056325, + "grad_norm": 16.169464111328125, + "learning_rate": 4.625178641028972e-07, + "loss": 1.0372, + "step": 3560 + }, + { + "epoch": 0.1391568730632053, + "grad_norm": 15.792376518249512, + "learning_rate": 4.638170715863323e-07, + "loss": 1.0882, + "step": 3570 + }, + { + "epoch": 0.13954666822584733, + "grad_norm": 16.077312469482422, + "learning_rate": 4.6511627906976743e-07, + "loss": 1.0413, + "step": 3580 + }, + { + "epoch": 0.13993646338848936, + "grad_norm": 14.990835189819336, + "learning_rate": 4.664154865532025e-07, + "loss": 0.9952, + "step": 3590 + }, + { + "epoch": 0.14032625855113137, + "grad_norm": 19.882553100585938, + "learning_rate": 4.677146940366376e-07, + "loss": 1.0156, + "step": 3600 + }, + { + "epoch": 0.1407160537137734, + "grad_norm": 17.026609420776367, + "learning_rate": 4.6901390152007273e-07, + "loss": 1.0349, + "step": 3610 + }, + { + "epoch": 0.14110584887641545, + "grad_norm": 22.95958137512207, + "learning_rate": 4.7031310900350786e-07, + "loss": 1.1261, + "step": 3620 + }, + { + "epoch": 0.14149564403905748, + "grad_norm": 29.056495666503906, + "learning_rate": 4.7161231648694295e-07, + "loss": 1.0194, + "step": 3630 + }, + { + "epoch": 0.14188543920169952, + "grad_norm": 23.535367965698242, + "learning_rate": 4.7291152397037803e-07, + "loss": 1.0546, + "step": 3640 + }, + { + "epoch": 0.14227523436434153, + "grad_norm": 20.27889060974121, + "learning_rate": 4.742107314538131e-07, + "loss": 1.0087, + "step": 3650 + }, + { + "epoch": 0.14266502952698357, + "grad_norm": 15.841670036315918, + "learning_rate": 4.755099389372483e-07, + "loss": 1.0677, + "step": 3660 + }, + { + "epoch": 0.1430548246896256, + "grad_norm": 23.247512817382812, + "learning_rate": 4.768091464206834e-07, + "loss": 1.1445, + "step": 3670 + }, + { + "epoch": 0.14344461985226764, + "grad_norm": 18.8807373046875, + "learning_rate": 4.781083539041185e-07, + "loss": 1.0909, + "step": 3680 + }, + { + "epoch": 0.14383441501490968, + "grad_norm": 15.810229301452637, + "learning_rate": 4.794075613875536e-07, + "loss": 0.9575, + "step": 3690 + }, + { + "epoch": 0.1442242101775517, + "grad_norm": 15.30854606628418, + "learning_rate": 4.807067688709886e-07, + "loss": 1.0425, + "step": 3700 + }, + { + "epoch": 0.14461400534019372, + "grad_norm": 17.152070999145508, + "learning_rate": 4.820059763544238e-07, + "loss": 1.0133, + "step": 3710 + }, + { + "epoch": 0.14500380050283576, + "grad_norm": 16.34093475341797, + "learning_rate": 4.833051838378589e-07, + "loss": 1.0039, + "step": 3720 + }, + { + "epoch": 0.1453935956654778, + "grad_norm": 16.455291748046875, + "learning_rate": 4.84604391321294e-07, + "loss": 1.0895, + "step": 3730 + }, + { + "epoch": 0.14578339082811984, + "grad_norm": 16.758512496948242, + "learning_rate": 4.859035988047291e-07, + "loss": 1.0855, + "step": 3740 + }, + { + "epoch": 0.14617318599076184, + "grad_norm": 17.82316017150879, + "learning_rate": 4.872028062881642e-07, + "loss": 0.9641, + "step": 3750 + }, + { + "epoch": 0.14656298115340388, + "grad_norm": 17.550748825073242, + "learning_rate": 4.885020137715993e-07, + "loss": 1.06, + "step": 3760 + }, + { + "epoch": 0.14695277631604592, + "grad_norm": 14.845223426818848, + "learning_rate": 4.898012212550345e-07, + "loss": 1.0204, + "step": 3770 + }, + { + "epoch": 0.14734257147868796, + "grad_norm": 16.990131378173828, + "learning_rate": 4.911004287384695e-07, + "loss": 1.0019, + "step": 3780 + }, + { + "epoch": 0.14773236664133, + "grad_norm": 15.07215404510498, + "learning_rate": 4.923996362219046e-07, + "loss": 1.0933, + "step": 3790 + }, + { + "epoch": 0.148122161803972, + "grad_norm": 18.679025650024414, + "learning_rate": 4.936988437053398e-07, + "loss": 1.0699, + "step": 3800 + }, + { + "epoch": 0.14851195696661404, + "grad_norm": 17.40754508972168, + "learning_rate": 4.949980511887748e-07, + "loss": 1.0267, + "step": 3810 + }, + { + "epoch": 0.14890175212925608, + "grad_norm": 18.919660568237305, + "learning_rate": 4.962972586722099e-07, + "loss": 1.1122, + "step": 3820 + }, + { + "epoch": 0.1492915472918981, + "grad_norm": 26.44634437561035, + "learning_rate": 4.975964661556451e-07, + "loss": 1.0921, + "step": 3830 + }, + { + "epoch": 0.14968134245454015, + "grad_norm": 14.579507827758789, + "learning_rate": 4.988956736390801e-07, + "loss": 0.987, + "step": 3840 + }, + { + "epoch": 0.15007113761718216, + "grad_norm": 17.81438446044922, + "learning_rate": 5.001948811225153e-07, + "loss": 1.0686, + "step": 3850 + }, + { + "epoch": 0.1504609327798242, + "grad_norm": 15.670595169067383, + "learning_rate": 5.014940886059504e-07, + "loss": 1.0943, + "step": 3860 + }, + { + "epoch": 0.15085072794246623, + "grad_norm": 16.167034149169922, + "learning_rate": 5.027932960893855e-07, + "loss": 1.0964, + "step": 3870 + }, + { + "epoch": 0.15124052310510827, + "grad_norm": 16.154050827026367, + "learning_rate": 5.040925035728205e-07, + "loss": 0.9993, + "step": 3880 + }, + { + "epoch": 0.1516303182677503, + "grad_norm": 17.288970947265625, + "learning_rate": 5.053917110562557e-07, + "loss": 1.0441, + "step": 3890 + }, + { + "epoch": 0.15202011343039232, + "grad_norm": 17.43962860107422, + "learning_rate": 5.066909185396908e-07, + "loss": 1.0818, + "step": 3900 + }, + { + "epoch": 0.15240990859303435, + "grad_norm": 14.26394271850586, + "learning_rate": 5.079901260231258e-07, + "loss": 0.9985, + "step": 3910 + }, + { + "epoch": 0.1527997037556764, + "grad_norm": 15.66629695892334, + "learning_rate": 5.09289333506561e-07, + "loss": 1.0944, + "step": 3920 + }, + { + "epoch": 0.15318949891831843, + "grad_norm": 15.555120468139648, + "learning_rate": 5.105885409899961e-07, + "loss": 0.9835, + "step": 3930 + }, + { + "epoch": 0.15357929408096047, + "grad_norm": 14.272843360900879, + "learning_rate": 5.118877484734312e-07, + "loss": 1.0597, + "step": 3940 + }, + { + "epoch": 0.15396908924360247, + "grad_norm": 22.521039962768555, + "learning_rate": 5.131869559568664e-07, + "loss": 1.0765, + "step": 3950 + }, + { + "epoch": 0.1543588844062445, + "grad_norm": 17.628700256347656, + "learning_rate": 5.144861634403014e-07, + "loss": 1.0502, + "step": 3960 + }, + { + "epoch": 0.15474867956888655, + "grad_norm": 18.74199867248535, + "learning_rate": 5.157853709237365e-07, + "loss": 1.0416, + "step": 3970 + }, + { + "epoch": 0.1551384747315286, + "grad_norm": 18.47379493713379, + "learning_rate": 5.170845784071715e-07, + "loss": 1.0052, + "step": 3980 + }, + { + "epoch": 0.15552826989417062, + "grad_norm": 22.22235870361328, + "learning_rate": 5.183837858906067e-07, + "loss": 1.0453, + "step": 3990 + }, + { + "epoch": 0.15591806505681263, + "grad_norm": 16.76043128967285, + "learning_rate": 5.196829933740417e-07, + "loss": 0.9797, + "step": 4000 + }, + { + "epoch": 0.15591806505681263, + "eval_loss": 1.0561199188232422, + "eval_runtime": 85.1508, + "eval_samples_per_second": 48.702, + "eval_steps_per_second": 6.095, + "step": 4000 + }, + { + "epoch": 0.15630786021945467, + "grad_norm": 16.90920639038086, + "learning_rate": 5.20982200857477e-07, + "loss": 1.0616, + "step": 4010 + }, + { + "epoch": 0.1566976553820967, + "grad_norm": 14.449828147888184, + "learning_rate": 5.222814083409121e-07, + "loss": 0.9975, + "step": 4020 + }, + { + "epoch": 0.15708745054473874, + "grad_norm": 19.242719650268555, + "learning_rate": 5.235806158243471e-07, + "loss": 0.979, + "step": 4030 + }, + { + "epoch": 0.15747724570738078, + "grad_norm": 17.882335662841797, + "learning_rate": 5.248798233077823e-07, + "loss": 1.2068, + "step": 4040 + }, + { + "epoch": 0.1578670408700228, + "grad_norm": 15.84154987335205, + "learning_rate": 5.261790307912173e-07, + "loss": 0.9958, + "step": 4050 + }, + { + "epoch": 0.15825683603266483, + "grad_norm": 17.563220977783203, + "learning_rate": 5.274782382746524e-07, + "loss": 0.9985, + "step": 4060 + }, + { + "epoch": 0.15864663119530686, + "grad_norm": 17.082080841064453, + "learning_rate": 5.287774457580875e-07, + "loss": 1.0012, + "step": 4070 + }, + { + "epoch": 0.1590364263579489, + "grad_norm": 15.663912773132324, + "learning_rate": 5.300766532415226e-07, + "loss": 1.0352, + "step": 4080 + }, + { + "epoch": 0.15942622152059094, + "grad_norm": 15.968063354492188, + "learning_rate": 5.313758607249578e-07, + "loss": 1.0592, + "step": 4090 + }, + { + "epoch": 0.15981601668323295, + "grad_norm": 16.00457000732422, + "learning_rate": 5.326750682083928e-07, + "loss": 0.9997, + "step": 4100 + }, + { + "epoch": 0.16020581184587498, + "grad_norm": 15.83539867401123, + "learning_rate": 5.33974275691828e-07, + "loss": 1.0632, + "step": 4110 + }, + { + "epoch": 0.16059560700851702, + "grad_norm": 14.713849067687988, + "learning_rate": 5.352734831752631e-07, + "loss": 1.0339, + "step": 4120 + }, + { + "epoch": 0.16098540217115906, + "grad_norm": 15.25914192199707, + "learning_rate": 5.365726906586981e-07, + "loss": 1.0279, + "step": 4130 + }, + { + "epoch": 0.1613751973338011, + "grad_norm": 15.835671424865723, + "learning_rate": 5.378718981421333e-07, + "loss": 1.0926, + "step": 4140 + }, + { + "epoch": 0.16176499249644313, + "grad_norm": 17.077091217041016, + "learning_rate": 5.391711056255683e-07, + "loss": 1.0802, + "step": 4150 + }, + { + "epoch": 0.16215478765908514, + "grad_norm": 15.163534164428711, + "learning_rate": 5.404703131090034e-07, + "loss": 1.0544, + "step": 4160 + }, + { + "epoch": 0.16254458282172718, + "grad_norm": 14.15582275390625, + "learning_rate": 5.417695205924387e-07, + "loss": 0.9581, + "step": 4170 + }, + { + "epoch": 0.16293437798436922, + "grad_norm": 14.383868217468262, + "learning_rate": 5.430687280758737e-07, + "loss": 1.1259, + "step": 4180 + }, + { + "epoch": 0.16332417314701125, + "grad_norm": 18.309288024902344, + "learning_rate": 5.443679355593088e-07, + "loss": 1.0352, + "step": 4190 + }, + { + "epoch": 0.1637139683096533, + "grad_norm": 15.44610595703125, + "learning_rate": 5.456671430427439e-07, + "loss": 1.0149, + "step": 4200 + }, + { + "epoch": 0.1641037634722953, + "grad_norm": 15.321208000183105, + "learning_rate": 5.46966350526179e-07, + "loss": 1.0534, + "step": 4210 + }, + { + "epoch": 0.16449355863493734, + "grad_norm": 14.682863235473633, + "learning_rate": 5.482655580096141e-07, + "loss": 1.1186, + "step": 4220 + }, + { + "epoch": 0.16488335379757937, + "grad_norm": 14.512438774108887, + "learning_rate": 5.495647654930492e-07, + "loss": 1.0367, + "step": 4230 + }, + { + "epoch": 0.1652731489602214, + "grad_norm": 19.255491256713867, + "learning_rate": 5.508639729764843e-07, + "loss": 1.0215, + "step": 4240 + }, + { + "epoch": 0.16566294412286345, + "grad_norm": 15.671401977539062, + "learning_rate": 5.521631804599193e-07, + "loss": 1.0818, + "step": 4250 + }, + { + "epoch": 0.16605273928550546, + "grad_norm": 14.380105018615723, + "learning_rate": 5.534623879433546e-07, + "loss": 1.0297, + "step": 4260 + }, + { + "epoch": 0.1664425344481475, + "grad_norm": 19.247676849365234, + "learning_rate": 5.547615954267897e-07, + "loss": 0.9842, + "step": 4270 + }, + { + "epoch": 0.16683232961078953, + "grad_norm": 14.556777954101562, + "learning_rate": 5.560608029102247e-07, + "loss": 1.0127, + "step": 4280 + }, + { + "epoch": 0.16722212477343157, + "grad_norm": 13.98294734954834, + "learning_rate": 5.573600103936599e-07, + "loss": 1.0543, + "step": 4290 + }, + { + "epoch": 0.1676119199360736, + "grad_norm": 15.801843643188477, + "learning_rate": 5.586592178770949e-07, + "loss": 1.0557, + "step": 4300 + }, + { + "epoch": 0.16800171509871561, + "grad_norm": 18.221302032470703, + "learning_rate": 5.5995842536053e-07, + "loss": 1.0671, + "step": 4310 + }, + { + "epoch": 0.16839151026135765, + "grad_norm": 17.9721736907959, + "learning_rate": 5.612576328439651e-07, + "loss": 1.0566, + "step": 4320 + }, + { + "epoch": 0.1687813054239997, + "grad_norm": 16.8494815826416, + "learning_rate": 5.625568403274002e-07, + "loss": 0.9981, + "step": 4330 + }, + { + "epoch": 0.16917110058664173, + "grad_norm": 15.758248329162598, + "learning_rate": 5.638560478108354e-07, + "loss": 1.0613, + "step": 4340 + }, + { + "epoch": 0.16956089574928376, + "grad_norm": 16.009477615356445, + "learning_rate": 5.651552552942705e-07, + "loss": 1.0302, + "step": 4350 + }, + { + "epoch": 0.16995069091192577, + "grad_norm": 18.980972290039062, + "learning_rate": 5.664544627777056e-07, + "loss": 1.0376, + "step": 4360 + }, + { + "epoch": 0.1703404860745678, + "grad_norm": 14.341987609863281, + "learning_rate": 5.677536702611406e-07, + "loss": 1.0191, + "step": 4370 + }, + { + "epoch": 0.17073028123720985, + "grad_norm": 16.15073013305664, + "learning_rate": 5.690528777445758e-07, + "loss": 1.0183, + "step": 4380 + }, + { + "epoch": 0.17112007639985188, + "grad_norm": 19.7049617767334, + "learning_rate": 5.703520852280109e-07, + "loss": 0.9588, + "step": 4390 + }, + { + "epoch": 0.17150987156249392, + "grad_norm": 14.053573608398438, + "learning_rate": 5.716512927114459e-07, + "loss": 1.0672, + "step": 4400 + }, + { + "epoch": 0.17189966672513593, + "grad_norm": 15.41024398803711, + "learning_rate": 5.729505001948811e-07, + "loss": 1.0104, + "step": 4410 + }, + { + "epoch": 0.17228946188777797, + "grad_norm": 14.522614479064941, + "learning_rate": 5.742497076783161e-07, + "loss": 1.0515, + "step": 4420 + }, + { + "epoch": 0.17267925705042, + "grad_norm": 15.333739280700684, + "learning_rate": 5.755489151617513e-07, + "loss": 1.0073, + "step": 4430 + }, + { + "epoch": 0.17306905221306204, + "grad_norm": 18.377424240112305, + "learning_rate": 5.768481226451865e-07, + "loss": 1.0678, + "step": 4440 + }, + { + "epoch": 0.17345884737570408, + "grad_norm": 14.466423034667969, + "learning_rate": 5.781473301286215e-07, + "loss": 1.0253, + "step": 4450 + }, + { + "epoch": 0.1738486425383461, + "grad_norm": 14.178470611572266, + "learning_rate": 5.794465376120566e-07, + "loss": 1.0693, + "step": 4460 + }, + { + "epoch": 0.17423843770098812, + "grad_norm": 13.55081844329834, + "learning_rate": 5.807457450954917e-07, + "loss": 1.026, + "step": 4470 + }, + { + "epoch": 0.17462823286363016, + "grad_norm": 14.8812255859375, + "learning_rate": 5.820449525789268e-07, + "loss": 1.0418, + "step": 4480 + }, + { + "epoch": 0.1750180280262722, + "grad_norm": 16.747859954833984, + "learning_rate": 5.833441600623619e-07, + "loss": 1.1218, + "step": 4490 + }, + { + "epoch": 0.17540782318891424, + "grad_norm": 19.470033645629883, + "learning_rate": 5.84643367545797e-07, + "loss": 0.9791, + "step": 4500 + }, + { + "epoch": 0.17540782318891424, + "eval_loss": 1.0384502410888672, + "eval_runtime": 83.1416, + "eval_samples_per_second": 49.879, + "eval_steps_per_second": 6.242, + "step": 4500 + }, + { + "epoch": 0.17579761835155625, + "grad_norm": 14.845098495483398, + "learning_rate": 5.859425750292322e-07, + "loss": 1.0336, + "step": 4510 + }, + { + "epoch": 0.17618741351419828, + "grad_norm": 13.312830924987793, + "learning_rate": 5.872417825126672e-07, + "loss": 1.1411, + "step": 4520 + }, + { + "epoch": 0.17657720867684032, + "grad_norm": 15.771671295166016, + "learning_rate": 5.885409899961024e-07, + "loss": 0.9958, + "step": 4530 + }, + { + "epoch": 0.17696700383948236, + "grad_norm": 15.051501274108887, + "learning_rate": 5.898401974795375e-07, + "loss": 1.1421, + "step": 4540 + }, + { + "epoch": 0.1773567990021244, + "grad_norm": 14.614441871643066, + "learning_rate": 5.911394049629725e-07, + "loss": 1.0324, + "step": 4550 + }, + { + "epoch": 0.1777465941647664, + "grad_norm": 21.041921615600586, + "learning_rate": 5.924386124464077e-07, + "loss": 1.0175, + "step": 4560 + }, + { + "epoch": 0.17813638932740844, + "grad_norm": 13.824596405029297, + "learning_rate": 5.937378199298427e-07, + "loss": 0.9563, + "step": 4570 + }, + { + "epoch": 0.17852618449005048, + "grad_norm": 16.64824104309082, + "learning_rate": 5.950370274132778e-07, + "loss": 1.0088, + "step": 4580 + }, + { + "epoch": 0.17891597965269251, + "grad_norm": 15.666703224182129, + "learning_rate": 5.963362348967131e-07, + "loss": 1.0211, + "step": 4590 + }, + { + "epoch": 0.17930577481533455, + "grad_norm": 14.569836616516113, + "learning_rate": 5.976354423801481e-07, + "loss": 1.0049, + "step": 4600 + }, + { + "epoch": 0.17969556997797656, + "grad_norm": 15.689733505249023, + "learning_rate": 5.989346498635832e-07, + "loss": 1.0271, + "step": 4610 + }, + { + "epoch": 0.1800853651406186, + "grad_norm": 16.67568588256836, + "learning_rate": 6.002338573470183e-07, + "loss": 1.0524, + "step": 4620 + }, + { + "epoch": 0.18047516030326063, + "grad_norm": 14.249446868896484, + "learning_rate": 6.015330648304534e-07, + "loss": 1.0627, + "step": 4630 + }, + { + "epoch": 0.18086495546590267, + "grad_norm": 15.069948196411133, + "learning_rate": 6.028322723138884e-07, + "loss": 1.0554, + "step": 4640 + }, + { + "epoch": 0.1812547506285447, + "grad_norm": 13.711263656616211, + "learning_rate": 6.041314797973236e-07, + "loss": 1.0597, + "step": 4650 + }, + { + "epoch": 0.18164454579118672, + "grad_norm": 15.280440330505371, + "learning_rate": 6.054306872807587e-07, + "loss": 1.0475, + "step": 4660 + }, + { + "epoch": 0.18203434095382875, + "grad_norm": 15.579916954040527, + "learning_rate": 6.067298947641938e-07, + "loss": 1.041, + "step": 4670 + }, + { + "epoch": 0.1824241361164708, + "grad_norm": 15.802238464355469, + "learning_rate": 6.08029102247629e-07, + "loss": 1.0761, + "step": 4680 + }, + { + "epoch": 0.18281393127911283, + "grad_norm": 16.816139221191406, + "learning_rate": 6.09328309731064e-07, + "loss": 1.0202, + "step": 4690 + }, + { + "epoch": 0.18320372644175487, + "grad_norm": 17.639631271362305, + "learning_rate": 6.106275172144991e-07, + "loss": 1.0208, + "step": 4700 + }, + { + "epoch": 0.1835935216043969, + "grad_norm": 13.672350883483887, + "learning_rate": 6.119267246979343e-07, + "loss": 0.9666, + "step": 4710 + }, + { + "epoch": 0.1839833167670389, + "grad_norm": 16.99921989440918, + "learning_rate": 6.132259321813693e-07, + "loss": 0.9716, + "step": 4720 + }, + { + "epoch": 0.18437311192968095, + "grad_norm": 15.034951210021973, + "learning_rate": 6.145251396648044e-07, + "loss": 1.0147, + "step": 4730 + }, + { + "epoch": 0.184762907092323, + "grad_norm": 18.509357452392578, + "learning_rate": 6.158243471482395e-07, + "loss": 0.9706, + "step": 4740 + }, + { + "epoch": 0.18515270225496502, + "grad_norm": 17.403804779052734, + "learning_rate": 6.171235546316747e-07, + "loss": 1.0508, + "step": 4750 + }, + { + "epoch": 0.18554249741760706, + "grad_norm": 16.30280303955078, + "learning_rate": 6.184227621151098e-07, + "loss": 1.0176, + "step": 4760 + }, + { + "epoch": 0.18593229258024907, + "grad_norm": 14.720854759216309, + "learning_rate": 6.197219695985449e-07, + "loss": 1.0495, + "step": 4770 + }, + { + "epoch": 0.1863220877428911, + "grad_norm": 15.4736909866333, + "learning_rate": 6.2102117708198e-07, + "loss": 1.0452, + "step": 4780 + }, + { + "epoch": 0.18671188290553314, + "grad_norm": 15.650922775268555, + "learning_rate": 6.22320384565415e-07, + "loss": 1.0479, + "step": 4790 + }, + { + "epoch": 0.18710167806817518, + "grad_norm": 13.406271934509277, + "learning_rate": 6.236195920488502e-07, + "loss": 1.0076, + "step": 4800 + }, + { + "epoch": 0.18749147323081722, + "grad_norm": 14.436541557312012, + "learning_rate": 6.249187995322853e-07, + "loss": 1.07, + "step": 4810 + }, + { + "epoch": 0.18788126839345923, + "grad_norm": 14.940513610839844, + "learning_rate": 6.262180070157203e-07, + "loss": 0.9739, + "step": 4820 + }, + { + "epoch": 0.18827106355610126, + "grad_norm": 16.77115249633789, + "learning_rate": 6.275172144991555e-07, + "loss": 1.0688, + "step": 4830 + }, + { + "epoch": 0.1886608587187433, + "grad_norm": 13.453080177307129, + "learning_rate": 6.288164219825906e-07, + "loss": 0.9579, + "step": 4840 + }, + { + "epoch": 0.18905065388138534, + "grad_norm": 16.892372131347656, + "learning_rate": 6.301156294660257e-07, + "loss": 1.0576, + "step": 4850 + }, + { + "epoch": 0.18944044904402738, + "grad_norm": 15.69309139251709, + "learning_rate": 6.314148369494609e-07, + "loss": 1.0117, + "step": 4860 + }, + { + "epoch": 0.18983024420666939, + "grad_norm": 17.763710021972656, + "learning_rate": 6.327140444328959e-07, + "loss": 1.0332, + "step": 4870 + }, + { + "epoch": 0.19022003936931142, + "grad_norm": 19.213150024414062, + "learning_rate": 6.34013251916331e-07, + "loss": 0.8974, + "step": 4880 + }, + { + "epoch": 0.19060983453195346, + "grad_norm": 16.0093936920166, + "learning_rate": 6.353124593997661e-07, + "loss": 1.0636, + "step": 4890 + }, + { + "epoch": 0.1909996296945955, + "grad_norm": 16.731304168701172, + "learning_rate": 6.366116668832012e-07, + "loss": 0.9851, + "step": 4900 + }, + { + "epoch": 0.19138942485723753, + "grad_norm": 14.75139045715332, + "learning_rate": 6.379108743666363e-07, + "loss": 0.9696, + "step": 4910 + }, + { + "epoch": 0.19177922001987954, + "grad_norm": 13.83337688446045, + "learning_rate": 6.392100818500715e-07, + "loss": 0.9957, + "step": 4920 + }, + { + "epoch": 0.19216901518252158, + "grad_norm": 15.691692352294922, + "learning_rate": 6.405092893335066e-07, + "loss": 1.0036, + "step": 4930 + }, + { + "epoch": 0.19255881034516362, + "grad_norm": 17.085634231567383, + "learning_rate": 6.418084968169416e-07, + "loss": 1.0462, + "step": 4940 + }, + { + "epoch": 0.19294860550780565, + "grad_norm": 15.70003604888916, + "learning_rate": 6.431077043003768e-07, + "loss": 1.0385, + "step": 4950 + }, + { + "epoch": 0.1933384006704477, + "grad_norm": 17.151500701904297, + "learning_rate": 6.444069117838118e-07, + "loss": 1.079, + "step": 4960 + }, + { + "epoch": 0.1937281958330897, + "grad_norm": 15.488120079040527, + "learning_rate": 6.457061192672469e-07, + "loss": 0.9941, + "step": 4970 + }, + { + "epoch": 0.19411799099573174, + "grad_norm": 16.171037673950195, + "learning_rate": 6.470053267506821e-07, + "loss": 1.0492, + "step": 4980 + }, + { + "epoch": 0.19450778615837377, + "grad_norm": 16.528167724609375, + "learning_rate": 6.483045342341171e-07, + "loss": 1.0377, + "step": 4990 + }, + { + "epoch": 0.1948975813210158, + "grad_norm": 13.529744148254395, + "learning_rate": 6.496037417175523e-07, + "loss": 1.0789, + "step": 5000 + }, + { + "epoch": 0.1948975813210158, + "eval_loss": 1.0238083600997925, + "eval_runtime": 83.059, + "eval_samples_per_second": 49.928, + "eval_steps_per_second": 6.249, + "step": 5000 + }, + { + "epoch": 0.19528737648365785, + "grad_norm": 15.47061538696289, + "learning_rate": 6.509029492009874e-07, + "loss": 0.9704, + "step": 5010 + }, + { + "epoch": 0.19567717164629986, + "grad_norm": 14.709036827087402, + "learning_rate": 6.522021566844225e-07, + "loss": 0.9886, + "step": 5020 + }, + { + "epoch": 0.1960669668089419, + "grad_norm": 19.647789001464844, + "learning_rate": 6.535013641678576e-07, + "loss": 1.0685, + "step": 5030 + }, + { + "epoch": 0.19645676197158393, + "grad_norm": 14.017085075378418, + "learning_rate": 6.548005716512927e-07, + "loss": 1.0543, + "step": 5040 + }, + { + "epoch": 0.19684655713422597, + "grad_norm": 15.966981887817383, + "learning_rate": 6.560997791347278e-07, + "loss": 0.9925, + "step": 5050 + }, + { + "epoch": 0.197236352296868, + "grad_norm": 15.69174861907959, + "learning_rate": 6.573989866181628e-07, + "loss": 1.0293, + "step": 5060 + }, + { + "epoch": 0.19762614745951002, + "grad_norm": 17.261995315551758, + "learning_rate": 6.58698194101598e-07, + "loss": 0.9999, + "step": 5070 + }, + { + "epoch": 0.19801594262215205, + "grad_norm": 17.29596710205078, + "learning_rate": 6.599974015850332e-07, + "loss": 0.9946, + "step": 5080 + }, + { + "epoch": 0.1984057377847941, + "grad_norm": 15.276107788085938, + "learning_rate": 6.612966090684682e-07, + "loss": 0.996, + "step": 5090 + }, + { + "epoch": 0.19879553294743613, + "grad_norm": 14.571300506591797, + "learning_rate": 6.625958165519034e-07, + "loss": 1.0553, + "step": 5100 + }, + { + "epoch": 0.19918532811007816, + "grad_norm": 14.274317741394043, + "learning_rate": 6.638950240353384e-07, + "loss": 0.9965, + "step": 5110 + }, + { + "epoch": 0.19957512327272017, + "grad_norm": 14.720294952392578, + "learning_rate": 6.651942315187735e-07, + "loss": 1.0313, + "step": 5120 + }, + { + "epoch": 0.1999649184353622, + "grad_norm": 13.559709548950195, + "learning_rate": 6.664934390022087e-07, + "loss": 0.9289, + "step": 5130 + }, + { + "epoch": 0.20035471359800425, + "grad_norm": 15.75797176361084, + "learning_rate": 6.677926464856437e-07, + "loss": 0.9778, + "step": 5140 + }, + { + "epoch": 0.20074450876064628, + "grad_norm": 13.492440223693848, + "learning_rate": 6.690918539690788e-07, + "loss": 0.9688, + "step": 5150 + }, + { + "epoch": 0.20113430392328832, + "grad_norm": 14.886689186096191, + "learning_rate": 6.703910614525138e-07, + "loss": 1.0263, + "step": 5160 + }, + { + "epoch": 0.20152409908593033, + "grad_norm": 14.970948219299316, + "learning_rate": 6.716902689359491e-07, + "loss": 1.0519, + "step": 5170 + }, + { + "epoch": 0.20191389424857237, + "grad_norm": 16.394268035888672, + "learning_rate": 6.729894764193842e-07, + "loss": 1.0284, + "step": 5180 + }, + { + "epoch": 0.2023036894112144, + "grad_norm": 15.220818519592285, + "learning_rate": 6.742886839028193e-07, + "loss": 1.1016, + "step": 5190 + }, + { + "epoch": 0.20269348457385644, + "grad_norm": 14.298688888549805, + "learning_rate": 6.755878913862544e-07, + "loss": 1.0636, + "step": 5200 + }, + { + "epoch": 0.20308327973649848, + "grad_norm": 17.60258674621582, + "learning_rate": 6.768870988696894e-07, + "loss": 1.0393, + "step": 5210 + }, + { + "epoch": 0.2034730748991405, + "grad_norm": 13.860746383666992, + "learning_rate": 6.781863063531245e-07, + "loss": 1.0051, + "step": 5220 + }, + { + "epoch": 0.20386287006178253, + "grad_norm": 16.627473831176758, + "learning_rate": 6.794855138365597e-07, + "loss": 0.981, + "step": 5230 + }, + { + "epoch": 0.20425266522442456, + "grad_norm": 19.569072723388672, + "learning_rate": 6.807847213199947e-07, + "loss": 1.0434, + "step": 5240 + }, + { + "epoch": 0.2046424603870666, + "grad_norm": 13.898557662963867, + "learning_rate": 6.8208392880343e-07, + "loss": 0.9202, + "step": 5250 + }, + { + "epoch": 0.20503225554970864, + "grad_norm": 13.643329620361328, + "learning_rate": 6.83383136286865e-07, + "loss": 1.0147, + "step": 5260 + }, + { + "epoch": 0.20542205071235067, + "grad_norm": 14.272087097167969, + "learning_rate": 6.846823437703001e-07, + "loss": 1.0325, + "step": 5270 + }, + { + "epoch": 0.20581184587499268, + "grad_norm": 13.641026496887207, + "learning_rate": 6.859815512537351e-07, + "loss": 1.0749, + "step": 5280 + }, + { + "epoch": 0.20620164103763472, + "grad_norm": 16.963171005249023, + "learning_rate": 6.872807587371703e-07, + "loss": 1.0102, + "step": 5290 + }, + { + "epoch": 0.20659143620027676, + "grad_norm": 13.869208335876465, + "learning_rate": 6.885799662206054e-07, + "loss": 0.9824, + "step": 5300 + }, + { + "epoch": 0.2069812313629188, + "grad_norm": 13.419792175292969, + "learning_rate": 6.898791737040404e-07, + "loss": 0.9711, + "step": 5310 + }, + { + "epoch": 0.20737102652556083, + "grad_norm": 21.119474411010742, + "learning_rate": 6.911783811874756e-07, + "loss": 1.0123, + "step": 5320 + }, + { + "epoch": 0.20776082168820284, + "grad_norm": 17.161096572875977, + "learning_rate": 6.924775886709107e-07, + "loss": 1.0465, + "step": 5330 + }, + { + "epoch": 0.20815061685084488, + "grad_norm": 16.594682693481445, + "learning_rate": 6.937767961543458e-07, + "loss": 1.0297, + "step": 5340 + }, + { + "epoch": 0.20854041201348691, + "grad_norm": 14.877861976623535, + "learning_rate": 6.95076003637781e-07, + "loss": 0.9755, + "step": 5350 + }, + { + "epoch": 0.20893020717612895, + "grad_norm": 13.702380180358887, + "learning_rate": 6.96375211121216e-07, + "loss": 1.0219, + "step": 5360 + }, + { + "epoch": 0.209320002338771, + "grad_norm": 15.334413528442383, + "learning_rate": 6.976744186046511e-07, + "loss": 1.0264, + "step": 5370 + }, + { + "epoch": 0.209709797501413, + "grad_norm": 14.141037940979004, + "learning_rate": 6.989736260880862e-07, + "loss": 0.9832, + "step": 5380 + }, + { + "epoch": 0.21009959266405503, + "grad_norm": 14.459108352661133, + "learning_rate": 7.002728335715213e-07, + "loss": 1.0234, + "step": 5390 + }, + { + "epoch": 0.21048938782669707, + "grad_norm": 15.051697731018066, + "learning_rate": 7.015720410549564e-07, + "loss": 0.9525, + "step": 5400 + }, + { + "epoch": 0.2108791829893391, + "grad_norm": 14.221467018127441, + "learning_rate": 7.028712485383915e-07, + "loss": 1.0288, + "step": 5410 + }, + { + "epoch": 0.21126897815198115, + "grad_norm": 19.185102462768555, + "learning_rate": 7.041704560218267e-07, + "loss": 0.973, + "step": 5420 + }, + { + "epoch": 0.21165877331462316, + "grad_norm": 14.803804397583008, + "learning_rate": 7.054696635052617e-07, + "loss": 0.9439, + "step": 5430 + }, + { + "epoch": 0.2120485684772652, + "grad_norm": 14.462215423583984, + "learning_rate": 7.067688709886969e-07, + "loss": 1.0284, + "step": 5440 + }, + { + "epoch": 0.21243836363990723, + "grad_norm": 14.7482328414917, + "learning_rate": 7.08068078472132e-07, + "loss": 1.005, + "step": 5450 + }, + { + "epoch": 0.21282815880254927, + "grad_norm": 15.092142105102539, + "learning_rate": 7.09367285955567e-07, + "loss": 1.0008, + "step": 5460 + }, + { + "epoch": 0.2132179539651913, + "grad_norm": 16.304113388061523, + "learning_rate": 7.106664934390022e-07, + "loss": 1.007, + "step": 5470 + }, + { + "epoch": 0.2136077491278333, + "grad_norm": 14.88569164276123, + "learning_rate": 7.119657009224372e-07, + "loss": 1.039, + "step": 5480 + }, + { + "epoch": 0.21399754429047535, + "grad_norm": 15.809123992919922, + "learning_rate": 7.132649084058723e-07, + "loss": 1.0021, + "step": 5490 + }, + { + "epoch": 0.2143873394531174, + "grad_norm": 16.1784610748291, + "learning_rate": 7.145641158893076e-07, + "loss": 0.9994, + "step": 5500 + }, + { + "epoch": 0.2143873394531174, + "eval_loss": 1.0179262161254883, + "eval_runtime": 84.4738, + "eval_samples_per_second": 49.092, + "eval_steps_per_second": 6.144, + "step": 5500 + }, + { + "epoch": 0.21477713461575942, + "grad_norm": 16.729164123535156, + "learning_rate": 7.158633233727426e-07, + "loss": 0.99, + "step": 5510 + }, + { + "epoch": 0.21516692977840146, + "grad_norm": 15.680314064025879, + "learning_rate": 7.171625308561777e-07, + "loss": 1.0946, + "step": 5520 + }, + { + "epoch": 0.21555672494104347, + "grad_norm": 13.532752990722656, + "learning_rate": 7.184617383396128e-07, + "loss": 0.9513, + "step": 5530 + }, + { + "epoch": 0.2159465201036855, + "grad_norm": 15.87139892578125, + "learning_rate": 7.197609458230479e-07, + "loss": 1.0091, + "step": 5540 + }, + { + "epoch": 0.21633631526632754, + "grad_norm": 14.46761417388916, + "learning_rate": 7.21060153306483e-07, + "loss": 0.9961, + "step": 5550 + }, + { + "epoch": 0.21672611042896958, + "grad_norm": 15.667901992797852, + "learning_rate": 7.223593607899181e-07, + "loss": 1.0183, + "step": 5560 + }, + { + "epoch": 0.21711590559161162, + "grad_norm": 15.079389572143555, + "learning_rate": 7.236585682733532e-07, + "loss": 1.0497, + "step": 5570 + }, + { + "epoch": 0.21750570075425363, + "grad_norm": 15.953119277954102, + "learning_rate": 7.249577757567883e-07, + "loss": 0.9893, + "step": 5580 + }, + { + "epoch": 0.21789549591689567, + "grad_norm": 16.65497398376465, + "learning_rate": 7.262569832402235e-07, + "loss": 0.9465, + "step": 5590 + }, + { + "epoch": 0.2182852910795377, + "grad_norm": 16.54023551940918, + "learning_rate": 7.275561907236585e-07, + "loss": 1.0256, + "step": 5600 + }, + { + "epoch": 0.21867508624217974, + "grad_norm": 16.5329647064209, + "learning_rate": 7.288553982070936e-07, + "loss": 1.0396, + "step": 5610 + }, + { + "epoch": 0.21906488140482178, + "grad_norm": 14.109983444213867, + "learning_rate": 7.301546056905288e-07, + "loss": 0.9909, + "step": 5620 + }, + { + "epoch": 0.21945467656746379, + "grad_norm": 14.571008682250977, + "learning_rate": 7.314538131739638e-07, + "loss": 1.0421, + "step": 5630 + }, + { + "epoch": 0.21984447173010582, + "grad_norm": 15.100592613220215, + "learning_rate": 7.327530206573989e-07, + "loss": 1.0256, + "step": 5640 + }, + { + "epoch": 0.22023426689274786, + "grad_norm": 15.60147762298584, + "learning_rate": 7.34052228140834e-07, + "loss": 0.9687, + "step": 5650 + }, + { + "epoch": 0.2206240620553899, + "grad_norm": 12.775625228881836, + "learning_rate": 7.353514356242692e-07, + "loss": 0.994, + "step": 5660 + }, + { + "epoch": 0.22101385721803193, + "grad_norm": 13.491575241088867, + "learning_rate": 7.366506431077043e-07, + "loss": 1.0215, + "step": 5670 + }, + { + "epoch": 0.22140365238067394, + "grad_norm": 13.057899475097656, + "learning_rate": 7.379498505911394e-07, + "loss": 0.9579, + "step": 5680 + }, + { + "epoch": 0.22179344754331598, + "grad_norm": 14.783434867858887, + "learning_rate": 7.392490580745745e-07, + "loss": 1.0579, + "step": 5690 + }, + { + "epoch": 0.22218324270595802, + "grad_norm": 16.33684539794922, + "learning_rate": 7.405482655580095e-07, + "loss": 1.0722, + "step": 5700 + }, + { + "epoch": 0.22257303786860005, + "grad_norm": 15.10921859741211, + "learning_rate": 7.418474730414447e-07, + "loss": 1.0023, + "step": 5710 + }, + { + "epoch": 0.2229628330312421, + "grad_norm": 15.505993843078613, + "learning_rate": 7.431466805248798e-07, + "loss": 1.0212, + "step": 5720 + }, + { + "epoch": 0.2233526281938841, + "grad_norm": 14.529448509216309, + "learning_rate": 7.444458880083148e-07, + "loss": 0.9869, + "step": 5730 + }, + { + "epoch": 0.22374242335652614, + "grad_norm": 13.311355590820312, + "learning_rate": 7.457450954917501e-07, + "loss": 0.8866, + "step": 5740 + }, + { + "epoch": 0.22413221851916817, + "grad_norm": 15.373902320861816, + "learning_rate": 7.470443029751851e-07, + "loss": 1.0244, + "step": 5750 + }, + { + "epoch": 0.2245220136818102, + "grad_norm": 14.613059043884277, + "learning_rate": 7.483435104586202e-07, + "loss": 1.0341, + "step": 5760 + }, + { + "epoch": 0.22491180884445225, + "grad_norm": 15.925219535827637, + "learning_rate": 7.496427179420554e-07, + "loss": 0.9762, + "step": 5770 + }, + { + "epoch": 0.22530160400709426, + "grad_norm": 13.845405578613281, + "learning_rate": 7.509419254254904e-07, + "loss": 0.9846, + "step": 5780 + }, + { + "epoch": 0.2256913991697363, + "grad_norm": 14.4869384765625, + "learning_rate": 7.522411329089255e-07, + "loss": 1.0153, + "step": 5790 + }, + { + "epoch": 0.22608119433237833, + "grad_norm": 13.37547492980957, + "learning_rate": 7.535403403923606e-07, + "loss": 1.0228, + "step": 5800 + }, + { + "epoch": 0.22647098949502037, + "grad_norm": 14.425969123840332, + "learning_rate": 7.548395478757957e-07, + "loss": 1.0111, + "step": 5810 + }, + { + "epoch": 0.2268607846576624, + "grad_norm": 13.448543548583984, + "learning_rate": 7.561387553592308e-07, + "loss": 1.0284, + "step": 5820 + }, + { + "epoch": 0.22725057982030444, + "grad_norm": 13.719442367553711, + "learning_rate": 7.57437962842666e-07, + "loss": 0.9912, + "step": 5830 + }, + { + "epoch": 0.22764037498294645, + "grad_norm": 15.14156723022461, + "learning_rate": 7.587371703261011e-07, + "loss": 1.0055, + "step": 5840 + }, + { + "epoch": 0.2280301701455885, + "grad_norm": 15.68527889251709, + "learning_rate": 7.600363778095361e-07, + "loss": 1.0243, + "step": 5850 + }, + { + "epoch": 0.22841996530823053, + "grad_norm": 13.674068450927734, + "learning_rate": 7.613355852929713e-07, + "loss": 0.98, + "step": 5860 + }, + { + "epoch": 0.22880976047087256, + "grad_norm": 15.385726928710938, + "learning_rate": 7.626347927764064e-07, + "loss": 0.9947, + "step": 5870 + }, + { + "epoch": 0.2291995556335146, + "grad_norm": 16.430587768554688, + "learning_rate": 7.639340002598414e-07, + "loss": 0.995, + "step": 5880 + }, + { + "epoch": 0.2295893507961566, + "grad_norm": 15.448851585388184, + "learning_rate": 7.652332077432766e-07, + "loss": 0.9726, + "step": 5890 + }, + { + "epoch": 0.22997914595879865, + "grad_norm": 14.246310234069824, + "learning_rate": 7.665324152267116e-07, + "loss": 1.0102, + "step": 5900 + }, + { + "epoch": 0.23036894112144068, + "grad_norm": 14.389453887939453, + "learning_rate": 7.678316227101468e-07, + "loss": 1.0575, + "step": 5910 + }, + { + "epoch": 0.23075873628408272, + "grad_norm": 16.1449031829834, + "learning_rate": 7.691308301935819e-07, + "loss": 0.9702, + "step": 5920 + }, + { + "epoch": 0.23114853144672476, + "grad_norm": 15.681499481201172, + "learning_rate": 7.70430037677017e-07, + "loss": 1.0253, + "step": 5930 + }, + { + "epoch": 0.23153832660936677, + "grad_norm": 17.637239456176758, + "learning_rate": 7.717292451604521e-07, + "loss": 1.0347, + "step": 5940 + }, + { + "epoch": 0.2319281217720088, + "grad_norm": 15.755040168762207, + "learning_rate": 7.730284526438872e-07, + "loss": 1.0479, + "step": 5950 + }, + { + "epoch": 0.23231791693465084, + "grad_norm": 14.450773239135742, + "learning_rate": 7.743276601273223e-07, + "loss": 1.0168, + "step": 5960 + }, + { + "epoch": 0.23270771209729288, + "grad_norm": 15.496644020080566, + "learning_rate": 7.756268676107573e-07, + "loss": 0.9085, + "step": 5970 + }, + { + "epoch": 0.23309750725993492, + "grad_norm": 15.421711921691895, + "learning_rate": 7.769260750941925e-07, + "loss": 1.0027, + "step": 5980 + }, + { + "epoch": 0.23348730242257693, + "grad_norm": 12.68456745147705, + "learning_rate": 7.782252825776277e-07, + "loss": 0.9732, + "step": 5990 + }, + { + "epoch": 0.23387709758521896, + "grad_norm": 15.658552169799805, + "learning_rate": 7.795244900610627e-07, + "loss": 0.9886, + "step": 6000 + }, + { + "epoch": 0.23387709758521896, + "eval_loss": 1.008408784866333, + "eval_runtime": 82.7433, + "eval_samples_per_second": 50.119, + "eval_steps_per_second": 6.272, + "step": 6000 + }, + { + "epoch": 0.234266892747861, + "grad_norm": 15.129308700561523, + "learning_rate": 7.808236975444979e-07, + "loss": 0.9321, + "step": 6010 + }, + { + "epoch": 0.23465668791050304, + "grad_norm": 17.187414169311523, + "learning_rate": 7.821229050279329e-07, + "loss": 0.9859, + "step": 6020 + }, + { + "epoch": 0.23504648307314507, + "grad_norm": 14.983357429504395, + "learning_rate": 7.83422112511368e-07, + "loss": 0.9746, + "step": 6030 + }, + { + "epoch": 0.23543627823578708, + "grad_norm": 15.761835098266602, + "learning_rate": 7.847213199948032e-07, + "loss": 0.9883, + "step": 6040 + }, + { + "epoch": 0.23582607339842912, + "grad_norm": 14.69127082824707, + "learning_rate": 7.860205274782382e-07, + "loss": 1.0073, + "step": 6050 + }, + { + "epoch": 0.23621586856107116, + "grad_norm": 14.32921028137207, + "learning_rate": 7.873197349616733e-07, + "loss": 1.0216, + "step": 6060 + }, + { + "epoch": 0.2366056637237132, + "grad_norm": 15.987038612365723, + "learning_rate": 7.886189424451084e-07, + "loss": 0.9576, + "step": 6070 + }, + { + "epoch": 0.23699545888635523, + "grad_norm": 15.02837085723877, + "learning_rate": 7.899181499285436e-07, + "loss": 1.0043, + "step": 6080 + }, + { + "epoch": 0.23738525404899724, + "grad_norm": 15.183786392211914, + "learning_rate": 7.912173574119787e-07, + "loss": 0.9972, + "step": 6090 + }, + { + "epoch": 0.23777504921163928, + "grad_norm": 13.752614974975586, + "learning_rate": 7.925165648954138e-07, + "loss": 0.9609, + "step": 6100 + }, + { + "epoch": 0.23816484437428131, + "grad_norm": 15.52311897277832, + "learning_rate": 7.938157723788489e-07, + "loss": 0.9418, + "step": 6110 + }, + { + "epoch": 0.23855463953692335, + "grad_norm": 13.700892448425293, + "learning_rate": 7.951149798622839e-07, + "loss": 0.9783, + "step": 6120 + }, + { + "epoch": 0.2389444346995654, + "grad_norm": 14.884476661682129, + "learning_rate": 7.964141873457191e-07, + "loss": 0.9368, + "step": 6130 + }, + { + "epoch": 0.2393342298622074, + "grad_norm": 14.45043659210205, + "learning_rate": 7.977133948291542e-07, + "loss": 1.0335, + "step": 6140 + }, + { + "epoch": 0.23972402502484944, + "grad_norm": 14.685229301452637, + "learning_rate": 7.990126023125892e-07, + "loss": 0.9836, + "step": 6150 + }, + { + "epoch": 0.24011382018749147, + "grad_norm": 15.066401481628418, + "learning_rate": 8.003118097960245e-07, + "loss": 0.9682, + "step": 6160 + }, + { + "epoch": 0.2405036153501335, + "grad_norm": 15.01012134552002, + "learning_rate": 8.016110172794595e-07, + "loss": 0.9509, + "step": 6170 + }, + { + "epoch": 0.24089341051277555, + "grad_norm": 13.661785125732422, + "learning_rate": 8.029102247628946e-07, + "loss": 0.997, + "step": 6180 + }, + { + "epoch": 0.24128320567541756, + "grad_norm": 15.887532234191895, + "learning_rate": 8.042094322463298e-07, + "loss": 0.9463, + "step": 6190 + }, + { + "epoch": 0.2416730008380596, + "grad_norm": 16.781452178955078, + "learning_rate": 8.055086397297648e-07, + "loss": 0.982, + "step": 6200 + }, + { + "epoch": 0.24206279600070163, + "grad_norm": 14.375569343566895, + "learning_rate": 8.068078472131999e-07, + "loss": 1.0074, + "step": 6210 + }, + { + "epoch": 0.24245259116334367, + "grad_norm": 14.924545288085938, + "learning_rate": 8.08107054696635e-07, + "loss": 1.0302, + "step": 6220 + }, + { + "epoch": 0.2428423863259857, + "grad_norm": 16.764692306518555, + "learning_rate": 8.094062621800701e-07, + "loss": 1.008, + "step": 6230 + }, + { + "epoch": 0.2432321814886277, + "grad_norm": 14.65231990814209, + "learning_rate": 8.107054696635052e-07, + "loss": 1.0206, + "step": 6240 + }, + { + "epoch": 0.24362197665126975, + "grad_norm": 12.948251724243164, + "learning_rate": 8.120046771469404e-07, + "loss": 0.9871, + "step": 6250 + }, + { + "epoch": 0.2440117718139118, + "grad_norm": 16.497608184814453, + "learning_rate": 8.133038846303755e-07, + "loss": 0.9903, + "step": 6260 + }, + { + "epoch": 0.24440156697655382, + "grad_norm": 14.26806640625, + "learning_rate": 8.146030921138105e-07, + "loss": 0.995, + "step": 6270 + }, + { + "epoch": 0.24479136213919586, + "grad_norm": 15.191696166992188, + "learning_rate": 8.159022995972457e-07, + "loss": 0.9465, + "step": 6280 + }, + { + "epoch": 0.24518115730183787, + "grad_norm": 13.746397972106934, + "learning_rate": 8.172015070806807e-07, + "loss": 0.992, + "step": 6290 + }, + { + "epoch": 0.2455709524644799, + "grad_norm": 12.811980247497559, + "learning_rate": 8.185007145641158e-07, + "loss": 0.9483, + "step": 6300 + }, + { + "epoch": 0.24596074762712195, + "grad_norm": 14.354470252990723, + "learning_rate": 8.19799922047551e-07, + "loss": 0.9671, + "step": 6310 + }, + { + "epoch": 0.24635054278976398, + "grad_norm": 17.096776962280273, + "learning_rate": 8.210991295309861e-07, + "loss": 0.9946, + "step": 6320 + }, + { + "epoch": 0.24674033795240602, + "grad_norm": 14.403414726257324, + "learning_rate": 8.223983370144212e-07, + "loss": 1.0161, + "step": 6330 + }, + { + "epoch": 0.24713013311504803, + "grad_norm": 14.489141464233398, + "learning_rate": 8.236975444978562e-07, + "loss": 0.9904, + "step": 6340 + }, + { + "epoch": 0.24751992827769007, + "grad_norm": 15.053180694580078, + "learning_rate": 8.249967519812914e-07, + "loss": 1.0291, + "step": 6350 + }, + { + "epoch": 0.2479097234403321, + "grad_norm": 16.551687240600586, + "learning_rate": 8.262959594647265e-07, + "loss": 1.0117, + "step": 6360 + }, + { + "epoch": 0.24829951860297414, + "grad_norm": 15.823965072631836, + "learning_rate": 8.275951669481615e-07, + "loss": 0.9695, + "step": 6370 + }, + { + "epoch": 0.24868931376561618, + "grad_norm": 15.22997760772705, + "learning_rate": 8.288943744315967e-07, + "loss": 1.0694, + "step": 6380 + }, + { + "epoch": 0.2490791089282582, + "grad_norm": 15.043110847473145, + "learning_rate": 8.301935819150317e-07, + "loss": 1.0284, + "step": 6390 + }, + { + "epoch": 0.24946890409090022, + "grad_norm": 13.130000114440918, + "learning_rate": 8.314927893984668e-07, + "loss": 0.9851, + "step": 6400 + }, + { + "epoch": 0.24985869925354226, + "grad_norm": 14.716496467590332, + "learning_rate": 8.327919968819021e-07, + "loss": 0.9617, + "step": 6410 + }, + { + "epoch": 0.2502484944161843, + "grad_norm": 13.478013038635254, + "learning_rate": 8.340912043653371e-07, + "loss": 0.95, + "step": 6420 + }, + { + "epoch": 0.25063828957882633, + "grad_norm": 15.811286926269531, + "learning_rate": 8.353904118487723e-07, + "loss": 0.9589, + "step": 6430 + }, + { + "epoch": 0.25102808474146837, + "grad_norm": 11.231595993041992, + "learning_rate": 8.366896193322073e-07, + "loss": 0.9803, + "step": 6440 + }, + { + "epoch": 0.2514178799041104, + "grad_norm": 14.72209358215332, + "learning_rate": 8.379888268156424e-07, + "loss": 0.9705, + "step": 6450 + }, + { + "epoch": 0.25180767506675245, + "grad_norm": 18.370105743408203, + "learning_rate": 8.392880342990775e-07, + "loss": 1.0107, + "step": 6460 + }, + { + "epoch": 0.2521974702293944, + "grad_norm": 16.057771682739258, + "learning_rate": 8.405872417825126e-07, + "loss": 0.978, + "step": 6470 + }, + { + "epoch": 0.25258726539203646, + "grad_norm": 14.759856224060059, + "learning_rate": 8.418864492659477e-07, + "loss": 0.9994, + "step": 6480 + }, + { + "epoch": 0.2529770605546785, + "grad_norm": 15.612014770507812, + "learning_rate": 8.431856567493828e-07, + "loss": 1.0099, + "step": 6490 + }, + { + "epoch": 0.25336685571732054, + "grad_norm": 16.094316482543945, + "learning_rate": 8.44484864232818e-07, + "loss": 0.9907, + "step": 6500 + }, + { + "epoch": 0.25336685571732054, + "eval_loss": 0.9982940554618835, + "eval_runtime": 82.8285, + "eval_samples_per_second": 50.067, + "eval_steps_per_second": 6.266, + "step": 6500 + }, + { + "epoch": 0.2537566508799626, + "grad_norm": 15.380989074707031, + "learning_rate": 8.457840717162531e-07, + "loss": 0.9658, + "step": 6510 + }, + { + "epoch": 0.2541464460426046, + "grad_norm": 16.016374588012695, + "learning_rate": 8.470832791996881e-07, + "loss": 0.9425, + "step": 6520 + }, + { + "epoch": 0.25453624120524665, + "grad_norm": 14.255965232849121, + "learning_rate": 8.483824866831233e-07, + "loss": 0.9322, + "step": 6530 + }, + { + "epoch": 0.2549260363678887, + "grad_norm": 16.5495662689209, + "learning_rate": 8.496816941665583e-07, + "loss": 0.9776, + "step": 6540 + }, + { + "epoch": 0.2553158315305307, + "grad_norm": 18.367450714111328, + "learning_rate": 8.509809016499934e-07, + "loss": 1.0039, + "step": 6550 + }, + { + "epoch": 0.25570562669317276, + "grad_norm": 15.758204460144043, + "learning_rate": 8.522801091334285e-07, + "loss": 1.0416, + "step": 6560 + }, + { + "epoch": 0.25609542185581474, + "grad_norm": 14.343899726867676, + "learning_rate": 8.535793166168637e-07, + "loss": 1.0402, + "step": 6570 + }, + { + "epoch": 0.2564852170184568, + "grad_norm": 12.901144981384277, + "learning_rate": 8.548785241002988e-07, + "loss": 1.0069, + "step": 6580 + }, + { + "epoch": 0.2568750121810988, + "grad_norm": 16.866046905517578, + "learning_rate": 8.561777315837339e-07, + "loss": 0.9626, + "step": 6590 + }, + { + "epoch": 0.25726480734374085, + "grad_norm": 13.181471824645996, + "learning_rate": 8.57476939067169e-07, + "loss": 1.0387, + "step": 6600 + }, + { + "epoch": 0.2576546025063829, + "grad_norm": 14.82888126373291, + "learning_rate": 8.58776146550604e-07, + "loss": 1.0033, + "step": 6610 + }, + { + "epoch": 0.2580443976690249, + "grad_norm": 15.603177070617676, + "learning_rate": 8.600753540340392e-07, + "loss": 1.0803, + "step": 6620 + }, + { + "epoch": 0.25843419283166696, + "grad_norm": 15.467535972595215, + "learning_rate": 8.613745615174743e-07, + "loss": 0.983, + "step": 6630 + }, + { + "epoch": 0.258823987994309, + "grad_norm": 12.85669231414795, + "learning_rate": 8.626737690009093e-07, + "loss": 0.9998, + "step": 6640 + }, + { + "epoch": 0.25921378315695104, + "grad_norm": 17.394739151000977, + "learning_rate": 8.639729764843446e-07, + "loss": 0.9543, + "step": 6650 + }, + { + "epoch": 0.2596035783195931, + "grad_norm": 15.160696029663086, + "learning_rate": 8.652721839677796e-07, + "loss": 1.0488, + "step": 6660 + }, + { + "epoch": 0.2599933734822351, + "grad_norm": 16.85416603088379, + "learning_rate": 8.665713914512147e-07, + "loss": 0.9224, + "step": 6670 + }, + { + "epoch": 0.2603831686448771, + "grad_norm": 15.190739631652832, + "learning_rate": 8.678705989346499e-07, + "loss": 0.9954, + "step": 6680 + }, + { + "epoch": 0.26077296380751913, + "grad_norm": 17.95996856689453, + "learning_rate": 8.691698064180849e-07, + "loss": 1.0229, + "step": 6690 + }, + { + "epoch": 0.26116275897016117, + "grad_norm": 13.480778694152832, + "learning_rate": 8.7046901390152e-07, + "loss": 1.0526, + "step": 6700 + }, + { + "epoch": 0.2615525541328032, + "grad_norm": 13.557330131530762, + "learning_rate": 8.717682213849551e-07, + "loss": 1.0048, + "step": 6710 + }, + { + "epoch": 0.26194234929544524, + "grad_norm": 17.34872817993164, + "learning_rate": 8.730674288683902e-07, + "loss": 0.9335, + "step": 6720 + }, + { + "epoch": 0.2623321444580873, + "grad_norm": 15.180060386657715, + "learning_rate": 8.743666363518254e-07, + "loss": 0.9386, + "step": 6730 + }, + { + "epoch": 0.2627219396207293, + "grad_norm": 14.60095500946045, + "learning_rate": 8.756658438352605e-07, + "loss": 1.025, + "step": 6740 + }, + { + "epoch": 0.26311173478337135, + "grad_norm": 14.170730590820312, + "learning_rate": 8.769650513186956e-07, + "loss": 0.9878, + "step": 6750 + }, + { + "epoch": 0.2635015299460134, + "grad_norm": 15.936722755432129, + "learning_rate": 8.782642588021306e-07, + "loss": 1.0003, + "step": 6760 + }, + { + "epoch": 0.26389132510865543, + "grad_norm": 15.04389762878418, + "learning_rate": 8.795634662855658e-07, + "loss": 0.9126, + "step": 6770 + }, + { + "epoch": 0.2642811202712974, + "grad_norm": 19.2154541015625, + "learning_rate": 8.808626737690009e-07, + "loss": 1.0655, + "step": 6780 + }, + { + "epoch": 0.26467091543393945, + "grad_norm": 14.865091323852539, + "learning_rate": 8.821618812524359e-07, + "loss": 1.0007, + "step": 6790 + }, + { + "epoch": 0.2650607105965815, + "grad_norm": 13.163175582885742, + "learning_rate": 8.834610887358711e-07, + "loss": 1.1411, + "step": 6800 + }, + { + "epoch": 0.2654505057592235, + "grad_norm": 15.74963665008545, + "learning_rate": 8.847602962193061e-07, + "loss": 0.9577, + "step": 6810 + }, + { + "epoch": 0.26584030092186556, + "grad_norm": 18.226572036743164, + "learning_rate": 8.860595037027413e-07, + "loss": 0.9918, + "step": 6820 + }, + { + "epoch": 0.2662300960845076, + "grad_norm": 14.836274147033691, + "learning_rate": 8.873587111861765e-07, + "loss": 0.9707, + "step": 6830 + }, + { + "epoch": 0.26661989124714963, + "grad_norm": 16.010282516479492, + "learning_rate": 8.886579186696115e-07, + "loss": 1.0282, + "step": 6840 + }, + { + "epoch": 0.26700968640979167, + "grad_norm": 14.209004402160645, + "learning_rate": 8.899571261530466e-07, + "loss": 0.9901, + "step": 6850 + }, + { + "epoch": 0.2673994815724337, + "grad_norm": 14.920477867126465, + "learning_rate": 8.912563336364817e-07, + "loss": 0.9963, + "step": 6860 + }, + { + "epoch": 0.26778927673507574, + "grad_norm": 12.969196319580078, + "learning_rate": 8.925555411199168e-07, + "loss": 1.0026, + "step": 6870 + }, + { + "epoch": 0.2681790718977177, + "grad_norm": 12.846720695495605, + "learning_rate": 8.938547486033518e-07, + "loss": 0.9223, + "step": 6880 + }, + { + "epoch": 0.26856886706035976, + "grad_norm": 13.77932357788086, + "learning_rate": 8.95153956086787e-07, + "loss": 1.0532, + "step": 6890 + }, + { + "epoch": 0.2689586622230018, + "grad_norm": 16.579750061035156, + "learning_rate": 8.964531635702222e-07, + "loss": 1.0682, + "step": 6900 + }, + { + "epoch": 0.26934845738564384, + "grad_norm": 13.730154037475586, + "learning_rate": 8.977523710536572e-07, + "loss": 1.037, + "step": 6910 + }, + { + "epoch": 0.2697382525482859, + "grad_norm": 14.519630432128906, + "learning_rate": 8.990515785370924e-07, + "loss": 1.0138, + "step": 6920 + }, + { + "epoch": 0.2701280477109279, + "grad_norm": 14.577149391174316, + "learning_rate": 9.003507860205274e-07, + "loss": 0.9768, + "step": 6930 + }, + { + "epoch": 0.27051784287356995, + "grad_norm": 15.694051742553711, + "learning_rate": 9.016499935039625e-07, + "loss": 0.9574, + "step": 6940 + }, + { + "epoch": 0.270907638036212, + "grad_norm": 14.747005462646484, + "learning_rate": 9.029492009873977e-07, + "loss": 0.9546, + "step": 6950 + }, + { + "epoch": 0.271297433198854, + "grad_norm": 13.21657657623291, + "learning_rate": 9.042484084708327e-07, + "loss": 1.0092, + "step": 6960 + }, + { + "epoch": 0.27168722836149606, + "grad_norm": 14.990208625793457, + "learning_rate": 9.055476159542678e-07, + "loss": 0.9834, + "step": 6970 + }, + { + "epoch": 0.27207702352413804, + "grad_norm": 13.693812370300293, + "learning_rate": 9.068468234377029e-07, + "loss": 1.0538, + "step": 6980 + }, + { + "epoch": 0.2724668186867801, + "grad_norm": 15.132277488708496, + "learning_rate": 9.081460309211381e-07, + "loss": 1.0271, + "step": 6990 + }, + { + "epoch": 0.2728566138494221, + "grad_norm": 12.762600898742676, + "learning_rate": 9.094452384045732e-07, + "loss": 1.0142, + "step": 7000 + }, + { + "epoch": 0.2728566138494221, + "eval_loss": 0.9905140399932861, + "eval_runtime": 82.7197, + "eval_samples_per_second": 50.133, + "eval_steps_per_second": 6.274, + "step": 7000 + }, + { + "epoch": 0.27324640901206415, + "grad_norm": 13.291237831115723, + "learning_rate": 9.107444458880083e-07, + "loss": 1.0392, + "step": 7010 + }, + { + "epoch": 0.2736362041747062, + "grad_norm": 14.18792724609375, + "learning_rate": 9.120436533714434e-07, + "loss": 0.9954, + "step": 7020 + }, + { + "epoch": 0.2740259993373482, + "grad_norm": 15.6013765335083, + "learning_rate": 9.133428608548784e-07, + "loss": 1.0525, + "step": 7030 + }, + { + "epoch": 0.27441579449999026, + "grad_norm": 14.35906982421875, + "learning_rate": 9.146420683383136e-07, + "loss": 0.9773, + "step": 7040 + }, + { + "epoch": 0.2748055896626323, + "grad_norm": 15.42289924621582, + "learning_rate": 9.159412758217487e-07, + "loss": 0.9999, + "step": 7050 + }, + { + "epoch": 0.27519538482527434, + "grad_norm": 13.793740272521973, + "learning_rate": 9.172404833051837e-07, + "loss": 1.0516, + "step": 7060 + }, + { + "epoch": 0.2755851799879164, + "grad_norm": 13.272202491760254, + "learning_rate": 9.18539690788619e-07, + "loss": 0.9667, + "step": 7070 + }, + { + "epoch": 0.27597497515055835, + "grad_norm": 15.208497047424316, + "learning_rate": 9.19838898272054e-07, + "loss": 1.0081, + "step": 7080 + }, + { + "epoch": 0.2763647703132004, + "grad_norm": 14.63493537902832, + "learning_rate": 9.211381057554891e-07, + "loss": 0.9458, + "step": 7090 + }, + { + "epoch": 0.27675456547584243, + "grad_norm": 14.298831939697266, + "learning_rate": 9.224373132389243e-07, + "loss": 0.9319, + "step": 7100 + }, + { + "epoch": 0.27714436063848447, + "grad_norm": 15.870024681091309, + "learning_rate": 9.237365207223593e-07, + "loss": 0.9832, + "step": 7110 + }, + { + "epoch": 0.2775341558011265, + "grad_norm": 12.755727767944336, + "learning_rate": 9.250357282057944e-07, + "loss": 0.9349, + "step": 7120 + }, + { + "epoch": 0.27792395096376854, + "grad_norm": 14.909743309020996, + "learning_rate": 9.263349356892295e-07, + "loss": 0.9397, + "step": 7130 + }, + { + "epoch": 0.2783137461264106, + "grad_norm": 15.338642120361328, + "learning_rate": 9.276341431726646e-07, + "loss": 1.0098, + "step": 7140 + }, + { + "epoch": 0.2787035412890526, + "grad_norm": 14.55411434173584, + "learning_rate": 9.289333506560998e-07, + "loss": 0.9771, + "step": 7150 + }, + { + "epoch": 0.27909333645169465, + "grad_norm": 14.71304988861084, + "learning_rate": 9.302325581395349e-07, + "loss": 1.0443, + "step": 7160 + }, + { + "epoch": 0.2794831316143367, + "grad_norm": 14.890127182006836, + "learning_rate": 9.3153176562297e-07, + "loss": 0.98, + "step": 7170 + }, + { + "epoch": 0.2798729267769787, + "grad_norm": 15.056764602661133, + "learning_rate": 9.32830973106405e-07, + "loss": 0.9396, + "step": 7180 + }, + { + "epoch": 0.2802627219396207, + "grad_norm": 13.5358304977417, + "learning_rate": 9.341301805898402e-07, + "loss": 1.035, + "step": 7190 + }, + { + "epoch": 0.28065251710226274, + "grad_norm": 15.333770751953125, + "learning_rate": 9.354293880732752e-07, + "loss": 0.9544, + "step": 7200 + }, + { + "epoch": 0.2810423122649048, + "grad_norm": 12.405569076538086, + "learning_rate": 9.367285955567103e-07, + "loss": 0.9625, + "step": 7210 + }, + { + "epoch": 0.2814321074275468, + "grad_norm": 14.385384559631348, + "learning_rate": 9.380278030401455e-07, + "loss": 0.9526, + "step": 7220 + }, + { + "epoch": 0.28182190259018886, + "grad_norm": 16.159231185913086, + "learning_rate": 9.393270105235806e-07, + "loss": 0.9417, + "step": 7230 + }, + { + "epoch": 0.2822116977528309, + "grad_norm": 12.691298484802246, + "learning_rate": 9.406262180070157e-07, + "loss": 0.951, + "step": 7240 + }, + { + "epoch": 0.28260149291547293, + "grad_norm": 14.884239196777344, + "learning_rate": 9.419254254904508e-07, + "loss": 0.9228, + "step": 7250 + }, + { + "epoch": 0.28299128807811497, + "grad_norm": 15.41154670715332, + "learning_rate": 9.432246329738859e-07, + "loss": 1.0061, + "step": 7260 + }, + { + "epoch": 0.283381083240757, + "grad_norm": 15.061527252197266, + "learning_rate": 9.44523840457321e-07, + "loss": 0.9852, + "step": 7270 + }, + { + "epoch": 0.28377087840339904, + "grad_norm": 12.765412330627441, + "learning_rate": 9.458230479407561e-07, + "loss": 0.9641, + "step": 7280 + }, + { + "epoch": 0.284160673566041, + "grad_norm": 20.124662399291992, + "learning_rate": 9.471222554241912e-07, + "loss": 0.9214, + "step": 7290 + }, + { + "epoch": 0.28455046872868306, + "grad_norm": 13.923117637634277, + "learning_rate": 9.484214629076262e-07, + "loss": 0.9656, + "step": 7300 + }, + { + "epoch": 0.2849402638913251, + "grad_norm": 14.351522445678711, + "learning_rate": 9.497206703910615e-07, + "loss": 0.9609, + "step": 7310 + }, + { + "epoch": 0.28533005905396713, + "grad_norm": 14.115310668945312, + "learning_rate": 9.510198778744966e-07, + "loss": 0.9873, + "step": 7320 + }, + { + "epoch": 0.28571985421660917, + "grad_norm": 13.593430519104004, + "learning_rate": 9.523190853579316e-07, + "loss": 1.0045, + "step": 7330 + }, + { + "epoch": 0.2861096493792512, + "grad_norm": 14.610860824584961, + "learning_rate": 9.536182928413668e-07, + "loss": 1.0003, + "step": 7340 + }, + { + "epoch": 0.28649944454189324, + "grad_norm": 13.852402687072754, + "learning_rate": 9.549175003248017e-07, + "loss": 0.9827, + "step": 7350 + }, + { + "epoch": 0.2868892397045353, + "grad_norm": 13.716668128967285, + "learning_rate": 9.56216707808237e-07, + "loss": 0.9662, + "step": 7360 + }, + { + "epoch": 0.2872790348671773, + "grad_norm": 14.240720748901367, + "learning_rate": 9.575159152916722e-07, + "loss": 0.9488, + "step": 7370 + }, + { + "epoch": 0.28766883002981936, + "grad_norm": 17.744895935058594, + "learning_rate": 9.588151227751072e-07, + "loss": 1.0037, + "step": 7380 + }, + { + "epoch": 0.28805862519246134, + "grad_norm": 15.923994064331055, + "learning_rate": 9.601143302585422e-07, + "loss": 1.0485, + "step": 7390 + }, + { + "epoch": 0.2884484203551034, + "grad_norm": 17.97040367126465, + "learning_rate": 9.614135377419772e-07, + "loss": 0.9723, + "step": 7400 + }, + { + "epoch": 0.2888382155177454, + "grad_norm": 15.043636322021484, + "learning_rate": 9.627127452254125e-07, + "loss": 0.9631, + "step": 7410 + }, + { + "epoch": 0.28922801068038745, + "grad_norm": 14.160172462463379, + "learning_rate": 9.640119527088475e-07, + "loss": 0.9509, + "step": 7420 + }, + { + "epoch": 0.2896178058430295, + "grad_norm": 15.685235023498535, + "learning_rate": 9.653111601922825e-07, + "loss": 0.9639, + "step": 7430 + }, + { + "epoch": 0.2900076010056715, + "grad_norm": 13.971315383911133, + "learning_rate": 9.666103676757178e-07, + "loss": 1.0537, + "step": 7440 + }, + { + "epoch": 0.29039739616831356, + "grad_norm": 13.192005157470703, + "learning_rate": 9.679095751591528e-07, + "loss": 0.9714, + "step": 7450 + }, + { + "epoch": 0.2907871913309556, + "grad_norm": 17.05986213684082, + "learning_rate": 9.69208782642588e-07, + "loss": 0.9594, + "step": 7460 + }, + { + "epoch": 0.29117698649359763, + "grad_norm": 14.035353660583496, + "learning_rate": 9.70507990126023e-07, + "loss": 1.0411, + "step": 7470 + }, + { + "epoch": 0.29156678165623967, + "grad_norm": 16.18226432800293, + "learning_rate": 9.718071976094581e-07, + "loss": 0.9906, + "step": 7480 + }, + { + "epoch": 0.29195657681888165, + "grad_norm": 12.675119400024414, + "learning_rate": 9.731064050928934e-07, + "loss": 0.9732, + "step": 7490 + }, + { + "epoch": 0.2923463719815237, + "grad_norm": 15.634903907775879, + "learning_rate": 9.744056125763284e-07, + "loss": 0.9507, + "step": 7500 + }, + { + "epoch": 0.2923463719815237, + "eval_loss": 0.9845738410949707, + "eval_runtime": 82.8775, + "eval_samples_per_second": 50.038, + "eval_steps_per_second": 6.262, + "step": 7500 + }, + { + "epoch": 0.2927361671441657, + "grad_norm": 13.629648208618164, + "learning_rate": 9.757048200597634e-07, + "loss": 0.8886, + "step": 7510 + }, + { + "epoch": 0.29312596230680776, + "grad_norm": 15.828271865844727, + "learning_rate": 9.770040275431987e-07, + "loss": 0.9761, + "step": 7520 + }, + { + "epoch": 0.2935157574694498, + "grad_norm": 13.541333198547363, + "learning_rate": 9.783032350266337e-07, + "loss": 0.9807, + "step": 7530 + }, + { + "epoch": 0.29390555263209184, + "grad_norm": 16.535472869873047, + "learning_rate": 9.79602442510069e-07, + "loss": 1.0101, + "step": 7540 + }, + { + "epoch": 0.2942953477947339, + "grad_norm": 13.762173652648926, + "learning_rate": 9.80901649993504e-07, + "loss": 0.9008, + "step": 7550 + }, + { + "epoch": 0.2946851429573759, + "grad_norm": 15.469985008239746, + "learning_rate": 9.82200857476939e-07, + "loss": 0.9762, + "step": 7560 + }, + { + "epoch": 0.29507493812001795, + "grad_norm": 13.669770240783691, + "learning_rate": 9.83500064960374e-07, + "loss": 0.9743, + "step": 7570 + }, + { + "epoch": 0.29546473328266, + "grad_norm": 15.509896278381348, + "learning_rate": 9.847992724438092e-07, + "loss": 1.0077, + "step": 7580 + }, + { + "epoch": 0.29585452844530197, + "grad_norm": 14.286842346191406, + "learning_rate": 9.860984799272443e-07, + "loss": 0.9822, + "step": 7590 + }, + { + "epoch": 0.296244323607944, + "grad_norm": 13.8685884475708, + "learning_rate": 9.873976874106795e-07, + "loss": 0.9559, + "step": 7600 + }, + { + "epoch": 0.29663411877058604, + "grad_norm": 14.825406074523926, + "learning_rate": 9.886968948941145e-07, + "loss": 1.0357, + "step": 7610 + }, + { + "epoch": 0.2970239139332281, + "grad_norm": 14.749993324279785, + "learning_rate": 9.899961023775496e-07, + "loss": 0.99, + "step": 7620 + }, + { + "epoch": 0.2974137090958701, + "grad_norm": 14.50264835357666, + "learning_rate": 9.912953098609848e-07, + "loss": 1.04, + "step": 7630 + }, + { + "epoch": 0.29780350425851215, + "grad_norm": 13.416096687316895, + "learning_rate": 9.925945173444198e-07, + "loss": 0.9476, + "step": 7640 + }, + { + "epoch": 0.2981932994211542, + "grad_norm": 14.255441665649414, + "learning_rate": 9.938937248278549e-07, + "loss": 0.9974, + "step": 7650 + }, + { + "epoch": 0.2985830945837962, + "grad_norm": 11.918822288513184, + "learning_rate": 9.951929323112901e-07, + "loss": 0.9315, + "step": 7660 + }, + { + "epoch": 0.29897288974643826, + "grad_norm": 14.827400207519531, + "learning_rate": 9.964921397947251e-07, + "loss": 0.9932, + "step": 7670 + }, + { + "epoch": 0.2993626849090803, + "grad_norm": 15.536322593688965, + "learning_rate": 9.977913472781602e-07, + "loss": 1.0211, + "step": 7680 + }, + { + "epoch": 0.2997524800717223, + "grad_norm": 15.216032028198242, + "learning_rate": 9.990905547615954e-07, + "loss": 0.9752, + "step": 7690 + }, + { + "epoch": 0.3001422752343643, + "grad_norm": 13.376747131347656, + "learning_rate": 9.999999953713473e-07, + "loss": 0.952, + "step": 7700 + }, + { + "epoch": 0.30053207039700636, + "grad_norm": 16.160545349121094, + "learning_rate": 9.999999130841925e-07, + "loss": 0.9582, + "step": 7710 + }, + { + "epoch": 0.3009218655596484, + "grad_norm": 14.782362937927246, + "learning_rate": 9.999997279381107e-07, + "loss": 1.0373, + "step": 7720 + }, + { + "epoch": 0.30131166072229043, + "grad_norm": 14.41512393951416, + "learning_rate": 9.9999943993314e-07, + "loss": 0.9625, + "step": 7730 + }, + { + "epoch": 0.30170145588493247, + "grad_norm": 14.634366989135742, + "learning_rate": 9.999990490693394e-07, + "loss": 0.9834, + "step": 7740 + }, + { + "epoch": 0.3020912510475745, + "grad_norm": 13.554277420043945, + "learning_rate": 9.999985553467896e-07, + "loss": 1.0223, + "step": 7750 + }, + { + "epoch": 0.30248104621021654, + "grad_norm": 14.325937271118164, + "learning_rate": 9.999979587655922e-07, + "loss": 1.0079, + "step": 7760 + }, + { + "epoch": 0.3028708413728586, + "grad_norm": 19.090686798095703, + "learning_rate": 9.999972593258697e-07, + "loss": 0.9486, + "step": 7770 + }, + { + "epoch": 0.3032606365355006, + "grad_norm": 15.524620056152344, + "learning_rate": 9.99996457027766e-07, + "loss": 0.8955, + "step": 7780 + }, + { + "epoch": 0.30365043169814265, + "grad_norm": 14.402090072631836, + "learning_rate": 9.999955518714464e-07, + "loss": 1.0195, + "step": 7790 + }, + { + "epoch": 0.30404022686078463, + "grad_norm": 14.509196281433105, + "learning_rate": 9.99994543857097e-07, + "loss": 1.0106, + "step": 7800 + }, + { + "epoch": 0.30443002202342667, + "grad_norm": 14.979307174682617, + "learning_rate": 9.999934329849249e-07, + "loss": 0.9319, + "step": 7810 + }, + { + "epoch": 0.3048198171860687, + "grad_norm": 15.558209419250488, + "learning_rate": 9.999922192551591e-07, + "loss": 0.9178, + "step": 7820 + }, + { + "epoch": 0.30520961234871075, + "grad_norm": 14.977529525756836, + "learning_rate": 9.99990902668049e-07, + "loss": 1.0122, + "step": 7830 + }, + { + "epoch": 0.3055994075113528, + "grad_norm": 15.117538452148438, + "learning_rate": 9.999894832238653e-07, + "loss": 0.9873, + "step": 7840 + }, + { + "epoch": 0.3059892026739948, + "grad_norm": 17.25786590576172, + "learning_rate": 9.999879609229001e-07, + "loss": 0.9664, + "step": 7850 + }, + { + "epoch": 0.30637899783663686, + "grad_norm": 15.586503982543945, + "learning_rate": 9.999863357654669e-07, + "loss": 0.9352, + "step": 7860 + }, + { + "epoch": 0.3067687929992789, + "grad_norm": 12.832935333251953, + "learning_rate": 9.999846077518996e-07, + "loss": 1.0465, + "step": 7870 + }, + { + "epoch": 0.30715858816192093, + "grad_norm": 12.577997207641602, + "learning_rate": 9.999827768825539e-07, + "loss": 0.9358, + "step": 7880 + }, + { + "epoch": 0.30754838332456297, + "grad_norm": 16.512418746948242, + "learning_rate": 9.999808431578065e-07, + "loss": 0.9778, + "step": 7890 + }, + { + "epoch": 0.30793817848720495, + "grad_norm": 13.769205093383789, + "learning_rate": 9.99978806578055e-07, + "loss": 1.0036, + "step": 7900 + }, + { + "epoch": 0.308327973649847, + "grad_norm": 14.481090545654297, + "learning_rate": 9.999766671437183e-07, + "loss": 0.969, + "step": 7910 + }, + { + "epoch": 0.308717768812489, + "grad_norm": 21.055606842041016, + "learning_rate": 9.99974424855237e-07, + "loss": 0.9618, + "step": 7920 + }, + { + "epoch": 0.30910756397513106, + "grad_norm": 15.131681442260742, + "learning_rate": 9.99972079713072e-07, + "loss": 0.9871, + "step": 7930 + }, + { + "epoch": 0.3094973591377731, + "grad_norm": 16.22330093383789, + "learning_rate": 9.999696317177056e-07, + "loss": 0.9626, + "step": 7940 + }, + { + "epoch": 0.30988715430041514, + "grad_norm": 13.578225135803223, + "learning_rate": 9.999670808696417e-07, + "loss": 0.9459, + "step": 7950 + }, + { + "epoch": 0.3102769494630572, + "grad_norm": 12.928533554077148, + "learning_rate": 9.99964427169405e-07, + "loss": 0.8895, + "step": 7960 + }, + { + "epoch": 0.3106667446256992, + "grad_norm": 15.491499900817871, + "learning_rate": 9.999616706175413e-07, + "loss": 0.9802, + "step": 7970 + }, + { + "epoch": 0.31105653978834125, + "grad_norm": 15.09556770324707, + "learning_rate": 9.999588112146178e-07, + "loss": 0.9507, + "step": 7980 + }, + { + "epoch": 0.3114463349509833, + "grad_norm": 15.550175666809082, + "learning_rate": 9.999558489612226e-07, + "loss": 0.9715, + "step": 7990 + }, + { + "epoch": 0.31183613011362527, + "grad_norm": 14.260639190673828, + "learning_rate": 9.99952783857965e-07, + "loss": 0.9342, + "step": 8000 + }, + { + "epoch": 0.31183613011362527, + "eval_loss": 0.9787659645080566, + "eval_runtime": 82.9474, + "eval_samples_per_second": 49.996, + "eval_steps_per_second": 6.257, + "step": 8000 + }, + { + "epoch": 0.3122259252762673, + "grad_norm": 13.445799827575684, + "learning_rate": 9.99949615905476e-07, + "loss": 0.9807, + "step": 8010 + }, + { + "epoch": 0.31261572043890934, + "grad_norm": 13.445015907287598, + "learning_rate": 9.999463451044066e-07, + "loss": 0.9963, + "step": 8020 + }, + { + "epoch": 0.3130055156015514, + "grad_norm": 15.397449493408203, + "learning_rate": 9.999429714554304e-07, + "loss": 0.9637, + "step": 8030 + }, + { + "epoch": 0.3133953107641934, + "grad_norm": 11.574400901794434, + "learning_rate": 9.99939494959241e-07, + "loss": 0.9412, + "step": 8040 + }, + { + "epoch": 0.31378510592683545, + "grad_norm": 15.374265670776367, + "learning_rate": 9.999359156165537e-07, + "loss": 0.9751, + "step": 8050 + }, + { + "epoch": 0.3141749010894775, + "grad_norm": 16.698551177978516, + "learning_rate": 9.999322334281047e-07, + "loss": 1.0106, + "step": 8060 + }, + { + "epoch": 0.3145646962521195, + "grad_norm": 14.370229721069336, + "learning_rate": 9.999284483946515e-07, + "loss": 0.9745, + "step": 8070 + }, + { + "epoch": 0.31495449141476156, + "grad_norm": 14.775843620300293, + "learning_rate": 9.999245605169731e-07, + "loss": 0.959, + "step": 8080 + }, + { + "epoch": 0.3153442865774036, + "grad_norm": 13.211127281188965, + "learning_rate": 9.99920569795869e-07, + "loss": 0.9598, + "step": 8090 + }, + { + "epoch": 0.3157340817400456, + "grad_norm": 12.708335876464844, + "learning_rate": 9.999164762321599e-07, + "loss": 0.9422, + "step": 8100 + }, + { + "epoch": 0.3161238769026876, + "grad_norm": 16.460819244384766, + "learning_rate": 9.999122798266884e-07, + "loss": 1.0382, + "step": 8110 + }, + { + "epoch": 0.31651367206532965, + "grad_norm": 14.602065086364746, + "learning_rate": 9.999079805803176e-07, + "loss": 1.0374, + "step": 8120 + }, + { + "epoch": 0.3169034672279717, + "grad_norm": 15.431829452514648, + "learning_rate": 9.99903578493932e-07, + "loss": 0.9789, + "step": 8130 + }, + { + "epoch": 0.31729326239061373, + "grad_norm": 12.524334907531738, + "learning_rate": 9.998990735684371e-07, + "loss": 0.9649, + "step": 8140 + }, + { + "epoch": 0.31768305755325577, + "grad_norm": 13.682332038879395, + "learning_rate": 9.998944658047597e-07, + "loss": 0.9634, + "step": 8150 + }, + { + "epoch": 0.3180728527158978, + "grad_norm": 13.69578742980957, + "learning_rate": 9.998897552038477e-07, + "loss": 0.9473, + "step": 8160 + }, + { + "epoch": 0.31846264787853984, + "grad_norm": 13.977960586547852, + "learning_rate": 9.9988494176667e-07, + "loss": 1.0084, + "step": 8170 + }, + { + "epoch": 0.3188524430411819, + "grad_norm": 17.768552780151367, + "learning_rate": 9.998800254942168e-07, + "loss": 1.014, + "step": 8180 + }, + { + "epoch": 0.3192422382038239, + "grad_norm": 15.443312644958496, + "learning_rate": 9.998750063875e-07, + "loss": 0.9787, + "step": 8190 + }, + { + "epoch": 0.3196320333664659, + "grad_norm": 14.104635238647461, + "learning_rate": 9.998698844475512e-07, + "loss": 1.0096, + "step": 8200 + }, + { + "epoch": 0.32002182852910793, + "grad_norm": 13.37528133392334, + "learning_rate": 9.99864659675425e-07, + "loss": 0.9613, + "step": 8210 + }, + { + "epoch": 0.32041162369174997, + "grad_norm": 14.541037559509277, + "learning_rate": 9.998593320721956e-07, + "loss": 1.0095, + "step": 8220 + }, + { + "epoch": 0.320801418854392, + "grad_norm": 18.974815368652344, + "learning_rate": 9.998539016389594e-07, + "loss": 0.9917, + "step": 8230 + }, + { + "epoch": 0.32119121401703404, + "grad_norm": 14.710959434509277, + "learning_rate": 9.998483683768332e-07, + "loss": 0.9429, + "step": 8240 + }, + { + "epoch": 0.3215810091796761, + "grad_norm": 11.879108428955078, + "learning_rate": 9.998427322869554e-07, + "loss": 0.9212, + "step": 8250 + }, + { + "epoch": 0.3219708043423181, + "grad_norm": 12.84335994720459, + "learning_rate": 9.998369933704856e-07, + "loss": 0.9728, + "step": 8260 + }, + { + "epoch": 0.32236059950496015, + "grad_norm": 15.80157470703125, + "learning_rate": 9.99831151628604e-07, + "loss": 0.9957, + "step": 8270 + }, + { + "epoch": 0.3227503946676022, + "grad_norm": 16.491865158081055, + "learning_rate": 9.998252070625127e-07, + "loss": 0.9828, + "step": 8280 + }, + { + "epoch": 0.32314018983024423, + "grad_norm": 13.716316223144531, + "learning_rate": 9.99819159673435e-07, + "loss": 1.0037, + "step": 8290 + }, + { + "epoch": 0.32352998499288627, + "grad_norm": 15.012218475341797, + "learning_rate": 9.998130094626138e-07, + "loss": 1.013, + "step": 8300 + }, + { + "epoch": 0.32391978015552825, + "grad_norm": 15.611296653747559, + "learning_rate": 9.998067564313154e-07, + "loss": 0.9627, + "step": 8310 + }, + { + "epoch": 0.3243095753181703, + "grad_norm": 14.732034683227539, + "learning_rate": 9.998004005808256e-07, + "loss": 0.9595, + "step": 8320 + }, + { + "epoch": 0.3246993704808123, + "grad_norm": 15.24380111694336, + "learning_rate": 9.997939419124521e-07, + "loss": 0.9556, + "step": 8330 + }, + { + "epoch": 0.32508916564345436, + "grad_norm": 16.97905731201172, + "learning_rate": 9.997873804275237e-07, + "loss": 0.941, + "step": 8340 + }, + { + "epoch": 0.3254789608060964, + "grad_norm": 18.051794052124023, + "learning_rate": 9.997807161273898e-07, + "loss": 0.9306, + "step": 8350 + }, + { + "epoch": 0.32586875596873843, + "grad_norm": 15.934016227722168, + "learning_rate": 9.997739490134217e-07, + "loss": 0.9751, + "step": 8360 + }, + { + "epoch": 0.32625855113138047, + "grad_norm": 14.868292808532715, + "learning_rate": 9.997670790870115e-07, + "loss": 0.9869, + "step": 8370 + }, + { + "epoch": 0.3266483462940225, + "grad_norm": 12.716612815856934, + "learning_rate": 9.997601063495723e-07, + "loss": 0.9923, + "step": 8380 + }, + { + "epoch": 0.32703814145666454, + "grad_norm": 16.77759552001953, + "learning_rate": 9.997530308025385e-07, + "loss": 0.9941, + "step": 8390 + }, + { + "epoch": 0.3274279366193066, + "grad_norm": 14.488628387451172, + "learning_rate": 9.99745852447366e-07, + "loss": 0.9538, + "step": 8400 + }, + { + "epoch": 0.32781773178194856, + "grad_norm": 13.485611915588379, + "learning_rate": 9.99738571285531e-07, + "loss": 0.9728, + "step": 8410 + }, + { + "epoch": 0.3282075269445906, + "grad_norm": 14.429534912109375, + "learning_rate": 9.99731187318532e-07, + "loss": 0.9591, + "step": 8420 + }, + { + "epoch": 0.32859732210723264, + "grad_norm": 14.72870922088623, + "learning_rate": 9.997237005478875e-07, + "loss": 0.9809, + "step": 8430 + }, + { + "epoch": 0.3289871172698747, + "grad_norm": 14.956521034240723, + "learning_rate": 9.997161109751377e-07, + "loss": 0.9847, + "step": 8440 + }, + { + "epoch": 0.3293769124325167, + "grad_norm": 14.083251953125, + "learning_rate": 9.997084186018442e-07, + "loss": 0.8973, + "step": 8450 + }, + { + "epoch": 0.32976670759515875, + "grad_norm": 14.14371395111084, + "learning_rate": 9.997006234295892e-07, + "loss": 0.9734, + "step": 8460 + }, + { + "epoch": 0.3301565027578008, + "grad_norm": 13.728975296020508, + "learning_rate": 9.996927254599767e-07, + "loss": 0.886, + "step": 8470 + }, + { + "epoch": 0.3305462979204428, + "grad_norm": 13.465234756469727, + "learning_rate": 9.99684724694631e-07, + "loss": 1.0051, + "step": 8480 + }, + { + "epoch": 0.33093609308308486, + "grad_norm": 14.268718719482422, + "learning_rate": 9.996766211351978e-07, + "loss": 0.97, + "step": 8490 + }, + { + "epoch": 0.3313258882457269, + "grad_norm": 13.960322380065918, + "learning_rate": 9.996684147833449e-07, + "loss": 0.9641, + "step": 8500 + }, + { + "epoch": 0.3313258882457269, + "eval_loss": 0.9757557511329651, + "eval_runtime": 83.183, + "eval_samples_per_second": 49.854, + "eval_steps_per_second": 6.239, + "step": 8500 + }, + { + "epoch": 0.3317156834083689, + "grad_norm": 13.617558479309082, + "learning_rate": 9.9966010564076e-07, + "loss": 0.9692, + "step": 8510 + }, + { + "epoch": 0.3321054785710109, + "grad_norm": 12.174407958984375, + "learning_rate": 9.996516937091526e-07, + "loss": 0.9252, + "step": 8520 + }, + { + "epoch": 0.33249527373365295, + "grad_norm": 14.942967414855957, + "learning_rate": 9.996431789902532e-07, + "loss": 0.9313, + "step": 8530 + }, + { + "epoch": 0.332885068896295, + "grad_norm": 16.221342086791992, + "learning_rate": 9.996345614858132e-07, + "loss": 1.0465, + "step": 8540 + }, + { + "epoch": 0.333274864058937, + "grad_norm": 14.790725708007812, + "learning_rate": 9.996258411976055e-07, + "loss": 1.0032, + "step": 8550 + }, + { + "epoch": 0.33366465922157906, + "grad_norm": 11.233137130737305, + "learning_rate": 9.99617018127424e-07, + "loss": 0.9528, + "step": 8560 + }, + { + "epoch": 0.3340544543842211, + "grad_norm": 15.187173843383789, + "learning_rate": 9.996080922770842e-07, + "loss": 0.982, + "step": 8570 + }, + { + "epoch": 0.33444424954686314, + "grad_norm": 12.738338470458984, + "learning_rate": 9.995990636484215e-07, + "loss": 0.9692, + "step": 8580 + }, + { + "epoch": 0.3348340447095052, + "grad_norm": 16.24250602722168, + "learning_rate": 9.99589932243294e-07, + "loss": 1.0566, + "step": 8590 + }, + { + "epoch": 0.3352238398721472, + "grad_norm": 14.907780647277832, + "learning_rate": 9.995806980635798e-07, + "loss": 0.9145, + "step": 8600 + }, + { + "epoch": 0.3356136350347892, + "grad_norm": 13.177349090576172, + "learning_rate": 9.995713611111783e-07, + "loss": 0.9717, + "step": 8610 + }, + { + "epoch": 0.33600343019743123, + "grad_norm": 15.739975929260254, + "learning_rate": 9.99561921388011e-07, + "loss": 0.9405, + "step": 8620 + }, + { + "epoch": 0.33639322536007327, + "grad_norm": 15.47493839263916, + "learning_rate": 9.99552378896019e-07, + "loss": 1.005, + "step": 8630 + }, + { + "epoch": 0.3367830205227153, + "grad_norm": 13.329784393310547, + "learning_rate": 9.99542733637166e-07, + "loss": 0.9634, + "step": 8640 + }, + { + "epoch": 0.33717281568535734, + "grad_norm": 13.230236053466797, + "learning_rate": 9.995329856134357e-07, + "loss": 0.9659, + "step": 8650 + }, + { + "epoch": 0.3375626108479994, + "grad_norm": 14.535469055175781, + "learning_rate": 9.99523134826834e-07, + "loss": 1.0086, + "step": 8660 + }, + { + "epoch": 0.3379524060106414, + "grad_norm": 12.412528991699219, + "learning_rate": 9.99513181279387e-07, + "loss": 0.911, + "step": 8670 + }, + { + "epoch": 0.33834220117328345, + "grad_norm": 15.425732612609863, + "learning_rate": 9.995031249731424e-07, + "loss": 0.9691, + "step": 8680 + }, + { + "epoch": 0.3387319963359255, + "grad_norm": 12.85866928100586, + "learning_rate": 9.99492965910169e-07, + "loss": 0.9619, + "step": 8690 + }, + { + "epoch": 0.3391217914985675, + "grad_norm": 15.06373119354248, + "learning_rate": 9.994827040925566e-07, + "loss": 0.9745, + "step": 8700 + }, + { + "epoch": 0.3395115866612095, + "grad_norm": 13.791409492492676, + "learning_rate": 9.994723395224163e-07, + "loss": 0.9882, + "step": 8710 + }, + { + "epoch": 0.33990138182385154, + "grad_norm": 14.661155700683594, + "learning_rate": 9.994618722018804e-07, + "loss": 1.0326, + "step": 8720 + }, + { + "epoch": 0.3402911769864936, + "grad_norm": 14.333767890930176, + "learning_rate": 9.994513021331022e-07, + "loss": 0.8941, + "step": 8730 + }, + { + "epoch": 0.3406809721491356, + "grad_norm": 15.299776077270508, + "learning_rate": 9.994406293182557e-07, + "loss": 1.013, + "step": 8740 + }, + { + "epoch": 0.34107076731177766, + "grad_norm": 14.523314476013184, + "learning_rate": 9.99429853759537e-07, + "loss": 0.9065, + "step": 8750 + }, + { + "epoch": 0.3414605624744197, + "grad_norm": 16.357574462890625, + "learning_rate": 9.994189754591628e-07, + "loss": 0.9896, + "step": 8760 + }, + { + "epoch": 0.34185035763706173, + "grad_norm": 16.647777557373047, + "learning_rate": 9.99407994419371e-07, + "loss": 1.0347, + "step": 8770 + }, + { + "epoch": 0.34224015279970377, + "grad_norm": 12.452442169189453, + "learning_rate": 9.993969106424202e-07, + "loss": 0.9773, + "step": 8780 + }, + { + "epoch": 0.3426299479623458, + "grad_norm": 12.598489761352539, + "learning_rate": 9.993857241305907e-07, + "loss": 0.9206, + "step": 8790 + }, + { + "epoch": 0.34301974312498784, + "grad_norm": 14.094327926635742, + "learning_rate": 9.993744348861838e-07, + "loss": 1.0111, + "step": 8800 + }, + { + "epoch": 0.3434095382876298, + "grad_norm": 15.226766586303711, + "learning_rate": 9.993630429115221e-07, + "loss": 0.957, + "step": 8810 + }, + { + "epoch": 0.34379933345027186, + "grad_norm": 15.444640159606934, + "learning_rate": 9.99351548208949e-07, + "loss": 1.0718, + "step": 8820 + }, + { + "epoch": 0.3441891286129139, + "grad_norm": 14.77653980255127, + "learning_rate": 9.993399507808288e-07, + "loss": 0.8917, + "step": 8830 + }, + { + "epoch": 0.34457892377555593, + "grad_norm": 14.704705238342285, + "learning_rate": 9.99328250629548e-07, + "loss": 0.9254, + "step": 8840 + }, + { + "epoch": 0.34496871893819797, + "grad_norm": 14.59453296661377, + "learning_rate": 9.99316447757513e-07, + "loss": 0.9446, + "step": 8850 + }, + { + "epoch": 0.34535851410084, + "grad_norm": 15.816579818725586, + "learning_rate": 9.99304542167152e-07, + "loss": 1.0491, + "step": 8860 + }, + { + "epoch": 0.34574830926348205, + "grad_norm": 15.151015281677246, + "learning_rate": 9.992925338609141e-07, + "loss": 0.9291, + "step": 8870 + }, + { + "epoch": 0.3461381044261241, + "grad_norm": 17.87434959411621, + "learning_rate": 9.9928042284127e-07, + "loss": 0.9191, + "step": 8880 + }, + { + "epoch": 0.3465278995887661, + "grad_norm": 14.73003101348877, + "learning_rate": 9.992682091107105e-07, + "loss": 0.9407, + "step": 8890 + }, + { + "epoch": 0.34691769475140816, + "grad_norm": 14.431510925292969, + "learning_rate": 9.99255892671749e-07, + "loss": 0.9095, + "step": 8900 + }, + { + "epoch": 0.3473074899140502, + "grad_norm": 14.205792427062988, + "learning_rate": 9.992434735269184e-07, + "loss": 0.9513, + "step": 8910 + }, + { + "epoch": 0.3476972850766922, + "grad_norm": 11.669910430908203, + "learning_rate": 9.992309516787743e-07, + "loss": 0.9475, + "step": 8920 + }, + { + "epoch": 0.3480870802393342, + "grad_norm": 14.825511932373047, + "learning_rate": 9.992183271298919e-07, + "loss": 0.9201, + "step": 8930 + }, + { + "epoch": 0.34847687540197625, + "grad_norm": 16.329303741455078, + "learning_rate": 9.99205599882869e-07, + "loss": 0.9801, + "step": 8940 + }, + { + "epoch": 0.3488666705646183, + "grad_norm": 14.861445426940918, + "learning_rate": 9.991927699403233e-07, + "loss": 0.9852, + "step": 8950 + }, + { + "epoch": 0.3492564657272603, + "grad_norm": 13.008413314819336, + "learning_rate": 9.991798373048945e-07, + "loss": 0.868, + "step": 8960 + }, + { + "epoch": 0.34964626088990236, + "grad_norm": 13.724481582641602, + "learning_rate": 9.991668019792428e-07, + "loss": 0.9608, + "step": 8970 + }, + { + "epoch": 0.3500360560525444, + "grad_norm": 12.937424659729004, + "learning_rate": 9.9915366396605e-07, + "loss": 0.9906, + "step": 8980 + }, + { + "epoch": 0.35042585121518643, + "grad_norm": 13.875147819519043, + "learning_rate": 9.991404232680187e-07, + "loss": 1.0002, + "step": 8990 + }, + { + "epoch": 0.35081564637782847, + "grad_norm": 13.55489730834961, + "learning_rate": 9.99127079887873e-07, + "loss": 0.947, + "step": 9000 + }, + { + "epoch": 0.35081564637782847, + "eval_loss": 0.9687345027923584, + "eval_runtime": 82.7416, + "eval_samples_per_second": 50.12, + "eval_steps_per_second": 6.273, + "step": 9000 + }, + { + "epoch": 0.3512054415404705, + "grad_norm": 12.241427421569824, + "learning_rate": 9.991136338283575e-07, + "loss": 0.8885, + "step": 9010 + }, + { + "epoch": 0.3515952367031125, + "grad_norm": 16.693248748779297, + "learning_rate": 9.991000850922384e-07, + "loss": 0.9976, + "step": 9020 + }, + { + "epoch": 0.3519850318657545, + "grad_norm": 15.039878845214844, + "learning_rate": 9.990864336823032e-07, + "loss": 0.9757, + "step": 9030 + }, + { + "epoch": 0.35237482702839656, + "grad_norm": 12.595972061157227, + "learning_rate": 9.990726796013597e-07, + "loss": 0.9861, + "step": 9040 + }, + { + "epoch": 0.3527646221910386, + "grad_norm": 14.765259742736816, + "learning_rate": 9.99058822852238e-07, + "loss": 0.8942, + "step": 9050 + }, + { + "epoch": 0.35315441735368064, + "grad_norm": 11.957147598266602, + "learning_rate": 9.990448634377884e-07, + "loss": 0.896, + "step": 9060 + }, + { + "epoch": 0.3535442125163227, + "grad_norm": 12.93525505065918, + "learning_rate": 9.990308013608823e-07, + "loss": 0.9565, + "step": 9070 + }, + { + "epoch": 0.3539340076789647, + "grad_norm": 13.205789566040039, + "learning_rate": 9.99016636624413e-07, + "loss": 0.9426, + "step": 9080 + }, + { + "epoch": 0.35432380284160675, + "grad_norm": 14.83141803741455, + "learning_rate": 9.99002369231294e-07, + "loss": 0.9448, + "step": 9090 + }, + { + "epoch": 0.3547135980042488, + "grad_norm": 13.877195358276367, + "learning_rate": 9.989879991844608e-07, + "loss": 1.0607, + "step": 9100 + }, + { + "epoch": 0.3551033931668908, + "grad_norm": 14.515678405761719, + "learning_rate": 9.989735264868694e-07, + "loss": 0.9607, + "step": 9110 + }, + { + "epoch": 0.3554931883295328, + "grad_norm": 14.510612487792969, + "learning_rate": 9.98958951141497e-07, + "loss": 1.0048, + "step": 9120 + }, + { + "epoch": 0.35588298349217484, + "grad_norm": 14.210322380065918, + "learning_rate": 9.98944273151342e-07, + "loss": 0.9284, + "step": 9130 + }, + { + "epoch": 0.3562727786548169, + "grad_norm": 14.737910270690918, + "learning_rate": 9.98929492519424e-07, + "loss": 0.941, + "step": 9140 + }, + { + "epoch": 0.3566625738174589, + "grad_norm": 13.264534950256348, + "learning_rate": 9.989146092487838e-07, + "loss": 0.9803, + "step": 9150 + }, + { + "epoch": 0.35705236898010095, + "grad_norm": 12.593497276306152, + "learning_rate": 9.988996233424826e-07, + "loss": 1.0313, + "step": 9160 + }, + { + "epoch": 0.357442164142743, + "grad_norm": 14.141003608703613, + "learning_rate": 9.98884534803604e-07, + "loss": 0.9557, + "step": 9170 + }, + { + "epoch": 0.35783195930538503, + "grad_norm": 13.678252220153809, + "learning_rate": 9.988693436352516e-07, + "loss": 0.9182, + "step": 9180 + }, + { + "epoch": 0.35822175446802706, + "grad_norm": 18.1361083984375, + "learning_rate": 9.988540498405503e-07, + "loss": 0.9495, + "step": 9190 + }, + { + "epoch": 0.3586115496306691, + "grad_norm": 13.725064277648926, + "learning_rate": 9.988386534226468e-07, + "loss": 0.9606, + "step": 9200 + }, + { + "epoch": 0.35900134479331114, + "grad_norm": 14.00694465637207, + "learning_rate": 9.988231543847082e-07, + "loss": 1.0131, + "step": 9210 + }, + { + "epoch": 0.3593911399559531, + "grad_norm": 14.68053913116455, + "learning_rate": 9.988075527299227e-07, + "loss": 0.9741, + "step": 9220 + }, + { + "epoch": 0.35978093511859516, + "grad_norm": 16.625001907348633, + "learning_rate": 9.987918484615e-07, + "loss": 0.9718, + "step": 9230 + }, + { + "epoch": 0.3601707302812372, + "grad_norm": 13.806562423706055, + "learning_rate": 9.987760415826709e-07, + "loss": 1.0256, + "step": 9240 + }, + { + "epoch": 0.36056052544387923, + "grad_norm": 12.68728256225586, + "learning_rate": 9.98760132096687e-07, + "loss": 0.921, + "step": 9250 + }, + { + "epoch": 0.36095032060652127, + "grad_norm": 16.069509506225586, + "learning_rate": 9.987441200068212e-07, + "loss": 1.0067, + "step": 9260 + }, + { + "epoch": 0.3613401157691633, + "grad_norm": 14.894481658935547, + "learning_rate": 9.987280053163673e-07, + "loss": 0.9666, + "step": 9270 + }, + { + "epoch": 0.36172991093180534, + "grad_norm": 13.980961799621582, + "learning_rate": 9.987117880286408e-07, + "loss": 0.9585, + "step": 9280 + }, + { + "epoch": 0.3621197060944474, + "grad_norm": 12.78778076171875, + "learning_rate": 9.986954681469775e-07, + "loss": 0.9525, + "step": 9290 + }, + { + "epoch": 0.3625095012570894, + "grad_norm": 13.58409309387207, + "learning_rate": 9.98679045674735e-07, + "loss": 0.8048, + "step": 9300 + }, + { + "epoch": 0.36289929641973145, + "grad_norm": 14.478632926940918, + "learning_rate": 9.986625206152914e-07, + "loss": 0.9881, + "step": 9310 + }, + { + "epoch": 0.36328909158237344, + "grad_norm": 13.584160804748535, + "learning_rate": 9.986458929720462e-07, + "loss": 0.8847, + "step": 9320 + }, + { + "epoch": 0.3636788867450155, + "grad_norm": 14.3297700881958, + "learning_rate": 9.986291627484205e-07, + "loss": 0.9214, + "step": 9330 + }, + { + "epoch": 0.3640686819076575, + "grad_norm": 14.053335189819336, + "learning_rate": 9.986123299478553e-07, + "loss": 1.0325, + "step": 9340 + }, + { + "epoch": 0.36445847707029955, + "grad_norm": 14.288161277770996, + "learning_rate": 9.985953945738139e-07, + "loss": 0.9135, + "step": 9350 + }, + { + "epoch": 0.3648482722329416, + "grad_norm": 15.358881950378418, + "learning_rate": 9.9857835662978e-07, + "loss": 1.0004, + "step": 9360 + }, + { + "epoch": 0.3652380673955836, + "grad_norm": 14.120654106140137, + "learning_rate": 9.985612161192586e-07, + "loss": 0.9935, + "step": 9370 + }, + { + "epoch": 0.36562786255822566, + "grad_norm": 14.477173805236816, + "learning_rate": 9.98543973045776e-07, + "loss": 0.9894, + "step": 9380 + }, + { + "epoch": 0.3660176577208677, + "grad_norm": 15.811548233032227, + "learning_rate": 9.985266274128792e-07, + "loss": 0.952, + "step": 9390 + }, + { + "epoch": 0.36640745288350973, + "grad_norm": 14.643311500549316, + "learning_rate": 9.985091792241368e-07, + "loss": 1.0142, + "step": 9400 + }, + { + "epoch": 0.36679724804615177, + "grad_norm": 14.838760375976562, + "learning_rate": 9.984916284831378e-07, + "loss": 0.9831, + "step": 9410 + }, + { + "epoch": 0.3671870432087938, + "grad_norm": 16.193405151367188, + "learning_rate": 9.984739751934928e-07, + "loss": 0.9343, + "step": 9420 + }, + { + "epoch": 0.3675768383714358, + "grad_norm": 16.387239456176758, + "learning_rate": 9.984562193588338e-07, + "loss": 0.9643, + "step": 9430 + }, + { + "epoch": 0.3679666335340778, + "grad_norm": 13.336305618286133, + "learning_rate": 9.984383609828128e-07, + "loss": 0.9982, + "step": 9440 + }, + { + "epoch": 0.36835642869671986, + "grad_norm": 13.11186695098877, + "learning_rate": 9.984204000691043e-07, + "loss": 0.8814, + "step": 9450 + }, + { + "epoch": 0.3687462238593619, + "grad_norm": 15.601489067077637, + "learning_rate": 9.984023366214027e-07, + "loss": 0.9526, + "step": 9460 + }, + { + "epoch": 0.36913601902200394, + "grad_norm": 14.682417869567871, + "learning_rate": 9.983841706434242e-07, + "loss": 0.951, + "step": 9470 + }, + { + "epoch": 0.369525814184646, + "grad_norm": 13.17271614074707, + "learning_rate": 9.983659021389058e-07, + "loss": 0.9374, + "step": 9480 + }, + { + "epoch": 0.369915609347288, + "grad_norm": 15.31359577178955, + "learning_rate": 9.983475311116055e-07, + "loss": 0.9334, + "step": 9490 + }, + { + "epoch": 0.37030540450993005, + "grad_norm": 17.01848030090332, + "learning_rate": 9.98329057565303e-07, + "loss": 0.976, + "step": 9500 + }, + { + "epoch": 0.37030540450993005, + "eval_loss": 0.9654324650764465, + "eval_runtime": 82.4072, + "eval_samples_per_second": 50.323, + "eval_steps_per_second": 6.298, + "step": 9500 + }, + { + "epoch": 0.3706951996725721, + "grad_norm": 15.624834060668945, + "learning_rate": 9.98310481503798e-07, + "loss": 0.9104, + "step": 9510 + }, + { + "epoch": 0.3710849948352141, + "grad_norm": 15.820115089416504, + "learning_rate": 9.982918029309124e-07, + "loss": 0.9827, + "step": 9520 + }, + { + "epoch": 0.3714747899978561, + "grad_norm": 14.124943733215332, + "learning_rate": 9.982730218504887e-07, + "loss": 0.9361, + "step": 9530 + }, + { + "epoch": 0.37186458516049814, + "grad_norm": 15.758829116821289, + "learning_rate": 9.982541382663903e-07, + "loss": 0.916, + "step": 9540 + }, + { + "epoch": 0.3722543803231402, + "grad_norm": 14.149874687194824, + "learning_rate": 9.98235152182502e-07, + "loss": 0.9172, + "step": 9550 + }, + { + "epoch": 0.3726441754857822, + "grad_norm": 13.352356910705566, + "learning_rate": 9.982160636027295e-07, + "loss": 0.9054, + "step": 9560 + }, + { + "epoch": 0.37303397064842425, + "grad_norm": 14.38839340209961, + "learning_rate": 9.981968725309998e-07, + "loss": 0.9249, + "step": 9570 + }, + { + "epoch": 0.3734237658110663, + "grad_norm": 13.713641166687012, + "learning_rate": 9.981775789712607e-07, + "loss": 0.9337, + "step": 9580 + }, + { + "epoch": 0.3738135609737083, + "grad_norm": 13.85941219329834, + "learning_rate": 9.981581829274813e-07, + "loss": 0.9268, + "step": 9590 + }, + { + "epoch": 0.37420335613635036, + "grad_norm": 13.492119789123535, + "learning_rate": 9.981386844036518e-07, + "loss": 0.9942, + "step": 9600 + }, + { + "epoch": 0.3745931512989924, + "grad_norm": 12.114486694335938, + "learning_rate": 9.981190834037835e-07, + "loss": 0.9432, + "step": 9610 + }, + { + "epoch": 0.37498294646163444, + "grad_norm": 12.689262390136719, + "learning_rate": 9.980993799319082e-07, + "loss": 0.8769, + "step": 9620 + }, + { + "epoch": 0.3753727416242764, + "grad_norm": 13.092058181762695, + "learning_rate": 9.980795739920794e-07, + "loss": 0.9906, + "step": 9630 + }, + { + "epoch": 0.37576253678691846, + "grad_norm": 12.944218635559082, + "learning_rate": 9.980596655883719e-07, + "loss": 0.9218, + "step": 9640 + }, + { + "epoch": 0.3761523319495605, + "grad_norm": 13.189135551452637, + "learning_rate": 9.980396547248811e-07, + "loss": 0.8972, + "step": 9650 + }, + { + "epoch": 0.37654212711220253, + "grad_norm": 14.054845809936523, + "learning_rate": 9.980195414057233e-07, + "loss": 1.0178, + "step": 9660 + }, + { + "epoch": 0.37693192227484457, + "grad_norm": 15.273857116699219, + "learning_rate": 9.979993256350365e-07, + "loss": 0.971, + "step": 9670 + }, + { + "epoch": 0.3773217174374866, + "grad_norm": 14.276952743530273, + "learning_rate": 9.97979007416979e-07, + "loss": 0.9903, + "step": 9680 + }, + { + "epoch": 0.37771151260012864, + "grad_norm": 12.734455108642578, + "learning_rate": 9.979585867557312e-07, + "loss": 0.9516, + "step": 9690 + }, + { + "epoch": 0.3781013077627707, + "grad_norm": 12.649968147277832, + "learning_rate": 9.979380636554937e-07, + "loss": 0.9789, + "step": 9700 + }, + { + "epoch": 0.3784911029254127, + "grad_norm": 15.669553756713867, + "learning_rate": 9.979174381204885e-07, + "loss": 0.9856, + "step": 9710 + }, + { + "epoch": 0.37888089808805475, + "grad_norm": 14.540860176086426, + "learning_rate": 9.978967101549585e-07, + "loss": 1.0336, + "step": 9720 + }, + { + "epoch": 0.37927069325069673, + "grad_norm": 13.739781379699707, + "learning_rate": 9.97875879763168e-07, + "loss": 0.9948, + "step": 9730 + }, + { + "epoch": 0.37966048841333877, + "grad_norm": 14.167993545532227, + "learning_rate": 9.97854946949402e-07, + "loss": 0.9424, + "step": 9740 + }, + { + "epoch": 0.3800502835759808, + "grad_norm": 15.163949966430664, + "learning_rate": 9.97833911717967e-07, + "loss": 0.9993, + "step": 9750 + }, + { + "epoch": 0.38044007873862284, + "grad_norm": 13.907435417175293, + "learning_rate": 9.978127740731904e-07, + "loss": 0.9564, + "step": 9760 + }, + { + "epoch": 0.3808298739012649, + "grad_norm": 15.621562004089355, + "learning_rate": 9.977915340194201e-07, + "loss": 1.0071, + "step": 9770 + }, + { + "epoch": 0.3812196690639069, + "grad_norm": 14.23133659362793, + "learning_rate": 9.977701915610259e-07, + "loss": 0.9353, + "step": 9780 + }, + { + "epoch": 0.38160946422654896, + "grad_norm": 16.496192932128906, + "learning_rate": 9.977487467023985e-07, + "loss": 1.0564, + "step": 9790 + }, + { + "epoch": 0.381999259389191, + "grad_norm": 15.796478271484375, + "learning_rate": 9.97727199447949e-07, + "loss": 0.9597, + "step": 9800 + }, + { + "epoch": 0.38238905455183303, + "grad_norm": 13.357057571411133, + "learning_rate": 9.977055498021105e-07, + "loss": 1.0066, + "step": 9810 + }, + { + "epoch": 0.38277884971447507, + "grad_norm": 12.660501480102539, + "learning_rate": 9.976837977693365e-07, + "loss": 0.9585, + "step": 9820 + }, + { + "epoch": 0.38316864487711705, + "grad_norm": 14.466283798217773, + "learning_rate": 9.976619433541016e-07, + "loss": 0.9422, + "step": 9830 + }, + { + "epoch": 0.3835584400397591, + "grad_norm": 14.77971076965332, + "learning_rate": 9.97639986560902e-07, + "loss": 0.9635, + "step": 9840 + }, + { + "epoch": 0.3839482352024011, + "grad_norm": 14.263030052185059, + "learning_rate": 9.976179273942546e-07, + "loss": 0.9633, + "step": 9850 + }, + { + "epoch": 0.38433803036504316, + "grad_norm": 13.228436470031738, + "learning_rate": 9.975957658586972e-07, + "loss": 0.9334, + "step": 9860 + }, + { + "epoch": 0.3847278255276852, + "grad_norm": 13.441778182983398, + "learning_rate": 9.975735019587889e-07, + "loss": 0.925, + "step": 9870 + }, + { + "epoch": 0.38511762069032723, + "grad_norm": 13.325193405151367, + "learning_rate": 9.975511356991097e-07, + "loss": 0.9232, + "step": 9880 + }, + { + "epoch": 0.38550741585296927, + "grad_norm": 16.357784271240234, + "learning_rate": 9.975286670842608e-07, + "loss": 1.0189, + "step": 9890 + }, + { + "epoch": 0.3858972110156113, + "grad_norm": 12.7966890335083, + "learning_rate": 9.975060961188642e-07, + "loss": 0.9789, + "step": 9900 + }, + { + "epoch": 0.38628700617825334, + "grad_norm": 13.032859802246094, + "learning_rate": 9.974834228075635e-07, + "loss": 0.9884, + "step": 9910 + }, + { + "epoch": 0.3866768013408954, + "grad_norm": 14.634812355041504, + "learning_rate": 9.97460647155023e-07, + "loss": 1.0022, + "step": 9920 + }, + { + "epoch": 0.38706659650353736, + "grad_norm": 16.04472541809082, + "learning_rate": 9.974377691659277e-07, + "loss": 0.9152, + "step": 9930 + }, + { + "epoch": 0.3874563916661794, + "grad_norm": 13.594503402709961, + "learning_rate": 9.974147888449842e-07, + "loss": 0.9533, + "step": 9940 + }, + { + "epoch": 0.38784618682882144, + "grad_norm": 14.798115730285645, + "learning_rate": 9.9739170619692e-07, + "loss": 0.9426, + "step": 9950 + }, + { + "epoch": 0.3882359819914635, + "grad_norm": 13.394013404846191, + "learning_rate": 9.973685212264837e-07, + "loss": 0.9153, + "step": 9960 + }, + { + "epoch": 0.3886257771541055, + "grad_norm": 16.3243408203125, + "learning_rate": 9.973452339384446e-07, + "loss": 0.9739, + "step": 9970 + }, + { + "epoch": 0.38901557231674755, + "grad_norm": 16.747888565063477, + "learning_rate": 9.973218443375935e-07, + "loss": 0.9822, + "step": 9980 + }, + { + "epoch": 0.3894053674793896, + "grad_norm": 15.241726875305176, + "learning_rate": 9.972983524287422e-07, + "loss": 0.929, + "step": 9990 + }, + { + "epoch": 0.3897951626420316, + "grad_norm": 14.312605857849121, + "learning_rate": 9.97274758216723e-07, + "loss": 0.8805, + "step": 10000 + }, + { + "epoch": 0.3897951626420316, + "eval_loss": 0.9638563394546509, + "eval_runtime": 82.9394, + "eval_samples_per_second": 50.0, + "eval_steps_per_second": 6.258, + "step": 10000 + }, + { + "epoch": 0.39018495780467366, + "grad_norm": 14.787019729614258, + "learning_rate": 9.9725106170639e-07, + "loss": 0.9157, + "step": 10010 + }, + { + "epoch": 0.3905747529673157, + "grad_norm": 11.424470901489258, + "learning_rate": 9.972272629026177e-07, + "loss": 0.8969, + "step": 10020 + }, + { + "epoch": 0.39096454812995773, + "grad_norm": 17.929676055908203, + "learning_rate": 9.972033618103024e-07, + "loss": 0.9622, + "step": 10030 + }, + { + "epoch": 0.3913543432925997, + "grad_norm": 17.2790584564209, + "learning_rate": 9.971793584343604e-07, + "loss": 0.8987, + "step": 10040 + }, + { + "epoch": 0.39174413845524175, + "grad_norm": 15.211067199707031, + "learning_rate": 9.971552527797303e-07, + "loss": 0.9533, + "step": 10050 + }, + { + "epoch": 0.3921339336178838, + "grad_norm": 15.00230884552002, + "learning_rate": 9.971310448513703e-07, + "loss": 0.9514, + "step": 10060 + }, + { + "epoch": 0.3925237287805258, + "grad_norm": 14.557380676269531, + "learning_rate": 9.97106734654261e-07, + "loss": 0.9487, + "step": 10070 + }, + { + "epoch": 0.39291352394316786, + "grad_norm": 15.573948860168457, + "learning_rate": 9.970823221934032e-07, + "loss": 0.9595, + "step": 10080 + }, + { + "epoch": 0.3933033191058099, + "grad_norm": 14.664349555969238, + "learning_rate": 9.97057807473819e-07, + "loss": 0.9512, + "step": 10090 + }, + { + "epoch": 0.39369311426845194, + "grad_norm": 14.361614227294922, + "learning_rate": 9.970331905005516e-07, + "loss": 0.9514, + "step": 10100 + }, + { + "epoch": 0.394082909431094, + "grad_norm": 16.898963928222656, + "learning_rate": 9.970084712786648e-07, + "loss": 0.956, + "step": 10110 + }, + { + "epoch": 0.394472704593736, + "grad_norm": 13.458127975463867, + "learning_rate": 9.969836498132444e-07, + "loss": 0.9806, + "step": 10120 + }, + { + "epoch": 0.39486249975637805, + "grad_norm": 14.404616355895996, + "learning_rate": 9.969587261093961e-07, + "loss": 0.9708, + "step": 10130 + }, + { + "epoch": 0.39525229491902003, + "grad_norm": 15.40008544921875, + "learning_rate": 9.969337001722473e-07, + "loss": 0.9314, + "step": 10140 + }, + { + "epoch": 0.39564209008166207, + "grad_norm": 13.285988807678223, + "learning_rate": 9.969085720069466e-07, + "loss": 0.8682, + "step": 10150 + }, + { + "epoch": 0.3960318852443041, + "grad_norm": 14.519221305847168, + "learning_rate": 9.968833416186629e-07, + "loss": 0.963, + "step": 10160 + }, + { + "epoch": 0.39642168040694614, + "grad_norm": 15.401004791259766, + "learning_rate": 9.968580090125865e-07, + "loss": 0.9646, + "step": 10170 + }, + { + "epoch": 0.3968114755695882, + "grad_norm": 13.503737449645996, + "learning_rate": 9.968325741939292e-07, + "loss": 0.9071, + "step": 10180 + }, + { + "epoch": 0.3972012707322302, + "grad_norm": 13.58029556274414, + "learning_rate": 9.96807037167923e-07, + "loss": 0.9237, + "step": 10190 + }, + { + "epoch": 0.39759106589487225, + "grad_norm": 16.54905128479004, + "learning_rate": 9.967813979398215e-07, + "loss": 0.92, + "step": 10200 + }, + { + "epoch": 0.3979808610575143, + "grad_norm": 14.636281967163086, + "learning_rate": 9.96755656514899e-07, + "loss": 0.9703, + "step": 10210 + }, + { + "epoch": 0.3983706562201563, + "grad_norm": 14.45918083190918, + "learning_rate": 9.967298128984513e-07, + "loss": 0.9068, + "step": 10220 + }, + { + "epoch": 0.39876045138279836, + "grad_norm": 16.344938278198242, + "learning_rate": 9.967038670957946e-07, + "loss": 0.9453, + "step": 10230 + }, + { + "epoch": 0.39915024654544035, + "grad_norm": 13.735774993896484, + "learning_rate": 9.966778191122664e-07, + "loss": 0.9763, + "step": 10240 + }, + { + "epoch": 0.3995400417080824, + "grad_norm": 12.888359069824219, + "learning_rate": 9.966516689532254e-07, + "loss": 0.9478, + "step": 10250 + }, + { + "epoch": 0.3999298368707244, + "grad_norm": 15.625178337097168, + "learning_rate": 9.96625416624051e-07, + "loss": 0.9117, + "step": 10260 + }, + { + "epoch": 0.40031963203336646, + "grad_norm": 14.189743995666504, + "learning_rate": 9.965990621301442e-07, + "loss": 0.889, + "step": 10270 + }, + { + "epoch": 0.4007094271960085, + "grad_norm": 13.041755676269531, + "learning_rate": 9.965726054769259e-07, + "loss": 0.8884, + "step": 10280 + }, + { + "epoch": 0.40109922235865053, + "grad_norm": 12.908567428588867, + "learning_rate": 9.965460466698392e-07, + "loss": 0.959, + "step": 10290 + }, + { + "epoch": 0.40148901752129257, + "grad_norm": 15.133548736572266, + "learning_rate": 9.965193857143474e-07, + "loss": 0.9814, + "step": 10300 + }, + { + "epoch": 0.4018788126839346, + "grad_norm": 13.856285095214844, + "learning_rate": 9.964926226159355e-07, + "loss": 0.9562, + "step": 10310 + }, + { + "epoch": 0.40226860784657664, + "grad_norm": 16.81253433227539, + "learning_rate": 9.964657573801089e-07, + "loss": 0.8939, + "step": 10320 + }, + { + "epoch": 0.4026584030092187, + "grad_norm": 15.102206230163574, + "learning_rate": 9.964387900123944e-07, + "loss": 0.9231, + "step": 10330 + }, + { + "epoch": 0.40304819817186066, + "grad_norm": 15.81601619720459, + "learning_rate": 9.964117205183395e-07, + "loss": 0.9883, + "step": 10340 + }, + { + "epoch": 0.4034379933345027, + "grad_norm": 13.626924514770508, + "learning_rate": 9.963845489035133e-07, + "loss": 1.0306, + "step": 10350 + }, + { + "epoch": 0.40382778849714474, + "grad_norm": 16.1990909576416, + "learning_rate": 9.963572751735048e-07, + "loss": 0.9634, + "step": 10360 + }, + { + "epoch": 0.40421758365978677, + "grad_norm": 13.489509582519531, + "learning_rate": 9.96329899333925e-07, + "loss": 0.9215, + "step": 10370 + }, + { + "epoch": 0.4046073788224288, + "grad_norm": 10.89586353302002, + "learning_rate": 9.963024213904058e-07, + "loss": 0.9926, + "step": 10380 + }, + { + "epoch": 0.40499717398507085, + "grad_norm": 14.685070991516113, + "learning_rate": 9.962748413485998e-07, + "loss": 0.8489, + "step": 10390 + }, + { + "epoch": 0.4053869691477129, + "grad_norm": 13.77380084991455, + "learning_rate": 9.962471592141807e-07, + "loss": 0.9209, + "step": 10400 + }, + { + "epoch": 0.4057767643103549, + "grad_norm": 14.390551567077637, + "learning_rate": 9.962193749928432e-07, + "loss": 0.9609, + "step": 10410 + }, + { + "epoch": 0.40616655947299696, + "grad_norm": 15.700751304626465, + "learning_rate": 9.96191488690303e-07, + "loss": 0.9442, + "step": 10420 + }, + { + "epoch": 0.406556354635639, + "grad_norm": 12.823718070983887, + "learning_rate": 9.961635003122966e-07, + "loss": 0.9023, + "step": 10430 + }, + { + "epoch": 0.406946149798281, + "grad_norm": 16.8238525390625, + "learning_rate": 9.96135409864582e-07, + "loss": 0.8809, + "step": 10440 + }, + { + "epoch": 0.407335944960923, + "grad_norm": 15.842167854309082, + "learning_rate": 9.961072173529378e-07, + "loss": 0.9609, + "step": 10450 + }, + { + "epoch": 0.40772574012356505, + "grad_norm": 12.16648006439209, + "learning_rate": 9.960789227831638e-07, + "loss": 0.9388, + "step": 10460 + }, + { + "epoch": 0.4081155352862071, + "grad_norm": 12.95360279083252, + "learning_rate": 9.960505261610804e-07, + "loss": 0.9789, + "step": 10470 + }, + { + "epoch": 0.4085053304488491, + "grad_norm": 13.693960189819336, + "learning_rate": 9.960220274925298e-07, + "loss": 0.9165, + "step": 10480 + }, + { + "epoch": 0.40889512561149116, + "grad_norm": 16.123159408569336, + "learning_rate": 9.959934267833743e-07, + "loss": 0.9244, + "step": 10490 + }, + { + "epoch": 0.4092849207741332, + "grad_norm": 12.895342826843262, + "learning_rate": 9.959647240394978e-07, + "loss": 1.0095, + "step": 10500 + }, + { + "epoch": 0.4092849207741332, + "eval_loss": 0.9569562077522278, + "eval_runtime": 83.135, + "eval_samples_per_second": 49.883, + "eval_steps_per_second": 6.243, + "step": 10500 + }, + { + "epoch": 0.40967471593677524, + "grad_norm": 11.670733451843262, + "learning_rate": 9.959359192668046e-07, + "loss": 0.9031, + "step": 10510 + }, + { + "epoch": 0.4100645110994173, + "grad_norm": 14.076419830322266, + "learning_rate": 9.959070124712208e-07, + "loss": 0.9534, + "step": 10520 + }, + { + "epoch": 0.4104543062620593, + "grad_norm": 12.594572067260742, + "learning_rate": 9.958780036586926e-07, + "loss": 0.9119, + "step": 10530 + }, + { + "epoch": 0.41084410142470135, + "grad_norm": 14.766462326049805, + "learning_rate": 9.958488928351882e-07, + "loss": 0.9273, + "step": 10540 + }, + { + "epoch": 0.41123389658734333, + "grad_norm": 14.75191593170166, + "learning_rate": 9.958196800066957e-07, + "loss": 0.9355, + "step": 10550 + }, + { + "epoch": 0.41162369174998537, + "grad_norm": 12.33820915222168, + "learning_rate": 9.95790365179225e-07, + "loss": 0.9138, + "step": 10560 + }, + { + "epoch": 0.4120134869126274, + "grad_norm": 15.523260116577148, + "learning_rate": 9.957609483588066e-07, + "loss": 0.944, + "step": 10570 + }, + { + "epoch": 0.41240328207526944, + "grad_norm": 13.736638069152832, + "learning_rate": 9.957314295514917e-07, + "loss": 0.9571, + "step": 10580 + }, + { + "epoch": 0.4127930772379115, + "grad_norm": 12.40423583984375, + "learning_rate": 9.957018087633536e-07, + "loss": 0.9624, + "step": 10590 + }, + { + "epoch": 0.4131828724005535, + "grad_norm": 14.28355598449707, + "learning_rate": 9.956720860004854e-07, + "loss": 0.9293, + "step": 10600 + }, + { + "epoch": 0.41357266756319555, + "grad_norm": 13.682143211364746, + "learning_rate": 9.956422612690015e-07, + "loss": 0.9792, + "step": 10610 + }, + { + "epoch": 0.4139624627258376, + "grad_norm": 11.843524932861328, + "learning_rate": 9.956123345750376e-07, + "loss": 0.9771, + "step": 10620 + }, + { + "epoch": 0.4143522578884796, + "grad_norm": 14.575690269470215, + "learning_rate": 9.955823059247498e-07, + "loss": 1.0649, + "step": 10630 + }, + { + "epoch": 0.41474205305112166, + "grad_norm": 12.649767875671387, + "learning_rate": 9.95552175324316e-07, + "loss": 0.9358, + "step": 10640 + }, + { + "epoch": 0.41513184821376364, + "grad_norm": 14.98377513885498, + "learning_rate": 9.955219427799346e-07, + "loss": 0.9986, + "step": 10650 + }, + { + "epoch": 0.4155216433764057, + "grad_norm": 12.254772186279297, + "learning_rate": 9.954916082978245e-07, + "loss": 0.8712, + "step": 10660 + }, + { + "epoch": 0.4159114385390477, + "grad_norm": 15.196223258972168, + "learning_rate": 9.954611718842265e-07, + "loss": 1.0354, + "step": 10670 + }, + { + "epoch": 0.41630123370168975, + "grad_norm": 15.497309684753418, + "learning_rate": 9.954306335454014e-07, + "loss": 0.9578, + "step": 10680 + }, + { + "epoch": 0.4166910288643318, + "grad_norm": 13.544538497924805, + "learning_rate": 9.953999932876322e-07, + "loss": 0.9333, + "step": 10690 + }, + { + "epoch": 0.41708082402697383, + "grad_norm": 12.884254455566406, + "learning_rate": 9.953692511172215e-07, + "loss": 0.968, + "step": 10700 + }, + { + "epoch": 0.41747061918961587, + "grad_norm": 14.25124454498291, + "learning_rate": 9.95338407040494e-07, + "loss": 0.9236, + "step": 10710 + }, + { + "epoch": 0.4178604143522579, + "grad_norm": 13.511805534362793, + "learning_rate": 9.953074610637946e-07, + "loss": 0.9523, + "step": 10720 + }, + { + "epoch": 0.41825020951489994, + "grad_norm": 15.67548656463623, + "learning_rate": 9.952764131934893e-07, + "loss": 0.9841, + "step": 10730 + }, + { + "epoch": 0.418640004677542, + "grad_norm": 14.852190971374512, + "learning_rate": 9.952452634359655e-07, + "loss": 0.9489, + "step": 10740 + }, + { + "epoch": 0.41902979984018396, + "grad_norm": 13.081192016601562, + "learning_rate": 9.952140117976314e-07, + "loss": 0.9682, + "step": 10750 + }, + { + "epoch": 0.419419595002826, + "grad_norm": 15.498686790466309, + "learning_rate": 9.951826582849154e-07, + "loss": 0.9596, + "step": 10760 + }, + { + "epoch": 0.41980939016546803, + "grad_norm": 18.637100219726562, + "learning_rate": 9.95151202904268e-07, + "loss": 1.0012, + "step": 10770 + }, + { + "epoch": 0.42019918532811007, + "grad_norm": 15.94414234161377, + "learning_rate": 9.951196456621602e-07, + "loss": 0.9112, + "step": 10780 + }, + { + "epoch": 0.4205889804907521, + "grad_norm": 16.570037841796875, + "learning_rate": 9.950879865650835e-07, + "loss": 0.959, + "step": 10790 + }, + { + "epoch": 0.42097877565339414, + "grad_norm": 14.854084968566895, + "learning_rate": 9.95056225619551e-07, + "loss": 0.9334, + "step": 10800 + }, + { + "epoch": 0.4213685708160362, + "grad_norm": 19.047744750976562, + "learning_rate": 9.950243628320964e-07, + "loss": 1.0121, + "step": 10810 + }, + { + "epoch": 0.4217583659786782, + "grad_norm": 14.521344184875488, + "learning_rate": 9.949923982092745e-07, + "loss": 0.9224, + "step": 10820 + }, + { + "epoch": 0.42214816114132026, + "grad_norm": 16.81005096435547, + "learning_rate": 9.949603317576608e-07, + "loss": 0.8822, + "step": 10830 + }, + { + "epoch": 0.4225379563039623, + "grad_norm": 13.175339698791504, + "learning_rate": 9.949281634838523e-07, + "loss": 0.8778, + "step": 10840 + }, + { + "epoch": 0.4229277514666043, + "grad_norm": 14.644318580627441, + "learning_rate": 9.948958933944662e-07, + "loss": 1.0214, + "step": 10850 + }, + { + "epoch": 0.4233175466292463, + "grad_norm": 14.797213554382324, + "learning_rate": 9.948635214961414e-07, + "loss": 0.9164, + "step": 10860 + }, + { + "epoch": 0.42370734179188835, + "grad_norm": 16.305530548095703, + "learning_rate": 9.948310477955372e-07, + "loss": 0.9902, + "step": 10870 + }, + { + "epoch": 0.4240971369545304, + "grad_norm": 12.49099063873291, + "learning_rate": 9.947984722993342e-07, + "loss": 0.9071, + "step": 10880 + }, + { + "epoch": 0.4244869321171724, + "grad_norm": 14.17330551147461, + "learning_rate": 9.947657950142332e-07, + "loss": 0.9282, + "step": 10890 + }, + { + "epoch": 0.42487672727981446, + "grad_norm": 14.673171043395996, + "learning_rate": 9.947330159469572e-07, + "loss": 0.941, + "step": 10900 + }, + { + "epoch": 0.4252665224424565, + "grad_norm": 13.55733871459961, + "learning_rate": 9.94700135104249e-07, + "loss": 0.9732, + "step": 10910 + }, + { + "epoch": 0.42565631760509853, + "grad_norm": 14.626479148864746, + "learning_rate": 9.946671524928728e-07, + "loss": 0.9086, + "step": 10920 + }, + { + "epoch": 0.42604611276774057, + "grad_norm": 13.657137870788574, + "learning_rate": 9.94634068119614e-07, + "loss": 0.9427, + "step": 10930 + }, + { + "epoch": 0.4264359079303826, + "grad_norm": 13.971735000610352, + "learning_rate": 9.946008819912785e-07, + "loss": 0.9926, + "step": 10940 + }, + { + "epoch": 0.4268257030930246, + "grad_norm": 13.487635612487793, + "learning_rate": 9.945675941146933e-07, + "loss": 0.9972, + "step": 10950 + }, + { + "epoch": 0.4272154982556666, + "grad_norm": 13.858904838562012, + "learning_rate": 9.945342044967062e-07, + "loss": 0.9386, + "step": 10960 + }, + { + "epoch": 0.42760529341830866, + "grad_norm": 14.918105125427246, + "learning_rate": 9.94500713144186e-07, + "loss": 0.8961, + "step": 10970 + }, + { + "epoch": 0.4279950885809507, + "grad_norm": 11.97497272491455, + "learning_rate": 9.944671200640227e-07, + "loss": 0.9735, + "step": 10980 + }, + { + "epoch": 0.42838488374359274, + "grad_norm": 15.019547462463379, + "learning_rate": 9.944334252631267e-07, + "loss": 0.9505, + "step": 10990 + }, + { + "epoch": 0.4287746789062348, + "grad_norm": 13.840620994567871, + "learning_rate": 9.9439962874843e-07, + "loss": 0.9331, + "step": 11000 + }, + { + "epoch": 0.4287746789062348, + "eval_loss": 0.9566273093223572, + "eval_runtime": 82.9466, + "eval_samples_per_second": 49.996, + "eval_steps_per_second": 6.257, + "step": 11000 + }, + { + "epoch": 0.4291644740688768, + "grad_norm": 14.678706169128418, + "learning_rate": 9.94365730526885e-07, + "loss": 0.8712, + "step": 11010 + }, + { + "epoch": 0.42955426923151885, + "grad_norm": 11.668639183044434, + "learning_rate": 9.943317306054647e-07, + "loss": 0.9444, + "step": 11020 + }, + { + "epoch": 0.4299440643941609, + "grad_norm": 13.721612930297852, + "learning_rate": 9.942976289911645e-07, + "loss": 0.9226, + "step": 11030 + }, + { + "epoch": 0.4303338595568029, + "grad_norm": 14.334944725036621, + "learning_rate": 9.942634256909988e-07, + "loss": 0.9056, + "step": 11040 + }, + { + "epoch": 0.43072365471944496, + "grad_norm": 13.231597900390625, + "learning_rate": 9.942291207120042e-07, + "loss": 0.9252, + "step": 11050 + }, + { + "epoch": 0.43111344988208694, + "grad_norm": 15.139326095581055, + "learning_rate": 9.941947140612378e-07, + "loss": 1.0025, + "step": 11060 + }, + { + "epoch": 0.431503245044729, + "grad_norm": 14.79094123840332, + "learning_rate": 9.941602057457777e-07, + "loss": 0.9344, + "step": 11070 + }, + { + "epoch": 0.431893040207371, + "grad_norm": 12.884871482849121, + "learning_rate": 9.941255957727227e-07, + "loss": 0.8296, + "step": 11080 + }, + { + "epoch": 0.43228283537001305, + "grad_norm": 14.696297645568848, + "learning_rate": 9.94090884149193e-07, + "loss": 0.9405, + "step": 11090 + }, + { + "epoch": 0.4326726305326551, + "grad_norm": 16.801464080810547, + "learning_rate": 9.94056070882329e-07, + "loss": 0.9565, + "step": 11100 + }, + { + "epoch": 0.4330624256952971, + "grad_norm": 13.599350929260254, + "learning_rate": 9.940211559792928e-07, + "loss": 0.9648, + "step": 11110 + }, + { + "epoch": 0.43345222085793916, + "grad_norm": 15.645402908325195, + "learning_rate": 9.93986139447267e-07, + "loss": 0.9192, + "step": 11120 + }, + { + "epoch": 0.4338420160205812, + "grad_norm": 13.688250541687012, + "learning_rate": 9.939510212934548e-07, + "loss": 0.9679, + "step": 11130 + }, + { + "epoch": 0.43423181118322324, + "grad_norm": 15.481486320495605, + "learning_rate": 9.939158015250808e-07, + "loss": 0.9394, + "step": 11140 + }, + { + "epoch": 0.4346216063458653, + "grad_norm": 15.153063774108887, + "learning_rate": 9.938804801493906e-07, + "loss": 0.9251, + "step": 11150 + }, + { + "epoch": 0.43501140150850726, + "grad_norm": 12.817085266113281, + "learning_rate": 9.9384505717365e-07, + "loss": 0.9418, + "step": 11160 + }, + { + "epoch": 0.4354011966711493, + "grad_norm": 14.47064208984375, + "learning_rate": 9.938095326051464e-07, + "loss": 0.8912, + "step": 11170 + }, + { + "epoch": 0.43579099183379133, + "grad_norm": 14.34460163116455, + "learning_rate": 9.937739064511879e-07, + "loss": 0.8733, + "step": 11180 + }, + { + "epoch": 0.43618078699643337, + "grad_norm": 12.969103813171387, + "learning_rate": 9.937381787191031e-07, + "loss": 0.9871, + "step": 11190 + }, + { + "epoch": 0.4365705821590754, + "grad_norm": 13.079099655151367, + "learning_rate": 9.937023494162422e-07, + "loss": 1.0151, + "step": 11200 + }, + { + "epoch": 0.43696037732171744, + "grad_norm": 13.834623336791992, + "learning_rate": 9.936664185499756e-07, + "loss": 0.9548, + "step": 11210 + }, + { + "epoch": 0.4373501724843595, + "grad_norm": 12.036235809326172, + "learning_rate": 9.936303861276953e-07, + "loss": 0.9219, + "step": 11220 + }, + { + "epoch": 0.4377399676470015, + "grad_norm": 11.52295970916748, + "learning_rate": 9.935942521568135e-07, + "loss": 0.9439, + "step": 11230 + }, + { + "epoch": 0.43812976280964355, + "grad_norm": 14.451292037963867, + "learning_rate": 9.935580166447639e-07, + "loss": 0.9319, + "step": 11240 + }, + { + "epoch": 0.4385195579722856, + "grad_norm": 14.200115203857422, + "learning_rate": 9.935216795990004e-07, + "loss": 0.9379, + "step": 11250 + }, + { + "epoch": 0.43890935313492757, + "grad_norm": 15.42430591583252, + "learning_rate": 9.934852410269986e-07, + "loss": 0.9074, + "step": 11260 + }, + { + "epoch": 0.4392991482975696, + "grad_norm": 13.139900207519531, + "learning_rate": 9.93448700936254e-07, + "loss": 0.9148, + "step": 11270 + }, + { + "epoch": 0.43968894346021165, + "grad_norm": 14.296747207641602, + "learning_rate": 9.934120593342842e-07, + "loss": 0.9193, + "step": 11280 + }, + { + "epoch": 0.4400787386228537, + "grad_norm": 14.010895729064941, + "learning_rate": 9.933753162286266e-07, + "loss": 0.9275, + "step": 11290 + }, + { + "epoch": 0.4404685337854957, + "grad_norm": 12.252301216125488, + "learning_rate": 9.933384716268402e-07, + "loss": 0.9191, + "step": 11300 + }, + { + "epoch": 0.44085832894813776, + "grad_norm": 13.19527530670166, + "learning_rate": 9.933015255365045e-07, + "loss": 0.8789, + "step": 11310 + }, + { + "epoch": 0.4412481241107798, + "grad_norm": 14.119636535644531, + "learning_rate": 9.932644779652197e-07, + "loss": 0.9777, + "step": 11320 + }, + { + "epoch": 0.44163791927342183, + "grad_norm": 12.882107734680176, + "learning_rate": 9.932273289206076e-07, + "loss": 1.0201, + "step": 11330 + }, + { + "epoch": 0.44202771443606387, + "grad_norm": 14.398632049560547, + "learning_rate": 9.9319007841031e-07, + "loss": 0.9401, + "step": 11340 + }, + { + "epoch": 0.4424175095987059, + "grad_norm": 14.666001319885254, + "learning_rate": 9.931527264419904e-07, + "loss": 0.9522, + "step": 11350 + }, + { + "epoch": 0.4428073047613479, + "grad_norm": 14.3677978515625, + "learning_rate": 9.931152730233325e-07, + "loss": 0.9283, + "step": 11360 + }, + { + "epoch": 0.4431970999239899, + "grad_norm": 15.663538932800293, + "learning_rate": 9.930777181620411e-07, + "loss": 0.9197, + "step": 11370 + }, + { + "epoch": 0.44358689508663196, + "grad_norm": 12.921825408935547, + "learning_rate": 9.93040061865842e-07, + "loss": 1.0132, + "step": 11380 + }, + { + "epoch": 0.443976690249274, + "grad_norm": 18.878551483154297, + "learning_rate": 9.930023041424817e-07, + "loss": 0.9372, + "step": 11390 + }, + { + "epoch": 0.44436648541191603, + "grad_norm": 14.169268608093262, + "learning_rate": 9.929644449997278e-07, + "loss": 0.8902, + "step": 11400 + }, + { + "epoch": 0.44475628057455807, + "grad_norm": 16.763944625854492, + "learning_rate": 9.929264844453686e-07, + "loss": 0.9162, + "step": 11410 + }, + { + "epoch": 0.4451460757372001, + "grad_norm": 15.66810417175293, + "learning_rate": 9.928884224872131e-07, + "loss": 0.9117, + "step": 11420 + }, + { + "epoch": 0.44553587089984215, + "grad_norm": 14.397140502929688, + "learning_rate": 9.928502591330917e-07, + "loss": 1.0235, + "step": 11430 + }, + { + "epoch": 0.4459256660624842, + "grad_norm": 12.73609733581543, + "learning_rate": 9.928119943908546e-07, + "loss": 0.8993, + "step": 11440 + }, + { + "epoch": 0.4463154612251262, + "grad_norm": 15.128432273864746, + "learning_rate": 9.92773628268374e-07, + "loss": 0.8694, + "step": 11450 + }, + { + "epoch": 0.4467052563877682, + "grad_norm": 11.959393501281738, + "learning_rate": 9.927351607735426e-07, + "loss": 0.9358, + "step": 11460 + }, + { + "epoch": 0.44709505155041024, + "grad_norm": 14.655324935913086, + "learning_rate": 9.926965919142737e-07, + "loss": 0.9507, + "step": 11470 + }, + { + "epoch": 0.4474848467130523, + "grad_norm": 10.762869834899902, + "learning_rate": 9.926579216985015e-07, + "loss": 0.8954, + "step": 11480 + }, + { + "epoch": 0.4478746418756943, + "grad_norm": 13.78454875946045, + "learning_rate": 9.926191501341813e-07, + "loss": 0.9709, + "step": 11490 + }, + { + "epoch": 0.44826443703833635, + "grad_norm": 14.93838119506836, + "learning_rate": 9.925802772292888e-07, + "loss": 0.9548, + "step": 11500 + }, + { + "epoch": 0.44826443703833635, + "eval_loss": 0.9491176605224609, + "eval_runtime": 83.0185, + "eval_samples_per_second": 49.953, + "eval_steps_per_second": 6.252, + "step": 11500 + }, + { + "epoch": 0.4486542322009784, + "grad_norm": 12.087103843688965, + "learning_rate": 9.925413029918213e-07, + "loss": 0.9188, + "step": 11510 + }, + { + "epoch": 0.4490440273636204, + "grad_norm": 13.578605651855469, + "learning_rate": 9.925022274297963e-07, + "loss": 0.9541, + "step": 11520 + }, + { + "epoch": 0.44943382252626246, + "grad_norm": 13.897317886352539, + "learning_rate": 9.924630505512523e-07, + "loss": 0.8494, + "step": 11530 + }, + { + "epoch": 0.4498236176889045, + "grad_norm": 13.473687171936035, + "learning_rate": 9.924237723642488e-07, + "loss": 0.9675, + "step": 11540 + }, + { + "epoch": 0.45021341285154653, + "grad_norm": 16.902502059936523, + "learning_rate": 9.92384392876866e-07, + "loss": 0.8753, + "step": 11550 + }, + { + "epoch": 0.4506032080141885, + "grad_norm": 13.871731758117676, + "learning_rate": 9.923449120972047e-07, + "loss": 0.9997, + "step": 11560 + }, + { + "epoch": 0.45099300317683055, + "grad_norm": 13.957497596740723, + "learning_rate": 9.923053300333872e-07, + "loss": 0.9537, + "step": 11570 + }, + { + "epoch": 0.4513827983394726, + "grad_norm": 14.42354679107666, + "learning_rate": 9.92265646693556e-07, + "loss": 0.9377, + "step": 11580 + }, + { + "epoch": 0.4517725935021146, + "grad_norm": 12.7905855178833, + "learning_rate": 9.922258620858748e-07, + "loss": 0.9222, + "step": 11590 + }, + { + "epoch": 0.45216238866475666, + "grad_norm": 14.887276649475098, + "learning_rate": 9.921859762185279e-07, + "loss": 0.9275, + "step": 11600 + }, + { + "epoch": 0.4525521838273987, + "grad_norm": 14.43886661529541, + "learning_rate": 9.921459890997206e-07, + "loss": 0.9585, + "step": 11610 + }, + { + "epoch": 0.45294197899004074, + "grad_norm": 15.415182113647461, + "learning_rate": 9.921059007376789e-07, + "loss": 0.9437, + "step": 11620 + }, + { + "epoch": 0.4533317741526828, + "grad_norm": 17.12296485900879, + "learning_rate": 9.920657111406496e-07, + "loss": 1.008, + "step": 11630 + }, + { + "epoch": 0.4537215693153248, + "grad_norm": 13.645906448364258, + "learning_rate": 9.920254203169007e-07, + "loss": 0.9561, + "step": 11640 + }, + { + "epoch": 0.45411136447796685, + "grad_norm": 14.468920707702637, + "learning_rate": 9.919850282747206e-07, + "loss": 0.9503, + "step": 11650 + }, + { + "epoch": 0.4545011596406089, + "grad_norm": 15.490865707397461, + "learning_rate": 9.919445350224187e-07, + "loss": 1.0126, + "step": 11660 + }, + { + "epoch": 0.45489095480325087, + "grad_norm": 12.488776206970215, + "learning_rate": 9.91903940568325e-07, + "loss": 0.9835, + "step": 11670 + }, + { + "epoch": 0.4552807499658929, + "grad_norm": 16.798051834106445, + "learning_rate": 9.91863244920791e-07, + "loss": 0.9587, + "step": 11680 + }, + { + "epoch": 0.45567054512853494, + "grad_norm": 15.690866470336914, + "learning_rate": 9.918224480881879e-07, + "loss": 0.9456, + "step": 11690 + }, + { + "epoch": 0.456060340291177, + "grad_norm": 13.093682289123535, + "learning_rate": 9.917815500789085e-07, + "loss": 0.928, + "step": 11700 + }, + { + "epoch": 0.456450135453819, + "grad_norm": 12.190604209899902, + "learning_rate": 9.917405509013666e-07, + "loss": 0.9799, + "step": 11710 + }, + { + "epoch": 0.45683993061646105, + "grad_norm": 15.838478088378906, + "learning_rate": 9.916994505639961e-07, + "loss": 0.9173, + "step": 11720 + }, + { + "epoch": 0.4572297257791031, + "grad_norm": 14.068618774414062, + "learning_rate": 9.916582490752521e-07, + "loss": 1.0008, + "step": 11730 + }, + { + "epoch": 0.45761952094174513, + "grad_norm": 14.173237800598145, + "learning_rate": 9.916169464436109e-07, + "loss": 1.0056, + "step": 11740 + }, + { + "epoch": 0.45800931610438717, + "grad_norm": 16.245668411254883, + "learning_rate": 9.915755426775687e-07, + "loss": 0.9068, + "step": 11750 + }, + { + "epoch": 0.4583991112670292, + "grad_norm": 13.108240127563477, + "learning_rate": 9.915340377856432e-07, + "loss": 0.9122, + "step": 11760 + }, + { + "epoch": 0.4587889064296712, + "grad_norm": 12.762725830078125, + "learning_rate": 9.914924317763727e-07, + "loss": 0.8988, + "step": 11770 + }, + { + "epoch": 0.4591787015923132, + "grad_norm": 13.523670196533203, + "learning_rate": 9.91450724658316e-07, + "loss": 0.9319, + "step": 11780 + }, + { + "epoch": 0.45956849675495526, + "grad_norm": 11.857429504394531, + "learning_rate": 9.914089164400536e-07, + "loss": 0.9757, + "step": 11790 + }, + { + "epoch": 0.4599582919175973, + "grad_norm": 16.054214477539062, + "learning_rate": 9.913670071301858e-07, + "loss": 0.9762, + "step": 11800 + }, + { + "epoch": 0.46034808708023933, + "grad_norm": 11.935985565185547, + "learning_rate": 9.913249967373341e-07, + "loss": 0.9279, + "step": 11810 + }, + { + "epoch": 0.46073788224288137, + "grad_norm": 13.908012390136719, + "learning_rate": 9.912828852701408e-07, + "loss": 0.9621, + "step": 11820 + }, + { + "epoch": 0.4611276774055234, + "grad_norm": 13.903159141540527, + "learning_rate": 9.912406727372692e-07, + "loss": 0.9004, + "step": 11830 + }, + { + "epoch": 0.46151747256816544, + "grad_norm": 14.316306114196777, + "learning_rate": 9.911983591474032e-07, + "loss": 0.9533, + "step": 11840 + }, + { + "epoch": 0.4619072677308075, + "grad_norm": 16.22847557067871, + "learning_rate": 9.911559445092468e-07, + "loss": 0.9228, + "step": 11850 + }, + { + "epoch": 0.4622970628934495, + "grad_norm": 12.131424903869629, + "learning_rate": 9.911134288315262e-07, + "loss": 0.9427, + "step": 11860 + }, + { + "epoch": 0.4626868580560915, + "grad_norm": 12.992269515991211, + "learning_rate": 9.910708121229874e-07, + "loss": 0.9858, + "step": 11870 + }, + { + "epoch": 0.46307665321873354, + "grad_norm": 13.075627326965332, + "learning_rate": 9.910280943923973e-07, + "loss": 0.9164, + "step": 11880 + }, + { + "epoch": 0.4634664483813756, + "grad_norm": 13.358198165893555, + "learning_rate": 9.909852756485437e-07, + "loss": 1.013, + "step": 11890 + }, + { + "epoch": 0.4638562435440176, + "grad_norm": 12.978425025939941, + "learning_rate": 9.909423559002354e-07, + "loss": 0.9899, + "step": 11900 + }, + { + "epoch": 0.46424603870665965, + "grad_norm": 13.389287948608398, + "learning_rate": 9.908993351563014e-07, + "loss": 0.9327, + "step": 11910 + }, + { + "epoch": 0.4646358338693017, + "grad_norm": 14.196331024169922, + "learning_rate": 9.908562134255923e-07, + "loss": 0.9382, + "step": 11920 + }, + { + "epoch": 0.4650256290319437, + "grad_norm": 15.30815315246582, + "learning_rate": 9.908129907169787e-07, + "loss": 1.0191, + "step": 11930 + }, + { + "epoch": 0.46541542419458576, + "grad_norm": 12.748794555664062, + "learning_rate": 9.907696670393524e-07, + "loss": 0.9353, + "step": 11940 + }, + { + "epoch": 0.4658052193572278, + "grad_norm": 12.333012580871582, + "learning_rate": 9.907262424016258e-07, + "loss": 0.9087, + "step": 11950 + }, + { + "epoch": 0.46619501451986983, + "grad_norm": 13.714804649353027, + "learning_rate": 9.90682716812732e-07, + "loss": 0.8293, + "step": 11960 + }, + { + "epoch": 0.4665848096825118, + "grad_norm": 13.154768943786621, + "learning_rate": 9.906390902816253e-07, + "loss": 0.9084, + "step": 11970 + }, + { + "epoch": 0.46697460484515385, + "grad_norm": 14.611852645874023, + "learning_rate": 9.9059536281728e-07, + "loss": 0.9256, + "step": 11980 + }, + { + "epoch": 0.4673644000077959, + "grad_norm": 11.699239730834961, + "learning_rate": 9.905515344286923e-07, + "loss": 0.9614, + "step": 11990 + }, + { + "epoch": 0.4677541951704379, + "grad_norm": 14.437845230102539, + "learning_rate": 9.90507605124878e-07, + "loss": 0.951, + "step": 12000 + }, + { + "epoch": 0.4677541951704379, + "eval_loss": 0.9485340714454651, + "eval_runtime": 83.1143, + "eval_samples_per_second": 49.895, + "eval_steps_per_second": 6.244, + "step": 12000 + }, + { + "epoch": 0.46814399033307996, + "grad_norm": 12.997838020324707, + "learning_rate": 9.90463574914874e-07, + "loss": 0.9201, + "step": 12010 + }, + { + "epoch": 0.468533785495722, + "grad_norm": 14.086441040039062, + "learning_rate": 9.904194438077384e-07, + "loss": 0.9559, + "step": 12020 + }, + { + "epoch": 0.46892358065836404, + "grad_norm": 15.288568496704102, + "learning_rate": 9.903752118125499e-07, + "loss": 0.9776, + "step": 12030 + }, + { + "epoch": 0.4693133758210061, + "grad_norm": 16.287843704223633, + "learning_rate": 9.903308789384073e-07, + "loss": 0.8745, + "step": 12040 + }, + { + "epoch": 0.4697031709836481, + "grad_norm": 14.52805233001709, + "learning_rate": 9.902864451944313e-07, + "loss": 0.9255, + "step": 12050 + }, + { + "epoch": 0.47009296614629015, + "grad_norm": 13.939203262329102, + "learning_rate": 9.902419105897622e-07, + "loss": 0.8248, + "step": 12060 + }, + { + "epoch": 0.47048276130893213, + "grad_norm": 14.166893005371094, + "learning_rate": 9.901972751335619e-07, + "loss": 0.9687, + "step": 12070 + }, + { + "epoch": 0.47087255647157417, + "grad_norm": 13.756606101989746, + "learning_rate": 9.901525388350123e-07, + "loss": 0.9741, + "step": 12080 + }, + { + "epoch": 0.4712623516342162, + "grad_norm": 15.817917823791504, + "learning_rate": 9.901077017033168e-07, + "loss": 0.9733, + "step": 12090 + }, + { + "epoch": 0.47165214679685824, + "grad_norm": 15.477527618408203, + "learning_rate": 9.900627637476993e-07, + "loss": 0.9205, + "step": 12100 + }, + { + "epoch": 0.4720419419595003, + "grad_norm": 13.244834899902344, + "learning_rate": 9.90017724977404e-07, + "loss": 0.9517, + "step": 12110 + }, + { + "epoch": 0.4724317371221423, + "grad_norm": 12.726200103759766, + "learning_rate": 9.899725854016965e-07, + "loss": 0.9772, + "step": 12120 + }, + { + "epoch": 0.47282153228478435, + "grad_norm": 18.40109634399414, + "learning_rate": 9.899273450298624e-07, + "loss": 0.867, + "step": 12130 + }, + { + "epoch": 0.4732113274474264, + "grad_norm": 14.58556079864502, + "learning_rate": 9.89882003871209e-07, + "loss": 0.9182, + "step": 12140 + }, + { + "epoch": 0.4736011226100684, + "grad_norm": 15.298309326171875, + "learning_rate": 9.898365619350634e-07, + "loss": 0.9922, + "step": 12150 + }, + { + "epoch": 0.47399091777271046, + "grad_norm": 12.31676959991455, + "learning_rate": 9.897910192307739e-07, + "loss": 0.9128, + "step": 12160 + }, + { + "epoch": 0.4743807129353525, + "grad_norm": 13.233631134033203, + "learning_rate": 9.897453757677094e-07, + "loss": 0.9105, + "step": 12170 + }, + { + "epoch": 0.4747705080979945, + "grad_norm": 12.956151008605957, + "learning_rate": 9.896996315552598e-07, + "loss": 0.9267, + "step": 12180 + }, + { + "epoch": 0.4751603032606365, + "grad_norm": 14.097780227661133, + "learning_rate": 9.896537866028354e-07, + "loss": 0.991, + "step": 12190 + }, + { + "epoch": 0.47555009842327856, + "grad_norm": 15.652153015136719, + "learning_rate": 9.89607840919867e-07, + "loss": 0.9572, + "step": 12200 + }, + { + "epoch": 0.4759398935859206, + "grad_norm": 13.64004898071289, + "learning_rate": 9.89561794515807e-07, + "loss": 0.8522, + "step": 12210 + }, + { + "epoch": 0.47632968874856263, + "grad_norm": 18.813194274902344, + "learning_rate": 9.895156474001277e-07, + "loss": 0.9452, + "step": 12220 + }, + { + "epoch": 0.47671948391120467, + "grad_norm": 14.575525283813477, + "learning_rate": 9.89469399582322e-07, + "loss": 0.9321, + "step": 12230 + }, + { + "epoch": 0.4771092790738467, + "grad_norm": 13.238840103149414, + "learning_rate": 9.894230510719045e-07, + "loss": 0.9567, + "step": 12240 + }, + { + "epoch": 0.47749907423648874, + "grad_norm": 11.960838317871094, + "learning_rate": 9.8937660187841e-07, + "loss": 0.9365, + "step": 12250 + }, + { + "epoch": 0.4778888693991308, + "grad_norm": 14.472798347473145, + "learning_rate": 9.893300520113934e-07, + "loss": 0.9758, + "step": 12260 + }, + { + "epoch": 0.4782786645617728, + "grad_norm": 13.932619094848633, + "learning_rate": 9.89283401480431e-07, + "loss": 0.8998, + "step": 12270 + }, + { + "epoch": 0.4786684597244148, + "grad_norm": 12.600906372070312, + "learning_rate": 9.892366502951198e-07, + "loss": 0.9519, + "step": 12280 + }, + { + "epoch": 0.47905825488705683, + "grad_norm": 12.410995483398438, + "learning_rate": 9.891897984650772e-07, + "loss": 0.8982, + "step": 12290 + }, + { + "epoch": 0.47944805004969887, + "grad_norm": 14.585325241088867, + "learning_rate": 9.891428459999417e-07, + "loss": 0.9117, + "step": 12300 + }, + { + "epoch": 0.4798378452123409, + "grad_norm": 15.682839393615723, + "learning_rate": 9.89095792909372e-07, + "loss": 0.8813, + "step": 12310 + }, + { + "epoch": 0.48022764037498294, + "grad_norm": 14.336636543273926, + "learning_rate": 9.89048639203048e-07, + "loss": 0.8298, + "step": 12320 + }, + { + "epoch": 0.480617435537625, + "grad_norm": 11.922830581665039, + "learning_rate": 9.890013848906697e-07, + "loss": 0.9285, + "step": 12330 + }, + { + "epoch": 0.481007230700267, + "grad_norm": 13.843091011047363, + "learning_rate": 9.889540299819585e-07, + "loss": 0.9096, + "step": 12340 + }, + { + "epoch": 0.48139702586290906, + "grad_norm": 12.643184661865234, + "learning_rate": 9.88906574486656e-07, + "loss": 0.8732, + "step": 12350 + }, + { + "epoch": 0.4817868210255511, + "grad_norm": 15.145201683044434, + "learning_rate": 9.888590184145248e-07, + "loss": 0.9924, + "step": 12360 + }, + { + "epoch": 0.48217661618819313, + "grad_norm": 13.51017951965332, + "learning_rate": 9.888113617753477e-07, + "loss": 0.9063, + "step": 12370 + }, + { + "epoch": 0.4825664113508351, + "grad_norm": 14.290228843688965, + "learning_rate": 9.887636045789289e-07, + "loss": 0.8808, + "step": 12380 + }, + { + "epoch": 0.48295620651347715, + "grad_norm": 13.749763488769531, + "learning_rate": 9.887157468350929e-07, + "loss": 0.8651, + "step": 12390 + }, + { + "epoch": 0.4833460016761192, + "grad_norm": 15.092280387878418, + "learning_rate": 9.886677885536845e-07, + "loss": 1.0083, + "step": 12400 + }, + { + "epoch": 0.4837357968387612, + "grad_norm": 13.192667961120605, + "learning_rate": 9.886197297445697e-07, + "loss": 0.9854, + "step": 12410 + }, + { + "epoch": 0.48412559200140326, + "grad_norm": 14.746283531188965, + "learning_rate": 9.885715704176355e-07, + "loss": 0.9821, + "step": 12420 + }, + { + "epoch": 0.4845153871640453, + "grad_norm": 14.168292045593262, + "learning_rate": 9.885233105827885e-07, + "loss": 0.987, + "step": 12430 + }, + { + "epoch": 0.48490518232668733, + "grad_norm": 14.055312156677246, + "learning_rate": 9.884749502499572e-07, + "loss": 0.9479, + "step": 12440 + }, + { + "epoch": 0.48529497748932937, + "grad_norm": 13.053544044494629, + "learning_rate": 9.884264894290897e-07, + "loss": 0.9718, + "step": 12450 + }, + { + "epoch": 0.4856847726519714, + "grad_norm": 12.523114204406738, + "learning_rate": 9.883779281301556e-07, + "loss": 0.9351, + "step": 12460 + }, + { + "epoch": 0.48607456781461345, + "grad_norm": 14.353793144226074, + "learning_rate": 9.883292663631446e-07, + "loss": 0.9752, + "step": 12470 + }, + { + "epoch": 0.4864643629772554, + "grad_norm": 15.509095191955566, + "learning_rate": 9.882805041380674e-07, + "loss": 1.0001, + "step": 12480 + }, + { + "epoch": 0.48685415813989746, + "grad_norm": 12.573677062988281, + "learning_rate": 9.882316414649553e-07, + "loss": 0.8838, + "step": 12490 + }, + { + "epoch": 0.4872439533025395, + "grad_norm": 12.953845024108887, + "learning_rate": 9.8818267835386e-07, + "loss": 0.8411, + "step": 12500 + }, + { + "epoch": 0.4872439533025395, + "eval_loss": 0.9466409087181091, + "eval_runtime": 82.8787, + "eval_samples_per_second": 50.037, + "eval_steps_per_second": 6.262, + "step": 12500 + }, + { + "epoch": 0.48763374846518154, + "grad_norm": 12.65576171875, + "learning_rate": 9.881336148148545e-07, + "loss": 0.9074, + "step": 12510 + }, + { + "epoch": 0.4880235436278236, + "grad_norm": 12.616167068481445, + "learning_rate": 9.880844508580317e-07, + "loss": 0.8688, + "step": 12520 + }, + { + "epoch": 0.4884133387904656, + "grad_norm": 12.640853881835938, + "learning_rate": 9.880351864935058e-07, + "loss": 0.9481, + "step": 12530 + }, + { + "epoch": 0.48880313395310765, + "grad_norm": 13.834686279296875, + "learning_rate": 9.879858217314109e-07, + "loss": 0.8972, + "step": 12540 + }, + { + "epoch": 0.4891929291157497, + "grad_norm": 13.401185035705566, + "learning_rate": 9.879363565819026e-07, + "loss": 0.9013, + "step": 12550 + }, + { + "epoch": 0.4895827242783917, + "grad_norm": 14.897149085998535, + "learning_rate": 9.878867910551566e-07, + "loss": 0.9232, + "step": 12560 + }, + { + "epoch": 0.48997251944103376, + "grad_norm": 15.785338401794434, + "learning_rate": 9.878371251613695e-07, + "loss": 0.8866, + "step": 12570 + }, + { + "epoch": 0.49036231460367574, + "grad_norm": 15.762842178344727, + "learning_rate": 9.877873589107586e-07, + "loss": 0.9019, + "step": 12580 + }, + { + "epoch": 0.4907521097663178, + "grad_norm": 16.25050163269043, + "learning_rate": 9.877374923135614e-07, + "loss": 0.9381, + "step": 12590 + }, + { + "epoch": 0.4911419049289598, + "grad_norm": 15.576542854309082, + "learning_rate": 9.876875253800366e-07, + "loss": 0.9268, + "step": 12600 + }, + { + "epoch": 0.49153170009160185, + "grad_norm": 15.194286346435547, + "learning_rate": 9.87637458120463e-07, + "loss": 0.9546, + "step": 12610 + }, + { + "epoch": 0.4919214952542439, + "grad_norm": 16.01327133178711, + "learning_rate": 9.875872905451406e-07, + "loss": 0.9628, + "step": 12620 + }, + { + "epoch": 0.4923112904168859, + "grad_norm": 17.05106544494629, + "learning_rate": 9.875370226643897e-07, + "loss": 0.8883, + "step": 12630 + }, + { + "epoch": 0.49270108557952796, + "grad_norm": 13.22525691986084, + "learning_rate": 9.874866544885514e-07, + "loss": 0.8909, + "step": 12640 + }, + { + "epoch": 0.49309088074217, + "grad_norm": 14.83056354522705, + "learning_rate": 9.87436186027987e-07, + "loss": 0.8651, + "step": 12650 + }, + { + "epoch": 0.49348067590481204, + "grad_norm": 12.368146896362305, + "learning_rate": 9.873856172930792e-07, + "loss": 0.9189, + "step": 12660 + }, + { + "epoch": 0.4938704710674541, + "grad_norm": 12.649965286254883, + "learning_rate": 9.873349482942306e-07, + "loss": 0.917, + "step": 12670 + }, + { + "epoch": 0.49426026623009606, + "grad_norm": 15.734679222106934, + "learning_rate": 9.87284179041865e-07, + "loss": 0.9056, + "step": 12680 + }, + { + "epoch": 0.4946500613927381, + "grad_norm": 12.972075462341309, + "learning_rate": 9.872333095464261e-07, + "loss": 0.9298, + "step": 12690 + }, + { + "epoch": 0.49503985655538013, + "grad_norm": 14.877608299255371, + "learning_rate": 9.87182339818379e-07, + "loss": 0.9827, + "step": 12700 + }, + { + "epoch": 0.49542965171802217, + "grad_norm": 15.838467597961426, + "learning_rate": 9.871312698682088e-07, + "loss": 0.9448, + "step": 12710 + }, + { + "epoch": 0.4958194468806642, + "grad_norm": 13.213595390319824, + "learning_rate": 9.870800997064218e-07, + "loss": 0.8956, + "step": 12720 + }, + { + "epoch": 0.49620924204330624, + "grad_norm": 14.162665367126465, + "learning_rate": 9.870288293435446e-07, + "loss": 0.8925, + "step": 12730 + }, + { + "epoch": 0.4965990372059483, + "grad_norm": 14.541419982910156, + "learning_rate": 9.869774587901242e-07, + "loss": 0.945, + "step": 12740 + }, + { + "epoch": 0.4969888323685903, + "grad_norm": 17.395889282226562, + "learning_rate": 9.869259880567288e-07, + "loss": 0.9497, + "step": 12750 + }, + { + "epoch": 0.49737862753123235, + "grad_norm": 16.58950424194336, + "learning_rate": 9.868744171539466e-07, + "loss": 0.9554, + "step": 12760 + }, + { + "epoch": 0.4977684226938744, + "grad_norm": 12.940457344055176, + "learning_rate": 9.868227460923867e-07, + "loss": 0.9215, + "step": 12770 + }, + { + "epoch": 0.4981582178565164, + "grad_norm": 17.77170753479004, + "learning_rate": 9.867709748826786e-07, + "loss": 0.8754, + "step": 12780 + }, + { + "epoch": 0.4985480130191584, + "grad_norm": 12.679247856140137, + "learning_rate": 9.86719103535473e-07, + "loss": 0.9524, + "step": 12790 + }, + { + "epoch": 0.49893780818180045, + "grad_norm": 15.573468208312988, + "learning_rate": 9.866671320614402e-07, + "loss": 0.9098, + "step": 12800 + }, + { + "epoch": 0.4993276033444425, + "grad_norm": 11.226459503173828, + "learning_rate": 9.86615060471272e-07, + "loss": 0.8686, + "step": 12810 + }, + { + "epoch": 0.4997173985070845, + "grad_norm": 12.711118698120117, + "learning_rate": 9.865628887756804e-07, + "loss": 0.9349, + "step": 12820 + }, + { + "epoch": 0.5001071936697266, + "grad_norm": 15.623577117919922, + "learning_rate": 9.865106169853983e-07, + "loss": 0.906, + "step": 12830 + }, + { + "epoch": 0.5004969888323686, + "grad_norm": 15.738816261291504, + "learning_rate": 9.864582451111784e-07, + "loss": 1.0495, + "step": 12840 + }, + { + "epoch": 0.5008867839950106, + "grad_norm": 14.449959754943848, + "learning_rate": 9.86405773163795e-07, + "loss": 0.9133, + "step": 12850 + }, + { + "epoch": 0.5012765791576527, + "grad_norm": 12.965691566467285, + "learning_rate": 9.863532011540422e-07, + "loss": 0.8689, + "step": 12860 + }, + { + "epoch": 0.5016663743202947, + "grad_norm": 16.520530700683594, + "learning_rate": 9.863005290927354e-07, + "loss": 0.8458, + "step": 12870 + }, + { + "epoch": 0.5020561694829367, + "grad_norm": 11.501152992248535, + "learning_rate": 9.862477569907096e-07, + "loss": 0.8817, + "step": 12880 + }, + { + "epoch": 0.5024459646455788, + "grad_norm": 12.430654525756836, + "learning_rate": 9.861948848588218e-07, + "loss": 0.9644, + "step": 12890 + }, + { + "epoch": 0.5028357598082208, + "grad_norm": 15.843810081481934, + "learning_rate": 9.86141912707948e-07, + "loss": 0.9442, + "step": 12900 + }, + { + "epoch": 0.5032255549708629, + "grad_norm": 15.430998802185059, + "learning_rate": 9.860888405489855e-07, + "loss": 0.8957, + "step": 12910 + }, + { + "epoch": 0.5036153501335049, + "grad_norm": 11.61874771118164, + "learning_rate": 9.860356683928528e-07, + "loss": 0.9777, + "step": 12920 + }, + { + "epoch": 0.5040051452961469, + "grad_norm": 11.458113670349121, + "learning_rate": 9.859823962504878e-07, + "loss": 0.9197, + "step": 12930 + }, + { + "epoch": 0.5043949404587889, + "grad_norm": 14.304285049438477, + "learning_rate": 9.8592902413285e-07, + "loss": 0.9152, + "step": 12940 + }, + { + "epoch": 0.5047847356214309, + "grad_norm": 14.106751441955566, + "learning_rate": 9.858755520509185e-07, + "loss": 0.9389, + "step": 12950 + }, + { + "epoch": 0.5051745307840729, + "grad_norm": 13.688305854797363, + "learning_rate": 9.858219800156939e-07, + "loss": 0.9104, + "step": 12960 + }, + { + "epoch": 0.505564325946715, + "grad_norm": 13.040285110473633, + "learning_rate": 9.857683080381967e-07, + "loss": 0.9474, + "step": 12970 + }, + { + "epoch": 0.505954121109357, + "grad_norm": 14.795648574829102, + "learning_rate": 9.85714536129468e-07, + "loss": 0.9755, + "step": 12980 + }, + { + "epoch": 0.506343916271999, + "grad_norm": 12.825078010559082, + "learning_rate": 9.856606643005702e-07, + "loss": 0.9265, + "step": 12990 + }, + { + "epoch": 0.5067337114346411, + "grad_norm": 15.840315818786621, + "learning_rate": 9.856066925625852e-07, + "loss": 0.9195, + "step": 13000 + }, + { + "epoch": 0.5067337114346411, + "eval_loss": 0.941493809223175, + "eval_runtime": 82.8247, + "eval_samples_per_second": 50.07, + "eval_steps_per_second": 6.266, + "step": 13000 + }, + { + "epoch": 0.5071235065972831, + "grad_norm": 14.2937650680542, + "learning_rate": 9.855526209266161e-07, + "loss": 0.9146, + "step": 13010 + }, + { + "epoch": 0.5075133017599252, + "grad_norm": 14.073277473449707, + "learning_rate": 9.854984494037867e-07, + "loss": 0.875, + "step": 13020 + }, + { + "epoch": 0.5079030969225672, + "grad_norm": 16.426881790161133, + "learning_rate": 9.854441780052405e-07, + "loss": 0.9399, + "step": 13030 + }, + { + "epoch": 0.5082928920852092, + "grad_norm": 14.828062057495117, + "learning_rate": 9.853898067421425e-07, + "loss": 0.9175, + "step": 13040 + }, + { + "epoch": 0.5086826872478513, + "grad_norm": 16.915422439575195, + "learning_rate": 9.853353356256777e-07, + "loss": 0.9715, + "step": 13050 + }, + { + "epoch": 0.5090724824104933, + "grad_norm": 14.544678688049316, + "learning_rate": 9.852807646670516e-07, + "loss": 0.849, + "step": 13060 + }, + { + "epoch": 0.5094622775731353, + "grad_norm": 14.735069274902344, + "learning_rate": 9.85226093877491e-07, + "loss": 0.9215, + "step": 13070 + }, + { + "epoch": 0.5098520727357774, + "grad_norm": 13.187628746032715, + "learning_rate": 9.85171323268242e-07, + "loss": 0.9216, + "step": 13080 + }, + { + "epoch": 0.5102418678984194, + "grad_norm": 14.750734329223633, + "learning_rate": 9.851164528505721e-07, + "loss": 0.9117, + "step": 13090 + }, + { + "epoch": 0.5106316630610614, + "grad_norm": 15.97279167175293, + "learning_rate": 9.850614826357695e-07, + "loss": 0.9225, + "step": 13100 + }, + { + "epoch": 0.5110214582237035, + "grad_norm": 16.17802619934082, + "learning_rate": 9.85006412635142e-07, + "loss": 0.9278, + "step": 13110 + }, + { + "epoch": 0.5114112533863455, + "grad_norm": 14.300265312194824, + "learning_rate": 9.849512428600189e-07, + "loss": 0.9253, + "step": 13120 + }, + { + "epoch": 0.5118010485489876, + "grad_norm": 13.042201042175293, + "learning_rate": 9.848959733217495e-07, + "loss": 0.9928, + "step": 13130 + }, + { + "epoch": 0.5121908437116295, + "grad_norm": 13.120453834533691, + "learning_rate": 9.848406040317036e-07, + "loss": 0.911, + "step": 13140 + }, + { + "epoch": 0.5125806388742715, + "grad_norm": 13.033644676208496, + "learning_rate": 9.847851350012717e-07, + "loss": 0.9026, + "step": 13150 + }, + { + "epoch": 0.5129704340369136, + "grad_norm": 14.109724998474121, + "learning_rate": 9.84729566241865e-07, + "loss": 0.99, + "step": 13160 + }, + { + "epoch": 0.5133602291995556, + "grad_norm": 14.272068977355957, + "learning_rate": 9.846738977649145e-07, + "loss": 0.969, + "step": 13170 + }, + { + "epoch": 0.5137500243621976, + "grad_norm": 13.893214225769043, + "learning_rate": 9.846181295818728e-07, + "loss": 0.9572, + "step": 13180 + }, + { + "epoch": 0.5141398195248397, + "grad_norm": 14.65523910522461, + "learning_rate": 9.84562261704212e-07, + "loss": 0.8989, + "step": 13190 + }, + { + "epoch": 0.5145296146874817, + "grad_norm": 15.863385200500488, + "learning_rate": 9.845062941434251e-07, + "loss": 0.9526, + "step": 13200 + }, + { + "epoch": 0.5149194098501237, + "grad_norm": 15.950799942016602, + "learning_rate": 9.84450226911026e-07, + "loss": 0.9614, + "step": 13210 + }, + { + "epoch": 0.5153092050127658, + "grad_norm": 13.954133033752441, + "learning_rate": 9.843940600185482e-07, + "loss": 0.9177, + "step": 13220 + }, + { + "epoch": 0.5156990001754078, + "grad_norm": 14.724565505981445, + "learning_rate": 9.843377934775468e-07, + "loss": 0.9756, + "step": 13230 + }, + { + "epoch": 0.5160887953380499, + "grad_norm": 12.840142250061035, + "learning_rate": 9.842814272995965e-07, + "loss": 0.9459, + "step": 13240 + }, + { + "epoch": 0.5164785905006919, + "grad_norm": 14.016914367675781, + "learning_rate": 9.84224961496293e-07, + "loss": 0.8883, + "step": 13250 + }, + { + "epoch": 0.5168683856633339, + "grad_norm": 14.933531761169434, + "learning_rate": 9.84168396079252e-07, + "loss": 0.905, + "step": 13260 + }, + { + "epoch": 0.517258180825976, + "grad_norm": 15.147429466247559, + "learning_rate": 9.841117310601105e-07, + "loss": 0.8957, + "step": 13270 + }, + { + "epoch": 0.517647975988618, + "grad_norm": 14.394329071044922, + "learning_rate": 9.84054966450525e-07, + "loss": 0.9696, + "step": 13280 + }, + { + "epoch": 0.51803777115126, + "grad_norm": 12.539999008178711, + "learning_rate": 9.839981022621733e-07, + "loss": 1.0065, + "step": 13290 + }, + { + "epoch": 0.5184275663139021, + "grad_norm": 14.674797058105469, + "learning_rate": 9.839411385067533e-07, + "loss": 0.9651, + "step": 13300 + }, + { + "epoch": 0.5188173614765441, + "grad_norm": 15.600942611694336, + "learning_rate": 9.838840751959837e-07, + "loss": 0.9385, + "step": 13310 + }, + { + "epoch": 0.5192071566391862, + "grad_norm": 12.952668190002441, + "learning_rate": 9.838269123416029e-07, + "loss": 0.9013, + "step": 13320 + }, + { + "epoch": 0.5195969518018282, + "grad_norm": 15.120955467224121, + "learning_rate": 9.83769649955371e-07, + "loss": 0.887, + "step": 13330 + }, + { + "epoch": 0.5199867469644702, + "grad_norm": 14.36095142364502, + "learning_rate": 9.837122880490675e-07, + "loss": 0.8935, + "step": 13340 + }, + { + "epoch": 0.5203765421271122, + "grad_norm": 15.690572738647461, + "learning_rate": 9.836548266344927e-07, + "loss": 0.9432, + "step": 13350 + }, + { + "epoch": 0.5207663372897542, + "grad_norm": 15.628254890441895, + "learning_rate": 9.835972657234675e-07, + "loss": 0.9539, + "step": 13360 + }, + { + "epoch": 0.5211561324523962, + "grad_norm": 13.378961563110352, + "learning_rate": 9.835396053278333e-07, + "loss": 0.8861, + "step": 13370 + }, + { + "epoch": 0.5215459276150383, + "grad_norm": 14.303468704223633, + "learning_rate": 9.83481845459452e-07, + "loss": 0.9469, + "step": 13380 + }, + { + "epoch": 0.5219357227776803, + "grad_norm": 11.6542387008667, + "learning_rate": 9.834239861302056e-07, + "loss": 0.9282, + "step": 13390 + }, + { + "epoch": 0.5223255179403223, + "grad_norm": 16.110191345214844, + "learning_rate": 9.83366027351997e-07, + "loss": 0.9751, + "step": 13400 + }, + { + "epoch": 0.5227153131029644, + "grad_norm": 13.52400016784668, + "learning_rate": 9.83307969136749e-07, + "loss": 0.907, + "step": 13410 + }, + { + "epoch": 0.5231051082656064, + "grad_norm": 13.94028091430664, + "learning_rate": 9.832498114964057e-07, + "loss": 0.9442, + "step": 13420 + }, + { + "epoch": 0.5234949034282484, + "grad_norm": 13.840215682983398, + "learning_rate": 9.831915544429307e-07, + "loss": 0.9271, + "step": 13430 + }, + { + "epoch": 0.5238846985908905, + "grad_norm": 12.236870765686035, + "learning_rate": 9.83133197988309e-07, + "loss": 0.8757, + "step": 13440 + }, + { + "epoch": 0.5242744937535325, + "grad_norm": 13.746227264404297, + "learning_rate": 9.830747421445451e-07, + "loss": 0.9122, + "step": 13450 + }, + { + "epoch": 0.5246642889161746, + "grad_norm": 14.083416938781738, + "learning_rate": 9.830161869236647e-07, + "loss": 0.9326, + "step": 13460 + }, + { + "epoch": 0.5250540840788166, + "grad_norm": 13.842083930969238, + "learning_rate": 9.829575323377136e-07, + "loss": 0.9448, + "step": 13470 + }, + { + "epoch": 0.5254438792414586, + "grad_norm": 13.058833122253418, + "learning_rate": 9.82898778398758e-07, + "loss": 0.9487, + "step": 13480 + }, + { + "epoch": 0.5258336744041007, + "grad_norm": 13.130695343017578, + "learning_rate": 9.82839925118885e-07, + "loss": 0.9972, + "step": 13490 + }, + { + "epoch": 0.5262234695667427, + "grad_norm": 14.080259323120117, + "learning_rate": 9.82780972510201e-07, + "loss": 0.9695, + "step": 13500 + }, + { + "epoch": 0.5262234695667427, + "eval_loss": 0.9390377402305603, + "eval_runtime": 82.957, + "eval_samples_per_second": 49.99, + "eval_steps_per_second": 6.256, + "step": 13500 + }, + { + "epoch": 0.5266132647293847, + "grad_norm": 15.0670804977417, + "learning_rate": 9.827219205848344e-07, + "loss": 1.0092, + "step": 13510 + }, + { + "epoch": 0.5270030598920268, + "grad_norm": 12.147978782653809, + "learning_rate": 9.826627693549327e-07, + "loss": 0.8906, + "step": 13520 + }, + { + "epoch": 0.5273928550546688, + "grad_norm": 12.986127853393555, + "learning_rate": 9.82603518832665e-07, + "loss": 0.8801, + "step": 13530 + }, + { + "epoch": 0.5277826502173109, + "grad_norm": 13.86558723449707, + "learning_rate": 9.825441690302193e-07, + "loss": 0.954, + "step": 13540 + }, + { + "epoch": 0.5281724453799528, + "grad_norm": 12.424535751342773, + "learning_rate": 9.824847199598057e-07, + "loss": 0.9276, + "step": 13550 + }, + { + "epoch": 0.5285622405425948, + "grad_norm": 14.492342948913574, + "learning_rate": 9.824251716336535e-07, + "loss": 0.9339, + "step": 13560 + }, + { + "epoch": 0.5289520357052369, + "grad_norm": 14.794229507446289, + "learning_rate": 9.823655240640132e-07, + "loss": 0.8827, + "step": 13570 + }, + { + "epoch": 0.5293418308678789, + "grad_norm": 15.856827735900879, + "learning_rate": 9.823057772631551e-07, + "loss": 0.9068, + "step": 13580 + }, + { + "epoch": 0.5297316260305209, + "grad_norm": 14.173441886901855, + "learning_rate": 9.8224593124337e-07, + "loss": 0.9596, + "step": 13590 + }, + { + "epoch": 0.530121421193163, + "grad_norm": 15.835366249084473, + "learning_rate": 9.8218598601697e-07, + "loss": 0.8875, + "step": 13600 + }, + { + "epoch": 0.530511216355805, + "grad_norm": 14.401946067810059, + "learning_rate": 9.82125941596286e-07, + "loss": 0.9778, + "step": 13610 + }, + { + "epoch": 0.530901011518447, + "grad_norm": 14.559992790222168, + "learning_rate": 9.82065797993671e-07, + "loss": 0.9318, + "step": 13620 + }, + { + "epoch": 0.5312908066810891, + "grad_norm": 12.489997863769531, + "learning_rate": 9.820055552214972e-07, + "loss": 0.9624, + "step": 13630 + }, + { + "epoch": 0.5316806018437311, + "grad_norm": 16.002044677734375, + "learning_rate": 9.819452132921578e-07, + "loss": 0.9549, + "step": 13640 + }, + { + "epoch": 0.5320703970063732, + "grad_norm": 12.650062561035156, + "learning_rate": 9.818847722180658e-07, + "loss": 0.9852, + "step": 13650 + }, + { + "epoch": 0.5324601921690152, + "grad_norm": 14.622328758239746, + "learning_rate": 9.818242320116557e-07, + "loss": 0.9398, + "step": 13660 + }, + { + "epoch": 0.5328499873316572, + "grad_norm": 12.61216926574707, + "learning_rate": 9.81763592685381e-07, + "loss": 0.9062, + "step": 13670 + }, + { + "epoch": 0.5332397824942993, + "grad_norm": 14.513840675354004, + "learning_rate": 9.817028542517166e-07, + "loss": 0.9185, + "step": 13680 + }, + { + "epoch": 0.5336295776569413, + "grad_norm": 13.670234680175781, + "learning_rate": 9.816420167231578e-07, + "loss": 0.8557, + "step": 13690 + }, + { + "epoch": 0.5340193728195833, + "grad_norm": 14.3908052444458, + "learning_rate": 9.815810801122196e-07, + "loss": 0.8896, + "step": 13700 + }, + { + "epoch": 0.5344091679822254, + "grad_norm": 12.547237396240234, + "learning_rate": 9.815200444314378e-07, + "loss": 0.9558, + "step": 13710 + }, + { + "epoch": 0.5347989631448674, + "grad_norm": 13.244909286499023, + "learning_rate": 9.814589096933684e-07, + "loss": 1.0023, + "step": 13720 + }, + { + "epoch": 0.5351887583075094, + "grad_norm": 14.939372062683105, + "learning_rate": 9.813976759105881e-07, + "loss": 0.8816, + "step": 13730 + }, + { + "epoch": 0.5355785534701515, + "grad_norm": 15.554043769836426, + "learning_rate": 9.81336343095694e-07, + "loss": 0.943, + "step": 13740 + }, + { + "epoch": 0.5359683486327934, + "grad_norm": 13.639766693115234, + "learning_rate": 9.812749112613029e-07, + "loss": 0.915, + "step": 13750 + }, + { + "epoch": 0.5363581437954354, + "grad_norm": 13.096029281616211, + "learning_rate": 9.812133804200526e-07, + "loss": 0.9119, + "step": 13760 + }, + { + "epoch": 0.5367479389580775, + "grad_norm": 11.36942195892334, + "learning_rate": 9.811517505846013e-07, + "loss": 0.8526, + "step": 13770 + }, + { + "epoch": 0.5371377341207195, + "grad_norm": 13.132712364196777, + "learning_rate": 9.81090021767627e-07, + "loss": 0.8763, + "step": 13780 + }, + { + "epoch": 0.5375275292833616, + "grad_norm": 10.508270263671875, + "learning_rate": 9.810281939818288e-07, + "loss": 0.8891, + "step": 13790 + }, + { + "epoch": 0.5379173244460036, + "grad_norm": 13.579071044921875, + "learning_rate": 9.809662672399254e-07, + "loss": 0.9449, + "step": 13800 + }, + { + "epoch": 0.5383071196086456, + "grad_norm": 13.286399841308594, + "learning_rate": 9.809042415546565e-07, + "loss": 0.9026, + "step": 13810 + }, + { + "epoch": 0.5386969147712877, + "grad_norm": 15.021225929260254, + "learning_rate": 9.80842116938782e-07, + "loss": 0.8878, + "step": 13820 + }, + { + "epoch": 0.5390867099339297, + "grad_norm": 13.898314476013184, + "learning_rate": 9.807798934050814e-07, + "loss": 0.963, + "step": 13830 + }, + { + "epoch": 0.5394765050965717, + "grad_norm": 12.925097465515137, + "learning_rate": 9.807175709663562e-07, + "loss": 1.0136, + "step": 13840 + }, + { + "epoch": 0.5398663002592138, + "grad_norm": 13.179591178894043, + "learning_rate": 9.806551496354264e-07, + "loss": 0.9624, + "step": 13850 + }, + { + "epoch": 0.5402560954218558, + "grad_norm": 12.433097839355469, + "learning_rate": 9.805926294251337e-07, + "loss": 0.9146, + "step": 13860 + }, + { + "epoch": 0.5406458905844979, + "grad_norm": 15.524016380310059, + "learning_rate": 9.805300103483391e-07, + "loss": 0.9537, + "step": 13870 + }, + { + "epoch": 0.5410356857471399, + "grad_norm": 14.448013305664062, + "learning_rate": 9.80467292417925e-07, + "loss": 0.8707, + "step": 13880 + }, + { + "epoch": 0.5414254809097819, + "grad_norm": 14.287160873413086, + "learning_rate": 9.804044756467933e-07, + "loss": 0.9401, + "step": 13890 + }, + { + "epoch": 0.541815276072424, + "grad_norm": 13.162546157836914, + "learning_rate": 9.803415600478666e-07, + "loss": 0.9985, + "step": 13900 + }, + { + "epoch": 0.542205071235066, + "grad_norm": 11.113014221191406, + "learning_rate": 9.802785456340878e-07, + "loss": 0.8959, + "step": 13910 + }, + { + "epoch": 0.542594866397708, + "grad_norm": 13.350449562072754, + "learning_rate": 9.802154324184201e-07, + "loss": 0.9075, + "step": 13920 + }, + { + "epoch": 0.5429846615603501, + "grad_norm": 13.880852699279785, + "learning_rate": 9.80152220413847e-07, + "loss": 0.9841, + "step": 13930 + }, + { + "epoch": 0.5433744567229921, + "grad_norm": 15.301618576049805, + "learning_rate": 9.800889096333723e-07, + "loss": 0.9257, + "step": 13940 + }, + { + "epoch": 0.5437642518856342, + "grad_norm": 15.594715118408203, + "learning_rate": 9.800255000900202e-07, + "loss": 0.9457, + "step": 13950 + }, + { + "epoch": 0.5441540470482761, + "grad_norm": 12.19060230255127, + "learning_rate": 9.79961991796835e-07, + "loss": 0.9164, + "step": 13960 + }, + { + "epoch": 0.5445438422109181, + "grad_norm": 12.633545875549316, + "learning_rate": 9.798983847668818e-07, + "loss": 0.8642, + "step": 13970 + }, + { + "epoch": 0.5449336373735602, + "grad_norm": 13.15356731414795, + "learning_rate": 9.798346790132457e-07, + "loss": 0.9292, + "step": 13980 + }, + { + "epoch": 0.5453234325362022, + "grad_norm": 14.212919235229492, + "learning_rate": 9.797708745490319e-07, + "loss": 0.9792, + "step": 13990 + }, + { + "epoch": 0.5457132276988442, + "grad_norm": 13.433027267456055, + "learning_rate": 9.79706971387366e-07, + "loss": 0.9082, + "step": 14000 + }, + { + "epoch": 0.5457132276988442, + "eval_loss": 0.9351343512535095, + "eval_runtime": 82.9257, + "eval_samples_per_second": 50.009, + "eval_steps_per_second": 6.259, + "step": 14000 + }, + { + "epoch": 0.5461030228614863, + "grad_norm": 12.65933609008789, + "learning_rate": 9.796429695413943e-07, + "loss": 0.9117, + "step": 14010 + }, + { + "epoch": 0.5464928180241283, + "grad_norm": 13.57572078704834, + "learning_rate": 9.795788690242832e-07, + "loss": 0.9936, + "step": 14020 + }, + { + "epoch": 0.5468826131867703, + "grad_norm": 13.902978897094727, + "learning_rate": 9.79514669849219e-07, + "loss": 0.9693, + "step": 14030 + }, + { + "epoch": 0.5472724083494124, + "grad_norm": 14.075730323791504, + "learning_rate": 9.794503720294086e-07, + "loss": 0.9829, + "step": 14040 + }, + { + "epoch": 0.5476622035120544, + "grad_norm": 16.02961540222168, + "learning_rate": 9.793859755780798e-07, + "loss": 0.9472, + "step": 14050 + }, + { + "epoch": 0.5480519986746965, + "grad_norm": 14.025467872619629, + "learning_rate": 9.793214805084795e-07, + "loss": 0.9177, + "step": 14060 + }, + { + "epoch": 0.5484417938373385, + "grad_norm": 12.083243370056152, + "learning_rate": 9.792568868338757e-07, + "loss": 0.9744, + "step": 14070 + }, + { + "epoch": 0.5488315889999805, + "grad_norm": 12.431770324707031, + "learning_rate": 9.791921945675565e-07, + "loss": 0.9508, + "step": 14080 + }, + { + "epoch": 0.5492213841626226, + "grad_norm": 11.956436157226562, + "learning_rate": 9.7912740372283e-07, + "loss": 0.8786, + "step": 14090 + }, + { + "epoch": 0.5496111793252646, + "grad_norm": 14.509554862976074, + "learning_rate": 9.790625143130253e-07, + "loss": 0.8649, + "step": 14100 + }, + { + "epoch": 0.5500009744879066, + "grad_norm": 12.980207443237305, + "learning_rate": 9.78997526351491e-07, + "loss": 0.9517, + "step": 14110 + }, + { + "epoch": 0.5503907696505487, + "grad_norm": 14.922877311706543, + "learning_rate": 9.789324398515964e-07, + "loss": 0.9196, + "step": 14120 + }, + { + "epoch": 0.5507805648131907, + "grad_norm": 15.891678810119629, + "learning_rate": 9.78867254826731e-07, + "loss": 0.9887, + "step": 14130 + }, + { + "epoch": 0.5511703599758327, + "grad_norm": 14.856050491333008, + "learning_rate": 9.788019712903041e-07, + "loss": 0.9296, + "step": 14140 + }, + { + "epoch": 0.5515601551384748, + "grad_norm": 14.890153884887695, + "learning_rate": 9.787365892557462e-07, + "loss": 0.9009, + "step": 14150 + }, + { + "epoch": 0.5519499503011167, + "grad_norm": 12.388936996459961, + "learning_rate": 9.786711087365073e-07, + "loss": 0.9627, + "step": 14160 + }, + { + "epoch": 0.5523397454637587, + "grad_norm": 12.587790489196777, + "learning_rate": 9.786055297460582e-07, + "loss": 0.9406, + "step": 14170 + }, + { + "epoch": 0.5527295406264008, + "grad_norm": 12.88075065612793, + "learning_rate": 9.785398522978894e-07, + "loss": 0.9255, + "step": 14180 + }, + { + "epoch": 0.5531193357890428, + "grad_norm": 13.306418418884277, + "learning_rate": 9.78474076405512e-07, + "loss": 0.951, + "step": 14190 + }, + { + "epoch": 0.5535091309516849, + "grad_norm": 12.390704154968262, + "learning_rate": 9.78408202082457e-07, + "loss": 0.9633, + "step": 14200 + }, + { + "epoch": 0.5538989261143269, + "grad_norm": 14.253022193908691, + "learning_rate": 9.783422293422765e-07, + "loss": 0.9474, + "step": 14210 + }, + { + "epoch": 0.5542887212769689, + "grad_norm": 15.160282135009766, + "learning_rate": 9.78276158198542e-07, + "loss": 0.9773, + "step": 14220 + }, + { + "epoch": 0.554678516439611, + "grad_norm": 13.04218578338623, + "learning_rate": 9.78209988664845e-07, + "loss": 0.9275, + "step": 14230 + }, + { + "epoch": 0.555068311602253, + "grad_norm": 13.458170890808105, + "learning_rate": 9.781437207547989e-07, + "loss": 0.8571, + "step": 14240 + }, + { + "epoch": 0.555458106764895, + "grad_norm": 15.279983520507812, + "learning_rate": 9.78077354482035e-07, + "loss": 0.9139, + "step": 14250 + }, + { + "epoch": 0.5558479019275371, + "grad_norm": 13.530505180358887, + "learning_rate": 9.78010889860207e-07, + "loss": 0.9253, + "step": 14260 + }, + { + "epoch": 0.5562376970901791, + "grad_norm": 13.807106018066406, + "learning_rate": 9.779443269029872e-07, + "loss": 0.9331, + "step": 14270 + }, + { + "epoch": 0.5566274922528212, + "grad_norm": 12.169309616088867, + "learning_rate": 9.77877665624069e-07, + "loss": 0.9306, + "step": 14280 + }, + { + "epoch": 0.5570172874154632, + "grad_norm": 11.47276496887207, + "learning_rate": 9.77810906037166e-07, + "loss": 0.923, + "step": 14290 + }, + { + "epoch": 0.5574070825781052, + "grad_norm": 13.763641357421875, + "learning_rate": 9.777440481560115e-07, + "loss": 0.9289, + "step": 14300 + }, + { + "epoch": 0.5577968777407473, + "grad_norm": 18.175439834594727, + "learning_rate": 9.776770919943595e-07, + "loss": 0.9634, + "step": 14310 + }, + { + "epoch": 0.5581866729033893, + "grad_norm": 13.934295654296875, + "learning_rate": 9.776100375659843e-07, + "loss": 0.9285, + "step": 14320 + }, + { + "epoch": 0.5585764680660313, + "grad_norm": 13.628682136535645, + "learning_rate": 9.7754288488468e-07, + "loss": 0.9681, + "step": 14330 + }, + { + "epoch": 0.5589662632286734, + "grad_norm": 12.14375114440918, + "learning_rate": 9.774756339642612e-07, + "loss": 0.8803, + "step": 14340 + }, + { + "epoch": 0.5593560583913154, + "grad_norm": 16.061145782470703, + "learning_rate": 9.774082848185622e-07, + "loss": 0.951, + "step": 14350 + }, + { + "epoch": 0.5597458535539575, + "grad_norm": 11.570030212402344, + "learning_rate": 9.773408374614385e-07, + "loss": 0.8707, + "step": 14360 + }, + { + "epoch": 0.5601356487165994, + "grad_norm": 12.731108665466309, + "learning_rate": 9.77273291906765e-07, + "loss": 0.9334, + "step": 14370 + }, + { + "epoch": 0.5605254438792414, + "grad_norm": 13.811582565307617, + "learning_rate": 9.77205648168437e-07, + "loss": 0.9898, + "step": 14380 + }, + { + "epoch": 0.5609152390418835, + "grad_norm": 14.607145309448242, + "learning_rate": 9.7713790626037e-07, + "loss": 0.8703, + "step": 14390 + }, + { + "epoch": 0.5613050342045255, + "grad_norm": 12.712349891662598, + "learning_rate": 9.770700661965e-07, + "loss": 0.8604, + "step": 14400 + }, + { + "epoch": 0.5616948293671675, + "grad_norm": 12.404290199279785, + "learning_rate": 9.770021279907824e-07, + "loss": 0.9375, + "step": 14410 + }, + { + "epoch": 0.5620846245298096, + "grad_norm": 14.441849708557129, + "learning_rate": 9.769340916571938e-07, + "loss": 0.9484, + "step": 14420 + }, + { + "epoch": 0.5624744196924516, + "grad_norm": 14.22604751586914, + "learning_rate": 9.7686595720973e-07, + "loss": 0.9795, + "step": 14430 + }, + { + "epoch": 0.5628642148550936, + "grad_norm": 13.021278381347656, + "learning_rate": 9.76797724662408e-07, + "loss": 0.9, + "step": 14440 + }, + { + "epoch": 0.5632540100177357, + "grad_norm": 15.578808784484863, + "learning_rate": 9.767293940292642e-07, + "loss": 0.8955, + "step": 14450 + }, + { + "epoch": 0.5636438051803777, + "grad_norm": 13.056351661682129, + "learning_rate": 9.766609653243555e-07, + "loss": 0.914, + "step": 14460 + }, + { + "epoch": 0.5640336003430197, + "grad_norm": 13.703203201293945, + "learning_rate": 9.765924385617586e-07, + "loss": 0.9186, + "step": 14470 + }, + { + "epoch": 0.5644233955056618, + "grad_norm": 14.69194221496582, + "learning_rate": 9.765238137555711e-07, + "loss": 0.9529, + "step": 14480 + }, + { + "epoch": 0.5648131906683038, + "grad_norm": 11.896600723266602, + "learning_rate": 9.764550909199102e-07, + "loss": 0.966, + "step": 14490 + }, + { + "epoch": 0.5652029858309459, + "grad_norm": 11.986417770385742, + "learning_rate": 9.763862700689134e-07, + "loss": 1.0205, + "step": 14500 + }, + { + "epoch": 0.5652029858309459, + "eval_loss": 0.9332651495933533, + "eval_runtime": 82.9497, + "eval_samples_per_second": 49.994, + "eval_steps_per_second": 6.257, + "step": 14500 + }, + { + "epoch": 0.5655927809935879, + "grad_norm": 14.664403915405273, + "learning_rate": 9.763173512167384e-07, + "loss": 0.96, + "step": 14510 + }, + { + "epoch": 0.5659825761562299, + "grad_norm": 13.48989486694336, + "learning_rate": 9.762483343775629e-07, + "loss": 0.9248, + "step": 14520 + }, + { + "epoch": 0.566372371318872, + "grad_norm": 14.038372993469238, + "learning_rate": 9.76179219565585e-07, + "loss": 0.9614, + "step": 14530 + }, + { + "epoch": 0.566762166481514, + "grad_norm": 12.835795402526855, + "learning_rate": 9.761100067950232e-07, + "loss": 0.9135, + "step": 14540 + }, + { + "epoch": 0.567151961644156, + "grad_norm": 13.142671585083008, + "learning_rate": 9.760406960801152e-07, + "loss": 0.928, + "step": 14550 + }, + { + "epoch": 0.5675417568067981, + "grad_norm": 13.363430976867676, + "learning_rate": 9.759712874351197e-07, + "loss": 0.8511, + "step": 14560 + }, + { + "epoch": 0.56793155196944, + "grad_norm": 13.882081985473633, + "learning_rate": 9.759017808743156e-07, + "loss": 0.9221, + "step": 14570 + }, + { + "epoch": 0.568321347132082, + "grad_norm": 12.949254989624023, + "learning_rate": 9.758321764120011e-07, + "loss": 0.9311, + "step": 14580 + }, + { + "epoch": 0.5687111422947241, + "grad_norm": 13.628260612487793, + "learning_rate": 9.757624740624954e-07, + "loss": 0.8656, + "step": 14590 + }, + { + "epoch": 0.5691009374573661, + "grad_norm": 13.459901809692383, + "learning_rate": 9.756926738401377e-07, + "loss": 0.9441, + "step": 14600 + }, + { + "epoch": 0.5694907326200082, + "grad_norm": 13.9908447265625, + "learning_rate": 9.756227757592869e-07, + "loss": 0.9483, + "step": 14610 + }, + { + "epoch": 0.5698805277826502, + "grad_norm": 11.984663963317871, + "learning_rate": 9.75552779834322e-07, + "loss": 0.9374, + "step": 14620 + }, + { + "epoch": 0.5702703229452922, + "grad_norm": 14.290929794311523, + "learning_rate": 9.75482686079643e-07, + "loss": 0.9553, + "step": 14630 + }, + { + "epoch": 0.5706601181079343, + "grad_norm": 12.609560012817383, + "learning_rate": 9.75412494509669e-07, + "loss": 0.9031, + "step": 14640 + }, + { + "epoch": 0.5710499132705763, + "grad_norm": 11.696626663208008, + "learning_rate": 9.7534220513884e-07, + "loss": 0.9523, + "step": 14650 + }, + { + "epoch": 0.5714397084332183, + "grad_norm": 13.471981048583984, + "learning_rate": 9.752718179816156e-07, + "loss": 0.9603, + "step": 14660 + }, + { + "epoch": 0.5718295035958604, + "grad_norm": 11.691251754760742, + "learning_rate": 9.752013330524757e-07, + "loss": 0.9209, + "step": 14670 + }, + { + "epoch": 0.5722192987585024, + "grad_norm": 16.88658905029297, + "learning_rate": 9.751307503659205e-07, + "loss": 1.0146, + "step": 14680 + }, + { + "epoch": 0.5726090939211445, + "grad_norm": 14.11522388458252, + "learning_rate": 9.750600699364698e-07, + "loss": 0.9527, + "step": 14690 + }, + { + "epoch": 0.5729988890837865, + "grad_norm": 13.263687133789062, + "learning_rate": 9.749892917786638e-07, + "loss": 0.877, + "step": 14700 + }, + { + "epoch": 0.5733886842464285, + "grad_norm": 11.954124450683594, + "learning_rate": 9.749184159070632e-07, + "loss": 0.9763, + "step": 14710 + }, + { + "epoch": 0.5737784794090706, + "grad_norm": 13.571649551391602, + "learning_rate": 9.74847442336248e-07, + "loss": 0.8891, + "step": 14720 + }, + { + "epoch": 0.5741682745717126, + "grad_norm": 13.065117835998535, + "learning_rate": 9.747763710808193e-07, + "loss": 0.8819, + "step": 14730 + }, + { + "epoch": 0.5745580697343546, + "grad_norm": 15.075433731079102, + "learning_rate": 9.747052021553975e-07, + "loss": 0.9353, + "step": 14740 + }, + { + "epoch": 0.5749478648969967, + "grad_norm": 13.564952850341797, + "learning_rate": 9.746339355746228e-07, + "loss": 0.9166, + "step": 14750 + }, + { + "epoch": 0.5753376600596387, + "grad_norm": 11.526659965515137, + "learning_rate": 9.745625713531566e-07, + "loss": 0.8548, + "step": 14760 + }, + { + "epoch": 0.5757274552222806, + "grad_norm": 13.383127212524414, + "learning_rate": 9.744911095056797e-07, + "loss": 0.8882, + "step": 14770 + }, + { + "epoch": 0.5761172503849227, + "grad_norm": 14.814136505126953, + "learning_rate": 9.74419550046893e-07, + "loss": 0.9855, + "step": 14780 + }, + { + "epoch": 0.5765070455475647, + "grad_norm": 14.393173217773438, + "learning_rate": 9.743478929915177e-07, + "loss": 0.9822, + "step": 14790 + }, + { + "epoch": 0.5768968407102067, + "grad_norm": 12.922545433044434, + "learning_rate": 9.742761383542946e-07, + "loss": 0.9142, + "step": 14800 + }, + { + "epoch": 0.5772866358728488, + "grad_norm": 13.655652046203613, + "learning_rate": 9.742042861499852e-07, + "loss": 0.9056, + "step": 14810 + }, + { + "epoch": 0.5776764310354908, + "grad_norm": 13.439281463623047, + "learning_rate": 9.741323363933707e-07, + "loss": 1.0104, + "step": 14820 + }, + { + "epoch": 0.5780662261981329, + "grad_norm": 13.212394714355469, + "learning_rate": 9.740602890992527e-07, + "loss": 0.8538, + "step": 14830 + }, + { + "epoch": 0.5784560213607749, + "grad_norm": 15.052072525024414, + "learning_rate": 9.73988144282452e-07, + "loss": 0.8838, + "step": 14840 + }, + { + "epoch": 0.5788458165234169, + "grad_norm": 14.627418518066406, + "learning_rate": 9.739159019578108e-07, + "loss": 0.8923, + "step": 14850 + }, + { + "epoch": 0.579235611686059, + "grad_norm": 14.643178939819336, + "learning_rate": 9.7384356214019e-07, + "loss": 0.9401, + "step": 14860 + }, + { + "epoch": 0.579625406848701, + "grad_norm": 13.664984703063965, + "learning_rate": 9.737711248444717e-07, + "loss": 0.948, + "step": 14870 + }, + { + "epoch": 0.580015202011343, + "grad_norm": 13.236997604370117, + "learning_rate": 9.736985900855572e-07, + "loss": 0.9362, + "step": 14880 + }, + { + "epoch": 0.5804049971739851, + "grad_norm": 13.343585014343262, + "learning_rate": 9.736259578783684e-07, + "loss": 0.8688, + "step": 14890 + }, + { + "epoch": 0.5807947923366271, + "grad_norm": 12.975387573242188, + "learning_rate": 9.735532282378469e-07, + "loss": 0.9811, + "step": 14900 + }, + { + "epoch": 0.5811845874992692, + "grad_norm": 14.161971092224121, + "learning_rate": 9.734804011789548e-07, + "loss": 0.9165, + "step": 14910 + }, + { + "epoch": 0.5815743826619112, + "grad_norm": 12.795123100280762, + "learning_rate": 9.734074767166736e-07, + "loss": 0.9463, + "step": 14920 + }, + { + "epoch": 0.5819641778245532, + "grad_norm": 13.976717948913574, + "learning_rate": 9.733344548660051e-07, + "loss": 0.8939, + "step": 14930 + }, + { + "epoch": 0.5823539729871953, + "grad_norm": 13.190685272216797, + "learning_rate": 9.732613356419715e-07, + "loss": 0.9431, + "step": 14940 + }, + { + "epoch": 0.5827437681498373, + "grad_norm": 14.345075607299805, + "learning_rate": 9.731881190596144e-07, + "loss": 0.9193, + "step": 14950 + }, + { + "epoch": 0.5831335633124793, + "grad_norm": 12.450021743774414, + "learning_rate": 9.731148051339965e-07, + "loss": 0.8871, + "step": 14960 + }, + { + "epoch": 0.5835233584751214, + "grad_norm": 14.945511817932129, + "learning_rate": 9.730413938801987e-07, + "loss": 0.8838, + "step": 14970 + }, + { + "epoch": 0.5839131536377633, + "grad_norm": 13.849363327026367, + "learning_rate": 9.72967885313324e-07, + "loss": 0.9472, + "step": 14980 + }, + { + "epoch": 0.5843029488004053, + "grad_norm": 14.438311576843262, + "learning_rate": 9.728942794484935e-07, + "loss": 0.9614, + "step": 14990 + }, + { + "epoch": 0.5846927439630474, + "grad_norm": 15.024739265441895, + "learning_rate": 9.7282057630085e-07, + "loss": 0.9035, + "step": 15000 + }, + { + "epoch": 0.5846927439630474, + "eval_loss": 0.9312300682067871, + "eval_runtime": 84.8415, + "eval_samples_per_second": 48.879, + "eval_steps_per_second": 6.117, + "step": 15000 + }, + { + "epoch": 0.5850825391256894, + "grad_norm": 13.873088836669922, + "learning_rate": 9.727467758855551e-07, + "loss": 0.9482, + "step": 15010 + }, + { + "epoch": 0.5854723342883315, + "grad_norm": 15.514175415039062, + "learning_rate": 9.726728782177913e-07, + "loss": 0.9242, + "step": 15020 + }, + { + "epoch": 0.5858621294509735, + "grad_norm": 14.53191089630127, + "learning_rate": 9.725988833127602e-07, + "loss": 0.9825, + "step": 15030 + }, + { + "epoch": 0.5862519246136155, + "grad_norm": 11.947699546813965, + "learning_rate": 9.725247911856843e-07, + "loss": 0.905, + "step": 15040 + }, + { + "epoch": 0.5866417197762576, + "grad_norm": 11.926911354064941, + "learning_rate": 9.724506018518054e-07, + "loss": 0.9598, + "step": 15050 + }, + { + "epoch": 0.5870315149388996, + "grad_norm": 14.044079780578613, + "learning_rate": 9.723763153263855e-07, + "loss": 0.9202, + "step": 15060 + }, + { + "epoch": 0.5874213101015416, + "grad_norm": 13.975722312927246, + "learning_rate": 9.72301931624707e-07, + "loss": 0.8824, + "step": 15070 + }, + { + "epoch": 0.5878111052641837, + "grad_norm": 11.088001251220703, + "learning_rate": 9.722274507620718e-07, + "loss": 0.9609, + "step": 15080 + }, + { + "epoch": 0.5882009004268257, + "grad_norm": 13.190027236938477, + "learning_rate": 9.721528727538019e-07, + "loss": 0.9561, + "step": 15090 + }, + { + "epoch": 0.5885906955894677, + "grad_norm": 10.689318656921387, + "learning_rate": 9.72078197615239e-07, + "loss": 0.9256, + "step": 15100 + }, + { + "epoch": 0.5889804907521098, + "grad_norm": 13.590693473815918, + "learning_rate": 9.720034253617458e-07, + "loss": 0.8957, + "step": 15110 + }, + { + "epoch": 0.5893702859147518, + "grad_norm": 13.708724975585938, + "learning_rate": 9.71928556008704e-07, + "loss": 0.9072, + "step": 15120 + }, + { + "epoch": 0.5897600810773939, + "grad_norm": 13.530667304992676, + "learning_rate": 9.718535895715152e-07, + "loss": 0.9833, + "step": 15130 + }, + { + "epoch": 0.5901498762400359, + "grad_norm": 15.110366821289062, + "learning_rate": 9.717785260656018e-07, + "loss": 0.9469, + "step": 15140 + }, + { + "epoch": 0.5905396714026779, + "grad_norm": 12.867822647094727, + "learning_rate": 9.717033655064056e-07, + "loss": 0.9107, + "step": 15150 + }, + { + "epoch": 0.59092946656532, + "grad_norm": 13.896025657653809, + "learning_rate": 9.716281079093885e-07, + "loss": 0.8865, + "step": 15160 + }, + { + "epoch": 0.591319261727962, + "grad_norm": 13.899539947509766, + "learning_rate": 9.71552753290032e-07, + "loss": 0.8806, + "step": 15170 + }, + { + "epoch": 0.5917090568906039, + "grad_norm": 12.652937889099121, + "learning_rate": 9.714773016638384e-07, + "loss": 0.946, + "step": 15180 + }, + { + "epoch": 0.592098852053246, + "grad_norm": 14.17641830444336, + "learning_rate": 9.714017530463291e-07, + "loss": 0.9832, + "step": 15190 + }, + { + "epoch": 0.592488647215888, + "grad_norm": 15.20641803741455, + "learning_rate": 9.71326107453046e-07, + "loss": 0.8682, + "step": 15200 + }, + { + "epoch": 0.59287844237853, + "grad_norm": 12.762015342712402, + "learning_rate": 9.712503648995505e-07, + "loss": 0.9543, + "step": 15210 + }, + { + "epoch": 0.5932682375411721, + "grad_norm": 13.665618896484375, + "learning_rate": 9.711745254014245e-07, + "loss": 0.9104, + "step": 15220 + }, + { + "epoch": 0.5936580327038141, + "grad_norm": 13.908102035522461, + "learning_rate": 9.710985889742694e-07, + "loss": 0.9, + "step": 15230 + }, + { + "epoch": 0.5940478278664562, + "grad_norm": 13.525138854980469, + "learning_rate": 9.710225556337065e-07, + "loss": 0.9442, + "step": 15240 + }, + { + "epoch": 0.5944376230290982, + "grad_norm": 12.227070808410645, + "learning_rate": 9.709464253953777e-07, + "loss": 0.9722, + "step": 15250 + }, + { + "epoch": 0.5948274181917402, + "grad_norm": 14.163692474365234, + "learning_rate": 9.70870198274944e-07, + "loss": 0.9333, + "step": 15260 + }, + { + "epoch": 0.5952172133543823, + "grad_norm": 14.082769393920898, + "learning_rate": 9.707938742880866e-07, + "loss": 0.8622, + "step": 15270 + }, + { + "epoch": 0.5956070085170243, + "grad_norm": 15.043621063232422, + "learning_rate": 9.707174534505067e-07, + "loss": 0.9066, + "step": 15280 + }, + { + "epoch": 0.5959968036796663, + "grad_norm": 13.600968360900879, + "learning_rate": 9.706409357779257e-07, + "loss": 0.8419, + "step": 15290 + }, + { + "epoch": 0.5963865988423084, + "grad_norm": 15.225130081176758, + "learning_rate": 9.705643212860847e-07, + "loss": 0.9828, + "step": 15300 + }, + { + "epoch": 0.5967763940049504, + "grad_norm": 14.643730163574219, + "learning_rate": 9.704876099907445e-07, + "loss": 0.9922, + "step": 15310 + }, + { + "epoch": 0.5971661891675925, + "grad_norm": 14.380305290222168, + "learning_rate": 9.704108019076858e-07, + "loss": 0.9559, + "step": 15320 + }, + { + "epoch": 0.5975559843302345, + "grad_norm": 15.331977844238281, + "learning_rate": 9.703338970527095e-07, + "loss": 0.9032, + "step": 15330 + }, + { + "epoch": 0.5979457794928765, + "grad_norm": 12.51699161529541, + "learning_rate": 9.702568954416365e-07, + "loss": 0.9268, + "step": 15340 + }, + { + "epoch": 0.5983355746555186, + "grad_norm": 18.02459144592285, + "learning_rate": 9.701797970903074e-07, + "loss": 0.8601, + "step": 15350 + }, + { + "epoch": 0.5987253698181606, + "grad_norm": 18.30661392211914, + "learning_rate": 9.701026020145827e-07, + "loss": 0.9346, + "step": 15360 + }, + { + "epoch": 0.5991151649808026, + "grad_norm": 13.292183876037598, + "learning_rate": 9.700253102303425e-07, + "loss": 0.919, + "step": 15370 + }, + { + "epoch": 0.5995049601434446, + "grad_norm": 13.807863235473633, + "learning_rate": 9.699479217534874e-07, + "loss": 0.95, + "step": 15380 + }, + { + "epoch": 0.5998947553060866, + "grad_norm": 14.404817581176758, + "learning_rate": 9.698704365999373e-07, + "loss": 0.8666, + "step": 15390 + }, + { + "epoch": 0.6002845504687286, + "grad_norm": 13.505146026611328, + "learning_rate": 9.697928547856327e-07, + "loss": 0.9689, + "step": 15400 + }, + { + "epoch": 0.6006743456313707, + "grad_norm": 13.610662460327148, + "learning_rate": 9.697151763265333e-07, + "loss": 0.8671, + "step": 15410 + }, + { + "epoch": 0.6010641407940127, + "grad_norm": 14.029986381530762, + "learning_rate": 9.69637401238619e-07, + "loss": 0.8984, + "step": 15420 + }, + { + "epoch": 0.6014539359566548, + "grad_norm": 12.155616760253906, + "learning_rate": 9.695595295378897e-07, + "loss": 0.9306, + "step": 15430 + }, + { + "epoch": 0.6018437311192968, + "grad_norm": 10.464404106140137, + "learning_rate": 9.694815612403648e-07, + "loss": 0.8636, + "step": 15440 + }, + { + "epoch": 0.6022335262819388, + "grad_norm": 13.606799125671387, + "learning_rate": 9.694034963620837e-07, + "loss": 0.9473, + "step": 15450 + }, + { + "epoch": 0.6026233214445809, + "grad_norm": 14.624902725219727, + "learning_rate": 9.69325334919106e-07, + "loss": 0.9452, + "step": 15460 + }, + { + "epoch": 0.6030131166072229, + "grad_norm": 15.783787727355957, + "learning_rate": 9.692470769275105e-07, + "loss": 0.9759, + "step": 15470 + }, + { + "epoch": 0.6034029117698649, + "grad_norm": 13.32320785522461, + "learning_rate": 9.691687224033967e-07, + "loss": 0.8984, + "step": 15480 + }, + { + "epoch": 0.603792706932507, + "grad_norm": 13.195551872253418, + "learning_rate": 9.690902713628835e-07, + "loss": 0.9504, + "step": 15490 + }, + { + "epoch": 0.604182502095149, + "grad_norm": 13.497276306152344, + "learning_rate": 9.690117238221092e-07, + "loss": 0.9884, + "step": 15500 + }, + { + "epoch": 0.604182502095149, + "eval_loss": 0.9286776781082153, + "eval_runtime": 82.7682, + "eval_samples_per_second": 50.104, + "eval_steps_per_second": 6.271, + "step": 15500 + }, + { + "epoch": 0.604572297257791, + "grad_norm": 12.89731502532959, + "learning_rate": 9.689330797972331e-07, + "loss": 0.9238, + "step": 15510 + }, + { + "epoch": 0.6049620924204331, + "grad_norm": 15.027881622314453, + "learning_rate": 9.688543393044332e-07, + "loss": 0.9132, + "step": 15520 + }, + { + "epoch": 0.6053518875830751, + "grad_norm": 11.512370109558105, + "learning_rate": 9.68775502359908e-07, + "loss": 0.8844, + "step": 15530 + }, + { + "epoch": 0.6057416827457172, + "grad_norm": 13.461048126220703, + "learning_rate": 9.686965689798756e-07, + "loss": 0.8432, + "step": 15540 + }, + { + "epoch": 0.6061314779083592, + "grad_norm": 15.977388381958008, + "learning_rate": 9.686175391805742e-07, + "loss": 0.9306, + "step": 15550 + }, + { + "epoch": 0.6065212730710012, + "grad_norm": 13.019350051879883, + "learning_rate": 9.685384129782614e-07, + "loss": 0.9088, + "step": 15560 + }, + { + "epoch": 0.6069110682336433, + "grad_norm": 13.332326889038086, + "learning_rate": 9.68459190389215e-07, + "loss": 0.8991, + "step": 15570 + }, + { + "epoch": 0.6073008633962853, + "grad_norm": 13.904509544372559, + "learning_rate": 9.683798714297325e-07, + "loss": 0.8896, + "step": 15580 + }, + { + "epoch": 0.6076906585589272, + "grad_norm": 11.58399772644043, + "learning_rate": 9.683004561161313e-07, + "loss": 0.9207, + "step": 15590 + }, + { + "epoch": 0.6080804537215693, + "grad_norm": 15.235713958740234, + "learning_rate": 9.682209444647484e-07, + "loss": 0.9597, + "step": 15600 + }, + { + "epoch": 0.6084702488842113, + "grad_norm": 11.916403770446777, + "learning_rate": 9.681413364919408e-07, + "loss": 0.8989, + "step": 15610 + }, + { + "epoch": 0.6088600440468533, + "grad_norm": 15.490662574768066, + "learning_rate": 9.680616322140853e-07, + "loss": 0.9197, + "step": 15620 + }, + { + "epoch": 0.6092498392094954, + "grad_norm": 14.118707656860352, + "learning_rate": 9.679818316475786e-07, + "loss": 0.9336, + "step": 15630 + }, + { + "epoch": 0.6096396343721374, + "grad_norm": 13.686227798461914, + "learning_rate": 9.679019348088372e-07, + "loss": 0.9126, + "step": 15640 + }, + { + "epoch": 0.6100294295347795, + "grad_norm": 13.507862091064453, + "learning_rate": 9.678219417142967e-07, + "loss": 0.9242, + "step": 15650 + }, + { + "epoch": 0.6104192246974215, + "grad_norm": 13.80568790435791, + "learning_rate": 9.677418523804138e-07, + "loss": 0.9883, + "step": 15660 + }, + { + "epoch": 0.6108090198600635, + "grad_norm": 13.460494995117188, + "learning_rate": 9.67661666823664e-07, + "loss": 0.9448, + "step": 15670 + }, + { + "epoch": 0.6111988150227056, + "grad_norm": 13.540401458740234, + "learning_rate": 9.67581385060543e-07, + "loss": 0.9035, + "step": 15680 + }, + { + "epoch": 0.6115886101853476, + "grad_norm": 13.659446716308594, + "learning_rate": 9.675010071075662e-07, + "loss": 0.9534, + "step": 15690 + }, + { + "epoch": 0.6119784053479896, + "grad_norm": 14.222783088684082, + "learning_rate": 9.674205329812686e-07, + "loss": 0.9487, + "step": 15700 + }, + { + "epoch": 0.6123682005106317, + "grad_norm": 16.12778091430664, + "learning_rate": 9.673399626982052e-07, + "loss": 0.8918, + "step": 15710 + }, + { + "epoch": 0.6127579956732737, + "grad_norm": 12.613432884216309, + "learning_rate": 9.672592962749511e-07, + "loss": 0.8443, + "step": 15720 + }, + { + "epoch": 0.6131477908359158, + "grad_norm": 14.32766342163086, + "learning_rate": 9.671785337281004e-07, + "loss": 0.8849, + "step": 15730 + }, + { + "epoch": 0.6135375859985578, + "grad_norm": 14.50903034210205, + "learning_rate": 9.670976750742676e-07, + "loss": 0.9591, + "step": 15740 + }, + { + "epoch": 0.6139273811611998, + "grad_norm": 13.125991821289062, + "learning_rate": 9.67016720330087e-07, + "loss": 0.8836, + "step": 15750 + }, + { + "epoch": 0.6143171763238419, + "grad_norm": 12.54773998260498, + "learning_rate": 9.669356695122117e-07, + "loss": 0.9041, + "step": 15760 + }, + { + "epoch": 0.6147069714864839, + "grad_norm": 12.715425491333008, + "learning_rate": 9.66854522637316e-07, + "loss": 0.9109, + "step": 15770 + }, + { + "epoch": 0.6150967666491259, + "grad_norm": 14.903458595275879, + "learning_rate": 9.667732797220934e-07, + "loss": 0.9503, + "step": 15780 + }, + { + "epoch": 0.6154865618117679, + "grad_norm": 11.897772789001465, + "learning_rate": 9.666919407832564e-07, + "loss": 0.9108, + "step": 15790 + }, + { + "epoch": 0.6158763569744099, + "grad_norm": 14.889135360717773, + "learning_rate": 9.66610505837538e-07, + "loss": 0.9142, + "step": 15800 + }, + { + "epoch": 0.6162661521370519, + "grad_norm": 16.108352661132812, + "learning_rate": 9.66528974901691e-07, + "loss": 0.9586, + "step": 15810 + }, + { + "epoch": 0.616655947299694, + "grad_norm": 14.593862533569336, + "learning_rate": 9.664473479924879e-07, + "loss": 0.8876, + "step": 15820 + }, + { + "epoch": 0.617045742462336, + "grad_norm": 15.885119438171387, + "learning_rate": 9.663656251267207e-07, + "loss": 0.9131, + "step": 15830 + }, + { + "epoch": 0.617435537624978, + "grad_norm": 14.007955551147461, + "learning_rate": 9.662838063212011e-07, + "loss": 0.8938, + "step": 15840 + }, + { + "epoch": 0.6178253327876201, + "grad_norm": 11.49051284790039, + "learning_rate": 9.662018915927608e-07, + "loss": 0.9216, + "step": 15850 + }, + { + "epoch": 0.6182151279502621, + "grad_norm": 13.827402114868164, + "learning_rate": 9.661198809582512e-07, + "loss": 1.002, + "step": 15860 + }, + { + "epoch": 0.6186049231129042, + "grad_norm": 16.070764541625977, + "learning_rate": 9.660377744345431e-07, + "loss": 0.8751, + "step": 15870 + }, + { + "epoch": 0.6189947182755462, + "grad_norm": 12.560461044311523, + "learning_rate": 9.659555720385276e-07, + "loss": 0.9001, + "step": 15880 + }, + { + "epoch": 0.6193845134381882, + "grad_norm": 15.653770446777344, + "learning_rate": 9.658732737871152e-07, + "loss": 0.9014, + "step": 15890 + }, + { + "epoch": 0.6197743086008303, + "grad_norm": 13.841830253601074, + "learning_rate": 9.657908796972358e-07, + "loss": 0.8766, + "step": 15900 + }, + { + "epoch": 0.6201641037634723, + "grad_norm": 11.792221069335938, + "learning_rate": 9.657083897858396e-07, + "loss": 0.9141, + "step": 15910 + }, + { + "epoch": 0.6205538989261143, + "grad_norm": 11.314265251159668, + "learning_rate": 9.656258040698962e-07, + "loss": 0.9024, + "step": 15920 + }, + { + "epoch": 0.6209436940887564, + "grad_norm": 12.276455879211426, + "learning_rate": 9.65543122566395e-07, + "loss": 0.8627, + "step": 15930 + }, + { + "epoch": 0.6213334892513984, + "grad_norm": 14.050057411193848, + "learning_rate": 9.65460345292345e-07, + "loss": 0.922, + "step": 15940 + }, + { + "epoch": 0.6217232844140405, + "grad_norm": 13.932723045349121, + "learning_rate": 9.653774722647748e-07, + "loss": 0.8926, + "step": 15950 + }, + { + "epoch": 0.6221130795766825, + "grad_norm": 14.086650848388672, + "learning_rate": 9.652945035007332e-07, + "loss": 0.9316, + "step": 15960 + }, + { + "epoch": 0.6225028747393245, + "grad_norm": 12.91126823425293, + "learning_rate": 9.652114390172881e-07, + "loss": 0.9548, + "step": 15970 + }, + { + "epoch": 0.6228926699019666, + "grad_norm": 16.955249786376953, + "learning_rate": 9.651282788315275e-07, + "loss": 0.9027, + "step": 15980 + }, + { + "epoch": 0.6232824650646085, + "grad_norm": 16.98932647705078, + "learning_rate": 9.65045022960559e-07, + "loss": 0.9187, + "step": 15990 + }, + { + "epoch": 0.6236722602272505, + "grad_norm": 12.131331443786621, + "learning_rate": 9.649616714215095e-07, + "loss": 0.8686, + "step": 16000 + }, + { + "epoch": 0.6236722602272505, + "eval_loss": 0.9259690046310425, + "eval_runtime": 82.7869, + "eval_samples_per_second": 50.092, + "eval_steps_per_second": 6.269, + "step": 16000 + }, + { + "epoch": 0.6240620553898926, + "grad_norm": 13.705302238464355, + "learning_rate": 9.648782242315261e-07, + "loss": 0.9617, + "step": 16010 + }, + { + "epoch": 0.6244518505525346, + "grad_norm": 12.093605041503906, + "learning_rate": 9.647946814077756e-07, + "loss": 0.8819, + "step": 16020 + }, + { + "epoch": 0.6248416457151766, + "grad_norm": 15.518138885498047, + "learning_rate": 9.647110429674439e-07, + "loss": 0.923, + "step": 16030 + }, + { + "epoch": 0.6252314408778187, + "grad_norm": 14.11060905456543, + "learning_rate": 9.64627308927737e-07, + "loss": 0.8766, + "step": 16040 + }, + { + "epoch": 0.6256212360404607, + "grad_norm": 12.744857788085938, + "learning_rate": 9.645434793058809e-07, + "loss": 0.8605, + "step": 16050 + }, + { + "epoch": 0.6260110312031028, + "grad_norm": 14.980656623840332, + "learning_rate": 9.6445955411912e-07, + "loss": 0.825, + "step": 16060 + }, + { + "epoch": 0.6264008263657448, + "grad_norm": 13.311100006103516, + "learning_rate": 9.643755333847201e-07, + "loss": 0.9215, + "step": 16070 + }, + { + "epoch": 0.6267906215283868, + "grad_norm": 12.126730918884277, + "learning_rate": 9.642914171199652e-07, + "loss": 0.873, + "step": 16080 + }, + { + "epoch": 0.6271804166910289, + "grad_norm": 14.918853759765625, + "learning_rate": 9.642072053421598e-07, + "loss": 0.9796, + "step": 16090 + }, + { + "epoch": 0.6275702118536709, + "grad_norm": 14.137699127197266, + "learning_rate": 9.641228980686277e-07, + "loss": 0.9217, + "step": 16100 + }, + { + "epoch": 0.6279600070163129, + "grad_norm": 14.747109413146973, + "learning_rate": 9.640384953167124e-07, + "loss": 0.9415, + "step": 16110 + }, + { + "epoch": 0.628349802178955, + "grad_norm": 14.193294525146484, + "learning_rate": 9.639539971037769e-07, + "loss": 0.9341, + "step": 16120 + }, + { + "epoch": 0.628739597341597, + "grad_norm": 14.41287612915039, + "learning_rate": 9.638694034472042e-07, + "loss": 0.9537, + "step": 16130 + }, + { + "epoch": 0.629129392504239, + "grad_norm": 14.716343879699707, + "learning_rate": 9.637847143643968e-07, + "loss": 0.9304, + "step": 16140 + }, + { + "epoch": 0.6295191876668811, + "grad_norm": 12.468219757080078, + "learning_rate": 9.636999298727766e-07, + "loss": 0.885, + "step": 16150 + }, + { + "epoch": 0.6299089828295231, + "grad_norm": 14.43801212310791, + "learning_rate": 9.636150499897851e-07, + "loss": 0.929, + "step": 16160 + }, + { + "epoch": 0.6302987779921652, + "grad_norm": 14.952789306640625, + "learning_rate": 9.63530074732884e-07, + "loss": 0.9512, + "step": 16170 + }, + { + "epoch": 0.6306885731548072, + "grad_norm": 13.220438003540039, + "learning_rate": 9.634450041195542e-07, + "loss": 0.8859, + "step": 16180 + }, + { + "epoch": 0.6310783683174492, + "grad_norm": 11.992671966552734, + "learning_rate": 9.633598381672957e-07, + "loss": 0.8911, + "step": 16190 + }, + { + "epoch": 0.6314681634800912, + "grad_norm": 13.544135093688965, + "learning_rate": 9.632745768936296e-07, + "loss": 0.8586, + "step": 16200 + }, + { + "epoch": 0.6318579586427332, + "grad_norm": 12.489459037780762, + "learning_rate": 9.631892203160947e-07, + "loss": 0.9116, + "step": 16210 + }, + { + "epoch": 0.6322477538053752, + "grad_norm": 14.536462783813477, + "learning_rate": 9.63103768452251e-07, + "loss": 0.8953, + "step": 16220 + }, + { + "epoch": 0.6326375489680173, + "grad_norm": 12.083584785461426, + "learning_rate": 9.630182213196773e-07, + "loss": 0.9196, + "step": 16230 + }, + { + "epoch": 0.6330273441306593, + "grad_norm": 13.038172721862793, + "learning_rate": 9.629325789359722e-07, + "loss": 0.8919, + "step": 16240 + }, + { + "epoch": 0.6334171392933013, + "grad_norm": 11.963274002075195, + "learning_rate": 9.628468413187536e-07, + "loss": 0.935, + "step": 16250 + }, + { + "epoch": 0.6338069344559434, + "grad_norm": 13.36107063293457, + "learning_rate": 9.627610084856596e-07, + "loss": 1.0302, + "step": 16260 + }, + { + "epoch": 0.6341967296185854, + "grad_norm": 14.64615249633789, + "learning_rate": 9.626750804543476e-07, + "loss": 0.9082, + "step": 16270 + }, + { + "epoch": 0.6345865247812275, + "grad_norm": 12.766172409057617, + "learning_rate": 9.625890572424942e-07, + "loss": 0.8899, + "step": 16280 + }, + { + "epoch": 0.6349763199438695, + "grad_norm": 13.398990631103516, + "learning_rate": 9.625029388677962e-07, + "loss": 0.8559, + "step": 16290 + }, + { + "epoch": 0.6353661151065115, + "grad_norm": 11.899115562438965, + "learning_rate": 9.624167253479695e-07, + "loss": 0.9243, + "step": 16300 + }, + { + "epoch": 0.6357559102691536, + "grad_norm": 12.475247383117676, + "learning_rate": 9.6233041670075e-07, + "loss": 0.8945, + "step": 16310 + }, + { + "epoch": 0.6361457054317956, + "grad_norm": 11.756570816040039, + "learning_rate": 9.622440129438927e-07, + "loss": 0.9147, + "step": 16320 + }, + { + "epoch": 0.6365355005944376, + "grad_norm": 13.796146392822266, + "learning_rate": 9.621575140951725e-07, + "loss": 0.917, + "step": 16330 + }, + { + "epoch": 0.6369252957570797, + "grad_norm": 12.747626304626465, + "learning_rate": 9.620709201723835e-07, + "loss": 0.8995, + "step": 16340 + }, + { + "epoch": 0.6373150909197217, + "grad_norm": 14.18416690826416, + "learning_rate": 9.6198423119334e-07, + "loss": 0.8875, + "step": 16350 + }, + { + "epoch": 0.6377048860823638, + "grad_norm": 13.69453239440918, + "learning_rate": 9.618974471758756e-07, + "loss": 0.8611, + "step": 16360 + }, + { + "epoch": 0.6380946812450058, + "grad_norm": 13.548929214477539, + "learning_rate": 9.618105681378428e-07, + "loss": 0.96, + "step": 16370 + }, + { + "epoch": 0.6384844764076478, + "grad_norm": 9.896097183227539, + "learning_rate": 9.617235940971145e-07, + "loss": 0.9082, + "step": 16380 + }, + { + "epoch": 0.6388742715702899, + "grad_norm": 13.362122535705566, + "learning_rate": 9.616365250715826e-07, + "loss": 0.9808, + "step": 16390 + }, + { + "epoch": 0.6392640667329318, + "grad_norm": 14.422347068786621, + "learning_rate": 9.615493610791589e-07, + "loss": 0.9172, + "step": 16400 + }, + { + "epoch": 0.6396538618955738, + "grad_norm": 12.475889205932617, + "learning_rate": 9.614621021377748e-07, + "loss": 0.9378, + "step": 16410 + }, + { + "epoch": 0.6400436570582159, + "grad_norm": 13.437970161437988, + "learning_rate": 9.613747482653806e-07, + "loss": 0.9848, + "step": 16420 + }, + { + "epoch": 0.6404334522208579, + "grad_norm": 12.905533790588379, + "learning_rate": 9.61287299479947e-07, + "loss": 0.9569, + "step": 16430 + }, + { + "epoch": 0.6408232473834999, + "grad_norm": 13.588510513305664, + "learning_rate": 9.611997557994634e-07, + "loss": 0.8712, + "step": 16440 + }, + { + "epoch": 0.641213042546142, + "grad_norm": 13.870532989501953, + "learning_rate": 9.61112117241939e-07, + "loss": 0.9351, + "step": 16450 + }, + { + "epoch": 0.641602837708784, + "grad_norm": 16.3858642578125, + "learning_rate": 9.610243838254034e-07, + "loss": 0.9331, + "step": 16460 + }, + { + "epoch": 0.641992632871426, + "grad_norm": 13.757356643676758, + "learning_rate": 9.609365555679042e-07, + "loss": 0.9322, + "step": 16470 + }, + { + "epoch": 0.6423824280340681, + "grad_norm": 12.890793800354004, + "learning_rate": 9.608486324875094e-07, + "loss": 0.9298, + "step": 16480 + }, + { + "epoch": 0.6427722231967101, + "grad_norm": 12.160136222839355, + "learning_rate": 9.607606146023063e-07, + "loss": 0.9629, + "step": 16490 + }, + { + "epoch": 0.6431620183593522, + "grad_norm": 14.943809509277344, + "learning_rate": 9.60672501930402e-07, + "loss": 0.9163, + "step": 16500 + }, + { + "epoch": 0.6431620183593522, + "eval_loss": 0.9235700368881226, + "eval_runtime": 83.3071, + "eval_samples_per_second": 49.78, + "eval_steps_per_second": 6.23, + "step": 16500 + }, + { + "epoch": 0.6435518135219942, + "grad_norm": 12.577781677246094, + "learning_rate": 9.605842944899227e-07, + "loss": 0.997, + "step": 16510 + }, + { + "epoch": 0.6439416086846362, + "grad_norm": 13.727602005004883, + "learning_rate": 9.604959922990143e-07, + "loss": 0.9319, + "step": 16520 + }, + { + "epoch": 0.6443314038472783, + "grad_norm": 15.657148361206055, + "learning_rate": 9.60407595375842e-07, + "loss": 0.8978, + "step": 16530 + }, + { + "epoch": 0.6447211990099203, + "grad_norm": 11.667567253112793, + "learning_rate": 9.603191037385908e-07, + "loss": 0.8425, + "step": 16540 + }, + { + "epoch": 0.6451109941725623, + "grad_norm": 15.091428756713867, + "learning_rate": 9.602305174054647e-07, + "loss": 0.9534, + "step": 16550 + }, + { + "epoch": 0.6455007893352044, + "grad_norm": 14.114659309387207, + "learning_rate": 9.60141836394688e-07, + "loss": 0.8448, + "step": 16560 + }, + { + "epoch": 0.6458905844978464, + "grad_norm": 13.17948055267334, + "learning_rate": 9.600530607245036e-07, + "loss": 0.9431, + "step": 16570 + }, + { + "epoch": 0.6462803796604885, + "grad_norm": 12.355938911437988, + "learning_rate": 9.599641904131742e-07, + "loss": 0.8883, + "step": 16580 + }, + { + "epoch": 0.6466701748231305, + "grad_norm": 13.770953178405762, + "learning_rate": 9.598752254789824e-07, + "loss": 0.8839, + "step": 16590 + }, + { + "epoch": 0.6470599699857725, + "grad_norm": 14.499216079711914, + "learning_rate": 9.597861659402294e-07, + "loss": 0.9217, + "step": 16600 + }, + { + "epoch": 0.6474497651484145, + "grad_norm": 15.786677360534668, + "learning_rate": 9.59697011815237e-07, + "loss": 0.9406, + "step": 16610 + }, + { + "epoch": 0.6478395603110565, + "grad_norm": 12.890243530273438, + "learning_rate": 9.59607763122345e-07, + "loss": 0.9404, + "step": 16620 + }, + { + "epoch": 0.6482293554736985, + "grad_norm": 12.807088851928711, + "learning_rate": 9.595184198799138e-07, + "loss": 1.0097, + "step": 16630 + }, + { + "epoch": 0.6486191506363406, + "grad_norm": 12.559128761291504, + "learning_rate": 9.594289821063231e-07, + "loss": 0.9801, + "step": 16640 + }, + { + "epoch": 0.6490089457989826, + "grad_norm": 14.907734870910645, + "learning_rate": 9.593394498199716e-07, + "loss": 0.8885, + "step": 16650 + }, + { + "epoch": 0.6493987409616246, + "grad_norm": 13.487298965454102, + "learning_rate": 9.592498230392778e-07, + "loss": 0.9405, + "step": 16660 + }, + { + "epoch": 0.6497885361242667, + "grad_norm": 13.930459976196289, + "learning_rate": 9.591601017826797e-07, + "loss": 0.9297, + "step": 16670 + }, + { + "epoch": 0.6501783312869087, + "grad_norm": 13.412724494934082, + "learning_rate": 9.59070286068634e-07, + "loss": 0.8674, + "step": 16680 + }, + { + "epoch": 0.6505681264495508, + "grad_norm": 14.93876838684082, + "learning_rate": 9.58980375915618e-07, + "loss": 0.962, + "step": 16690 + }, + { + "epoch": 0.6509579216121928, + "grad_norm": 12.107748985290527, + "learning_rate": 9.588903713421276e-07, + "loss": 0.9175, + "step": 16700 + }, + { + "epoch": 0.6513477167748348, + "grad_norm": 15.81786060333252, + "learning_rate": 9.588002723666782e-07, + "loss": 0.9428, + "step": 16710 + }, + { + "epoch": 0.6517375119374769, + "grad_norm": 16.012392044067383, + "learning_rate": 9.58710079007805e-07, + "loss": 0.9474, + "step": 16720 + }, + { + "epoch": 0.6521273071001189, + "grad_norm": 13.123269081115723, + "learning_rate": 9.586197912840624e-07, + "loss": 0.958, + "step": 16730 + }, + { + "epoch": 0.6525171022627609, + "grad_norm": 13.346026420593262, + "learning_rate": 9.585294092140241e-07, + "loss": 0.9053, + "step": 16740 + }, + { + "epoch": 0.652906897425403, + "grad_norm": 13.002252578735352, + "learning_rate": 9.58438932816283e-07, + "loss": 0.97, + "step": 16750 + }, + { + "epoch": 0.653296692588045, + "grad_norm": 13.559731483459473, + "learning_rate": 9.583483621094524e-07, + "loss": 0.9222, + "step": 16760 + }, + { + "epoch": 0.653686487750687, + "grad_norm": 12.334734916687012, + "learning_rate": 9.582576971121639e-07, + "loss": 0.8631, + "step": 16770 + }, + { + "epoch": 0.6540762829133291, + "grad_norm": 11.824180603027344, + "learning_rate": 9.581669378430688e-07, + "loss": 0.9012, + "step": 16780 + }, + { + "epoch": 0.6544660780759711, + "grad_norm": 12.998884201049805, + "learning_rate": 9.580760843208381e-07, + "loss": 0.9139, + "step": 16790 + }, + { + "epoch": 0.6548558732386132, + "grad_norm": 13.493663787841797, + "learning_rate": 9.579851365641623e-07, + "loss": 0.8829, + "step": 16800 + }, + { + "epoch": 0.6552456684012551, + "grad_norm": 15.175875663757324, + "learning_rate": 9.578940945917503e-07, + "loss": 0.8859, + "step": 16810 + }, + { + "epoch": 0.6556354635638971, + "grad_norm": 13.716005325317383, + "learning_rate": 9.578029584223314e-07, + "loss": 0.9091, + "step": 16820 + }, + { + "epoch": 0.6560252587265392, + "grad_norm": 12.267144203186035, + "learning_rate": 9.57711728074654e-07, + "loss": 0.8786, + "step": 16830 + }, + { + "epoch": 0.6564150538891812, + "grad_norm": 12.85669994354248, + "learning_rate": 9.576204035674858e-07, + "loss": 0.9697, + "step": 16840 + }, + { + "epoch": 0.6568048490518232, + "grad_norm": 13.846673011779785, + "learning_rate": 9.575289849196138e-07, + "loss": 0.9066, + "step": 16850 + }, + { + "epoch": 0.6571946442144653, + "grad_norm": 12.92249870300293, + "learning_rate": 9.574374721498445e-07, + "loss": 0.809, + "step": 16860 + }, + { + "epoch": 0.6575844393771073, + "grad_norm": 14.544699668884277, + "learning_rate": 9.573458652770037e-07, + "loss": 0.8684, + "step": 16870 + }, + { + "epoch": 0.6579742345397493, + "grad_norm": 12.984567642211914, + "learning_rate": 9.572541643199365e-07, + "loss": 0.9423, + "step": 16880 + }, + { + "epoch": 0.6583640297023914, + "grad_norm": 16.34952735900879, + "learning_rate": 9.571623692975075e-07, + "loss": 0.9524, + "step": 16890 + }, + { + "epoch": 0.6587538248650334, + "grad_norm": 12.762922286987305, + "learning_rate": 9.570704802286006e-07, + "loss": 0.8945, + "step": 16900 + }, + { + "epoch": 0.6591436200276755, + "grad_norm": 13.30054759979248, + "learning_rate": 9.56978497132119e-07, + "loss": 0.8957, + "step": 16910 + }, + { + "epoch": 0.6595334151903175, + "grad_norm": 14.314189910888672, + "learning_rate": 9.568864200269856e-07, + "loss": 0.9009, + "step": 16920 + }, + { + "epoch": 0.6599232103529595, + "grad_norm": 15.504403114318848, + "learning_rate": 9.567942489321417e-07, + "loss": 0.936, + "step": 16930 + }, + { + "epoch": 0.6603130055156016, + "grad_norm": 12.443195343017578, + "learning_rate": 9.567019838665488e-07, + "loss": 0.8972, + "step": 16940 + }, + { + "epoch": 0.6607028006782436, + "grad_norm": 13.669306755065918, + "learning_rate": 9.566096248491874e-07, + "loss": 0.9396, + "step": 16950 + }, + { + "epoch": 0.6610925958408856, + "grad_norm": 15.044306755065918, + "learning_rate": 9.565171718990577e-07, + "loss": 0.8828, + "step": 16960 + }, + { + "epoch": 0.6614823910035277, + "grad_norm": 14.140836715698242, + "learning_rate": 9.56424625035179e-07, + "loss": 0.9089, + "step": 16970 + }, + { + "epoch": 0.6618721861661697, + "grad_norm": 13.341310501098633, + "learning_rate": 9.563319842765891e-07, + "loss": 0.8819, + "step": 16980 + }, + { + "epoch": 0.6622619813288118, + "grad_norm": 12.516045570373535, + "learning_rate": 9.562392496423464e-07, + "loss": 0.8674, + "step": 16990 + }, + { + "epoch": 0.6626517764914538, + "grad_norm": 13.560593605041504, + "learning_rate": 9.561464211515282e-07, + "loss": 0.9598, + "step": 17000 + }, + { + "epoch": 0.6626517764914538, + "eval_loss": 0.9236935377120972, + "eval_runtime": 82.7692, + "eval_samples_per_second": 50.103, + "eval_steps_per_second": 6.27, + "step": 17000 + }, + { + "epoch": 0.6630415716540957, + "grad_norm": 11.84150505065918, + "learning_rate": 9.560534988232308e-07, + "loss": 0.8669, + "step": 17010 + }, + { + "epoch": 0.6634313668167378, + "grad_norm": 14.494362831115723, + "learning_rate": 9.559604826765697e-07, + "loss": 0.8982, + "step": 17020 + }, + { + "epoch": 0.6638211619793798, + "grad_norm": 15.241474151611328, + "learning_rate": 9.558673727306805e-07, + "loss": 0.8975, + "step": 17030 + }, + { + "epoch": 0.6642109571420218, + "grad_norm": 13.842434883117676, + "learning_rate": 9.557741690047174e-07, + "loss": 0.9862, + "step": 17040 + }, + { + "epoch": 0.6646007523046639, + "grad_norm": 13.226613998413086, + "learning_rate": 9.55680871517854e-07, + "loss": 0.9644, + "step": 17050 + }, + { + "epoch": 0.6649905474673059, + "grad_norm": 14.46308422088623, + "learning_rate": 9.555874802892833e-07, + "loss": 0.8819, + "step": 17060 + }, + { + "epoch": 0.6653803426299479, + "grad_norm": 13.117613792419434, + "learning_rate": 9.554939953382172e-07, + "loss": 0.8968, + "step": 17070 + }, + { + "epoch": 0.66577013779259, + "grad_norm": 12.502833366394043, + "learning_rate": 9.554004166838879e-07, + "loss": 0.8916, + "step": 17080 + }, + { + "epoch": 0.666159932955232, + "grad_norm": 11.338464736938477, + "learning_rate": 9.553067443455457e-07, + "loss": 0.8579, + "step": 17090 + }, + { + "epoch": 0.666549728117874, + "grad_norm": 13.71585464477539, + "learning_rate": 9.55212978342461e-07, + "loss": 0.9036, + "step": 17100 + }, + { + "epoch": 0.6669395232805161, + "grad_norm": 12.597564697265625, + "learning_rate": 9.551191186939227e-07, + "loss": 0.8982, + "step": 17110 + }, + { + "epoch": 0.6673293184431581, + "grad_norm": 12.157960891723633, + "learning_rate": 9.550251654192399e-07, + "loss": 0.9119, + "step": 17120 + }, + { + "epoch": 0.6677191136058002, + "grad_norm": 13.683663368225098, + "learning_rate": 9.549311185377403e-07, + "loss": 0.8441, + "step": 17130 + }, + { + "epoch": 0.6681089087684422, + "grad_norm": 14.56511116027832, + "learning_rate": 9.548369780687709e-07, + "loss": 0.7984, + "step": 17140 + }, + { + "epoch": 0.6684987039310842, + "grad_norm": 12.064804077148438, + "learning_rate": 9.547427440316979e-07, + "loss": 0.9111, + "step": 17150 + }, + { + "epoch": 0.6688884990937263, + "grad_norm": 12.042222023010254, + "learning_rate": 9.546484164459076e-07, + "loss": 0.8888, + "step": 17160 + }, + { + "epoch": 0.6692782942563683, + "grad_norm": 12.001788139343262, + "learning_rate": 9.54553995330804e-07, + "loss": 0.9969, + "step": 17170 + }, + { + "epoch": 0.6696680894190103, + "grad_norm": 14.792008399963379, + "learning_rate": 9.544594807058121e-07, + "loss": 0.9677, + "step": 17180 + }, + { + "epoch": 0.6700578845816524, + "grad_norm": 15.773207664489746, + "learning_rate": 9.543648725903747e-07, + "loss": 0.9061, + "step": 17190 + }, + { + "epoch": 0.6704476797442944, + "grad_norm": 13.00118637084961, + "learning_rate": 9.542701710039546e-07, + "loss": 0.8571, + "step": 17200 + }, + { + "epoch": 0.6708374749069365, + "grad_norm": 13.957091331481934, + "learning_rate": 9.541753759660333e-07, + "loss": 0.9902, + "step": 17210 + }, + { + "epoch": 0.6712272700695784, + "grad_norm": 13.113454818725586, + "learning_rate": 9.540804874961122e-07, + "loss": 0.8882, + "step": 17220 + }, + { + "epoch": 0.6716170652322204, + "grad_norm": 16.45529556274414, + "learning_rate": 9.539855056137113e-07, + "loss": 0.9547, + "step": 17230 + }, + { + "epoch": 0.6720068603948625, + "grad_norm": 13.820405960083008, + "learning_rate": 9.538904303383703e-07, + "loss": 0.9102, + "step": 17240 + }, + { + "epoch": 0.6723966555575045, + "grad_norm": 14.125528335571289, + "learning_rate": 9.537952616896478e-07, + "loss": 0.9144, + "step": 17250 + }, + { + "epoch": 0.6727864507201465, + "grad_norm": 12.063582420349121, + "learning_rate": 9.536999996871216e-07, + "loss": 0.9411, + "step": 17260 + }, + { + "epoch": 0.6731762458827886, + "grad_norm": 15.52597427368164, + "learning_rate": 9.536046443503887e-07, + "loss": 0.9912, + "step": 17270 + }, + { + "epoch": 0.6735660410454306, + "grad_norm": 15.073016166687012, + "learning_rate": 9.535091956990658e-07, + "loss": 0.9643, + "step": 17280 + }, + { + "epoch": 0.6739558362080726, + "grad_norm": 12.604642868041992, + "learning_rate": 9.53413653752788e-07, + "loss": 0.9005, + "step": 17290 + }, + { + "epoch": 0.6743456313707147, + "grad_norm": 14.373453140258789, + "learning_rate": 9.5331801853121e-07, + "loss": 0.955, + "step": 17300 + }, + { + "epoch": 0.6747354265333567, + "grad_norm": 13.770384788513184, + "learning_rate": 9.532222900540059e-07, + "loss": 0.9448, + "step": 17310 + }, + { + "epoch": 0.6751252216959988, + "grad_norm": 16.96061897277832, + "learning_rate": 9.531264683408688e-07, + "loss": 0.8823, + "step": 17320 + }, + { + "epoch": 0.6755150168586408, + "grad_norm": 13.57363510131836, + "learning_rate": 9.530305534115107e-07, + "loss": 0.8818, + "step": 17330 + }, + { + "epoch": 0.6759048120212828, + "grad_norm": 13.790176391601562, + "learning_rate": 9.529345452856632e-07, + "loss": 0.8858, + "step": 17340 + }, + { + "epoch": 0.6762946071839249, + "grad_norm": 15.075521469116211, + "learning_rate": 9.528384439830767e-07, + "loss": 0.9466, + "step": 17350 + }, + { + "epoch": 0.6766844023465669, + "grad_norm": 13.67878246307373, + "learning_rate": 9.527422495235214e-07, + "loss": 0.9024, + "step": 17360 + }, + { + "epoch": 0.6770741975092089, + "grad_norm": 14.266810417175293, + "learning_rate": 9.526459619267855e-07, + "loss": 0.8801, + "step": 17370 + }, + { + "epoch": 0.677463992671851, + "grad_norm": 13.549795150756836, + "learning_rate": 9.525495812126777e-07, + "loss": 0.946, + "step": 17380 + }, + { + "epoch": 0.677853787834493, + "grad_norm": 12.832058906555176, + "learning_rate": 9.524531074010249e-07, + "loss": 0.8959, + "step": 17390 + }, + { + "epoch": 0.678243582997135, + "grad_norm": 12.825282096862793, + "learning_rate": 9.523565405116737e-07, + "loss": 0.9013, + "step": 17400 + }, + { + "epoch": 0.6786333781597771, + "grad_norm": 13.458733558654785, + "learning_rate": 9.522598805644895e-07, + "loss": 0.9639, + "step": 17410 + }, + { + "epoch": 0.679023173322419, + "grad_norm": 12.963126182556152, + "learning_rate": 9.52163127579357e-07, + "loss": 0.9278, + "step": 17420 + }, + { + "epoch": 0.679412968485061, + "grad_norm": 16.81829261779785, + "learning_rate": 9.520662815761801e-07, + "loss": 0.9445, + "step": 17430 + }, + { + "epoch": 0.6798027636477031, + "grad_norm": 15.79819393157959, + "learning_rate": 9.519693425748816e-07, + "loss": 0.9445, + "step": 17440 + }, + { + "epoch": 0.6801925588103451, + "grad_norm": 15.548563003540039, + "learning_rate": 9.518723105954039e-07, + "loss": 0.9334, + "step": 17450 + }, + { + "epoch": 0.6805823539729872, + "grad_norm": 13.461705207824707, + "learning_rate": 9.517751856577078e-07, + "loss": 0.9249, + "step": 17460 + }, + { + "epoch": 0.6809721491356292, + "grad_norm": 12.633861541748047, + "learning_rate": 9.516779677817739e-07, + "loss": 0.9267, + "step": 17470 + }, + { + "epoch": 0.6813619442982712, + "grad_norm": 12.074872970581055, + "learning_rate": 9.515806569876015e-07, + "loss": 0.9385, + "step": 17480 + }, + { + "epoch": 0.6817517394609133, + "grad_norm": 13.09652328491211, + "learning_rate": 9.514832532952092e-07, + "loss": 0.9266, + "step": 17490 + }, + { + "epoch": 0.6821415346235553, + "grad_norm": 16.11394500732422, + "learning_rate": 9.513857567246351e-07, + "loss": 0.9233, + "step": 17500 + }, + { + "epoch": 0.6821415346235553, + "eval_loss": 0.9204753041267395, + "eval_runtime": 83.4548, + "eval_samples_per_second": 49.692, + "eval_steps_per_second": 6.219, + "step": 17500 + }, + { + "epoch": 0.6825313297861973, + "grad_norm": 14.533645629882812, + "learning_rate": 9.512881672959353e-07, + "loss": 0.8807, + "step": 17510 + }, + { + "epoch": 0.6829211249488394, + "grad_norm": 12.568058013916016, + "learning_rate": 9.511904850291861e-07, + "loss": 0.9278, + "step": 17520 + }, + { + "epoch": 0.6833109201114814, + "grad_norm": 14.265005111694336, + "learning_rate": 9.510927099444825e-07, + "loss": 0.909, + "step": 17530 + }, + { + "epoch": 0.6837007152741235, + "grad_norm": 11.950000762939453, + "learning_rate": 9.509948420619384e-07, + "loss": 0.9437, + "step": 17540 + }, + { + "epoch": 0.6840905104367655, + "grad_norm": 15.43362808227539, + "learning_rate": 9.508968814016872e-07, + "loss": 0.918, + "step": 17550 + }, + { + "epoch": 0.6844803055994075, + "grad_norm": 14.831604957580566, + "learning_rate": 9.507988279838809e-07, + "loss": 0.9309, + "step": 17560 + }, + { + "epoch": 0.6848701007620496, + "grad_norm": 13.463624000549316, + "learning_rate": 9.50700681828691e-07, + "loss": 0.822, + "step": 17570 + }, + { + "epoch": 0.6852598959246916, + "grad_norm": 15.086277961730957, + "learning_rate": 9.506024429563079e-07, + "loss": 0.8972, + "step": 17580 + }, + { + "epoch": 0.6856496910873336, + "grad_norm": 14.212367057800293, + "learning_rate": 9.505041113869412e-07, + "loss": 0.9506, + "step": 17590 + }, + { + "epoch": 0.6860394862499757, + "grad_norm": 14.685442924499512, + "learning_rate": 9.504056871408193e-07, + "loss": 0.9033, + "step": 17600 + }, + { + "epoch": 0.6864292814126177, + "grad_norm": 12.53986930847168, + "learning_rate": 9.503071702381898e-07, + "loss": 0.9611, + "step": 17610 + }, + { + "epoch": 0.6868190765752596, + "grad_norm": 15.18563461303711, + "learning_rate": 9.502085606993196e-07, + "loss": 0.9912, + "step": 17620 + }, + { + "epoch": 0.6872088717379017, + "grad_norm": 15.497118949890137, + "learning_rate": 9.50109858544494e-07, + "loss": 0.941, + "step": 17630 + }, + { + "epoch": 0.6875986669005437, + "grad_norm": 12.265838623046875, + "learning_rate": 9.500110637940184e-07, + "loss": 0.8449, + "step": 17640 + }, + { + "epoch": 0.6879884620631858, + "grad_norm": 12.894959449768066, + "learning_rate": 9.499121764682163e-07, + "loss": 0.8832, + "step": 17650 + }, + { + "epoch": 0.6883782572258278, + "grad_norm": 13.931320190429688, + "learning_rate": 9.498131965874307e-07, + "loss": 0.9543, + "step": 17660 + }, + { + "epoch": 0.6887680523884698, + "grad_norm": 12.854257583618164, + "learning_rate": 9.497141241720235e-07, + "loss": 0.9291, + "step": 17670 + }, + { + "epoch": 0.6891578475511119, + "grad_norm": 12.485031127929688, + "learning_rate": 9.496149592423754e-07, + "loss": 0.9886, + "step": 17680 + }, + { + "epoch": 0.6895476427137539, + "grad_norm": 13.397436141967773, + "learning_rate": 9.49515701818887e-07, + "loss": 0.8727, + "step": 17690 + }, + { + "epoch": 0.6899374378763959, + "grad_norm": 13.317821502685547, + "learning_rate": 9.494163519219768e-07, + "loss": 0.926, + "step": 17700 + }, + { + "epoch": 0.690327233039038, + "grad_norm": 12.028900146484375, + "learning_rate": 9.49316909572083e-07, + "loss": 0.8672, + "step": 17710 + }, + { + "epoch": 0.69071702820168, + "grad_norm": 12.836874961853027, + "learning_rate": 9.492173747896627e-07, + "loss": 0.8514, + "step": 17720 + }, + { + "epoch": 0.691106823364322, + "grad_norm": 14.655769348144531, + "learning_rate": 9.491177475951921e-07, + "loss": 0.9344, + "step": 17730 + }, + { + "epoch": 0.6914966185269641, + "grad_norm": 15.439092636108398, + "learning_rate": 9.490180280091662e-07, + "loss": 0.8979, + "step": 17740 + }, + { + "epoch": 0.6918864136896061, + "grad_norm": 12.526187896728516, + "learning_rate": 9.48918216052099e-07, + "loss": 0.8634, + "step": 17750 + }, + { + "epoch": 0.6922762088522482, + "grad_norm": 15.269088745117188, + "learning_rate": 9.488183117445237e-07, + "loss": 0.8876, + "step": 17760 + }, + { + "epoch": 0.6926660040148902, + "grad_norm": 15.298846244812012, + "learning_rate": 9.487183151069924e-07, + "loss": 0.9425, + "step": 17770 + }, + { + "epoch": 0.6930557991775322, + "grad_norm": 12.00965690612793, + "learning_rate": 9.486182261600764e-07, + "loss": 0.8948, + "step": 17780 + }, + { + "epoch": 0.6934455943401743, + "grad_norm": 14.4827241897583, + "learning_rate": 9.485180449243653e-07, + "loss": 0.9466, + "step": 17790 + }, + { + "epoch": 0.6938353895028163, + "grad_norm": 12.976679801940918, + "learning_rate": 9.484177714204685e-07, + "loss": 0.9114, + "step": 17800 + }, + { + "epoch": 0.6942251846654584, + "grad_norm": 12.978883743286133, + "learning_rate": 9.483174056690141e-07, + "loss": 0.8922, + "step": 17810 + }, + { + "epoch": 0.6946149798281004, + "grad_norm": 15.376219749450684, + "learning_rate": 9.48216947690649e-07, + "loss": 0.9912, + "step": 17820 + }, + { + "epoch": 0.6950047749907423, + "grad_norm": 14.338665008544922, + "learning_rate": 9.481163975060392e-07, + "loss": 0.9037, + "step": 17830 + }, + { + "epoch": 0.6953945701533844, + "grad_norm": 12.906991004943848, + "learning_rate": 9.480157551358697e-07, + "loss": 0.8519, + "step": 17840 + }, + { + "epoch": 0.6957843653160264, + "grad_norm": 12.501875877380371, + "learning_rate": 9.479150206008446e-07, + "loss": 0.8671, + "step": 17850 + }, + { + "epoch": 0.6961741604786684, + "grad_norm": 13.642354965209961, + "learning_rate": 9.478141939216865e-07, + "loss": 0.8834, + "step": 17860 + }, + { + "epoch": 0.6965639556413105, + "grad_norm": 13.558080673217773, + "learning_rate": 9.477132751191374e-07, + "loss": 0.9245, + "step": 17870 + }, + { + "epoch": 0.6969537508039525, + "grad_norm": 12.826229095458984, + "learning_rate": 9.476122642139581e-07, + "loss": 0.8653, + "step": 17880 + }, + { + "epoch": 0.6973435459665945, + "grad_norm": 13.182852745056152, + "learning_rate": 9.475111612269285e-07, + "loss": 0.8696, + "step": 17890 + }, + { + "epoch": 0.6977333411292366, + "grad_norm": 13.866589546203613, + "learning_rate": 9.474099661788471e-07, + "loss": 0.8898, + "step": 17900 + }, + { + "epoch": 0.6981231362918786, + "grad_norm": 15.524595260620117, + "learning_rate": 9.473086790905314e-07, + "loss": 0.9259, + "step": 17910 + }, + { + "epoch": 0.6985129314545206, + "grad_norm": 15.280961036682129, + "learning_rate": 9.472072999828183e-07, + "loss": 0.8929, + "step": 17920 + }, + { + "epoch": 0.6989027266171627, + "grad_norm": 13.863781929016113, + "learning_rate": 9.47105828876563e-07, + "loss": 0.9273, + "step": 17930 + }, + { + "epoch": 0.6992925217798047, + "grad_norm": 13.694204330444336, + "learning_rate": 9.470042657926401e-07, + "loss": 0.9328, + "step": 17940 + }, + { + "epoch": 0.6996823169424468, + "grad_norm": 13.050552368164062, + "learning_rate": 9.469026107519428e-07, + "loss": 0.8991, + "step": 17950 + }, + { + "epoch": 0.7000721121050888, + "grad_norm": 13.632086753845215, + "learning_rate": 9.468008637753837e-07, + "loss": 0.924, + "step": 17960 + }, + { + "epoch": 0.7004619072677308, + "grad_norm": 13.500877380371094, + "learning_rate": 9.466990248838937e-07, + "loss": 0.9128, + "step": 17970 + }, + { + "epoch": 0.7008517024303729, + "grad_norm": 14.369331359863281, + "learning_rate": 9.465970940984229e-07, + "loss": 0.9016, + "step": 17980 + }, + { + "epoch": 0.7012414975930149, + "grad_norm": 14.703322410583496, + "learning_rate": 9.464950714399402e-07, + "loss": 0.9416, + "step": 17990 + }, + { + "epoch": 0.7016312927556569, + "grad_norm": 12.711353302001953, + "learning_rate": 9.463929569294337e-07, + "loss": 0.8742, + "step": 18000 + }, + { + "epoch": 0.7016312927556569, + "eval_loss": 0.9186906814575195, + "eval_runtime": 85.8141, + "eval_samples_per_second": 48.325, + "eval_steps_per_second": 6.048, + "step": 18000 + }, + { + "epoch": 0.702021087918299, + "grad_norm": 15.155229568481445, + "learning_rate": 9.4629075058791e-07, + "loss": 0.9324, + "step": 18010 + }, + { + "epoch": 0.702410883080941, + "grad_norm": 11.613571166992188, + "learning_rate": 9.461884524363947e-07, + "loss": 0.904, + "step": 18020 + }, + { + "epoch": 0.7028006782435829, + "grad_norm": 15.31238079071045, + "learning_rate": 9.460860624959328e-07, + "loss": 0.9067, + "step": 18030 + }, + { + "epoch": 0.703190473406225, + "grad_norm": 14.714792251586914, + "learning_rate": 9.459835807875873e-07, + "loss": 0.9698, + "step": 18040 + }, + { + "epoch": 0.703580268568867, + "grad_norm": 13.542590141296387, + "learning_rate": 9.458810073324407e-07, + "loss": 0.9469, + "step": 18050 + }, + { + "epoch": 0.703970063731509, + "grad_norm": 12.487536430358887, + "learning_rate": 9.457783421515941e-07, + "loss": 0.9551, + "step": 18060 + }, + { + "epoch": 0.7043598588941511, + "grad_norm": 12.740382194519043, + "learning_rate": 9.456755852661678e-07, + "loss": 0.9453, + "step": 18070 + }, + { + "epoch": 0.7047496540567931, + "grad_norm": 15.064069747924805, + "learning_rate": 9.455727366973003e-07, + "loss": 0.9062, + "step": 18080 + }, + { + "epoch": 0.7051394492194352, + "grad_norm": 12.857431411743164, + "learning_rate": 9.454697964661498e-07, + "loss": 0.902, + "step": 18090 + }, + { + "epoch": 0.7055292443820772, + "grad_norm": 11.534789085388184, + "learning_rate": 9.453667645938927e-07, + "loss": 0.8873, + "step": 18100 + }, + { + "epoch": 0.7059190395447192, + "grad_norm": 14.811248779296875, + "learning_rate": 9.452636411017247e-07, + "loss": 0.8436, + "step": 18110 + }, + { + "epoch": 0.7063088347073613, + "grad_norm": 12.110944747924805, + "learning_rate": 9.451604260108601e-07, + "loss": 0.8713, + "step": 18120 + }, + { + "epoch": 0.7066986298700033, + "grad_norm": 13.708393096923828, + "learning_rate": 9.450571193425318e-07, + "loss": 0.9639, + "step": 18130 + }, + { + "epoch": 0.7070884250326454, + "grad_norm": 12.887459754943848, + "learning_rate": 9.449537211179923e-07, + "loss": 0.889, + "step": 18140 + }, + { + "epoch": 0.7074782201952874, + "grad_norm": 14.25551700592041, + "learning_rate": 9.44850231358512e-07, + "loss": 0.9498, + "step": 18150 + }, + { + "epoch": 0.7078680153579294, + "grad_norm": 12.876201629638672, + "learning_rate": 9.44746650085381e-07, + "loss": 0.8974, + "step": 18160 + }, + { + "epoch": 0.7082578105205715, + "grad_norm": 16.51074981689453, + "learning_rate": 9.446429773199076e-07, + "loss": 0.9672, + "step": 18170 + }, + { + "epoch": 0.7086476056832135, + "grad_norm": 12.210908889770508, + "learning_rate": 9.445392130834192e-07, + "loss": 0.9339, + "step": 18180 + }, + { + "epoch": 0.7090374008458555, + "grad_norm": 13.727279663085938, + "learning_rate": 9.444353573972619e-07, + "loss": 0.9402, + "step": 18190 + }, + { + "epoch": 0.7094271960084976, + "grad_norm": 13.921587944030762, + "learning_rate": 9.443314102828006e-07, + "loss": 0.9398, + "step": 18200 + }, + { + "epoch": 0.7098169911711396, + "grad_norm": 15.07577133178711, + "learning_rate": 9.442273717614193e-07, + "loss": 0.868, + "step": 18210 + }, + { + "epoch": 0.7102067863337816, + "grad_norm": 14.538484573364258, + "learning_rate": 9.441232418545206e-07, + "loss": 0.9301, + "step": 18220 + }, + { + "epoch": 0.7105965814964237, + "grad_norm": 13.210702896118164, + "learning_rate": 9.440190205835257e-07, + "loss": 0.9371, + "step": 18230 + }, + { + "epoch": 0.7109863766590656, + "grad_norm": 13.034502029418945, + "learning_rate": 9.439147079698748e-07, + "loss": 0.9023, + "step": 18240 + }, + { + "epoch": 0.7113761718217076, + "grad_norm": 12.76980972290039, + "learning_rate": 9.438103040350269e-07, + "loss": 0.8906, + "step": 18250 + }, + { + "epoch": 0.7117659669843497, + "grad_norm": 12.216375350952148, + "learning_rate": 9.437058088004598e-07, + "loss": 0.9479, + "step": 18260 + }, + { + "epoch": 0.7121557621469917, + "grad_norm": 15.908778190612793, + "learning_rate": 9.436012222876699e-07, + "loss": 0.9045, + "step": 18270 + }, + { + "epoch": 0.7125455573096338, + "grad_norm": 13.615480422973633, + "learning_rate": 9.434965445181728e-07, + "loss": 0.9785, + "step": 18280 + }, + { + "epoch": 0.7129353524722758, + "grad_norm": 13.56826400756836, + "learning_rate": 9.433917755135024e-07, + "loss": 0.9326, + "step": 18290 + }, + { + "epoch": 0.7133251476349178, + "grad_norm": 14.542903900146484, + "learning_rate": 9.432869152952115e-07, + "loss": 0.9147, + "step": 18300 + }, + { + "epoch": 0.7137149427975599, + "grad_norm": 12.940265655517578, + "learning_rate": 9.43181963884872e-07, + "loss": 0.9049, + "step": 18310 + }, + { + "epoch": 0.7141047379602019, + "grad_norm": 14.337098121643066, + "learning_rate": 9.430769213040739e-07, + "loss": 0.9186, + "step": 18320 + }, + { + "epoch": 0.7144945331228439, + "grad_norm": 13.68740177154541, + "learning_rate": 9.429717875744266e-07, + "loss": 0.9596, + "step": 18330 + }, + { + "epoch": 0.714884328285486, + "grad_norm": 14.048835754394531, + "learning_rate": 9.428665627175579e-07, + "loss": 0.9326, + "step": 18340 + }, + { + "epoch": 0.715274123448128, + "grad_norm": 13.147500991821289, + "learning_rate": 9.427612467551142e-07, + "loss": 0.8618, + "step": 18350 + }, + { + "epoch": 0.7156639186107701, + "grad_norm": 14.001333236694336, + "learning_rate": 9.426558397087614e-07, + "loss": 0.9315, + "step": 18360 + }, + { + "epoch": 0.7160537137734121, + "grad_norm": 11.26240062713623, + "learning_rate": 9.425503416001834e-07, + "loss": 0.9199, + "step": 18370 + }, + { + "epoch": 0.7164435089360541, + "grad_norm": 13.852128982543945, + "learning_rate": 9.424447524510828e-07, + "loss": 0.8923, + "step": 18380 + }, + { + "epoch": 0.7168333040986962, + "grad_norm": 14.085286140441895, + "learning_rate": 9.423390722831815e-07, + "loss": 0.9718, + "step": 18390 + }, + { + "epoch": 0.7172230992613382, + "grad_norm": 14.071269989013672, + "learning_rate": 9.422333011182195e-07, + "loss": 0.8941, + "step": 18400 + }, + { + "epoch": 0.7176128944239802, + "grad_norm": 11.598502159118652, + "learning_rate": 9.421274389779561e-07, + "loss": 0.8494, + "step": 18410 + }, + { + "epoch": 0.7180026895866223, + "grad_norm": 14.628556251525879, + "learning_rate": 9.420214858841688e-07, + "loss": 0.9231, + "step": 18420 + }, + { + "epoch": 0.7183924847492643, + "grad_norm": 15.391176223754883, + "learning_rate": 9.419154418586544e-07, + "loss": 0.9216, + "step": 18430 + }, + { + "epoch": 0.7187822799119062, + "grad_norm": 14.181211471557617, + "learning_rate": 9.418093069232277e-07, + "loss": 0.8199, + "step": 18440 + }, + { + "epoch": 0.7191720750745483, + "grad_norm": 15.12315845489502, + "learning_rate": 9.417030810997226e-07, + "loss": 0.9165, + "step": 18450 + }, + { + "epoch": 0.7195618702371903, + "grad_norm": 11.516173362731934, + "learning_rate": 9.415967644099918e-07, + "loss": 0.9489, + "step": 18460 + }, + { + "epoch": 0.7199516653998324, + "grad_norm": 13.253766059875488, + "learning_rate": 9.414903568759065e-07, + "loss": 0.9456, + "step": 18470 + }, + { + "epoch": 0.7203414605624744, + "grad_norm": 13.267916679382324, + "learning_rate": 9.413838585193567e-07, + "loss": 0.9113, + "step": 18480 + }, + { + "epoch": 0.7207312557251164, + "grad_norm": 13.633246421813965, + "learning_rate": 9.412772693622508e-07, + "loss": 0.8732, + "step": 18490 + }, + { + "epoch": 0.7211210508877585, + "grad_norm": 13.96886920928955, + "learning_rate": 9.411705894265163e-07, + "loss": 0.9601, + "step": 18500 + }, + { + "epoch": 0.7211210508877585, + "eval_loss": 0.917121410369873, + "eval_runtime": 86.9805, + "eval_samples_per_second": 47.677, + "eval_steps_per_second": 5.967, + "step": 18500 + }, + { + "epoch": 0.7215108460504005, + "grad_norm": 10.76214599609375, + "learning_rate": 9.410638187340989e-07, + "loss": 0.8193, + "step": 18510 + }, + { + "epoch": 0.7219006412130425, + "grad_norm": 11.74148178100586, + "learning_rate": 9.409569573069637e-07, + "loss": 0.909, + "step": 18520 + }, + { + "epoch": 0.7222904363756846, + "grad_norm": 11.194693565368652, + "learning_rate": 9.408500051670935e-07, + "loss": 0.8551, + "step": 18530 + }, + { + "epoch": 0.7226802315383266, + "grad_norm": 14.40573787689209, + "learning_rate": 9.407429623364907e-07, + "loss": 0.924, + "step": 18540 + }, + { + "epoch": 0.7230700267009686, + "grad_norm": 13.219156265258789, + "learning_rate": 9.406358288371756e-07, + "loss": 0.8993, + "step": 18550 + }, + { + "epoch": 0.7234598218636107, + "grad_norm": 13.959362030029297, + "learning_rate": 9.405286046911878e-07, + "loss": 0.9558, + "step": 18560 + }, + { + "epoch": 0.7238496170262527, + "grad_norm": 15.848827362060547, + "learning_rate": 9.404212899205848e-07, + "loss": 0.9908, + "step": 18570 + }, + { + "epoch": 0.7242394121888948, + "grad_norm": 14.441815376281738, + "learning_rate": 9.403138845474436e-07, + "loss": 0.8839, + "step": 18580 + }, + { + "epoch": 0.7246292073515368, + "grad_norm": 14.135734558105469, + "learning_rate": 9.402063885938591e-07, + "loss": 0.9003, + "step": 18590 + }, + { + "epoch": 0.7250190025141788, + "grad_norm": 13.13001537322998, + "learning_rate": 9.400988020819454e-07, + "loss": 0.8661, + "step": 18600 + }, + { + "epoch": 0.7254087976768209, + "grad_norm": 14.594408988952637, + "learning_rate": 9.399911250338346e-07, + "loss": 0.9191, + "step": 18610 + }, + { + "epoch": 0.7257985928394629, + "grad_norm": 13.872284889221191, + "learning_rate": 9.398833574716781e-07, + "loss": 1.0312, + "step": 18620 + }, + { + "epoch": 0.726188388002105, + "grad_norm": 14.01760482788086, + "learning_rate": 9.397754994176456e-07, + "loss": 0.8905, + "step": 18630 + }, + { + "epoch": 0.7265781831647469, + "grad_norm": 12.384340286254883, + "learning_rate": 9.396675508939253e-07, + "loss": 0.8665, + "step": 18640 + }, + { + "epoch": 0.7269679783273889, + "grad_norm": 14.185029029846191, + "learning_rate": 9.395595119227243e-07, + "loss": 0.9264, + "step": 18650 + }, + { + "epoch": 0.727357773490031, + "grad_norm": 13.3076171875, + "learning_rate": 9.39451382526268e-07, + "loss": 0.9161, + "step": 18660 + }, + { + "epoch": 0.727747568652673, + "grad_norm": 11.385509490966797, + "learning_rate": 9.393431627268007e-07, + "loss": 0.9174, + "step": 18670 + }, + { + "epoch": 0.728137363815315, + "grad_norm": 12.866765022277832, + "learning_rate": 9.39234852546585e-07, + "loss": 0.9336, + "step": 18680 + }, + { + "epoch": 0.7285271589779571, + "grad_norm": 13.161572456359863, + "learning_rate": 9.39126452007902e-07, + "loss": 0.932, + "step": 18690 + }, + { + "epoch": 0.7289169541405991, + "grad_norm": 13.99316692352295, + "learning_rate": 9.390179611330524e-07, + "loss": 0.8738, + "step": 18700 + }, + { + "epoch": 0.7293067493032411, + "grad_norm": 13.508529663085938, + "learning_rate": 9.389093799443539e-07, + "loss": 0.9602, + "step": 18710 + }, + { + "epoch": 0.7296965444658832, + "grad_norm": 13.332401275634766, + "learning_rate": 9.388007084641442e-07, + "loss": 0.8366, + "step": 18720 + }, + { + "epoch": 0.7300863396285252, + "grad_norm": 13.641983032226562, + "learning_rate": 9.386919467147786e-07, + "loss": 0.9168, + "step": 18730 + }, + { + "epoch": 0.7304761347911672, + "grad_norm": 11.790929794311523, + "learning_rate": 9.385830947186314e-07, + "loss": 0.926, + "step": 18740 + }, + { + "epoch": 0.7308659299538093, + "grad_norm": 13.041055679321289, + "learning_rate": 9.384741524980954e-07, + "loss": 0.8536, + "step": 18750 + }, + { + "epoch": 0.7312557251164513, + "grad_norm": 13.78531551361084, + "learning_rate": 9.383651200755822e-07, + "loss": 0.9274, + "step": 18760 + }, + { + "epoch": 0.7316455202790934, + "grad_norm": 15.34023380279541, + "learning_rate": 9.382559974735214e-07, + "loss": 0.914, + "step": 18770 + }, + { + "epoch": 0.7320353154417354, + "grad_norm": 13.642807006835938, + "learning_rate": 9.381467847143616e-07, + "loss": 0.9169, + "step": 18780 + }, + { + "epoch": 0.7324251106043774, + "grad_norm": 16.112722396850586, + "learning_rate": 9.380374818205698e-07, + "loss": 0.9422, + "step": 18790 + }, + { + "epoch": 0.7328149057670195, + "grad_norm": 12.96888542175293, + "learning_rate": 9.379280888146317e-07, + "loss": 0.9909, + "step": 18800 + }, + { + "epoch": 0.7332047009296615, + "grad_norm": 15.229576110839844, + "learning_rate": 9.378186057190511e-07, + "loss": 0.9312, + "step": 18810 + }, + { + "epoch": 0.7335944960923035, + "grad_norm": 13.720773696899414, + "learning_rate": 9.377090325563509e-07, + "loss": 0.928, + "step": 18820 + }, + { + "epoch": 0.7339842912549456, + "grad_norm": 12.475160598754883, + "learning_rate": 9.375993693490722e-07, + "loss": 0.9677, + "step": 18830 + }, + { + "epoch": 0.7343740864175876, + "grad_norm": 13.881220817565918, + "learning_rate": 9.374896161197746e-07, + "loss": 0.8824, + "step": 18840 + }, + { + "epoch": 0.7347638815802295, + "grad_norm": 12.278969764709473, + "learning_rate": 9.373797728910364e-07, + "loss": 0.872, + "step": 18850 + }, + { + "epoch": 0.7351536767428716, + "grad_norm": 12.957053184509277, + "learning_rate": 9.372698396854542e-07, + "loss": 0.9531, + "step": 18860 + }, + { + "epoch": 0.7355434719055136, + "grad_norm": 14.161120414733887, + "learning_rate": 9.371598165256434e-07, + "loss": 0.8799, + "step": 18870 + }, + { + "epoch": 0.7359332670681556, + "grad_norm": 12.473468780517578, + "learning_rate": 9.370497034342376e-07, + "loss": 0.8807, + "step": 18880 + }, + { + "epoch": 0.7363230622307977, + "grad_norm": 14.540212631225586, + "learning_rate": 9.369395004338889e-07, + "loss": 0.8751, + "step": 18890 + }, + { + "epoch": 0.7367128573934397, + "grad_norm": 12.239325523376465, + "learning_rate": 9.368292075472683e-07, + "loss": 0.9329, + "step": 18900 + }, + { + "epoch": 0.7371026525560818, + "grad_norm": 13.429837226867676, + "learning_rate": 9.367188247970649e-07, + "loss": 0.9765, + "step": 18910 + }, + { + "epoch": 0.7374924477187238, + "grad_norm": 15.97394847869873, + "learning_rate": 9.366083522059866e-07, + "loss": 0.9737, + "step": 18920 + }, + { + "epoch": 0.7378822428813658, + "grad_norm": 14.044434547424316, + "learning_rate": 9.364977897967592e-07, + "loss": 0.8691, + "step": 18930 + }, + { + "epoch": 0.7382720380440079, + "grad_norm": 12.20175552368164, + "learning_rate": 9.363871375921278e-07, + "loss": 0.8375, + "step": 18940 + }, + { + "epoch": 0.7386618332066499, + "grad_norm": 11.886333465576172, + "learning_rate": 9.36276395614855e-07, + "loss": 0.8953, + "step": 18950 + }, + { + "epoch": 0.739051628369292, + "grad_norm": 15.096920013427734, + "learning_rate": 9.36165563887723e-07, + "loss": 0.86, + "step": 18960 + }, + { + "epoch": 0.739441423531934, + "grad_norm": 13.679168701171875, + "learning_rate": 9.360546424335314e-07, + "loss": 0.9491, + "step": 18970 + }, + { + "epoch": 0.739831218694576, + "grad_norm": 15.265667915344238, + "learning_rate": 9.35943631275099e-07, + "loss": 0.8998, + "step": 18980 + }, + { + "epoch": 0.7402210138572181, + "grad_norm": 12.423198699951172, + "learning_rate": 9.358325304352627e-07, + "loss": 0.9444, + "step": 18990 + }, + { + "epoch": 0.7406108090198601, + "grad_norm": 12.197076797485352, + "learning_rate": 9.357213399368777e-07, + "loss": 0.9452, + "step": 19000 + }, + { + "epoch": 0.7406108090198601, + "eval_loss": 0.9147753119468689, + "eval_runtime": 82.9629, + "eval_samples_per_second": 49.986, + "eval_steps_per_second": 6.256, + "step": 19000 + }, + { + "epoch": 0.7410006041825021, + "grad_norm": 14.954702377319336, + "learning_rate": 9.356100598028184e-07, + "loss": 0.8781, + "step": 19010 + }, + { + "epoch": 0.7413903993451442, + "grad_norm": 10.800849914550781, + "learning_rate": 9.354986900559767e-07, + "loss": 0.8847, + "step": 19020 + }, + { + "epoch": 0.7417801945077862, + "grad_norm": 15.346233367919922, + "learning_rate": 9.353872307192636e-07, + "loss": 0.9626, + "step": 19030 + }, + { + "epoch": 0.7421699896704282, + "grad_norm": 13.1574125289917, + "learning_rate": 9.352756818156079e-07, + "loss": 0.9027, + "step": 19040 + }, + { + "epoch": 0.7425597848330702, + "grad_norm": 13.372720718383789, + "learning_rate": 9.351640433679575e-07, + "loss": 0.8419, + "step": 19050 + }, + { + "epoch": 0.7429495799957122, + "grad_norm": 13.584761619567871, + "learning_rate": 9.350523153992784e-07, + "loss": 0.9255, + "step": 19060 + }, + { + "epoch": 0.7433393751583542, + "grad_norm": 12.77706527709961, + "learning_rate": 9.349404979325551e-07, + "loss": 0.9213, + "step": 19070 + }, + { + "epoch": 0.7437291703209963, + "grad_norm": 13.08869743347168, + "learning_rate": 9.348285909907904e-07, + "loss": 0.8767, + "step": 19080 + }, + { + "epoch": 0.7441189654836383, + "grad_norm": 12.139748573303223, + "learning_rate": 9.347165945970054e-07, + "loss": 0.8635, + "step": 19090 + }, + { + "epoch": 0.7445087606462804, + "grad_norm": 11.889470100402832, + "learning_rate": 9.3460450877424e-07, + "loss": 0.9342, + "step": 19100 + }, + { + "epoch": 0.7448985558089224, + "grad_norm": 13.867616653442383, + "learning_rate": 9.344923335455522e-07, + "loss": 0.9254, + "step": 19110 + }, + { + "epoch": 0.7452883509715644, + "grad_norm": 12.905143737792969, + "learning_rate": 9.343800689340181e-07, + "loss": 0.9935, + "step": 19120 + }, + { + "epoch": 0.7456781461342065, + "grad_norm": 11.400679588317871, + "learning_rate": 9.342677149627331e-07, + "loss": 0.9398, + "step": 19130 + }, + { + "epoch": 0.7460679412968485, + "grad_norm": 16.164587020874023, + "learning_rate": 9.3415527165481e-07, + "loss": 0.9262, + "step": 19140 + }, + { + "epoch": 0.7464577364594905, + "grad_norm": 13.474817276000977, + "learning_rate": 9.340427390333807e-07, + "loss": 0.9047, + "step": 19150 + }, + { + "epoch": 0.7468475316221326, + "grad_norm": 12.797599792480469, + "learning_rate": 9.33930117121595e-07, + "loss": 0.8533, + "step": 19160 + }, + { + "epoch": 0.7472373267847746, + "grad_norm": 11.769448280334473, + "learning_rate": 9.338174059426212e-07, + "loss": 0.8656, + "step": 19170 + }, + { + "epoch": 0.7476271219474167, + "grad_norm": 14.429923057556152, + "learning_rate": 9.337046055196463e-07, + "loss": 0.8874, + "step": 19180 + }, + { + "epoch": 0.7480169171100587, + "grad_norm": 11.549078941345215, + "learning_rate": 9.335917158758749e-07, + "loss": 0.9153, + "step": 19190 + }, + { + "epoch": 0.7484067122727007, + "grad_norm": 14.501729011535645, + "learning_rate": 9.334787370345308e-07, + "loss": 0.906, + "step": 19200 + }, + { + "epoch": 0.7487965074353428, + "grad_norm": 12.641839027404785, + "learning_rate": 9.333656690188556e-07, + "loss": 0.9664, + "step": 19210 + }, + { + "epoch": 0.7491863025979848, + "grad_norm": 15.005885124206543, + "learning_rate": 9.332525118521096e-07, + "loss": 0.9291, + "step": 19220 + }, + { + "epoch": 0.7495760977606268, + "grad_norm": 13.34351921081543, + "learning_rate": 9.331392655575707e-07, + "loss": 0.9566, + "step": 19230 + }, + { + "epoch": 0.7499658929232689, + "grad_norm": 12.92685604095459, + "learning_rate": 9.330259301585363e-07, + "loss": 0.9367, + "step": 19240 + }, + { + "epoch": 0.7503556880859108, + "grad_norm": 14.045212745666504, + "learning_rate": 9.329125056783216e-07, + "loss": 0.9053, + "step": 19250 + }, + { + "epoch": 0.7507454832485528, + "grad_norm": 12.083708763122559, + "learning_rate": 9.327989921402595e-07, + "loss": 0.8451, + "step": 19260 + }, + { + "epoch": 0.7511352784111949, + "grad_norm": 14.150635719299316, + "learning_rate": 9.32685389567702e-07, + "loss": 0.8872, + "step": 19270 + }, + { + "epoch": 0.7515250735738369, + "grad_norm": 11.30999755859375, + "learning_rate": 9.325716979840193e-07, + "loss": 0.9316, + "step": 19280 + }, + { + "epoch": 0.751914868736479, + "grad_norm": 13.655519485473633, + "learning_rate": 9.324579174125995e-07, + "loss": 0.9462, + "step": 19290 + }, + { + "epoch": 0.752304663899121, + "grad_norm": 12.355594635009766, + "learning_rate": 9.323440478768497e-07, + "loss": 0.9294, + "step": 19300 + }, + { + "epoch": 0.752694459061763, + "grad_norm": 12.897870063781738, + "learning_rate": 9.322300894001946e-07, + "loss": 0.8911, + "step": 19310 + }, + { + "epoch": 0.7530842542244051, + "grad_norm": 12.15866756439209, + "learning_rate": 9.321160420060777e-07, + "loss": 0.8899, + "step": 19320 + }, + { + "epoch": 0.7534740493870471, + "grad_norm": 12.254719734191895, + "learning_rate": 9.320019057179603e-07, + "loss": 0.9228, + "step": 19330 + }, + { + "epoch": 0.7538638445496891, + "grad_norm": 13.468976974487305, + "learning_rate": 9.318876805593227e-07, + "loss": 0.9264, + "step": 19340 + }, + { + "epoch": 0.7542536397123312, + "grad_norm": 14.903093338012695, + "learning_rate": 9.317733665536626e-07, + "loss": 0.9496, + "step": 19350 + }, + { + "epoch": 0.7546434348749732, + "grad_norm": 13.869974136352539, + "learning_rate": 9.316589637244967e-07, + "loss": 0.8995, + "step": 19360 + }, + { + "epoch": 0.7550332300376152, + "grad_norm": 12.288444519042969, + "learning_rate": 9.315444720953598e-07, + "loss": 0.9087, + "step": 19370 + }, + { + "epoch": 0.7554230252002573, + "grad_norm": 12.031224250793457, + "learning_rate": 9.314298916898046e-07, + "loss": 0.8912, + "step": 19380 + }, + { + "epoch": 0.7558128203628993, + "grad_norm": 14.640643119812012, + "learning_rate": 9.313152225314025e-07, + "loss": 0.8714, + "step": 19390 + }, + { + "epoch": 0.7562026155255414, + "grad_norm": 14.473215103149414, + "learning_rate": 9.31200464643743e-07, + "loss": 0.9188, + "step": 19400 + }, + { + "epoch": 0.7565924106881834, + "grad_norm": 12.073697090148926, + "learning_rate": 9.310856180504338e-07, + "loss": 0.8528, + "step": 19410 + }, + { + "epoch": 0.7569822058508254, + "grad_norm": 16.40866470336914, + "learning_rate": 9.309706827751009e-07, + "loss": 0.8329, + "step": 19420 + }, + { + "epoch": 0.7573720010134675, + "grad_norm": 14.862870216369629, + "learning_rate": 9.308556588413886e-07, + "loss": 0.9588, + "step": 19430 + }, + { + "epoch": 0.7577617961761095, + "grad_norm": 12.290288925170898, + "learning_rate": 9.307405462729594e-07, + "loss": 0.9238, + "step": 19440 + }, + { + "epoch": 0.7581515913387515, + "grad_norm": 13.600674629211426, + "learning_rate": 9.306253450934939e-07, + "loss": 0.9159, + "step": 19450 + }, + { + "epoch": 0.7585413865013935, + "grad_norm": 14.994470596313477, + "learning_rate": 9.305100553266912e-07, + "loss": 0.9043, + "step": 19460 + }, + { + "epoch": 0.7589311816640355, + "grad_norm": 12.173378944396973, + "learning_rate": 9.303946769962683e-07, + "loss": 0.916, + "step": 19470 + }, + { + "epoch": 0.7593209768266775, + "grad_norm": 13.707518577575684, + "learning_rate": 9.302792101259606e-07, + "loss": 0.8592, + "step": 19480 + }, + { + "epoch": 0.7597107719893196, + "grad_norm": 13.981630325317383, + "learning_rate": 9.301636547395218e-07, + "loss": 0.8996, + "step": 19490 + }, + { + "epoch": 0.7601005671519616, + "grad_norm": 14.698176383972168, + "learning_rate": 9.300480108607236e-07, + "loss": 0.9052, + "step": 19500 + }, + { + "epoch": 0.7601005671519616, + "eval_loss": 0.9126564264297485, + "eval_runtime": 82.7427, + "eval_samples_per_second": 50.119, + "eval_steps_per_second": 6.272, + "step": 19500 + }, + { + "epoch": 0.7604903623146037, + "grad_norm": 14.25434398651123, + "learning_rate": 9.299322785133563e-07, + "loss": 0.9005, + "step": 19510 + }, + { + "epoch": 0.7608801574772457, + "grad_norm": 13.68078899383545, + "learning_rate": 9.298164577212278e-07, + "loss": 0.9209, + "step": 19520 + }, + { + "epoch": 0.7612699526398877, + "grad_norm": 14.647333145141602, + "learning_rate": 9.297005485081643e-07, + "loss": 0.8825, + "step": 19530 + }, + { + "epoch": 0.7616597478025298, + "grad_norm": 12.906424522399902, + "learning_rate": 9.295845508980111e-07, + "loss": 0.8754, + "step": 19540 + }, + { + "epoch": 0.7620495429651718, + "grad_norm": 12.834390640258789, + "learning_rate": 9.294684649146305e-07, + "loss": 0.8587, + "step": 19550 + }, + { + "epoch": 0.7624393381278138, + "grad_norm": 12.515497207641602, + "learning_rate": 9.293522905819036e-07, + "loss": 0.917, + "step": 19560 + }, + { + "epoch": 0.7628291332904559, + "grad_norm": 13.566902160644531, + "learning_rate": 9.292360279237293e-07, + "loss": 0.8872, + "step": 19570 + }, + { + "epoch": 0.7632189284530979, + "grad_norm": 13.772652626037598, + "learning_rate": 9.291196769640253e-07, + "loss": 0.9263, + "step": 19580 + }, + { + "epoch": 0.76360872361574, + "grad_norm": 14.990094184875488, + "learning_rate": 9.290032377267268e-07, + "loss": 0.9543, + "step": 19590 + }, + { + "epoch": 0.763998518778382, + "grad_norm": 12.206477165222168, + "learning_rate": 9.288867102357876e-07, + "loss": 0.927, + "step": 19600 + }, + { + "epoch": 0.764388313941024, + "grad_norm": 13.191361427307129, + "learning_rate": 9.287700945151793e-07, + "loss": 0.9004, + "step": 19610 + }, + { + "epoch": 0.7647781091036661, + "grad_norm": 14.486326217651367, + "learning_rate": 9.286533905888918e-07, + "loss": 0.9024, + "step": 19620 + }, + { + "epoch": 0.7651679042663081, + "grad_norm": 11.5840482711792, + "learning_rate": 9.285365984809335e-07, + "loss": 0.8684, + "step": 19630 + }, + { + "epoch": 0.7655576994289501, + "grad_norm": 12.343940734863281, + "learning_rate": 9.284197182153305e-07, + "loss": 0.9495, + "step": 19640 + }, + { + "epoch": 0.7659474945915922, + "grad_norm": 13.822680473327637, + "learning_rate": 9.283027498161271e-07, + "loss": 0.8879, + "step": 19650 + }, + { + "epoch": 0.7663372897542341, + "grad_norm": 11.845915794372559, + "learning_rate": 9.281856933073858e-07, + "loss": 0.9031, + "step": 19660 + }, + { + "epoch": 0.7667270849168761, + "grad_norm": 12.172077178955078, + "learning_rate": 9.280685487131869e-07, + "loss": 0.8089, + "step": 19670 + }, + { + "epoch": 0.7671168800795182, + "grad_norm": 12.389348030090332, + "learning_rate": 9.279513160576298e-07, + "loss": 0.8614, + "step": 19680 + }, + { + "epoch": 0.7675066752421602, + "grad_norm": 13.184268951416016, + "learning_rate": 9.278339953648311e-07, + "loss": 0.9781, + "step": 19690 + }, + { + "epoch": 0.7678964704048022, + "grad_norm": 11.1494140625, + "learning_rate": 9.277165866589254e-07, + "loss": 0.9533, + "step": 19700 + }, + { + "epoch": 0.7682862655674443, + "grad_norm": 10.947586059570312, + "learning_rate": 9.275990899640662e-07, + "loss": 0.8876, + "step": 19710 + }, + { + "epoch": 0.7686760607300863, + "grad_norm": 14.029844284057617, + "learning_rate": 9.274815053044244e-07, + "loss": 0.8938, + "step": 19720 + }, + { + "epoch": 0.7690658558927284, + "grad_norm": 11.988734245300293, + "learning_rate": 9.273638327041896e-07, + "loss": 0.9611, + "step": 19730 + }, + { + "epoch": 0.7694556510553704, + "grad_norm": 12.956838607788086, + "learning_rate": 9.272460721875689e-07, + "loss": 0.9125, + "step": 19740 + }, + { + "epoch": 0.7698454462180124, + "grad_norm": 13.493463516235352, + "learning_rate": 9.271282237787876e-07, + "loss": 0.9208, + "step": 19750 + }, + { + "epoch": 0.7702352413806545, + "grad_norm": 13.741813659667969, + "learning_rate": 9.270102875020898e-07, + "loss": 0.9085, + "step": 19760 + }, + { + "epoch": 0.7706250365432965, + "grad_norm": 12.525851249694824, + "learning_rate": 9.268922633817364e-07, + "loss": 0.9366, + "step": 19770 + }, + { + "epoch": 0.7710148317059385, + "grad_norm": 12.234073638916016, + "learning_rate": 9.267741514420075e-07, + "loss": 0.9121, + "step": 19780 + }, + { + "epoch": 0.7714046268685806, + "grad_norm": 13.043913841247559, + "learning_rate": 9.266559517072007e-07, + "loss": 0.8985, + "step": 19790 + }, + { + "epoch": 0.7717944220312226, + "grad_norm": 12.160224914550781, + "learning_rate": 9.265376642016318e-07, + "loss": 0.8419, + "step": 19800 + }, + { + "epoch": 0.7721842171938647, + "grad_norm": 13.695694923400879, + "learning_rate": 9.264192889496348e-07, + "loss": 0.9227, + "step": 19810 + }, + { + "epoch": 0.7725740123565067, + "grad_norm": 14.185380935668945, + "learning_rate": 9.263008259755615e-07, + "loss": 0.9358, + "step": 19820 + }, + { + "epoch": 0.7729638075191487, + "grad_norm": 12.702988624572754, + "learning_rate": 9.261822753037818e-07, + "loss": 0.8914, + "step": 19830 + }, + { + "epoch": 0.7733536026817908, + "grad_norm": 12.881747245788574, + "learning_rate": 9.260636369586839e-07, + "loss": 0.862, + "step": 19840 + }, + { + "epoch": 0.7737433978444328, + "grad_norm": 16.136011123657227, + "learning_rate": 9.259449109646733e-07, + "loss": 0.8966, + "step": 19850 + }, + { + "epoch": 0.7741331930070747, + "grad_norm": 14.277297973632812, + "learning_rate": 9.258260973461747e-07, + "loss": 0.9779, + "step": 19860 + }, + { + "epoch": 0.7745229881697168, + "grad_norm": 13.848838806152344, + "learning_rate": 9.257071961276299e-07, + "loss": 0.903, + "step": 19870 + }, + { + "epoch": 0.7749127833323588, + "grad_norm": 12.536938667297363, + "learning_rate": 9.255882073334991e-07, + "loss": 0.8458, + "step": 19880 + }, + { + "epoch": 0.7753025784950008, + "grad_norm": 13.086567878723145, + "learning_rate": 9.254691309882602e-07, + "loss": 1.0031, + "step": 19890 + }, + { + "epoch": 0.7756923736576429, + "grad_norm": 11.609024047851562, + "learning_rate": 9.253499671164097e-07, + "loss": 0.9331, + "step": 19900 + }, + { + "epoch": 0.7760821688202849, + "grad_norm": 13.876961708068848, + "learning_rate": 9.252307157424614e-07, + "loss": 0.8992, + "step": 19910 + }, + { + "epoch": 0.776471963982927, + "grad_norm": 15.557445526123047, + "learning_rate": 9.251113768909475e-07, + "loss": 0.9253, + "step": 19920 + }, + { + "epoch": 0.776861759145569, + "grad_norm": 17.407400131225586, + "learning_rate": 9.249919505864183e-07, + "loss": 0.9281, + "step": 19930 + }, + { + "epoch": 0.777251554308211, + "grad_norm": 13.150190353393555, + "learning_rate": 9.248724368534418e-07, + "loss": 0.9168, + "step": 19940 + }, + { + "epoch": 0.7776413494708531, + "grad_norm": 17.447084426879883, + "learning_rate": 9.247528357166042e-07, + "loss": 0.9612, + "step": 19950 + }, + { + "epoch": 0.7780311446334951, + "grad_norm": 11.970006942749023, + "learning_rate": 9.246331472005095e-07, + "loss": 0.9378, + "step": 19960 + }, + { + "epoch": 0.7784209397961371, + "grad_norm": 13.703678131103516, + "learning_rate": 9.245133713297799e-07, + "loss": 0.9016, + "step": 19970 + }, + { + "epoch": 0.7788107349587792, + "grad_norm": 15.60241985321045, + "learning_rate": 9.243935081290553e-07, + "loss": 0.988, + "step": 19980 + }, + { + "epoch": 0.7792005301214212, + "grad_norm": 13.60238265991211, + "learning_rate": 9.242735576229937e-07, + "loss": 0.8687, + "step": 19990 + }, + { + "epoch": 0.7795903252840632, + "grad_norm": 12.76761531829834, + "learning_rate": 9.241535198362711e-07, + "loss": 0.9263, + "step": 20000 + }, + { + "epoch": 0.7795903252840632, + "eval_loss": 0.91310054063797, + "eval_runtime": 85.9372, + "eval_samples_per_second": 48.256, + "eval_steps_per_second": 6.039, + "step": 20000 + }, + { + "epoch": 0.7799801204467053, + "grad_norm": 11.687899589538574, + "learning_rate": 9.240333947935815e-07, + "loss": 0.8619, + "step": 20010 + }, + { + "epoch": 0.7803699156093473, + "grad_norm": 15.681601524353027, + "learning_rate": 9.239131825196366e-07, + "loss": 0.94, + "step": 20020 + }, + { + "epoch": 0.7807597107719894, + "grad_norm": 13.092780113220215, + "learning_rate": 9.237928830391666e-07, + "loss": 0.9102, + "step": 20030 + }, + { + "epoch": 0.7811495059346314, + "grad_norm": 12.14647388458252, + "learning_rate": 9.236724963769188e-07, + "loss": 0.8792, + "step": 20040 + }, + { + "epoch": 0.7815393010972734, + "grad_norm": 13.535216331481934, + "learning_rate": 9.235520225576591e-07, + "loss": 0.9865, + "step": 20050 + }, + { + "epoch": 0.7819290962599155, + "grad_norm": 14.095476150512695, + "learning_rate": 9.23431461606171e-07, + "loss": 0.9124, + "step": 20060 + }, + { + "epoch": 0.7823188914225574, + "grad_norm": 11.460887908935547, + "learning_rate": 9.233108135472563e-07, + "loss": 0.8617, + "step": 20070 + }, + { + "epoch": 0.7827086865851994, + "grad_norm": 12.302130699157715, + "learning_rate": 9.231900784057343e-07, + "loss": 0.8986, + "step": 20080 + }, + { + "epoch": 0.7830984817478415, + "grad_norm": 14.067779541015625, + "learning_rate": 9.230692562064421e-07, + "loss": 0.8964, + "step": 20090 + }, + { + "epoch": 0.7834882769104835, + "grad_norm": 13.95444393157959, + "learning_rate": 9.229483469742355e-07, + "loss": 0.9082, + "step": 20100 + }, + { + "epoch": 0.7838780720731255, + "grad_norm": 11.962923049926758, + "learning_rate": 9.228273507339874e-07, + "loss": 0.8984, + "step": 20110 + }, + { + "epoch": 0.7842678672357676, + "grad_norm": 12.327075004577637, + "learning_rate": 9.22706267510589e-07, + "loss": 0.8583, + "step": 20120 + }, + { + "epoch": 0.7846576623984096, + "grad_norm": 12.342303276062012, + "learning_rate": 9.225850973289493e-07, + "loss": 0.8755, + "step": 20130 + }, + { + "epoch": 0.7850474575610517, + "grad_norm": 11.734973907470703, + "learning_rate": 9.22463840213995e-07, + "loss": 0.8805, + "step": 20140 + }, + { + "epoch": 0.7854372527236937, + "grad_norm": 14.594986915588379, + "learning_rate": 9.22342496190671e-07, + "loss": 0.9522, + "step": 20150 + }, + { + "epoch": 0.7858270478863357, + "grad_norm": 13.11673641204834, + "learning_rate": 9.222210652839399e-07, + "loss": 0.9275, + "step": 20160 + }, + { + "epoch": 0.7862168430489778, + "grad_norm": 13.20789909362793, + "learning_rate": 9.22099547518782e-07, + "loss": 0.9384, + "step": 20170 + }, + { + "epoch": 0.7866066382116198, + "grad_norm": 14.688589096069336, + "learning_rate": 9.219779429201961e-07, + "loss": 0.8694, + "step": 20180 + }, + { + "epoch": 0.7869964333742618, + "grad_norm": 11.104736328125, + "learning_rate": 9.218562515131982e-07, + "loss": 0.924, + "step": 20190 + }, + { + "epoch": 0.7873862285369039, + "grad_norm": 14.288511276245117, + "learning_rate": 9.217344733228225e-07, + "loss": 0.8987, + "step": 20200 + }, + { + "epoch": 0.7877760236995459, + "grad_norm": 11.828682899475098, + "learning_rate": 9.216126083741209e-07, + "loss": 0.8749, + "step": 20210 + }, + { + "epoch": 0.788165818862188, + "grad_norm": 14.456504821777344, + "learning_rate": 9.214906566921632e-07, + "loss": 0.9912, + "step": 20220 + }, + { + "epoch": 0.78855561402483, + "grad_norm": 12.314278602600098, + "learning_rate": 9.213686183020369e-07, + "loss": 0.9131, + "step": 20230 + }, + { + "epoch": 0.788945409187472, + "grad_norm": 14.36974811553955, + "learning_rate": 9.212464932288476e-07, + "loss": 0.8744, + "step": 20240 + }, + { + "epoch": 0.7893352043501141, + "grad_norm": 12.443222045898438, + "learning_rate": 9.211242814977187e-07, + "loss": 0.8489, + "step": 20250 + }, + { + "epoch": 0.7897249995127561, + "grad_norm": 13.215991020202637, + "learning_rate": 9.210019831337913e-07, + "loss": 0.9113, + "step": 20260 + }, + { + "epoch": 0.790114794675398, + "grad_norm": 14.286725997924805, + "learning_rate": 9.208795981622241e-07, + "loss": 0.9381, + "step": 20270 + }, + { + "epoch": 0.7905045898380401, + "grad_norm": 13.993630409240723, + "learning_rate": 9.207571266081944e-07, + "loss": 0.8829, + "step": 20280 + }, + { + "epoch": 0.7908943850006821, + "grad_norm": 10.700719833374023, + "learning_rate": 9.206345684968964e-07, + "loss": 0.8836, + "step": 20290 + }, + { + "epoch": 0.7912841801633241, + "grad_norm": 15.926965713500977, + "learning_rate": 9.205119238535425e-07, + "loss": 0.9147, + "step": 20300 + }, + { + "epoch": 0.7916739753259662, + "grad_norm": 11.549506187438965, + "learning_rate": 9.203891927033629e-07, + "loss": 0.9214, + "step": 20310 + }, + { + "epoch": 0.7920637704886082, + "grad_norm": 14.642057418823242, + "learning_rate": 9.202663750716058e-07, + "loss": 0.9321, + "step": 20320 + }, + { + "epoch": 0.7924535656512502, + "grad_norm": 12.292379379272461, + "learning_rate": 9.20143470983537e-07, + "loss": 0.9449, + "step": 20330 + }, + { + "epoch": 0.7928433608138923, + "grad_norm": 13.839735984802246, + "learning_rate": 9.200204804644396e-07, + "loss": 0.9511, + "step": 20340 + }, + { + "epoch": 0.7932331559765343, + "grad_norm": 13.694084167480469, + "learning_rate": 9.198974035396155e-07, + "loss": 0.864, + "step": 20350 + }, + { + "epoch": 0.7936229511391764, + "grad_norm": 13.255352020263672, + "learning_rate": 9.197742402343836e-07, + "loss": 0.8891, + "step": 20360 + }, + { + "epoch": 0.7940127463018184, + "grad_norm": 12.083122253417969, + "learning_rate": 9.196509905740808e-07, + "loss": 0.9372, + "step": 20370 + }, + { + "epoch": 0.7944025414644604, + "grad_norm": 13.733275413513184, + "learning_rate": 9.195276545840617e-07, + "loss": 0.9249, + "step": 20380 + }, + { + "epoch": 0.7947923366271025, + "grad_norm": 17.653289794921875, + "learning_rate": 9.194042322896989e-07, + "loss": 0.9282, + "step": 20390 + }, + { + "epoch": 0.7951821317897445, + "grad_norm": 12.25571060180664, + "learning_rate": 9.192807237163822e-07, + "loss": 0.8426, + "step": 20400 + }, + { + "epoch": 0.7955719269523865, + "grad_norm": 14.491943359375, + "learning_rate": 9.1915712888952e-07, + "loss": 0.9446, + "step": 20410 + }, + { + "epoch": 0.7959617221150286, + "grad_norm": 11.114202499389648, + "learning_rate": 9.190334478345376e-07, + "loss": 0.8466, + "step": 20420 + }, + { + "epoch": 0.7963515172776706, + "grad_norm": 12.82889175415039, + "learning_rate": 9.189096805768785e-07, + "loss": 0.8474, + "step": 20430 + }, + { + "epoch": 0.7967413124403127, + "grad_norm": 14.98286247253418, + "learning_rate": 9.187858271420042e-07, + "loss": 0.9042, + "step": 20440 + }, + { + "epoch": 0.7971311076029547, + "grad_norm": 12.709452629089355, + "learning_rate": 9.186618875553931e-07, + "loss": 0.8892, + "step": 20450 + }, + { + "epoch": 0.7975209027655967, + "grad_norm": 15.785135269165039, + "learning_rate": 9.185378618425419e-07, + "loss": 0.9488, + "step": 20460 + }, + { + "epoch": 0.7979106979282388, + "grad_norm": 13.66340446472168, + "learning_rate": 9.184137500289648e-07, + "loss": 0.9202, + "step": 20470 + }, + { + "epoch": 0.7983004930908807, + "grad_norm": 10.74239730834961, + "learning_rate": 9.182895521401942e-07, + "loss": 0.9112, + "step": 20480 + }, + { + "epoch": 0.7986902882535227, + "grad_norm": 13.011300086975098, + "learning_rate": 9.181652682017795e-07, + "loss": 0.9153, + "step": 20490 + }, + { + "epoch": 0.7990800834161648, + "grad_norm": 13.968279838562012, + "learning_rate": 9.180408982392884e-07, + "loss": 0.9247, + "step": 20500 + }, + { + "epoch": 0.7990800834161648, + "eval_loss": 0.9111083149909973, + "eval_runtime": 82.8337, + "eval_samples_per_second": 50.064, + "eval_steps_per_second": 6.266, + "step": 20500 + }, + { + "epoch": 0.7994698785788068, + "grad_norm": 13.574003219604492, + "learning_rate": 9.179164422783057e-07, + "loss": 0.8148, + "step": 20510 + }, + { + "epoch": 0.7998596737414488, + "grad_norm": 13.820855140686035, + "learning_rate": 9.177919003444343e-07, + "loss": 0.8428, + "step": 20520 + }, + { + "epoch": 0.8002494689040909, + "grad_norm": 13.888851165771484, + "learning_rate": 9.176672724632949e-07, + "loss": 0.9602, + "step": 20530 + }, + { + "epoch": 0.8006392640667329, + "grad_norm": 17.454608917236328, + "learning_rate": 9.175425586605255e-07, + "loss": 0.9204, + "step": 20540 + }, + { + "epoch": 0.801029059229375, + "grad_norm": 12.34350872039795, + "learning_rate": 9.174177589617821e-07, + "loss": 0.8873, + "step": 20550 + }, + { + "epoch": 0.801418854392017, + "grad_norm": 13.83306884765625, + "learning_rate": 9.172928733927381e-07, + "loss": 0.9076, + "step": 20560 + }, + { + "epoch": 0.801808649554659, + "grad_norm": 13.16793441772461, + "learning_rate": 9.171679019790848e-07, + "loss": 0.9019, + "step": 20570 + }, + { + "epoch": 0.8021984447173011, + "grad_norm": 12.778064727783203, + "learning_rate": 9.170428447465311e-07, + "loss": 0.9312, + "step": 20580 + }, + { + "epoch": 0.8025882398799431, + "grad_norm": 14.788809776306152, + "learning_rate": 9.169177017208032e-07, + "loss": 0.927, + "step": 20590 + }, + { + "epoch": 0.8029780350425851, + "grad_norm": 12.880694389343262, + "learning_rate": 9.167924729276455e-07, + "loss": 0.9, + "step": 20600 + }, + { + "epoch": 0.8033678302052272, + "grad_norm": 14.928827285766602, + "learning_rate": 9.166671583928198e-07, + "loss": 0.9042, + "step": 20610 + }, + { + "epoch": 0.8037576253678692, + "grad_norm": 12.287346839904785, + "learning_rate": 9.165417581421056e-07, + "loss": 0.8961, + "step": 20620 + }, + { + "epoch": 0.8041474205305112, + "grad_norm": 12.641127586364746, + "learning_rate": 9.164162722012998e-07, + "loss": 0.9432, + "step": 20630 + }, + { + "epoch": 0.8045372156931533, + "grad_norm": 13.225910186767578, + "learning_rate": 9.162907005962172e-07, + "loss": 0.9531, + "step": 20640 + }, + { + "epoch": 0.8049270108557953, + "grad_norm": 14.7459077835083, + "learning_rate": 9.161650433526902e-07, + "loss": 0.87, + "step": 20650 + }, + { + "epoch": 0.8053168060184374, + "grad_norm": 11.91112995147705, + "learning_rate": 9.160393004965688e-07, + "loss": 0.8206, + "step": 20660 + }, + { + "epoch": 0.8057066011810794, + "grad_norm": 13.079111099243164, + "learning_rate": 9.159134720537201e-07, + "loss": 0.9419, + "step": 20670 + }, + { + "epoch": 0.8060963963437213, + "grad_norm": 12.377565383911133, + "learning_rate": 9.157875580500298e-07, + "loss": 0.9179, + "step": 20680 + }, + { + "epoch": 0.8064861915063634, + "grad_norm": 11.951333045959473, + "learning_rate": 9.156615585114003e-07, + "loss": 0.8727, + "step": 20690 + }, + { + "epoch": 0.8068759866690054, + "grad_norm": 12.269471168518066, + "learning_rate": 9.155354734637522e-07, + "loss": 0.8427, + "step": 20700 + }, + { + "epoch": 0.8072657818316474, + "grad_norm": 14.251818656921387, + "learning_rate": 9.154093029330234e-07, + "loss": 0.8727, + "step": 20710 + }, + { + "epoch": 0.8076555769942895, + "grad_norm": 11.36607837677002, + "learning_rate": 9.152830469451693e-07, + "loss": 0.8861, + "step": 20720 + }, + { + "epoch": 0.8080453721569315, + "grad_norm": 12.314285278320312, + "learning_rate": 9.151567055261632e-07, + "loss": 0.8613, + "step": 20730 + }, + { + "epoch": 0.8084351673195735, + "grad_norm": 13.008698463439941, + "learning_rate": 9.150302787019957e-07, + "loss": 0.8971, + "step": 20740 + }, + { + "epoch": 0.8088249624822156, + "grad_norm": 12.286422729492188, + "learning_rate": 9.14903766498675e-07, + "loss": 0.9391, + "step": 20750 + }, + { + "epoch": 0.8092147576448576, + "grad_norm": 13.820294380187988, + "learning_rate": 9.147771689422269e-07, + "loss": 0.8861, + "step": 20760 + }, + { + "epoch": 0.8096045528074997, + "grad_norm": 13.752324104309082, + "learning_rate": 9.14650486058695e-07, + "loss": 0.8532, + "step": 20770 + }, + { + "epoch": 0.8099943479701417, + "grad_norm": 12.568449974060059, + "learning_rate": 9.1452371787414e-07, + "loss": 0.8595, + "step": 20780 + }, + { + "epoch": 0.8103841431327837, + "grad_norm": 13.179365158081055, + "learning_rate": 9.143968644146405e-07, + "loss": 0.9302, + "step": 20790 + }, + { + "epoch": 0.8107739382954258, + "grad_norm": 11.456853866577148, + "learning_rate": 9.142699257062924e-07, + "loss": 0.8847, + "step": 20800 + }, + { + "epoch": 0.8111637334580678, + "grad_norm": 14.048447608947754, + "learning_rate": 9.141429017752096e-07, + "loss": 0.9279, + "step": 20810 + }, + { + "epoch": 0.8115535286207098, + "grad_norm": 12.1907958984375, + "learning_rate": 9.140157926475229e-07, + "loss": 0.9314, + "step": 20820 + }, + { + "epoch": 0.8119433237833519, + "grad_norm": 13.158321380615234, + "learning_rate": 9.13888598349381e-07, + "loss": 0.8372, + "step": 20830 + }, + { + "epoch": 0.8123331189459939, + "grad_norm": 13.925211906433105, + "learning_rate": 9.137613189069497e-07, + "loss": 0.8956, + "step": 20840 + }, + { + "epoch": 0.812722914108636, + "grad_norm": 14.971929550170898, + "learning_rate": 9.136339543464133e-07, + "loss": 0.9354, + "step": 20850 + }, + { + "epoch": 0.813112709271278, + "grad_norm": 12.120024681091309, + "learning_rate": 9.135065046939726e-07, + "loss": 0.9298, + "step": 20860 + }, + { + "epoch": 0.81350250443392, + "grad_norm": 13.305877685546875, + "learning_rate": 9.133789699758464e-07, + "loss": 0.9169, + "step": 20870 + }, + { + "epoch": 0.813892299596562, + "grad_norm": 13.102115631103516, + "learning_rate": 9.132513502182707e-07, + "loss": 0.8729, + "step": 20880 + }, + { + "epoch": 0.814282094759204, + "grad_norm": 13.103133201599121, + "learning_rate": 9.131236454474994e-07, + "loss": 0.8856, + "step": 20890 + }, + { + "epoch": 0.814671889921846, + "grad_norm": 12.148378372192383, + "learning_rate": 9.129958556898034e-07, + "loss": 0.923, + "step": 20900 + }, + { + "epoch": 0.8150616850844881, + "grad_norm": 13.564702033996582, + "learning_rate": 9.128679809714715e-07, + "loss": 0.9477, + "step": 20910 + }, + { + "epoch": 0.8154514802471301, + "grad_norm": 13.826949119567871, + "learning_rate": 9.127400213188096e-07, + "loss": 0.8538, + "step": 20920 + }, + { + "epoch": 0.8158412754097721, + "grad_norm": 13.877846717834473, + "learning_rate": 9.126119767581417e-07, + "loss": 0.8242, + "step": 20930 + }, + { + "epoch": 0.8162310705724142, + "grad_norm": 12.27348518371582, + "learning_rate": 9.124838473158086e-07, + "loss": 0.9256, + "step": 20940 + }, + { + "epoch": 0.8166208657350562, + "grad_norm": 12.847216606140137, + "learning_rate": 9.123556330181688e-07, + "loss": 0.9627, + "step": 20950 + }, + { + "epoch": 0.8170106608976982, + "grad_norm": 13.764994621276855, + "learning_rate": 9.122273338915985e-07, + "loss": 0.9056, + "step": 20960 + }, + { + "epoch": 0.8174004560603403, + "grad_norm": 14.677112579345703, + "learning_rate": 9.120989499624907e-07, + "loss": 0.8778, + "step": 20970 + }, + { + "epoch": 0.8177902512229823, + "grad_norm": 14.006379127502441, + "learning_rate": 9.119704812572566e-07, + "loss": 0.8982, + "step": 20980 + }, + { + "epoch": 0.8181800463856244, + "grad_norm": 13.276876449584961, + "learning_rate": 9.118419278023244e-07, + "loss": 0.9446, + "step": 20990 + }, + { + "epoch": 0.8185698415482664, + "grad_norm": 11.384350776672363, + "learning_rate": 9.1171328962414e-07, + "loss": 0.915, + "step": 21000 + }, + { + "epoch": 0.8185698415482664, + "eval_loss": 0.9099260568618774, + "eval_runtime": 83.7792, + "eval_samples_per_second": 49.499, + "eval_steps_per_second": 6.195, + "step": 21000 + }, + { + "epoch": 0.8189596367109084, + "grad_norm": 12.838311195373535, + "learning_rate": 9.115845667491662e-07, + "loss": 0.9287, + "step": 21010 + }, + { + "epoch": 0.8193494318735505, + "grad_norm": 12.870182037353516, + "learning_rate": 9.114557592038841e-07, + "loss": 0.8591, + "step": 21020 + }, + { + "epoch": 0.8197392270361925, + "grad_norm": 14.571227073669434, + "learning_rate": 9.113268670147913e-07, + "loss": 0.9063, + "step": 21030 + }, + { + "epoch": 0.8201290221988345, + "grad_norm": 13.172647476196289, + "learning_rate": 9.111978902084034e-07, + "loss": 0.8549, + "step": 21040 + }, + { + "epoch": 0.8205188173614766, + "grad_norm": 11.640169143676758, + "learning_rate": 9.110688288112534e-07, + "loss": 0.846, + "step": 21050 + }, + { + "epoch": 0.8209086125241186, + "grad_norm": 14.199711799621582, + "learning_rate": 9.109396828498912e-07, + "loss": 0.9114, + "step": 21060 + }, + { + "epoch": 0.8212984076867607, + "grad_norm": 13.395323753356934, + "learning_rate": 9.108104523508847e-07, + "loss": 0.9661, + "step": 21070 + }, + { + "epoch": 0.8216882028494027, + "grad_norm": 11.446035385131836, + "learning_rate": 9.106811373408187e-07, + "loss": 0.878, + "step": 21080 + }, + { + "epoch": 0.8220779980120446, + "grad_norm": 14.23696231842041, + "learning_rate": 9.105517378462959e-07, + "loss": 0.8749, + "step": 21090 + }, + { + "epoch": 0.8224677931746867, + "grad_norm": 15.392660140991211, + "learning_rate": 9.104222538939359e-07, + "loss": 0.969, + "step": 21100 + }, + { + "epoch": 0.8228575883373287, + "grad_norm": 14.680463790893555, + "learning_rate": 9.102926855103758e-07, + "loss": 0.8795, + "step": 21110 + }, + { + "epoch": 0.8232473834999707, + "grad_norm": 10.10588550567627, + "learning_rate": 9.101630327222702e-07, + "loss": 0.8006, + "step": 21120 + }, + { + "epoch": 0.8236371786626128, + "grad_norm": 14.661775588989258, + "learning_rate": 9.100332955562912e-07, + "loss": 0.9604, + "step": 21130 + }, + { + "epoch": 0.8240269738252548, + "grad_norm": 11.952764511108398, + "learning_rate": 9.099034740391278e-07, + "loss": 0.876, + "step": 21140 + }, + { + "epoch": 0.8244167689878968, + "grad_norm": 12.32952880859375, + "learning_rate": 9.097735681974866e-07, + "loss": 0.8565, + "step": 21150 + }, + { + "epoch": 0.8248065641505389, + "grad_norm": 14.418686866760254, + "learning_rate": 9.096435780580918e-07, + "loss": 0.891, + "step": 21160 + }, + { + "epoch": 0.8251963593131809, + "grad_norm": 14.218153953552246, + "learning_rate": 9.095135036476844e-07, + "loss": 0.8804, + "step": 21170 + }, + { + "epoch": 0.825586154475823, + "grad_norm": 13.520167350769043, + "learning_rate": 9.093833449930233e-07, + "loss": 0.8186, + "step": 21180 + }, + { + "epoch": 0.825975949638465, + "grad_norm": 12.958057403564453, + "learning_rate": 9.092531021208842e-07, + "loss": 0.9498, + "step": 21190 + }, + { + "epoch": 0.826365744801107, + "grad_norm": 12.97569751739502, + "learning_rate": 9.091227750580608e-07, + "loss": 0.9608, + "step": 21200 + }, + { + "epoch": 0.8267555399637491, + "grad_norm": 13.833818435668945, + "learning_rate": 9.089923638313632e-07, + "loss": 0.9275, + "step": 21210 + }, + { + "epoch": 0.8271453351263911, + "grad_norm": 14.584861755371094, + "learning_rate": 9.088618684676196e-07, + "loss": 0.8537, + "step": 21220 + }, + { + "epoch": 0.8275351302890331, + "grad_norm": 14.388948440551758, + "learning_rate": 9.087312889936751e-07, + "loss": 0.8949, + "step": 21230 + }, + { + "epoch": 0.8279249254516752, + "grad_norm": 17.273820877075195, + "learning_rate": 9.086006254363924e-07, + "loss": 0.8803, + "step": 21240 + }, + { + "epoch": 0.8283147206143172, + "grad_norm": 12.837658882141113, + "learning_rate": 9.084698778226514e-07, + "loss": 0.8664, + "step": 21250 + }, + { + "epoch": 0.8287045157769592, + "grad_norm": 14.537212371826172, + "learning_rate": 9.08339046179349e-07, + "loss": 0.8885, + "step": 21260 + }, + { + "epoch": 0.8290943109396013, + "grad_norm": 12.475194931030273, + "learning_rate": 9.082081305333997e-07, + "loss": 0.8861, + "step": 21270 + }, + { + "epoch": 0.8294841061022433, + "grad_norm": 13.453390121459961, + "learning_rate": 9.080771309117352e-07, + "loss": 0.8722, + "step": 21280 + }, + { + "epoch": 0.8298739012648852, + "grad_norm": 14.019271850585938, + "learning_rate": 9.079460473413043e-07, + "loss": 0.9561, + "step": 21290 + }, + { + "epoch": 0.8302636964275273, + "grad_norm": 16.61382484436035, + "learning_rate": 9.078148798490734e-07, + "loss": 0.938, + "step": 21300 + }, + { + "epoch": 0.8306534915901693, + "grad_norm": 12.695673942565918, + "learning_rate": 9.076836284620262e-07, + "loss": 0.9207, + "step": 21310 + }, + { + "epoch": 0.8310432867528114, + "grad_norm": 13.421342849731445, + "learning_rate": 9.075522932071631e-07, + "loss": 0.9027, + "step": 21320 + }, + { + "epoch": 0.8314330819154534, + "grad_norm": 13.82275676727295, + "learning_rate": 9.074208741115021e-07, + "loss": 0.9324, + "step": 21330 + }, + { + "epoch": 0.8318228770780954, + "grad_norm": 13.796899795532227, + "learning_rate": 9.072893712020788e-07, + "loss": 0.8511, + "step": 21340 + }, + { + "epoch": 0.8322126722407375, + "grad_norm": 12.66252326965332, + "learning_rate": 9.071577845059455e-07, + "loss": 0.9076, + "step": 21350 + }, + { + "epoch": 0.8326024674033795, + "grad_norm": 14.82443618774414, + "learning_rate": 9.070261140501719e-07, + "loss": 0.8976, + "step": 21360 + }, + { + "epoch": 0.8329922625660215, + "grad_norm": 12.440718650817871, + "learning_rate": 9.06894359861845e-07, + "loss": 0.9369, + "step": 21370 + }, + { + "epoch": 0.8333820577286636, + "grad_norm": 11.333327293395996, + "learning_rate": 9.067625219680691e-07, + "loss": 0.9503, + "step": 21380 + }, + { + "epoch": 0.8337718528913056, + "grad_norm": 13.797769546508789, + "learning_rate": 9.066306003959654e-07, + "loss": 0.9401, + "step": 21390 + }, + { + "epoch": 0.8341616480539477, + "grad_norm": 13.083796501159668, + "learning_rate": 9.064985951726728e-07, + "loss": 0.8984, + "step": 21400 + }, + { + "epoch": 0.8345514432165897, + "grad_norm": 13.291300773620605, + "learning_rate": 9.063665063253469e-07, + "loss": 0.9344, + "step": 21410 + }, + { + "epoch": 0.8349412383792317, + "grad_norm": 14.713855743408203, + "learning_rate": 9.062343338811609e-07, + "loss": 0.9077, + "step": 21420 + }, + { + "epoch": 0.8353310335418738, + "grad_norm": 11.754354476928711, + "learning_rate": 9.061020778673049e-07, + "loss": 0.7929, + "step": 21430 + }, + { + "epoch": 0.8357208287045158, + "grad_norm": 14.496849060058594, + "learning_rate": 9.059697383109864e-07, + "loss": 0.8841, + "step": 21440 + }, + { + "epoch": 0.8361106238671578, + "grad_norm": 13.750931739807129, + "learning_rate": 9.058373152394299e-07, + "loss": 0.9528, + "step": 21450 + }, + { + "epoch": 0.8365004190297999, + "grad_norm": 13.394865989685059, + "learning_rate": 9.057048086798774e-07, + "loss": 0.861, + "step": 21460 + }, + { + "epoch": 0.8368902141924419, + "grad_norm": 12.10695743560791, + "learning_rate": 9.055722186595878e-07, + "loss": 0.8709, + "step": 21470 + }, + { + "epoch": 0.837280009355084, + "grad_norm": 13.600817680358887, + "learning_rate": 9.054395452058372e-07, + "loss": 0.902, + "step": 21480 + }, + { + "epoch": 0.8376698045177259, + "grad_norm": 14.576142311096191, + "learning_rate": 9.053067883459189e-07, + "loss": 0.9084, + "step": 21490 + }, + { + "epoch": 0.8380595996803679, + "grad_norm": 13.02922248840332, + "learning_rate": 9.051739481071433e-07, + "loss": 0.9226, + "step": 21500 + }, + { + "epoch": 0.8380595996803679, + "eval_loss": 0.9090902209281921, + "eval_runtime": 83.0784, + "eval_samples_per_second": 49.917, + "eval_steps_per_second": 6.247, + "step": 21500 + }, + { + "epoch": 0.83844939484301, + "grad_norm": 13.008776664733887, + "learning_rate": 9.050410245168381e-07, + "loss": 0.8857, + "step": 21510 + }, + { + "epoch": 0.838839190005652, + "grad_norm": 13.534220695495605, + "learning_rate": 9.049080176023482e-07, + "loss": 0.8745, + "step": 21520 + }, + { + "epoch": 0.839228985168294, + "grad_norm": 13.479632377624512, + "learning_rate": 9.047749273910352e-07, + "loss": 0.9257, + "step": 21530 + }, + { + "epoch": 0.8396187803309361, + "grad_norm": 14.95528793334961, + "learning_rate": 9.046417539102783e-07, + "loss": 0.8719, + "step": 21540 + }, + { + "epoch": 0.8400085754935781, + "grad_norm": 20.018394470214844, + "learning_rate": 9.045084971874737e-07, + "loss": 0.9463, + "step": 21550 + }, + { + "epoch": 0.8403983706562201, + "grad_norm": 13.088545799255371, + "learning_rate": 9.043751572500347e-07, + "loss": 0.9364, + "step": 21560 + }, + { + "epoch": 0.8407881658188622, + "grad_norm": 11.355172157287598, + "learning_rate": 9.042417341253914e-07, + "loss": 0.8293, + "step": 21570 + }, + { + "epoch": 0.8411779609815042, + "grad_norm": 13.545299530029297, + "learning_rate": 9.041082278409918e-07, + "loss": 0.8701, + "step": 21580 + }, + { + "epoch": 0.8415677561441463, + "grad_norm": 13.51905632019043, + "learning_rate": 9.039746384243004e-07, + "loss": 0.9034, + "step": 21590 + }, + { + "epoch": 0.8419575513067883, + "grad_norm": 14.087303161621094, + "learning_rate": 9.038409659027986e-07, + "loss": 0.9455, + "step": 21600 + }, + { + "epoch": 0.8423473464694303, + "grad_norm": 14.358713150024414, + "learning_rate": 9.037072103039855e-07, + "loss": 0.8791, + "step": 21610 + }, + { + "epoch": 0.8427371416320724, + "grad_norm": 14.497900009155273, + "learning_rate": 9.035733716553772e-07, + "loss": 0.9724, + "step": 21620 + }, + { + "epoch": 0.8431269367947144, + "grad_norm": 14.16992473602295, + "learning_rate": 9.034394499845063e-07, + "loss": 0.9292, + "step": 21630 + }, + { + "epoch": 0.8435167319573564, + "grad_norm": 13.40927791595459, + "learning_rate": 9.033054453189233e-07, + "loss": 0.9186, + "step": 21640 + }, + { + "epoch": 0.8439065271199985, + "grad_norm": 14.81359577178955, + "learning_rate": 9.031713576861949e-07, + "loss": 0.8912, + "step": 21650 + }, + { + "epoch": 0.8442963222826405, + "grad_norm": 12.390361785888672, + "learning_rate": 9.030371871139057e-07, + "loss": 0.8784, + "step": 21660 + }, + { + "epoch": 0.8446861174452825, + "grad_norm": 14.579010009765625, + "learning_rate": 9.029029336296568e-07, + "loss": 0.9294, + "step": 21670 + }, + { + "epoch": 0.8450759126079246, + "grad_norm": 15.339632987976074, + "learning_rate": 9.027685972610665e-07, + "loss": 0.899, + "step": 21680 + }, + { + "epoch": 0.8454657077705666, + "grad_norm": 13.82767105102539, + "learning_rate": 9.026341780357705e-07, + "loss": 0.9112, + "step": 21690 + }, + { + "epoch": 0.8458555029332085, + "grad_norm": 15.720512390136719, + "learning_rate": 9.024996759814209e-07, + "loss": 0.8851, + "step": 21700 + }, + { + "epoch": 0.8462452980958506, + "grad_norm": 12.607199668884277, + "learning_rate": 9.023650911256872e-07, + "loss": 0.8702, + "step": 21710 + }, + { + "epoch": 0.8466350932584926, + "grad_norm": 11.14775276184082, + "learning_rate": 9.022304234962562e-07, + "loss": 0.9483, + "step": 21720 + }, + { + "epoch": 0.8470248884211347, + "grad_norm": 10.135401725769043, + "learning_rate": 9.020956731208312e-07, + "loss": 0.8326, + "step": 21730 + }, + { + "epoch": 0.8474146835837767, + "grad_norm": 13.74143123626709, + "learning_rate": 9.019608400271328e-07, + "loss": 0.8543, + "step": 21740 + }, + { + "epoch": 0.8478044787464187, + "grad_norm": 11.97140884399414, + "learning_rate": 9.018259242428986e-07, + "loss": 0.8804, + "step": 21750 + }, + { + "epoch": 0.8481942739090608, + "grad_norm": 12.63471508026123, + "learning_rate": 9.016909257958833e-07, + "loss": 0.8953, + "step": 21760 + }, + { + "epoch": 0.8485840690717028, + "grad_norm": 12.790428161621094, + "learning_rate": 9.015558447138583e-07, + "loss": 0.8818, + "step": 21770 + }, + { + "epoch": 0.8489738642343448, + "grad_norm": 11.905839920043945, + "learning_rate": 9.014206810246121e-07, + "loss": 0.9265, + "step": 21780 + }, + { + "epoch": 0.8493636593969869, + "grad_norm": 13.096185684204102, + "learning_rate": 9.012854347559506e-07, + "loss": 0.8614, + "step": 21790 + }, + { + "epoch": 0.8497534545596289, + "grad_norm": 10.801603317260742, + "learning_rate": 9.011501059356964e-07, + "loss": 0.8096, + "step": 21800 + }, + { + "epoch": 0.850143249722271, + "grad_norm": 12.584639549255371, + "learning_rate": 9.010146945916888e-07, + "loss": 0.9457, + "step": 21810 + }, + { + "epoch": 0.850533044884913, + "grad_norm": 10.216394424438477, + "learning_rate": 9.008792007517845e-07, + "loss": 0.8088, + "step": 21820 + }, + { + "epoch": 0.850922840047555, + "grad_norm": 12.994810104370117, + "learning_rate": 9.007436244438569e-07, + "loss": 0.9237, + "step": 21830 + }, + { + "epoch": 0.8513126352101971, + "grad_norm": 13.801116943359375, + "learning_rate": 9.006079656957965e-07, + "loss": 0.9097, + "step": 21840 + }, + { + "epoch": 0.8517024303728391, + "grad_norm": 12.289114952087402, + "learning_rate": 9.004722245355108e-07, + "loss": 0.8972, + "step": 21850 + }, + { + "epoch": 0.8520922255354811, + "grad_norm": 13.461929321289062, + "learning_rate": 9.00336400990924e-07, + "loss": 1.0001, + "step": 21860 + }, + { + "epoch": 0.8524820206981232, + "grad_norm": 13.124207496643066, + "learning_rate": 9.002004950899776e-07, + "loss": 0.9262, + "step": 21870 + }, + { + "epoch": 0.8528718158607652, + "grad_norm": 13.562239646911621, + "learning_rate": 9.000645068606301e-07, + "loss": 0.9102, + "step": 21880 + }, + { + "epoch": 0.8532616110234073, + "grad_norm": 12.1265287399292, + "learning_rate": 8.999284363308561e-07, + "loss": 0.8884, + "step": 21890 + }, + { + "epoch": 0.8536514061860492, + "grad_norm": 13.701234817504883, + "learning_rate": 8.997922835286484e-07, + "loss": 0.91, + "step": 21900 + }, + { + "epoch": 0.8540412013486912, + "grad_norm": 14.029929161071777, + "learning_rate": 8.996560484820156e-07, + "loss": 0.9024, + "step": 21910 + }, + { + "epoch": 0.8544309965113333, + "grad_norm": 11.016921043395996, + "learning_rate": 8.99519731218984e-07, + "loss": 0.8556, + "step": 21920 + }, + { + "epoch": 0.8548207916739753, + "grad_norm": 14.105498313903809, + "learning_rate": 8.993833317675962e-07, + "loss": 0.9125, + "step": 21930 + }, + { + "epoch": 0.8552105868366173, + "grad_norm": 13.061525344848633, + "learning_rate": 8.992468501559122e-07, + "loss": 0.8939, + "step": 21940 + }, + { + "epoch": 0.8556003819992594, + "grad_norm": 14.45816707611084, + "learning_rate": 8.991102864120086e-07, + "loss": 0.9076, + "step": 21950 + }, + { + "epoch": 0.8559901771619014, + "grad_norm": 13.855520248413086, + "learning_rate": 8.989736405639792e-07, + "loss": 0.8948, + "step": 21960 + }, + { + "epoch": 0.8563799723245434, + "grad_norm": 13.307405471801758, + "learning_rate": 8.988369126399343e-07, + "loss": 0.868, + "step": 21970 + }, + { + "epoch": 0.8567697674871855, + "grad_norm": 12.386114120483398, + "learning_rate": 8.987001026680014e-07, + "loss": 0.9069, + "step": 21980 + }, + { + "epoch": 0.8571595626498275, + "grad_norm": 12.267435073852539, + "learning_rate": 8.985632106763248e-07, + "loss": 0.9244, + "step": 21990 + }, + { + "epoch": 0.8575493578124695, + "grad_norm": 15.029688835144043, + "learning_rate": 8.984262366930653e-07, + "loss": 0.8633, + "step": 22000 + }, + { + "epoch": 0.8575493578124695, + "eval_loss": 0.9064081311225891, + "eval_runtime": 82.8649, + "eval_samples_per_second": 50.045, + "eval_steps_per_second": 6.263, + "step": 22000 + }, + { + "epoch": 0.8579391529751116, + "grad_norm": 12.505367279052734, + "learning_rate": 8.982891807464013e-07, + "loss": 0.8827, + "step": 22010 + }, + { + "epoch": 0.8583289481377536, + "grad_norm": 15.047062873840332, + "learning_rate": 8.981520428645275e-07, + "loss": 0.88, + "step": 22020 + }, + { + "epoch": 0.8587187433003957, + "grad_norm": 15.56449031829834, + "learning_rate": 8.980148230756555e-07, + "loss": 0.8381, + "step": 22030 + }, + { + "epoch": 0.8591085384630377, + "grad_norm": 12.24487590789795, + "learning_rate": 8.978775214080141e-07, + "loss": 0.8718, + "step": 22040 + }, + { + "epoch": 0.8594983336256797, + "grad_norm": 12.147867202758789, + "learning_rate": 8.977401378898486e-07, + "loss": 0.9495, + "step": 22050 + }, + { + "epoch": 0.8598881287883218, + "grad_norm": 10.67481803894043, + "learning_rate": 8.976026725494212e-07, + "loss": 0.9091, + "step": 22060 + }, + { + "epoch": 0.8602779239509638, + "grad_norm": 12.627367973327637, + "learning_rate": 8.97465125415011e-07, + "loss": 0.9011, + "step": 22070 + }, + { + "epoch": 0.8606677191136058, + "grad_norm": 13.642681121826172, + "learning_rate": 8.97327496514914e-07, + "loss": 0.9358, + "step": 22080 + }, + { + "epoch": 0.8610575142762479, + "grad_norm": 15.911971092224121, + "learning_rate": 8.971897858774427e-07, + "loss": 0.9004, + "step": 22090 + }, + { + "epoch": 0.8614473094388899, + "grad_norm": 13.642032623291016, + "learning_rate": 8.970519935309269e-07, + "loss": 0.9567, + "step": 22100 + }, + { + "epoch": 0.8618371046015318, + "grad_norm": 12.386261940002441, + "learning_rate": 8.969141195037127e-07, + "loss": 0.8793, + "step": 22110 + }, + { + "epoch": 0.8622268997641739, + "grad_norm": 13.471172332763672, + "learning_rate": 8.967761638241636e-07, + "loss": 0.9462, + "step": 22120 + }, + { + "epoch": 0.8626166949268159, + "grad_norm": 11.691216468811035, + "learning_rate": 8.966381265206593e-07, + "loss": 0.9965, + "step": 22130 + }, + { + "epoch": 0.863006490089458, + "grad_norm": 16.051721572875977, + "learning_rate": 8.965000076215965e-07, + "loss": 0.9214, + "step": 22140 + }, + { + "epoch": 0.8633962852521, + "grad_norm": 12.66292667388916, + "learning_rate": 8.963618071553889e-07, + "loss": 0.9162, + "step": 22150 + }, + { + "epoch": 0.863786080414742, + "grad_norm": 13.350120544433594, + "learning_rate": 8.962235251504666e-07, + "loss": 0.8916, + "step": 22160 + }, + { + "epoch": 0.8641758755773841, + "grad_norm": 13.573288917541504, + "learning_rate": 8.960851616352769e-07, + "loss": 0.9638, + "step": 22170 + }, + { + "epoch": 0.8645656707400261, + "grad_norm": 14.271215438842773, + "learning_rate": 8.959467166382837e-07, + "loss": 0.9342, + "step": 22180 + }, + { + "epoch": 0.8649554659026681, + "grad_norm": 13.345029830932617, + "learning_rate": 8.958081901879672e-07, + "loss": 0.9333, + "step": 22190 + }, + { + "epoch": 0.8653452610653102, + "grad_norm": 11.451812744140625, + "learning_rate": 8.956695823128253e-07, + "loss": 0.9212, + "step": 22200 + }, + { + "epoch": 0.8657350562279522, + "grad_norm": 14.416302680969238, + "learning_rate": 8.955308930413717e-07, + "loss": 0.8647, + "step": 22210 + }, + { + "epoch": 0.8661248513905943, + "grad_norm": 13.178296089172363, + "learning_rate": 8.953921224021374e-07, + "loss": 0.8745, + "step": 22220 + }, + { + "epoch": 0.8665146465532363, + "grad_norm": 12.618408203125, + "learning_rate": 8.952532704236701e-07, + "loss": 0.9026, + "step": 22230 + }, + { + "epoch": 0.8669044417158783, + "grad_norm": 15.165433883666992, + "learning_rate": 8.951143371345342e-07, + "loss": 0.9918, + "step": 22240 + }, + { + "epoch": 0.8672942368785204, + "grad_norm": 12.498310089111328, + "learning_rate": 8.949753225633105e-07, + "loss": 0.9203, + "step": 22250 + }, + { + "epoch": 0.8676840320411624, + "grad_norm": 12.744924545288086, + "learning_rate": 8.94836226738597e-07, + "loss": 0.8567, + "step": 22260 + }, + { + "epoch": 0.8680738272038044, + "grad_norm": 13.908757209777832, + "learning_rate": 8.946970496890079e-07, + "loss": 0.8645, + "step": 22270 + }, + { + "epoch": 0.8684636223664465, + "grad_norm": 10.994424819946289, + "learning_rate": 8.945577914431748e-07, + "loss": 0.91, + "step": 22280 + }, + { + "epoch": 0.8688534175290885, + "grad_norm": 13.063972473144531, + "learning_rate": 8.944184520297453e-07, + "loss": 0.9443, + "step": 22290 + }, + { + "epoch": 0.8692432126917305, + "grad_norm": 13.354785919189453, + "learning_rate": 8.942790314773843e-07, + "loss": 0.8288, + "step": 22300 + }, + { + "epoch": 0.8696330078543725, + "grad_norm": 13.410421371459961, + "learning_rate": 8.941395298147728e-07, + "loss": 0.8615, + "step": 22310 + }, + { + "epoch": 0.8700228030170145, + "grad_norm": 11.427786827087402, + "learning_rate": 8.93999947070609e-07, + "loss": 0.8748, + "step": 22320 + }, + { + "epoch": 0.8704125981796565, + "grad_norm": 13.338788986206055, + "learning_rate": 8.938602832736075e-07, + "loss": 0.9738, + "step": 22330 + }, + { + "epoch": 0.8708023933422986, + "grad_norm": 11.633015632629395, + "learning_rate": 8.937205384524997e-07, + "loss": 0.8677, + "step": 22340 + }, + { + "epoch": 0.8711921885049406, + "grad_norm": 13.497509956359863, + "learning_rate": 8.935807126360335e-07, + "loss": 0.8592, + "step": 22350 + }, + { + "epoch": 0.8715819836675827, + "grad_norm": 14.914027214050293, + "learning_rate": 8.934408058529735e-07, + "loss": 0.8859, + "step": 22360 + }, + { + "epoch": 0.8719717788302247, + "grad_norm": 14.125609397888184, + "learning_rate": 8.933008181321013e-07, + "loss": 0.9091, + "step": 22370 + }, + { + "epoch": 0.8723615739928667, + "grad_norm": 14.816338539123535, + "learning_rate": 8.931607495022147e-07, + "loss": 0.9599, + "step": 22380 + }, + { + "epoch": 0.8727513691555088, + "grad_norm": 11.920924186706543, + "learning_rate": 8.930205999921285e-07, + "loss": 0.9655, + "step": 22390 + }, + { + "epoch": 0.8731411643181508, + "grad_norm": 14.017515182495117, + "learning_rate": 8.928803696306735e-07, + "loss": 0.8414, + "step": 22400 + }, + { + "epoch": 0.8735309594807928, + "grad_norm": 12.449407577514648, + "learning_rate": 8.927400584466982e-07, + "loss": 1.0218, + "step": 22410 + }, + { + "epoch": 0.8739207546434349, + "grad_norm": 10.87103271484375, + "learning_rate": 8.925996664690667e-07, + "loss": 0.8879, + "step": 22420 + }, + { + "epoch": 0.8743105498060769, + "grad_norm": 13.074341773986816, + "learning_rate": 8.924591937266603e-07, + "loss": 0.8949, + "step": 22430 + }, + { + "epoch": 0.874700344968719, + "grad_norm": 13.620491981506348, + "learning_rate": 8.923186402483766e-07, + "loss": 0.8785, + "step": 22440 + }, + { + "epoch": 0.875090140131361, + "grad_norm": 13.167861938476562, + "learning_rate": 8.921780060631302e-07, + "loss": 0.8664, + "step": 22450 + }, + { + "epoch": 0.875479935294003, + "grad_norm": 13.330260276794434, + "learning_rate": 8.92037291199852e-07, + "loss": 0.8585, + "step": 22460 + }, + { + "epoch": 0.8758697304566451, + "grad_norm": 11.642390251159668, + "learning_rate": 8.918964956874895e-07, + "loss": 0.9295, + "step": 22470 + }, + { + "epoch": 0.8762595256192871, + "grad_norm": 14.286483764648438, + "learning_rate": 8.917556195550069e-07, + "loss": 0.903, + "step": 22480 + }, + { + "epoch": 0.8766493207819291, + "grad_norm": 11.244755744934082, + "learning_rate": 8.916146628313848e-07, + "loss": 0.8579, + "step": 22490 + }, + { + "epoch": 0.8770391159445712, + "grad_norm": 12.591062545776367, + "learning_rate": 8.914736255456208e-07, + "loss": 0.9307, + "step": 22500 + }, + { + "epoch": 0.8770391159445712, + "eval_loss": 0.9050285816192627, + "eval_runtime": 83.4511, + "eval_samples_per_second": 49.694, + "eval_steps_per_second": 6.219, + "step": 22500 + }, + { + "epoch": 0.8774289111072131, + "grad_norm": 14.885220527648926, + "learning_rate": 8.913325077267285e-07, + "loss": 0.9509, + "step": 22510 + }, + { + "epoch": 0.8778187062698551, + "grad_norm": 12.175387382507324, + "learning_rate": 8.911913094037385e-07, + "loss": 0.9069, + "step": 22520 + }, + { + "epoch": 0.8782085014324972, + "grad_norm": 13.45309829711914, + "learning_rate": 8.910500306056978e-07, + "loss": 0.8667, + "step": 22530 + }, + { + "epoch": 0.8785982965951392, + "grad_norm": 12.692883491516113, + "learning_rate": 8.909086713616702e-07, + "loss": 0.8453, + "step": 22540 + }, + { + "epoch": 0.8789880917577813, + "grad_norm": 12.696819305419922, + "learning_rate": 8.907672317007355e-07, + "loss": 0.8543, + "step": 22550 + }, + { + "epoch": 0.8793778869204233, + "grad_norm": 13.598355293273926, + "learning_rate": 8.906257116519904e-07, + "loss": 0.9375, + "step": 22560 + }, + { + "epoch": 0.8797676820830653, + "grad_norm": 13.774086952209473, + "learning_rate": 8.904841112445482e-07, + "loss": 0.8965, + "step": 22570 + }, + { + "epoch": 0.8801574772457074, + "grad_norm": 13.858778953552246, + "learning_rate": 8.903424305075387e-07, + "loss": 0.9107, + "step": 22580 + }, + { + "epoch": 0.8805472724083494, + "grad_norm": 13.999808311462402, + "learning_rate": 8.902006694701081e-07, + "loss": 0.8813, + "step": 22590 + }, + { + "epoch": 0.8809370675709914, + "grad_norm": 11.351675033569336, + "learning_rate": 8.900588281614191e-07, + "loss": 0.8473, + "step": 22600 + }, + { + "epoch": 0.8813268627336335, + "grad_norm": 12.952067375183105, + "learning_rate": 8.899169066106511e-07, + "loss": 0.9582, + "step": 22610 + }, + { + "epoch": 0.8817166578962755, + "grad_norm": 12.525954246520996, + "learning_rate": 8.897749048469998e-07, + "loss": 0.8686, + "step": 22620 + }, + { + "epoch": 0.8821064530589176, + "grad_norm": 13.05552864074707, + "learning_rate": 8.896328228996777e-07, + "loss": 0.9249, + "step": 22630 + }, + { + "epoch": 0.8824962482215596, + "grad_norm": 13.277647018432617, + "learning_rate": 8.894906607979133e-07, + "loss": 0.9006, + "step": 22640 + }, + { + "epoch": 0.8828860433842016, + "grad_norm": 12.067602157592773, + "learning_rate": 8.893484185709523e-07, + "loss": 0.8969, + "step": 22650 + }, + { + "epoch": 0.8832758385468437, + "grad_norm": 15.378790855407715, + "learning_rate": 8.892060962480561e-07, + "loss": 0.9168, + "step": 22660 + }, + { + "epoch": 0.8836656337094857, + "grad_norm": 12.63527774810791, + "learning_rate": 8.89063693858503e-07, + "loss": 0.9442, + "step": 22670 + }, + { + "epoch": 0.8840554288721277, + "grad_norm": 12.48078441619873, + "learning_rate": 8.889212114315879e-07, + "loss": 0.8988, + "step": 22680 + }, + { + "epoch": 0.8844452240347698, + "grad_norm": 13.223127365112305, + "learning_rate": 8.887786489966218e-07, + "loss": 0.8245, + "step": 22690 + }, + { + "epoch": 0.8848350191974118, + "grad_norm": 12.528448104858398, + "learning_rate": 8.886360065829325e-07, + "loss": 0.8349, + "step": 22700 + }, + { + "epoch": 0.8852248143600538, + "grad_norm": 12.33480167388916, + "learning_rate": 8.884932842198641e-07, + "loss": 0.8496, + "step": 22710 + }, + { + "epoch": 0.8856146095226958, + "grad_norm": 13.84949016571045, + "learning_rate": 8.883504819367768e-07, + "loss": 0.9397, + "step": 22720 + }, + { + "epoch": 0.8860044046853378, + "grad_norm": 14.53945541381836, + "learning_rate": 8.88207599763048e-07, + "loss": 0.891, + "step": 22730 + }, + { + "epoch": 0.8863941998479798, + "grad_norm": 13.971746444702148, + "learning_rate": 8.88064637728071e-07, + "loss": 0.9171, + "step": 22740 + }, + { + "epoch": 0.8867839950106219, + "grad_norm": 13.187960624694824, + "learning_rate": 8.879215958612556e-07, + "loss": 0.8711, + "step": 22750 + }, + { + "epoch": 0.8871737901732639, + "grad_norm": 16.219009399414062, + "learning_rate": 8.87778474192028e-07, + "loss": 0.8482, + "step": 22760 + }, + { + "epoch": 0.887563585335906, + "grad_norm": 14.09238338470459, + "learning_rate": 8.876352727498311e-07, + "loss": 0.8592, + "step": 22770 + }, + { + "epoch": 0.887953380498548, + "grad_norm": 13.13365364074707, + "learning_rate": 8.874919915641238e-07, + "loss": 0.8976, + "step": 22780 + }, + { + "epoch": 0.88834317566119, + "grad_norm": 13.46047592163086, + "learning_rate": 8.873486306643817e-07, + "loss": 0.9079, + "step": 22790 + }, + { + "epoch": 0.8887329708238321, + "grad_norm": 13.740199089050293, + "learning_rate": 8.872051900800966e-07, + "loss": 0.9824, + "step": 22800 + }, + { + "epoch": 0.8891227659864741, + "grad_norm": 13.782218933105469, + "learning_rate": 8.870616698407769e-07, + "loss": 0.9241, + "step": 22810 + }, + { + "epoch": 0.8895125611491161, + "grad_norm": 12.052614212036133, + "learning_rate": 8.869180699759473e-07, + "loss": 0.9279, + "step": 22820 + }, + { + "epoch": 0.8899023563117582, + "grad_norm": 13.61395263671875, + "learning_rate": 8.867743905151488e-07, + "loss": 0.9295, + "step": 22830 + }, + { + "epoch": 0.8902921514744002, + "grad_norm": 10.535191535949707, + "learning_rate": 8.866306314879388e-07, + "loss": 0.8681, + "step": 22840 + }, + { + "epoch": 0.8906819466370423, + "grad_norm": 12.016164779663086, + "learning_rate": 8.864867929238913e-07, + "loss": 0.8652, + "step": 22850 + }, + { + "epoch": 0.8910717417996843, + "grad_norm": 13.213897705078125, + "learning_rate": 8.863428748525961e-07, + "loss": 0.8722, + "step": 22860 + }, + { + "epoch": 0.8914615369623263, + "grad_norm": 12.421365737915039, + "learning_rate": 8.861988773036602e-07, + "loss": 0.9207, + "step": 22870 + }, + { + "epoch": 0.8918513321249684, + "grad_norm": 13.187318801879883, + "learning_rate": 8.86054800306706e-07, + "loss": 0.9443, + "step": 22880 + }, + { + "epoch": 0.8922411272876104, + "grad_norm": 12.807977676391602, + "learning_rate": 8.859106438913729e-07, + "loss": 0.9073, + "step": 22890 + }, + { + "epoch": 0.8926309224502524, + "grad_norm": 12.861021995544434, + "learning_rate": 8.857664080873166e-07, + "loss": 0.8516, + "step": 22900 + }, + { + "epoch": 0.8930207176128945, + "grad_norm": 13.972962379455566, + "learning_rate": 8.856220929242089e-07, + "loss": 0.9144, + "step": 22910 + }, + { + "epoch": 0.8934105127755364, + "grad_norm": 13.490053176879883, + "learning_rate": 8.854776984317378e-07, + "loss": 0.9178, + "step": 22920 + }, + { + "epoch": 0.8938003079381784, + "grad_norm": 12.621403694152832, + "learning_rate": 8.853332246396082e-07, + "loss": 0.9117, + "step": 22930 + }, + { + "epoch": 0.8941901031008205, + "grad_norm": 11.63735294342041, + "learning_rate": 8.851886715775406e-07, + "loss": 0.8768, + "step": 22940 + }, + { + "epoch": 0.8945798982634625, + "grad_norm": 14.49917984008789, + "learning_rate": 8.850440392752723e-07, + "loss": 0.9101, + "step": 22950 + }, + { + "epoch": 0.8949696934261046, + "grad_norm": 13.012141227722168, + "learning_rate": 8.84899327762557e-07, + "loss": 0.9413, + "step": 22960 + }, + { + "epoch": 0.8953594885887466, + "grad_norm": 16.267471313476562, + "learning_rate": 8.84754537069164e-07, + "loss": 0.9088, + "step": 22970 + }, + { + "epoch": 0.8957492837513886, + "grad_norm": 10.475727081298828, + "learning_rate": 8.846096672248795e-07, + "loss": 0.862, + "step": 22980 + }, + { + "epoch": 0.8961390789140307, + "grad_norm": 11.38765811920166, + "learning_rate": 8.844647182595059e-07, + "loss": 0.9222, + "step": 22990 + }, + { + "epoch": 0.8965288740766727, + "grad_norm": 14.402395248413086, + "learning_rate": 8.843196902028618e-07, + "loss": 0.8596, + "step": 23000 + }, + { + "epoch": 0.8965288740766727, + "eval_loss": 0.9019464254379272, + "eval_runtime": 82.6992, + "eval_samples_per_second": 50.146, + "eval_steps_per_second": 6.276, + "step": 23000 + }, + { + "epoch": 0.8969186692393147, + "grad_norm": 14.14964771270752, + "learning_rate": 8.84174583084782e-07, + "loss": 0.937, + "step": 23010 + }, + { + "epoch": 0.8973084644019568, + "grad_norm": 14.001018524169922, + "learning_rate": 8.840293969351175e-07, + "loss": 0.8099, + "step": 23020 + }, + { + "epoch": 0.8976982595645988, + "grad_norm": 14.234670639038086, + "learning_rate": 8.83884131783736e-07, + "loss": 0.8994, + "step": 23030 + }, + { + "epoch": 0.8980880547272408, + "grad_norm": 14.878447532653809, + "learning_rate": 8.837387876605209e-07, + "loss": 0.9152, + "step": 23040 + }, + { + "epoch": 0.8984778498898829, + "grad_norm": 13.215280532836914, + "learning_rate": 8.835933645953722e-07, + "loss": 0.8884, + "step": 23050 + }, + { + "epoch": 0.8988676450525249, + "grad_norm": 12.790202140808105, + "learning_rate": 8.83447862618206e-07, + "loss": 0.8963, + "step": 23060 + }, + { + "epoch": 0.899257440215167, + "grad_norm": 11.240697860717773, + "learning_rate": 8.833022817589546e-07, + "loss": 0.8789, + "step": 23070 + }, + { + "epoch": 0.899647235377809, + "grad_norm": 11.788422584533691, + "learning_rate": 8.831566220475668e-07, + "loss": 0.8598, + "step": 23080 + }, + { + "epoch": 0.900037030540451, + "grad_norm": 11.10592269897461, + "learning_rate": 8.830108835140071e-07, + "loss": 0.8554, + "step": 23090 + }, + { + "epoch": 0.9004268257030931, + "grad_norm": 11.270769119262695, + "learning_rate": 8.828650661882566e-07, + "loss": 0.8293, + "step": 23100 + }, + { + "epoch": 0.9008166208657351, + "grad_norm": 15.193254470825195, + "learning_rate": 8.827191701003127e-07, + "loss": 0.8837, + "step": 23110 + }, + { + "epoch": 0.901206416028377, + "grad_norm": 12.982580184936523, + "learning_rate": 8.825731952801887e-07, + "loss": 0.9082, + "step": 23120 + }, + { + "epoch": 0.9015962111910191, + "grad_norm": 14.397028923034668, + "learning_rate": 8.824271417579143e-07, + "loss": 0.9272, + "step": 23130 + }, + { + "epoch": 0.9019860063536611, + "grad_norm": 11.427946090698242, + "learning_rate": 8.822810095635352e-07, + "loss": 0.7883, + "step": 23140 + }, + { + "epoch": 0.9023758015163031, + "grad_norm": 12.576859474182129, + "learning_rate": 8.821347987271135e-07, + "loss": 0.9725, + "step": 23150 + }, + { + "epoch": 0.9027655966789452, + "grad_norm": 14.795187950134277, + "learning_rate": 8.819885092787275e-07, + "loss": 0.8669, + "step": 23160 + }, + { + "epoch": 0.9031553918415872, + "grad_norm": 11.001447677612305, + "learning_rate": 8.818421412484713e-07, + "loss": 0.863, + "step": 23170 + }, + { + "epoch": 0.9035451870042293, + "grad_norm": 12.653565406799316, + "learning_rate": 8.816956946664558e-07, + "loss": 0.9263, + "step": 23180 + }, + { + "epoch": 0.9039349821668713, + "grad_norm": 11.365368843078613, + "learning_rate": 8.815491695628072e-07, + "loss": 0.8714, + "step": 23190 + }, + { + "epoch": 0.9043247773295133, + "grad_norm": 15.443947792053223, + "learning_rate": 8.814025659676687e-07, + "loss": 0.9362, + "step": 23200 + }, + { + "epoch": 0.9047145724921554, + "grad_norm": 11.974738121032715, + "learning_rate": 8.812558839111991e-07, + "loss": 0.9154, + "step": 23210 + }, + { + "epoch": 0.9051043676547974, + "grad_norm": 13.998259544372559, + "learning_rate": 8.811091234235736e-07, + "loss": 0.8609, + "step": 23220 + }, + { + "epoch": 0.9054941628174394, + "grad_norm": 12.091790199279785, + "learning_rate": 8.809622845349834e-07, + "loss": 0.9755, + "step": 23230 + }, + { + "epoch": 0.9058839579800815, + "grad_norm": 13.624860763549805, + "learning_rate": 8.80815367275636e-07, + "loss": 0.8564, + "step": 23240 + }, + { + "epoch": 0.9062737531427235, + "grad_norm": 12.440071105957031, + "learning_rate": 8.806683716757548e-07, + "loss": 0.8528, + "step": 23250 + }, + { + "epoch": 0.9066635483053656, + "grad_norm": 13.891864776611328, + "learning_rate": 8.805212977655793e-07, + "loss": 0.9163, + "step": 23260 + }, + { + "epoch": 0.9070533434680076, + "grad_norm": 14.447443008422852, + "learning_rate": 8.803741455753658e-07, + "loss": 0.9874, + "step": 23270 + }, + { + "epoch": 0.9074431386306496, + "grad_norm": 12.00536060333252, + "learning_rate": 8.802269151353853e-07, + "loss": 0.88, + "step": 23280 + }, + { + "epoch": 0.9078329337932917, + "grad_norm": 11.880412101745605, + "learning_rate": 8.800796064759264e-07, + "loss": 0.956, + "step": 23290 + }, + { + "epoch": 0.9082227289559337, + "grad_norm": 11.96114730834961, + "learning_rate": 8.799322196272928e-07, + "loss": 0.8902, + "step": 23300 + }, + { + "epoch": 0.9086125241185757, + "grad_norm": 12.576797485351562, + "learning_rate": 8.797847546198047e-07, + "loss": 0.8762, + "step": 23310 + }, + { + "epoch": 0.9090023192812178, + "grad_norm": 12.739381790161133, + "learning_rate": 8.796372114837983e-07, + "loss": 0.8901, + "step": 23320 + }, + { + "epoch": 0.9093921144438597, + "grad_norm": 11.353896141052246, + "learning_rate": 8.79489590249626e-07, + "loss": 0.9274, + "step": 23330 + }, + { + "epoch": 0.9097819096065017, + "grad_norm": 12.88727855682373, + "learning_rate": 8.793418909476558e-07, + "loss": 0.8754, + "step": 23340 + }, + { + "epoch": 0.9101717047691438, + "grad_norm": 11.915215492248535, + "learning_rate": 8.791941136082724e-07, + "loss": 0.8634, + "step": 23350 + }, + { + "epoch": 0.9105614999317858, + "grad_norm": 13.04912281036377, + "learning_rate": 8.790462582618761e-07, + "loss": 0.814, + "step": 23360 + }, + { + "epoch": 0.9109512950944278, + "grad_norm": 12.962596893310547, + "learning_rate": 8.788983249388833e-07, + "loss": 0.8783, + "step": 23370 + }, + { + "epoch": 0.9113410902570699, + "grad_norm": 13.146615028381348, + "learning_rate": 8.787503136697267e-07, + "loss": 0.9085, + "step": 23380 + }, + { + "epoch": 0.9117308854197119, + "grad_norm": 11.63349437713623, + "learning_rate": 8.786022244848547e-07, + "loss": 0.911, + "step": 23390 + }, + { + "epoch": 0.912120680582354, + "grad_norm": 14.532453536987305, + "learning_rate": 8.784540574147324e-07, + "loss": 0.9551, + "step": 23400 + }, + { + "epoch": 0.912510475744996, + "grad_norm": 12.701743125915527, + "learning_rate": 8.783058124898396e-07, + "loss": 0.9114, + "step": 23410 + }, + { + "epoch": 0.912900270907638, + "grad_norm": 16.137697219848633, + "learning_rate": 8.781574897406734e-07, + "loss": 0.83, + "step": 23420 + }, + { + "epoch": 0.9132900660702801, + "grad_norm": 13.568126678466797, + "learning_rate": 8.780090891977466e-07, + "loss": 0.8851, + "step": 23430 + }, + { + "epoch": 0.9136798612329221, + "grad_norm": 14.178144454956055, + "learning_rate": 8.778606108915875e-07, + "loss": 0.8664, + "step": 23440 + }, + { + "epoch": 0.9140696563955641, + "grad_norm": 11.996467590332031, + "learning_rate": 8.777120548527408e-07, + "loss": 0.8289, + "step": 23450 + }, + { + "epoch": 0.9144594515582062, + "grad_norm": 14.306840896606445, + "learning_rate": 8.775634211117673e-07, + "loss": 0.9326, + "step": 23460 + }, + { + "epoch": 0.9148492467208482, + "grad_norm": 13.991084098815918, + "learning_rate": 8.774147096992435e-07, + "loss": 0.8765, + "step": 23470 + }, + { + "epoch": 0.9152390418834903, + "grad_norm": 12.609655380249023, + "learning_rate": 8.772659206457622e-07, + "loss": 0.9144, + "step": 23480 + }, + { + "epoch": 0.9156288370461323, + "grad_norm": 12.509888648986816, + "learning_rate": 8.771170539819317e-07, + "loss": 0.8882, + "step": 23490 + }, + { + "epoch": 0.9160186322087743, + "grad_norm": 12.97066879272461, + "learning_rate": 8.769681097383766e-07, + "loss": 0.9306, + "step": 23500 + }, + { + "epoch": 0.9160186322087743, + "eval_loss": 0.9026774168014526, + "eval_runtime": 84.8414, + "eval_samples_per_second": 48.879, + "eval_steps_per_second": 6.117, + "step": 23500 + }, + { + "epoch": 0.9164084273714164, + "grad_norm": 13.108624458312988, + "learning_rate": 8.768190879457375e-07, + "loss": 0.8968, + "step": 23510 + }, + { + "epoch": 0.9167982225340584, + "grad_norm": 13.553518295288086, + "learning_rate": 8.766699886346707e-07, + "loss": 0.9577, + "step": 23520 + }, + { + "epoch": 0.9171880176967003, + "grad_norm": 13.645200729370117, + "learning_rate": 8.765208118358487e-07, + "loss": 0.8312, + "step": 23530 + }, + { + "epoch": 0.9175778128593424, + "grad_norm": 13.920916557312012, + "learning_rate": 8.763715575799599e-07, + "loss": 0.8899, + "step": 23540 + }, + { + "epoch": 0.9179676080219844, + "grad_norm": 13.411571502685547, + "learning_rate": 8.762222258977084e-07, + "loss": 0.9868, + "step": 23550 + }, + { + "epoch": 0.9183574031846264, + "grad_norm": 12.825125694274902, + "learning_rate": 8.760728168198146e-07, + "loss": 0.9237, + "step": 23560 + }, + { + "epoch": 0.9187471983472685, + "grad_norm": 11.639894485473633, + "learning_rate": 8.759233303770143e-07, + "loss": 0.8412, + "step": 23570 + }, + { + "epoch": 0.9191369935099105, + "grad_norm": 14.388006210327148, + "learning_rate": 8.7577376660006e-07, + "loss": 0.9246, + "step": 23580 + }, + { + "epoch": 0.9195267886725526, + "grad_norm": 12.244664192199707, + "learning_rate": 8.756241255197192e-07, + "loss": 0.9541, + "step": 23590 + }, + { + "epoch": 0.9199165838351946, + "grad_norm": 13.692625999450684, + "learning_rate": 8.75474407166776e-07, + "loss": 0.9534, + "step": 23600 + }, + { + "epoch": 0.9203063789978366, + "grad_norm": 11.20402717590332, + "learning_rate": 8.753246115720301e-07, + "loss": 0.8607, + "step": 23610 + }, + { + "epoch": 0.9206961741604787, + "grad_norm": 12.736128807067871, + "learning_rate": 8.751747387662971e-07, + "loss": 0.9294, + "step": 23620 + }, + { + "epoch": 0.9210859693231207, + "grad_norm": 15.177579879760742, + "learning_rate": 8.750247887804084e-07, + "loss": 0.8437, + "step": 23630 + }, + { + "epoch": 0.9214757644857627, + "grad_norm": 12.69448471069336, + "learning_rate": 8.748747616452117e-07, + "loss": 0.8679, + "step": 23640 + }, + { + "epoch": 0.9218655596484048, + "grad_norm": 13.18442153930664, + "learning_rate": 8.747246573915699e-07, + "loss": 0.8943, + "step": 23650 + }, + { + "epoch": 0.9222553548110468, + "grad_norm": 12.31995677947998, + "learning_rate": 8.745744760503626e-07, + "loss": 0.8876, + "step": 23660 + }, + { + "epoch": 0.9226451499736888, + "grad_norm": 13.388740539550781, + "learning_rate": 8.744242176524843e-07, + "loss": 0.8856, + "step": 23670 + }, + { + "epoch": 0.9230349451363309, + "grad_norm": 10.530184745788574, + "learning_rate": 8.74273882228846e-07, + "loss": 0.9026, + "step": 23680 + }, + { + "epoch": 0.9234247402989729, + "grad_norm": 12.49325180053711, + "learning_rate": 8.741234698103746e-07, + "loss": 0.8442, + "step": 23690 + }, + { + "epoch": 0.923814535461615, + "grad_norm": 13.253352165222168, + "learning_rate": 8.739729804280124e-07, + "loss": 0.8851, + "step": 23700 + }, + { + "epoch": 0.924204330624257, + "grad_norm": 12.856823921203613, + "learning_rate": 8.738224141127177e-07, + "loss": 0.8386, + "step": 23710 + }, + { + "epoch": 0.924594125786899, + "grad_norm": 12.306761741638184, + "learning_rate": 8.73671770895465e-07, + "loss": 0.8074, + "step": 23720 + }, + { + "epoch": 0.924983920949541, + "grad_norm": 12.310993194580078, + "learning_rate": 8.735210508072439e-07, + "loss": 0.8598, + "step": 23730 + }, + { + "epoch": 0.925373716112183, + "grad_norm": 11.304727554321289, + "learning_rate": 8.733702538790606e-07, + "loss": 0.8786, + "step": 23740 + }, + { + "epoch": 0.925763511274825, + "grad_norm": 11.426965713500977, + "learning_rate": 8.732193801419365e-07, + "loss": 0.835, + "step": 23750 + }, + { + "epoch": 0.9261533064374671, + "grad_norm": 14.6702241897583, + "learning_rate": 8.730684296269091e-07, + "loss": 0.8711, + "step": 23760 + }, + { + "epoch": 0.9265431016001091, + "grad_norm": 13.58132266998291, + "learning_rate": 8.729174023650316e-07, + "loss": 0.9256, + "step": 23770 + }, + { + "epoch": 0.9269328967627511, + "grad_norm": 12.344805717468262, + "learning_rate": 8.727662983873729e-07, + "loss": 0.8683, + "step": 23780 + }, + { + "epoch": 0.9273226919253932, + "grad_norm": 13.053298950195312, + "learning_rate": 8.726151177250181e-07, + "loss": 0.9204, + "step": 23790 + }, + { + "epoch": 0.9277124870880352, + "grad_norm": 11.30885124206543, + "learning_rate": 8.724638604090674e-07, + "loss": 0.867, + "step": 23800 + }, + { + "epoch": 0.9281022822506773, + "grad_norm": 13.741878509521484, + "learning_rate": 8.723125264706372e-07, + "loss": 0.8872, + "step": 23810 + }, + { + "epoch": 0.9284920774133193, + "grad_norm": 11.18321704864502, + "learning_rate": 8.721611159408599e-07, + "loss": 0.8895, + "step": 23820 + }, + { + "epoch": 0.9288818725759613, + "grad_norm": 12.678994178771973, + "learning_rate": 8.720096288508831e-07, + "loss": 0.8595, + "step": 23830 + }, + { + "epoch": 0.9292716677386034, + "grad_norm": 12.444290161132812, + "learning_rate": 8.718580652318706e-07, + "loss": 0.9212, + "step": 23840 + }, + { + "epoch": 0.9296614629012454, + "grad_norm": 11.66925048828125, + "learning_rate": 8.717064251150014e-07, + "loss": 0.8961, + "step": 23850 + }, + { + "epoch": 0.9300512580638874, + "grad_norm": 15.284521102905273, + "learning_rate": 8.715547085314708e-07, + "loss": 0.8938, + "step": 23860 + }, + { + "epoch": 0.9304410532265295, + "grad_norm": 12.77971363067627, + "learning_rate": 8.714029155124898e-07, + "loss": 0.8869, + "step": 23870 + }, + { + "epoch": 0.9308308483891715, + "grad_norm": 12.705084800720215, + "learning_rate": 8.712510460892846e-07, + "loss": 0.886, + "step": 23880 + }, + { + "epoch": 0.9312206435518136, + "grad_norm": 14.007545471191406, + "learning_rate": 8.710991002930977e-07, + "loss": 0.8346, + "step": 23890 + }, + { + "epoch": 0.9316104387144556, + "grad_norm": 13.194957733154297, + "learning_rate": 8.709470781551869e-07, + "loss": 0.9218, + "step": 23900 + }, + { + "epoch": 0.9320002338770976, + "grad_norm": 12.742473602294922, + "learning_rate": 8.707949797068259e-07, + "loss": 0.9341, + "step": 23910 + }, + { + "epoch": 0.9323900290397397, + "grad_norm": 12.699479103088379, + "learning_rate": 8.706428049793042e-07, + "loss": 0.8705, + "step": 23920 + }, + { + "epoch": 0.9327798242023817, + "grad_norm": 12.078168869018555, + "learning_rate": 8.704905540039266e-07, + "loss": 0.9053, + "step": 23930 + }, + { + "epoch": 0.9331696193650236, + "grad_norm": 13.103363037109375, + "learning_rate": 8.703382268120143e-07, + "loss": 0.9416, + "step": 23940 + }, + { + "epoch": 0.9335594145276657, + "grad_norm": 13.202085494995117, + "learning_rate": 8.701858234349033e-07, + "loss": 0.9123, + "step": 23950 + }, + { + "epoch": 0.9339492096903077, + "grad_norm": 13.232192993164062, + "learning_rate": 8.70033343903946e-07, + "loss": 0.9468, + "step": 23960 + }, + { + "epoch": 0.9343390048529497, + "grad_norm": 13.784865379333496, + "learning_rate": 8.698807882505099e-07, + "loss": 0.8361, + "step": 23970 + }, + { + "epoch": 0.9347288000155918, + "grad_norm": 11.770886421203613, + "learning_rate": 8.697281565059786e-07, + "loss": 0.9303, + "step": 23980 + }, + { + "epoch": 0.9351185951782338, + "grad_norm": 12.482429504394531, + "learning_rate": 8.695754487017511e-07, + "loss": 0.8764, + "step": 23990 + }, + { + "epoch": 0.9355083903408759, + "grad_norm": 13.466660499572754, + "learning_rate": 8.694226648692422e-07, + "loss": 0.8799, + "step": 24000 + }, + { + "epoch": 0.9355083903408759, + "eval_loss": 0.9010300636291504, + "eval_runtime": 82.8818, + "eval_samples_per_second": 50.035, + "eval_steps_per_second": 6.262, + "step": 24000 + }, + { + "epoch": 0.9358981855035179, + "grad_norm": 11.523694038391113, + "learning_rate": 8.692698050398824e-07, + "loss": 0.8735, + "step": 24010 + }, + { + "epoch": 0.9362879806661599, + "grad_norm": 10.836953163146973, + "learning_rate": 8.691168692451173e-07, + "loss": 0.8102, + "step": 24020 + }, + { + "epoch": 0.936677775828802, + "grad_norm": 13.500554084777832, + "learning_rate": 8.689638575164089e-07, + "loss": 0.9136, + "step": 24030 + }, + { + "epoch": 0.937067570991444, + "grad_norm": 16.357952117919922, + "learning_rate": 8.688107698852342e-07, + "loss": 0.881, + "step": 24040 + }, + { + "epoch": 0.937457366154086, + "grad_norm": 11.878355026245117, + "learning_rate": 8.686576063830862e-07, + "loss": 0.8814, + "step": 24050 + }, + { + "epoch": 0.9378471613167281, + "grad_norm": 12.411444664001465, + "learning_rate": 8.685043670414734e-07, + "loss": 0.8975, + "step": 24060 + }, + { + "epoch": 0.9382369564793701, + "grad_norm": 14.331562042236328, + "learning_rate": 8.683510518919197e-07, + "loss": 0.8711, + "step": 24070 + }, + { + "epoch": 0.9386267516420121, + "grad_norm": 13.71038818359375, + "learning_rate": 8.681976609659651e-07, + "loss": 0.8991, + "step": 24080 + }, + { + "epoch": 0.9390165468046542, + "grad_norm": 12.733404159545898, + "learning_rate": 8.680441942951645e-07, + "loss": 0.8841, + "step": 24090 + }, + { + "epoch": 0.9394063419672962, + "grad_norm": 12.526485443115234, + "learning_rate": 8.678906519110889e-07, + "loss": 0.9551, + "step": 24100 + }, + { + "epoch": 0.9397961371299383, + "grad_norm": 14.007279396057129, + "learning_rate": 8.677370338453248e-07, + "loss": 0.9454, + "step": 24110 + }, + { + "epoch": 0.9401859322925803, + "grad_norm": 15.950329780578613, + "learning_rate": 8.67583340129474e-07, + "loss": 0.7943, + "step": 24120 + }, + { + "epoch": 0.9405757274552223, + "grad_norm": 11.185551643371582, + "learning_rate": 8.674295707951543e-07, + "loss": 0.907, + "step": 24130 + }, + { + "epoch": 0.9409655226178643, + "grad_norm": 15.121079444885254, + "learning_rate": 8.672757258739984e-07, + "loss": 1.0362, + "step": 24140 + }, + { + "epoch": 0.9413553177805063, + "grad_norm": 11.652565002441406, + "learning_rate": 8.671218053976552e-07, + "loss": 0.8383, + "step": 24150 + }, + { + "epoch": 0.9417451129431483, + "grad_norm": 12.16205883026123, + "learning_rate": 8.66967809397789e-07, + "loss": 0.9111, + "step": 24160 + }, + { + "epoch": 0.9421349081057904, + "grad_norm": 13.082060813903809, + "learning_rate": 8.668137379060795e-07, + "loss": 0.9253, + "step": 24170 + }, + { + "epoch": 0.9425247032684324, + "grad_norm": 13.473897933959961, + "learning_rate": 8.666595909542218e-07, + "loss": 0.9292, + "step": 24180 + }, + { + "epoch": 0.9429144984310744, + "grad_norm": 13.610812187194824, + "learning_rate": 8.665053685739267e-07, + "loss": 0.9218, + "step": 24190 + }, + { + "epoch": 0.9433042935937165, + "grad_norm": 11.609959602355957, + "learning_rate": 8.663510707969206e-07, + "loss": 0.8476, + "step": 24200 + }, + { + "epoch": 0.9436940887563585, + "grad_norm": 10.758955001831055, + "learning_rate": 8.661966976549453e-07, + "loss": 0.8707, + "step": 24210 + }, + { + "epoch": 0.9440838839190006, + "grad_norm": 13.520397186279297, + "learning_rate": 8.660422491797581e-07, + "loss": 0.8596, + "step": 24220 + }, + { + "epoch": 0.9444736790816426, + "grad_norm": 12.471734046936035, + "learning_rate": 8.658877254031319e-07, + "loss": 0.8202, + "step": 24230 + }, + { + "epoch": 0.9448634742442846, + "grad_norm": 14.344609260559082, + "learning_rate": 8.657331263568548e-07, + "loss": 0.9432, + "step": 24240 + }, + { + "epoch": 0.9452532694069267, + "grad_norm": 14.137019157409668, + "learning_rate": 8.655784520727307e-07, + "loss": 0.9175, + "step": 24250 + }, + { + "epoch": 0.9456430645695687, + "grad_norm": 13.851522445678711, + "learning_rate": 8.65423702582579e-07, + "loss": 0.9366, + "step": 24260 + }, + { + "epoch": 0.9460328597322107, + "grad_norm": 12.297534942626953, + "learning_rate": 8.652688779182342e-07, + "loss": 0.8837, + "step": 24270 + }, + { + "epoch": 0.9464226548948528, + "grad_norm": 12.714227676391602, + "learning_rate": 8.651139781115467e-07, + "loss": 0.92, + "step": 24280 + }, + { + "epoch": 0.9468124500574948, + "grad_norm": 14.571334838867188, + "learning_rate": 8.649590031943821e-07, + "loss": 0.8366, + "step": 24290 + }, + { + "epoch": 0.9472022452201369, + "grad_norm": 11.872702598571777, + "learning_rate": 8.648039531986214e-07, + "loss": 0.8892, + "step": 24300 + }, + { + "epoch": 0.9475920403827789, + "grad_norm": 14.444818496704102, + "learning_rate": 8.646488281561613e-07, + "loss": 0.9393, + "step": 24310 + }, + { + "epoch": 0.9479818355454209, + "grad_norm": 13.260595321655273, + "learning_rate": 8.644936280989137e-07, + "loss": 0.9027, + "step": 24320 + }, + { + "epoch": 0.948371630708063, + "grad_norm": 15.092535018920898, + "learning_rate": 8.643383530588062e-07, + "loss": 0.8826, + "step": 24330 + }, + { + "epoch": 0.948761425870705, + "grad_norm": 12.755534172058105, + "learning_rate": 8.641830030677814e-07, + "loss": 0.8913, + "step": 24340 + }, + { + "epoch": 0.9491512210333469, + "grad_norm": 12.230353355407715, + "learning_rate": 8.640275781577977e-07, + "loss": 0.9396, + "step": 24350 + }, + { + "epoch": 0.949541016195989, + "grad_norm": 15.140388488769531, + "learning_rate": 8.638720783608289e-07, + "loss": 0.8676, + "step": 24360 + }, + { + "epoch": 0.949930811358631, + "grad_norm": 11.981585502624512, + "learning_rate": 8.637165037088638e-07, + "loss": 0.9287, + "step": 24370 + }, + { + "epoch": 0.950320606521273, + "grad_norm": 12.562039375305176, + "learning_rate": 8.635608542339073e-07, + "loss": 0.9154, + "step": 24380 + }, + { + "epoch": 0.9507104016839151, + "grad_norm": 11.155308723449707, + "learning_rate": 8.634051299679788e-07, + "loss": 0.8628, + "step": 24390 + }, + { + "epoch": 0.9511001968465571, + "grad_norm": 12.650090217590332, + "learning_rate": 8.63249330943114e-07, + "loss": 0.824, + "step": 24400 + }, + { + "epoch": 0.9514899920091991, + "grad_norm": 11.540812492370605, + "learning_rate": 8.630934571913633e-07, + "loss": 0.884, + "step": 24410 + }, + { + "epoch": 0.9518797871718412, + "grad_norm": 14.70451545715332, + "learning_rate": 8.629375087447927e-07, + "loss": 0.9445, + "step": 24420 + }, + { + "epoch": 0.9522695823344832, + "grad_norm": 10.911174774169922, + "learning_rate": 8.627814856354838e-07, + "loss": 0.8865, + "step": 24430 + }, + { + "epoch": 0.9526593774971253, + "grad_norm": 13.402850151062012, + "learning_rate": 8.626253878955332e-07, + "loss": 0.8582, + "step": 24440 + }, + { + "epoch": 0.9530491726597673, + "grad_norm": 14.46366024017334, + "learning_rate": 8.624692155570528e-07, + "loss": 0.9073, + "step": 24450 + }, + { + "epoch": 0.9534389678224093, + "grad_norm": 13.013025283813477, + "learning_rate": 8.623129686521705e-07, + "loss": 0.9051, + "step": 24460 + }, + { + "epoch": 0.9538287629850514, + "grad_norm": 13.172517776489258, + "learning_rate": 8.621566472130287e-07, + "loss": 0.9382, + "step": 24470 + }, + { + "epoch": 0.9542185581476934, + "grad_norm": 13.994763374328613, + "learning_rate": 8.620002512717857e-07, + "loss": 0.8596, + "step": 24480 + }, + { + "epoch": 0.9546083533103354, + "grad_norm": 12.222554206848145, + "learning_rate": 8.61843780860615e-07, + "loss": 0.8752, + "step": 24490 + }, + { + "epoch": 0.9549981484729775, + "grad_norm": 11.988141059875488, + "learning_rate": 8.616872360117052e-07, + "loss": 0.9265, + "step": 24500 + }, + { + "epoch": 0.9549981484729775, + "eval_loss": 0.9000971913337708, + "eval_runtime": 82.8686, + "eval_samples_per_second": 50.043, + "eval_steps_per_second": 6.263, + "step": 24500 + }, + { + "epoch": 0.9553879436356195, + "grad_norm": 15.157137870788574, + "learning_rate": 8.615306167572603e-07, + "loss": 0.8667, + "step": 24510 + }, + { + "epoch": 0.9557777387982616, + "grad_norm": 11.83220100402832, + "learning_rate": 8.613739231294999e-07, + "loss": 0.9404, + "step": 24520 + }, + { + "epoch": 0.9561675339609036, + "grad_norm": 12.993494033813477, + "learning_rate": 8.612171551606586e-07, + "loss": 0.8404, + "step": 24530 + }, + { + "epoch": 0.9565573291235456, + "grad_norm": 12.881315231323242, + "learning_rate": 8.610603128829863e-07, + "loss": 0.8959, + "step": 24540 + }, + { + "epoch": 0.9569471242861876, + "grad_norm": 12.451800346374512, + "learning_rate": 8.609033963287484e-07, + "loss": 0.935, + "step": 24550 + }, + { + "epoch": 0.9573369194488296, + "grad_norm": 13.672378540039062, + "learning_rate": 8.607464055302253e-07, + "loss": 0.9024, + "step": 24560 + }, + { + "epoch": 0.9577267146114716, + "grad_norm": 13.266074180603027, + "learning_rate": 8.605893405197131e-07, + "loss": 0.9002, + "step": 24570 + }, + { + "epoch": 0.9581165097741137, + "grad_norm": 15.62568187713623, + "learning_rate": 8.604322013295227e-07, + "loss": 0.8407, + "step": 24580 + }, + { + "epoch": 0.9585063049367557, + "grad_norm": 14.83586597442627, + "learning_rate": 8.602749879919801e-07, + "loss": 0.8958, + "step": 24590 + }, + { + "epoch": 0.9588961000993977, + "grad_norm": 13.40401840209961, + "learning_rate": 8.601177005394274e-07, + "loss": 0.8567, + "step": 24600 + }, + { + "epoch": 0.9592858952620398, + "grad_norm": 13.62450122833252, + "learning_rate": 8.599603390042212e-07, + "loss": 0.8769, + "step": 24610 + }, + { + "epoch": 0.9596756904246818, + "grad_norm": 15.232893943786621, + "learning_rate": 8.598029034187338e-07, + "loss": 0.8729, + "step": 24620 + }, + { + "epoch": 0.9600654855873239, + "grad_norm": 11.46220588684082, + "learning_rate": 8.596453938153522e-07, + "loss": 0.8417, + "step": 24630 + }, + { + "epoch": 0.9604552807499659, + "grad_norm": 12.134919166564941, + "learning_rate": 8.594878102264792e-07, + "loss": 0.8571, + "step": 24640 + }, + { + "epoch": 0.9608450759126079, + "grad_norm": 14.451122283935547, + "learning_rate": 8.593301526845325e-07, + "loss": 0.9296, + "step": 24650 + }, + { + "epoch": 0.96123487107525, + "grad_norm": 13.465937614440918, + "learning_rate": 8.591724212219448e-07, + "loss": 0.8861, + "step": 24660 + }, + { + "epoch": 0.961624666237892, + "grad_norm": 12.748940467834473, + "learning_rate": 8.590146158711648e-07, + "loss": 0.891, + "step": 24670 + }, + { + "epoch": 0.962014461400534, + "grad_norm": 13.085394859313965, + "learning_rate": 8.588567366646556e-07, + "loss": 0.8951, + "step": 24680 + }, + { + "epoch": 0.9624042565631761, + "grad_norm": 13.225638389587402, + "learning_rate": 8.586987836348955e-07, + "loss": 0.8483, + "step": 24690 + }, + { + "epoch": 0.9627940517258181, + "grad_norm": 11.684745788574219, + "learning_rate": 8.585407568143787e-07, + "loss": 0.8243, + "step": 24700 + }, + { + "epoch": 0.9631838468884601, + "grad_norm": 13.252849578857422, + "learning_rate": 8.583826562356143e-07, + "loss": 0.8947, + "step": 24710 + }, + { + "epoch": 0.9635736420511022, + "grad_norm": 16.545886993408203, + "learning_rate": 8.582244819311257e-07, + "loss": 0.91, + "step": 24720 + }, + { + "epoch": 0.9639634372137442, + "grad_norm": 13.425329208374023, + "learning_rate": 8.580662339334528e-07, + "loss": 0.8402, + "step": 24730 + }, + { + "epoch": 0.9643532323763863, + "grad_norm": 10.943960189819336, + "learning_rate": 8.579079122751498e-07, + "loss": 0.8049, + "step": 24740 + }, + { + "epoch": 0.9647430275390282, + "grad_norm": 12.36875057220459, + "learning_rate": 8.577495169887862e-07, + "loss": 0.8613, + "step": 24750 + }, + { + "epoch": 0.9651328227016702, + "grad_norm": 14.046988487243652, + "learning_rate": 8.57591048106947e-07, + "loss": 0.8891, + "step": 24760 + }, + { + "epoch": 0.9655226178643123, + "grad_norm": 14.464166641235352, + "learning_rate": 8.574325056622321e-07, + "loss": 0.8832, + "step": 24770 + }, + { + "epoch": 0.9659124130269543, + "grad_norm": 12.866965293884277, + "learning_rate": 8.572738896872562e-07, + "loss": 0.8357, + "step": 24780 + }, + { + "epoch": 0.9663022081895963, + "grad_norm": 14.262253761291504, + "learning_rate": 8.571152002146497e-07, + "loss": 0.8628, + "step": 24790 + }, + { + "epoch": 0.9666920033522384, + "grad_norm": 12.69200611114502, + "learning_rate": 8.569564372770577e-07, + "loss": 0.8116, + "step": 24800 + }, + { + "epoch": 0.9670817985148804, + "grad_norm": 12.65813159942627, + "learning_rate": 8.567976009071407e-07, + "loss": 0.8669, + "step": 24810 + }, + { + "epoch": 0.9674715936775224, + "grad_norm": 15.225479125976562, + "learning_rate": 8.566386911375743e-07, + "loss": 0.8853, + "step": 24820 + }, + { + "epoch": 0.9678613888401645, + "grad_norm": 14.536882400512695, + "learning_rate": 8.564797080010489e-07, + "loss": 0.8966, + "step": 24830 + }, + { + "epoch": 0.9682511840028065, + "grad_norm": 13.043333053588867, + "learning_rate": 8.5632065153027e-07, + "loss": 0.9147, + "step": 24840 + }, + { + "epoch": 0.9686409791654486, + "grad_norm": 13.560032844543457, + "learning_rate": 8.561615217579588e-07, + "loss": 0.8453, + "step": 24850 + }, + { + "epoch": 0.9690307743280906, + "grad_norm": 14.036949157714844, + "learning_rate": 8.560023187168507e-07, + "loss": 0.9316, + "step": 24860 + }, + { + "epoch": 0.9694205694907326, + "grad_norm": 12.701640129089355, + "learning_rate": 8.558430424396969e-07, + "loss": 0.8419, + "step": 24870 + }, + { + "epoch": 0.9698103646533747, + "grad_norm": 12.011168479919434, + "learning_rate": 8.556836929592634e-07, + "loss": 0.9841, + "step": 24880 + }, + { + "epoch": 0.9702001598160167, + "grad_norm": 12.033271789550781, + "learning_rate": 8.55524270308331e-07, + "loss": 0.8062, + "step": 24890 + }, + { + "epoch": 0.9705899549786587, + "grad_norm": 13.15074634552002, + "learning_rate": 8.55364774519696e-07, + "loss": 0.8939, + "step": 24900 + }, + { + "epoch": 0.9709797501413008, + "grad_norm": 12.342111587524414, + "learning_rate": 8.552052056261694e-07, + "loss": 0.8978, + "step": 24910 + }, + { + "epoch": 0.9713695453039428, + "grad_norm": 16.423892974853516, + "learning_rate": 8.550455636605776e-07, + "loss": 0.8724, + "step": 24920 + }, + { + "epoch": 0.9717593404665849, + "grad_norm": 12.308789253234863, + "learning_rate": 8.548858486557616e-07, + "loss": 0.9605, + "step": 24930 + }, + { + "epoch": 0.9721491356292269, + "grad_norm": 13.975433349609375, + "learning_rate": 8.547260606445776e-07, + "loss": 0.9294, + "step": 24940 + }, + { + "epoch": 0.9725389307918689, + "grad_norm": 12.259604454040527, + "learning_rate": 8.54566199659897e-07, + "loss": 0.9092, + "step": 24950 + }, + { + "epoch": 0.9729287259545109, + "grad_norm": 11.402291297912598, + "learning_rate": 8.54406265734606e-07, + "loss": 0.9351, + "step": 24960 + }, + { + "epoch": 0.9733185211171529, + "grad_norm": 12.762960433959961, + "learning_rate": 8.54246258901606e-07, + "loss": 0.9425, + "step": 24970 + }, + { + "epoch": 0.9737083162797949, + "grad_norm": 12.814912796020508, + "learning_rate": 8.540861791938129e-07, + "loss": 0.9155, + "step": 24980 + }, + { + "epoch": 0.974098111442437, + "grad_norm": 11.953180313110352, + "learning_rate": 8.539260266441585e-07, + "loss": 0.8635, + "step": 24990 + }, + { + "epoch": 0.974487906605079, + "grad_norm": 11.66627025604248, + "learning_rate": 8.537658012855885e-07, + "loss": 0.8305, + "step": 25000 + }, + { + "epoch": 0.974487906605079, + "eval_loss": 0.8995266556739807, + "eval_runtime": 82.4712, + "eval_samples_per_second": 50.284, + "eval_steps_per_second": 6.293, + "step": 25000 + }, + { + "epoch": 0.974877701767721, + "grad_norm": 13.037186622619629, + "learning_rate": 8.536055031510645e-07, + "loss": 0.8578, + "step": 25010 + }, + { + "epoch": 0.9752674969303631, + "grad_norm": 13.866379737854004, + "learning_rate": 8.534451322735624e-07, + "loss": 0.9588, + "step": 25020 + }, + { + "epoch": 0.9756572920930051, + "grad_norm": 12.873283386230469, + "learning_rate": 8.532846886860736e-07, + "loss": 0.9742, + "step": 25030 + }, + { + "epoch": 0.9760470872556471, + "grad_norm": 10.88054370880127, + "learning_rate": 8.531241724216042e-07, + "loss": 0.8815, + "step": 25040 + }, + { + "epoch": 0.9764368824182892, + "grad_norm": 12.10800838470459, + "learning_rate": 8.529635835131752e-07, + "loss": 0.9378, + "step": 25050 + }, + { + "epoch": 0.9768266775809312, + "grad_norm": 12.612936973571777, + "learning_rate": 8.528029219938226e-07, + "loss": 0.8707, + "step": 25060 + }, + { + "epoch": 0.9772164727435733, + "grad_norm": 11.916803359985352, + "learning_rate": 8.526421878965974e-07, + "loss": 0.9018, + "step": 25070 + }, + { + "epoch": 0.9776062679062153, + "grad_norm": 12.767949104309082, + "learning_rate": 8.524813812545655e-07, + "loss": 0.8921, + "step": 25080 + }, + { + "epoch": 0.9779960630688573, + "grad_norm": 12.895742416381836, + "learning_rate": 8.523205021008075e-07, + "loss": 0.9213, + "step": 25090 + }, + { + "epoch": 0.9783858582314994, + "grad_norm": 13.18868637084961, + "learning_rate": 8.521595504684194e-07, + "loss": 0.8423, + "step": 25100 + }, + { + "epoch": 0.9787756533941414, + "grad_norm": 13.184746742248535, + "learning_rate": 8.519985263905116e-07, + "loss": 0.8537, + "step": 25110 + }, + { + "epoch": 0.9791654485567834, + "grad_norm": 12.919042587280273, + "learning_rate": 8.518374299002098e-07, + "loss": 0.8674, + "step": 25120 + }, + { + "epoch": 0.9795552437194255, + "grad_norm": 11.637657165527344, + "learning_rate": 8.516762610306542e-07, + "loss": 0.8519, + "step": 25130 + }, + { + "epoch": 0.9799450388820675, + "grad_norm": 14.409613609313965, + "learning_rate": 8.515150198150004e-07, + "loss": 0.929, + "step": 25140 + }, + { + "epoch": 0.9803348340447096, + "grad_norm": 13.78165340423584, + "learning_rate": 8.513537062864185e-07, + "loss": 0.7949, + "step": 25150 + }, + { + "epoch": 0.9807246292073515, + "grad_norm": 13.699678421020508, + "learning_rate": 8.511923204780935e-07, + "loss": 0.833, + "step": 25160 + }, + { + "epoch": 0.9811144243699935, + "grad_norm": 11.948232650756836, + "learning_rate": 8.510308624232255e-07, + "loss": 1.0114, + "step": 25170 + }, + { + "epoch": 0.9815042195326356, + "grad_norm": 12.935140609741211, + "learning_rate": 8.508693321550292e-07, + "loss": 0.8209, + "step": 25180 + }, + { + "epoch": 0.9818940146952776, + "grad_norm": 14.824143409729004, + "learning_rate": 8.507077297067342e-07, + "loss": 0.8289, + "step": 25190 + }, + { + "epoch": 0.9822838098579196, + "grad_norm": 12.251105308532715, + "learning_rate": 8.505460551115853e-07, + "loss": 0.9013, + "step": 25200 + }, + { + "epoch": 0.9826736050205617, + "grad_norm": 14.520829200744629, + "learning_rate": 8.503843084028414e-07, + "loss": 0.9177, + "step": 25210 + }, + { + "epoch": 0.9830634001832037, + "grad_norm": 10.954856872558594, + "learning_rate": 8.502224896137771e-07, + "loss": 0.8934, + "step": 25220 + }, + { + "epoch": 0.9834531953458457, + "grad_norm": 10.250443458557129, + "learning_rate": 8.500605987776813e-07, + "loss": 0.8934, + "step": 25230 + }, + { + "epoch": 0.9838429905084878, + "grad_norm": 11.334168434143066, + "learning_rate": 8.498986359278577e-07, + "loss": 0.8619, + "step": 25240 + }, + { + "epoch": 0.9842327856711298, + "grad_norm": 13.912988662719727, + "learning_rate": 8.497366010976252e-07, + "loss": 0.9207, + "step": 25250 + }, + { + "epoch": 0.9846225808337719, + "grad_norm": 13.912152290344238, + "learning_rate": 8.49574494320317e-07, + "loss": 0.84, + "step": 25260 + }, + { + "epoch": 0.9850123759964139, + "grad_norm": 13.134868621826172, + "learning_rate": 8.494123156292816e-07, + "loss": 0.8869, + "step": 25270 + }, + { + "epoch": 0.9854021711590559, + "grad_norm": 12.96687126159668, + "learning_rate": 8.492500650578819e-07, + "loss": 0.8398, + "step": 25280 + }, + { + "epoch": 0.985791966321698, + "grad_norm": 12.796394348144531, + "learning_rate": 8.490877426394957e-07, + "loss": 0.9695, + "step": 25290 + }, + { + "epoch": 0.98618176148434, + "grad_norm": 15.445104598999023, + "learning_rate": 8.48925348407516e-07, + "loss": 0.9317, + "step": 25300 + }, + { + "epoch": 0.986571556646982, + "grad_norm": 10.339613914489746, + "learning_rate": 8.487628823953496e-07, + "loss": 0.8937, + "step": 25310 + }, + { + "epoch": 0.9869613518096241, + "grad_norm": 13.730612754821777, + "learning_rate": 8.486003446364192e-07, + "loss": 0.8379, + "step": 25320 + }, + { + "epoch": 0.9873511469722661, + "grad_norm": 12.501172065734863, + "learning_rate": 8.484377351641615e-07, + "loss": 0.8922, + "step": 25330 + }, + { + "epoch": 0.9877409421349082, + "grad_norm": 10.061121940612793, + "learning_rate": 8.48275054012028e-07, + "loss": 0.824, + "step": 25340 + }, + { + "epoch": 0.9881307372975502, + "grad_norm": 14.377124786376953, + "learning_rate": 8.481123012134857e-07, + "loss": 0.9416, + "step": 25350 + }, + { + "epoch": 0.9885205324601921, + "grad_norm": 12.18248462677002, + "learning_rate": 8.47949476802015e-07, + "loss": 0.9225, + "step": 25360 + }, + { + "epoch": 0.9889103276228342, + "grad_norm": 11.721858024597168, + "learning_rate": 8.477865808111123e-07, + "loss": 0.8848, + "step": 25370 + }, + { + "epoch": 0.9893001227854762, + "grad_norm": 15.619644165039062, + "learning_rate": 8.47623613274288e-07, + "loss": 0.8638, + "step": 25380 + }, + { + "epoch": 0.9896899179481182, + "grad_norm": 12.25162124633789, + "learning_rate": 8.474605742250676e-07, + "loss": 0.8609, + "step": 25390 + }, + { + "epoch": 0.9900797131107603, + "grad_norm": 15.34259033203125, + "learning_rate": 8.472974636969912e-07, + "loss": 0.9108, + "step": 25400 + }, + { + "epoch": 0.9904695082734023, + "grad_norm": 12.894454956054688, + "learning_rate": 8.471342817236133e-07, + "loss": 0.8283, + "step": 25410 + }, + { + "epoch": 0.9908593034360443, + "grad_norm": 12.529272079467773, + "learning_rate": 8.469710283385034e-07, + "loss": 0.9177, + "step": 25420 + }, + { + "epoch": 0.9912490985986864, + "grad_norm": 13.199865341186523, + "learning_rate": 8.468077035752459e-07, + "loss": 0.9092, + "step": 25430 + }, + { + "epoch": 0.9916388937613284, + "grad_norm": 15.86511516571045, + "learning_rate": 8.466443074674394e-07, + "loss": 0.9112, + "step": 25440 + }, + { + "epoch": 0.9920286889239704, + "grad_norm": 13.194429397583008, + "learning_rate": 8.464808400486973e-07, + "loss": 0.8577, + "step": 25450 + }, + { + "epoch": 0.9924184840866125, + "grad_norm": 11.913835525512695, + "learning_rate": 8.463173013526481e-07, + "loss": 0.914, + "step": 25460 + }, + { + "epoch": 0.9928082792492545, + "grad_norm": 13.460233688354492, + "learning_rate": 8.461536914129345e-07, + "loss": 0.8934, + "step": 25470 + }, + { + "epoch": 0.9931980744118966, + "grad_norm": 13.704668045043945, + "learning_rate": 8.459900102632138e-07, + "loss": 0.9128, + "step": 25480 + }, + { + "epoch": 0.9935878695745386, + "grad_norm": 14.611261367797852, + "learning_rate": 8.458262579371582e-07, + "loss": 0.8594, + "step": 25490 + }, + { + "epoch": 0.9939776647371806, + "grad_norm": 13.090282440185547, + "learning_rate": 8.456624344684549e-07, + "loss": 0.8714, + "step": 25500 + }, + { + "epoch": 0.9939776647371806, + "eval_loss": 0.8985117077827454, + "eval_runtime": 82.7138, + "eval_samples_per_second": 50.137, + "eval_steps_per_second": 6.275, + "step": 25500 + }, + { + "epoch": 0.9943674598998227, + "grad_norm": 14.951517105102539, + "learning_rate": 8.454985398908047e-07, + "loss": 0.9046, + "step": 25510 + }, + { + "epoch": 0.9947572550624647, + "grad_norm": 13.595707893371582, + "learning_rate": 8.453345742379241e-07, + "loss": 0.8253, + "step": 25520 + }, + { + "epoch": 0.9951470502251067, + "grad_norm": 13.271056175231934, + "learning_rate": 8.451705375435435e-07, + "loss": 0.9554, + "step": 25530 + }, + { + "epoch": 0.9955368453877488, + "grad_norm": 13.979809761047363, + "learning_rate": 8.450064298414083e-07, + "loss": 0.9284, + "step": 25540 + }, + { + "epoch": 0.9959266405503908, + "grad_norm": 13.42162799835205, + "learning_rate": 8.448422511652785e-07, + "loss": 0.8668, + "step": 25550 + }, + { + "epoch": 0.9963164357130329, + "grad_norm": 12.989556312561035, + "learning_rate": 8.446780015489283e-07, + "loss": 0.8743, + "step": 25560 + }, + { + "epoch": 0.9967062308756748, + "grad_norm": 12.966590881347656, + "learning_rate": 8.445136810261471e-07, + "loss": 0.8297, + "step": 25570 + }, + { + "epoch": 0.9970960260383168, + "grad_norm": 14.850770950317383, + "learning_rate": 8.443492896307384e-07, + "loss": 0.9835, + "step": 25580 + }, + { + "epoch": 0.9974858212009589, + "grad_norm": 11.702842712402344, + "learning_rate": 8.441848273965204e-07, + "loss": 0.9063, + "step": 25590 + }, + { + "epoch": 0.9978756163636009, + "grad_norm": 12.99045181274414, + "learning_rate": 8.440202943573261e-07, + "loss": 0.787, + "step": 25600 + }, + { + "epoch": 0.9982654115262429, + "grad_norm": 13.557534217834473, + "learning_rate": 8.438556905470026e-07, + "loss": 0.9082, + "step": 25610 + }, + { + "epoch": 0.998655206688885, + "grad_norm": 13.030200004577637, + "learning_rate": 8.436910159994124e-07, + "loss": 0.9301, + "step": 25620 + }, + { + "epoch": 0.999045001851527, + "grad_norm": 11.152798652648926, + "learning_rate": 8.435262707484313e-07, + "loss": 0.9523, + "step": 25630 + }, + { + "epoch": 0.999434797014169, + "grad_norm": 12.724126815795898, + "learning_rate": 8.433614548279509e-07, + "loss": 0.8576, + "step": 25640 + }, + { + "epoch": 0.9998245921768111, + "grad_norm": 13.679668426513672, + "learning_rate": 8.431965682718765e-07, + "loss": 0.8056, + "step": 25650 + }, + { + "epoch": 1.000194897581321, + "grad_norm": 12.230389595031738, + "learning_rate": 8.430316111141282e-07, + "loss": 0.7886, + "step": 25660 + }, + { + "epoch": 1.000584692743963, + "grad_norm": 12.312921524047852, + "learning_rate": 8.428665833886407e-07, + "loss": 0.8268, + "step": 25670 + }, + { + "epoch": 1.0009744879066051, + "grad_norm": 13.638102531433105, + "learning_rate": 8.427014851293632e-07, + "loss": 0.8921, + "step": 25680 + }, + { + "epoch": 1.001364283069247, + "grad_norm": 13.919229507446289, + "learning_rate": 8.425363163702595e-07, + "loss": 0.8024, + "step": 25690 + }, + { + "epoch": 1.0017540782318892, + "grad_norm": 10.034284591674805, + "learning_rate": 8.423710771453074e-07, + "loss": 0.7747, + "step": 25700 + }, + { + "epoch": 1.0021438733945311, + "grad_norm": 12.720212936401367, + "learning_rate": 8.422057674884999e-07, + "loss": 0.809, + "step": 25710 + }, + { + "epoch": 1.0025336685571733, + "grad_norm": 12.131061553955078, + "learning_rate": 8.42040387433844e-07, + "loss": 0.8253, + "step": 25720 + }, + { + "epoch": 1.0029234637198152, + "grad_norm": 11.912078857421875, + "learning_rate": 8.418749370153613e-07, + "loss": 0.8066, + "step": 25730 + }, + { + "epoch": 1.0033132588824574, + "grad_norm": 11.983098030090332, + "learning_rate": 8.417094162670881e-07, + "loss": 0.8386, + "step": 25740 + }, + { + "epoch": 1.0037030540450993, + "grad_norm": 13.018217086791992, + "learning_rate": 8.415438252230749e-07, + "loss": 0.7857, + "step": 25750 + }, + { + "epoch": 1.0040928492077412, + "grad_norm": 14.735390663146973, + "learning_rate": 8.413781639173865e-07, + "loss": 0.816, + "step": 25760 + }, + { + "epoch": 1.0044826443703834, + "grad_norm": 13.064722061157227, + "learning_rate": 8.412124323841027e-07, + "loss": 0.7942, + "step": 25770 + }, + { + "epoch": 1.0048724395330253, + "grad_norm": 14.078709602355957, + "learning_rate": 8.410466306573174e-07, + "loss": 0.8471, + "step": 25780 + }, + { + "epoch": 1.0052622346956674, + "grad_norm": 11.342695236206055, + "learning_rate": 8.408807587711388e-07, + "loss": 0.758, + "step": 25790 + }, + { + "epoch": 1.0056520298583094, + "grad_norm": 13.141053199768066, + "learning_rate": 8.407148167596899e-07, + "loss": 0.8138, + "step": 25800 + }, + { + "epoch": 1.0060418250209515, + "grad_norm": 13.689034461975098, + "learning_rate": 8.405488046571078e-07, + "loss": 0.7499, + "step": 25810 + }, + { + "epoch": 1.0064316201835934, + "grad_norm": 13.734099388122559, + "learning_rate": 8.403827224975442e-07, + "loss": 0.8105, + "step": 25820 + }, + { + "epoch": 1.0068214153462356, + "grad_norm": 14.513954162597656, + "learning_rate": 8.402165703151654e-07, + "loss": 0.8038, + "step": 25830 + }, + { + "epoch": 1.0072112105088775, + "grad_norm": 12.727503776550293, + "learning_rate": 8.400503481441515e-07, + "loss": 0.813, + "step": 25840 + }, + { + "epoch": 1.0076010056715197, + "grad_norm": 14.995710372924805, + "learning_rate": 8.398840560186975e-07, + "loss": 0.8475, + "step": 25850 + }, + { + "epoch": 1.0079908008341616, + "grad_norm": 11.71441650390625, + "learning_rate": 8.397176939730128e-07, + "loss": 0.773, + "step": 25860 + }, + { + "epoch": 1.0083805959968037, + "grad_norm": 11.987086296081543, + "learning_rate": 8.39551262041321e-07, + "loss": 0.7801, + "step": 25870 + }, + { + "epoch": 1.0087703911594457, + "grad_norm": 14.022822380065918, + "learning_rate": 8.393847602578599e-07, + "loss": 0.8064, + "step": 25880 + }, + { + "epoch": 1.0091601863220878, + "grad_norm": 13.233519554138184, + "learning_rate": 8.392181886568823e-07, + "loss": 0.8286, + "step": 25890 + }, + { + "epoch": 1.0095499814847297, + "grad_norm": 12.001953125, + "learning_rate": 8.390515472726547e-07, + "loss": 0.8299, + "step": 25900 + }, + { + "epoch": 1.009939776647372, + "grad_norm": 15.382742881774902, + "learning_rate": 8.38884836139458e-07, + "loss": 0.8152, + "step": 25910 + }, + { + "epoch": 1.0103295718100138, + "grad_norm": 10.791077613830566, + "learning_rate": 8.387180552915882e-07, + "loss": 0.7579, + "step": 25920 + }, + { + "epoch": 1.010719366972656, + "grad_norm": 13.976044654846191, + "learning_rate": 8.385512047633548e-07, + "loss": 0.81, + "step": 25930 + }, + { + "epoch": 1.011109162135298, + "grad_norm": 13.754496574401855, + "learning_rate": 8.383842845890819e-07, + "loss": 0.8023, + "step": 25940 + }, + { + "epoch": 1.01149895729794, + "grad_norm": 12.652153015136719, + "learning_rate": 8.382172948031081e-07, + "loss": 0.8615, + "step": 25950 + }, + { + "epoch": 1.011888752460582, + "grad_norm": 13.221951484680176, + "learning_rate": 8.38050235439786e-07, + "loss": 0.8211, + "step": 25960 + }, + { + "epoch": 1.012278547623224, + "grad_norm": 11.235804557800293, + "learning_rate": 8.378831065334829e-07, + "loss": 0.8233, + "step": 25970 + }, + { + "epoch": 1.012668342785866, + "grad_norm": 12.431456565856934, + "learning_rate": 8.377159081185801e-07, + "loss": 0.81, + "step": 25980 + }, + { + "epoch": 1.013058137948508, + "grad_norm": 11.836759567260742, + "learning_rate": 8.375486402294734e-07, + "loss": 0.8451, + "step": 25990 + }, + { + "epoch": 1.0134479331111501, + "grad_norm": 10.193984985351562, + "learning_rate": 8.373813029005725e-07, + "loss": 0.8383, + "step": 26000 + }, + { + "epoch": 1.0134479331111501, + "eval_loss": 0.8999340534210205, + "eval_runtime": 82.9724, + "eval_samples_per_second": 49.98, + "eval_steps_per_second": 6.255, + "step": 26000 + }, + { + "epoch": 1.013837728273792, + "grad_norm": 13.110339164733887, + "learning_rate": 8.372138961663021e-07, + "loss": 0.8374, + "step": 26010 + }, + { + "epoch": 1.0142275234364342, + "grad_norm": 11.832721710205078, + "learning_rate": 8.370464200611005e-07, + "loss": 0.7692, + "step": 26020 + }, + { + "epoch": 1.0146173185990761, + "grad_norm": 13.495882034301758, + "learning_rate": 8.368788746194206e-07, + "loss": 0.7821, + "step": 26030 + }, + { + "epoch": 1.0150071137617183, + "grad_norm": 12.255905151367188, + "learning_rate": 8.367112598757295e-07, + "loss": 0.7626, + "step": 26040 + }, + { + "epoch": 1.0153969089243602, + "grad_norm": 13.247872352600098, + "learning_rate": 8.365435758645086e-07, + "loss": 0.7384, + "step": 26050 + }, + { + "epoch": 1.0157867040870023, + "grad_norm": 12.817907333374023, + "learning_rate": 8.363758226202533e-07, + "loss": 0.7804, + "step": 26060 + }, + { + "epoch": 1.0161764992496443, + "grad_norm": 12.655926704406738, + "learning_rate": 8.362080001774736e-07, + "loss": 0.835, + "step": 26070 + }, + { + "epoch": 1.0165662944122864, + "grad_norm": 12.620603561401367, + "learning_rate": 8.360401085706937e-07, + "loss": 0.7947, + "step": 26080 + }, + { + "epoch": 1.0169560895749283, + "grad_norm": 13.926952362060547, + "learning_rate": 8.358721478344516e-07, + "loss": 0.8073, + "step": 26090 + }, + { + "epoch": 1.0173458847375705, + "grad_norm": 11.627435684204102, + "learning_rate": 8.357041180033e-07, + "loss": 0.7551, + "step": 26100 + }, + { + "epoch": 1.0177356799002124, + "grad_norm": 12.883454322814941, + "learning_rate": 8.355360191118058e-07, + "loss": 0.7962, + "step": 26110 + }, + { + "epoch": 1.0181254750628546, + "grad_norm": 11.53410816192627, + "learning_rate": 8.353678511945497e-07, + "loss": 0.7997, + "step": 26120 + }, + { + "epoch": 1.0185152702254965, + "grad_norm": 13.038854598999023, + "learning_rate": 8.35199614286127e-07, + "loss": 0.769, + "step": 26130 + }, + { + "epoch": 1.0189050653881386, + "grad_norm": 12.371731758117676, + "learning_rate": 8.350313084211469e-07, + "loss": 0.7756, + "step": 26140 + }, + { + "epoch": 1.0192948605507806, + "grad_norm": 13.076275825500488, + "learning_rate": 8.348629336342331e-07, + "loss": 0.8289, + "step": 26150 + }, + { + "epoch": 1.0196846557134227, + "grad_norm": 12.196067810058594, + "learning_rate": 8.346944899600232e-07, + "loss": 0.7849, + "step": 26160 + }, + { + "epoch": 1.0200744508760646, + "grad_norm": 12.81905460357666, + "learning_rate": 8.34525977433169e-07, + "loss": 0.8386, + "step": 26170 + }, + { + "epoch": 1.0204642460387066, + "grad_norm": 12.867860794067383, + "learning_rate": 8.343573960883368e-07, + "loss": 0.8222, + "step": 26180 + }, + { + "epoch": 1.0208540412013487, + "grad_norm": 13.581961631774902, + "learning_rate": 8.341887459602064e-07, + "loss": 0.8256, + "step": 26190 + }, + { + "epoch": 1.0212438363639906, + "grad_norm": 13.526277542114258, + "learning_rate": 8.340200270834725e-07, + "loss": 0.8099, + "step": 26200 + }, + { + "epoch": 1.0216336315266328, + "grad_norm": 12.352663040161133, + "learning_rate": 8.338512394928434e-07, + "loss": 0.8167, + "step": 26210 + }, + { + "epoch": 1.0220234266892747, + "grad_norm": 11.71937084197998, + "learning_rate": 8.33682383223042e-07, + "loss": 0.7471, + "step": 26220 + }, + { + "epoch": 1.0224132218519169, + "grad_norm": 11.957476615905762, + "learning_rate": 8.335134583088045e-07, + "loss": 0.7689, + "step": 26230 + }, + { + "epoch": 1.0228030170145588, + "grad_norm": 10.258593559265137, + "learning_rate": 8.333444647848825e-07, + "loss": 0.8237, + "step": 26240 + }, + { + "epoch": 1.023192812177201, + "grad_norm": 12.674856185913086, + "learning_rate": 8.331754026860405e-07, + "loss": 0.7664, + "step": 26250 + }, + { + "epoch": 1.0235826073398429, + "grad_norm": 14.840815544128418, + "learning_rate": 8.330062720470578e-07, + "loss": 0.9019, + "step": 26260 + }, + { + "epoch": 1.023972402502485, + "grad_norm": 12.925127029418945, + "learning_rate": 8.328370729027273e-07, + "loss": 0.8355, + "step": 26270 + }, + { + "epoch": 1.024362197665127, + "grad_norm": 12.326546669006348, + "learning_rate": 8.326678052878566e-07, + "loss": 0.8293, + "step": 26280 + }, + { + "epoch": 1.024751992827769, + "grad_norm": 14.481454849243164, + "learning_rate": 8.324984692372671e-07, + "loss": 0.809, + "step": 26290 + }, + { + "epoch": 1.025141787990411, + "grad_norm": 14.834517478942871, + "learning_rate": 8.32329064785794e-07, + "loss": 0.8426, + "step": 26300 + }, + { + "epoch": 1.0255315831530532, + "grad_norm": 13.77194881439209, + "learning_rate": 8.321595919682871e-07, + "loss": 0.8248, + "step": 26310 + }, + { + "epoch": 1.025921378315695, + "grad_norm": 13.203896522521973, + "learning_rate": 8.319900508196097e-07, + "loss": 0.8225, + "step": 26320 + }, + { + "epoch": 1.0263111734783372, + "grad_norm": 12.071765899658203, + "learning_rate": 8.318204413746397e-07, + "loss": 0.7881, + "step": 26330 + }, + { + "epoch": 1.0267009686409792, + "grad_norm": 12.606271743774414, + "learning_rate": 8.316507636682686e-07, + "loss": 0.7566, + "step": 26340 + }, + { + "epoch": 1.0270907638036213, + "grad_norm": 14.735746383666992, + "learning_rate": 8.314810177354023e-07, + "loss": 0.7998, + "step": 26350 + }, + { + "epoch": 1.0274805589662632, + "grad_norm": 12.664928436279297, + "learning_rate": 8.313112036109606e-07, + "loss": 0.807, + "step": 26360 + }, + { + "epoch": 1.0278703541289054, + "grad_norm": 14.861725807189941, + "learning_rate": 8.311413213298772e-07, + "loss": 0.8155, + "step": 26370 + }, + { + "epoch": 1.0282601492915473, + "grad_norm": 14.595712661743164, + "learning_rate": 8.309713709270999e-07, + "loss": 0.817, + "step": 26380 + }, + { + "epoch": 1.0286499444541892, + "grad_norm": 13.959186553955078, + "learning_rate": 8.308013524375906e-07, + "loss": 0.7874, + "step": 26390 + }, + { + "epoch": 1.0290397396168314, + "grad_norm": 13.790559768676758, + "learning_rate": 8.306312658963252e-07, + "loss": 0.7515, + "step": 26400 + }, + { + "epoch": 1.0294295347794733, + "grad_norm": 11.547865867614746, + "learning_rate": 8.304611113382933e-07, + "loss": 0.7702, + "step": 26410 + }, + { + "epoch": 1.0298193299421154, + "grad_norm": 13.873056411743164, + "learning_rate": 8.302908887984989e-07, + "loss": 0.776, + "step": 26420 + }, + { + "epoch": 1.0302091251047574, + "grad_norm": 13.07530689239502, + "learning_rate": 8.301205983119601e-07, + "loss": 0.8167, + "step": 26430 + }, + { + "epoch": 1.0305989202673995, + "grad_norm": 14.523664474487305, + "learning_rate": 8.299502399137081e-07, + "loss": 0.8364, + "step": 26440 + }, + { + "epoch": 1.0309887154300414, + "grad_norm": 12.750444412231445, + "learning_rate": 8.297798136387893e-07, + "loss": 0.8464, + "step": 26450 + }, + { + "epoch": 1.0313785105926836, + "grad_norm": 13.524201393127441, + "learning_rate": 8.296093195222629e-07, + "loss": 0.8449, + "step": 26460 + }, + { + "epoch": 1.0317683057553255, + "grad_norm": 12.466487884521484, + "learning_rate": 8.29438757599203e-07, + "loss": 0.7847, + "step": 26470 + }, + { + "epoch": 1.0321581009179677, + "grad_norm": 11.444717407226562, + "learning_rate": 8.292681279046969e-07, + "loss": 0.7927, + "step": 26480 + }, + { + "epoch": 1.0325478960806096, + "grad_norm": 12.384919166564941, + "learning_rate": 8.290974304738465e-07, + "loss": 0.8145, + "step": 26490 + }, + { + "epoch": 1.0329376912432517, + "grad_norm": 13.145297050476074, + "learning_rate": 8.289266653417672e-07, + "loss": 0.8306, + "step": 26500 + }, + { + "epoch": 1.0329376912432517, + "eval_loss": 0.9041900634765625, + "eval_runtime": 82.9916, + "eval_samples_per_second": 49.969, + "eval_steps_per_second": 6.254, + "step": 26500 + }, + { + "epoch": 1.0333274864058937, + "grad_norm": 13.99440860748291, + "learning_rate": 8.287558325435881e-07, + "loss": 0.8188, + "step": 26510 + }, + { + "epoch": 1.0337172815685358, + "grad_norm": 14.72019100189209, + "learning_rate": 8.285849321144532e-07, + "loss": 0.8128, + "step": 26520 + }, + { + "epoch": 1.0341070767311777, + "grad_norm": 13.388384819030762, + "learning_rate": 8.284139640895192e-07, + "loss": 0.836, + "step": 26530 + }, + { + "epoch": 1.03449687189382, + "grad_norm": 12.018630981445312, + "learning_rate": 8.282429285039576e-07, + "loss": 0.7736, + "step": 26540 + }, + { + "epoch": 1.0348866670564618, + "grad_norm": 11.513049125671387, + "learning_rate": 8.280718253929534e-07, + "loss": 0.8036, + "step": 26550 + }, + { + "epoch": 1.035276462219104, + "grad_norm": 12.40732479095459, + "learning_rate": 8.279006547917056e-07, + "loss": 0.8064, + "step": 26560 + }, + { + "epoch": 1.035666257381746, + "grad_norm": 12.888972282409668, + "learning_rate": 8.27729416735427e-07, + "loss": 0.8813, + "step": 26570 + }, + { + "epoch": 1.0360560525443878, + "grad_norm": 11.713201522827148, + "learning_rate": 8.275581112593443e-07, + "loss": 0.7809, + "step": 26580 + }, + { + "epoch": 1.03644584770703, + "grad_norm": 14.21922492980957, + "learning_rate": 8.273867383986983e-07, + "loss": 0.8344, + "step": 26590 + }, + { + "epoch": 1.036835642869672, + "grad_norm": 11.363812446594238, + "learning_rate": 8.272152981887433e-07, + "loss": 0.8495, + "step": 26600 + }, + { + "epoch": 1.037225438032314, + "grad_norm": 11.974915504455566, + "learning_rate": 8.270437906647477e-07, + "loss": 0.8458, + "step": 26610 + }, + { + "epoch": 1.037615233194956, + "grad_norm": 11.189030647277832, + "learning_rate": 8.268722158619934e-07, + "loss": 0.7811, + "step": 26620 + }, + { + "epoch": 1.0380050283575981, + "grad_norm": 13.42796802520752, + "learning_rate": 8.267005738157767e-07, + "loss": 0.8272, + "step": 26630 + }, + { + "epoch": 1.03839482352024, + "grad_norm": 11.45654582977295, + "learning_rate": 8.265288645614073e-07, + "loss": 0.8118, + "step": 26640 + }, + { + "epoch": 1.0387846186828822, + "grad_norm": 15.133028984069824, + "learning_rate": 8.26357088134209e-07, + "loss": 0.8038, + "step": 26650 + }, + { + "epoch": 1.0391744138455241, + "grad_norm": 12.039899826049805, + "learning_rate": 8.261852445695192e-07, + "loss": 0.8645, + "step": 26660 + }, + { + "epoch": 1.0395642090081663, + "grad_norm": 11.357874870300293, + "learning_rate": 8.260133339026892e-07, + "loss": 0.7526, + "step": 26670 + }, + { + "epoch": 1.0399540041708082, + "grad_norm": 10.882299423217773, + "learning_rate": 8.258413561690841e-07, + "loss": 0.8211, + "step": 26680 + }, + { + "epoch": 1.0403437993334503, + "grad_norm": 13.363794326782227, + "learning_rate": 8.256693114040827e-07, + "loss": 0.8151, + "step": 26690 + }, + { + "epoch": 1.0407335944960923, + "grad_norm": 11.659010887145996, + "learning_rate": 8.25497199643078e-07, + "loss": 0.7808, + "step": 26700 + }, + { + "epoch": 1.0411233896587344, + "grad_norm": 14.34227466583252, + "learning_rate": 8.25325020921476e-07, + "loss": 0.8036, + "step": 26710 + }, + { + "epoch": 1.0415131848213763, + "grad_norm": 11.672528266906738, + "learning_rate": 8.251527752746973e-07, + "loss": 0.7795, + "step": 26720 + }, + { + "epoch": 1.0419029799840185, + "grad_norm": 12.94571590423584, + "learning_rate": 8.249804627381758e-07, + "loss": 0.7964, + "step": 26730 + }, + { + "epoch": 1.0422927751466604, + "grad_norm": 12.850025177001953, + "learning_rate": 8.248080833473592e-07, + "loss": 0.7756, + "step": 26740 + }, + { + "epoch": 1.0426825703093026, + "grad_norm": 12.717275619506836, + "learning_rate": 8.246356371377092e-07, + "loss": 0.9165, + "step": 26750 + }, + { + "epoch": 1.0430723654719445, + "grad_norm": 13.126909255981445, + "learning_rate": 8.244631241447008e-07, + "loss": 0.8316, + "step": 26760 + }, + { + "epoch": 1.0434621606345864, + "grad_norm": 11.642067909240723, + "learning_rate": 8.242905444038235e-07, + "loss": 0.7573, + "step": 26770 + }, + { + "epoch": 1.0438519557972286, + "grad_norm": 12.134324073791504, + "learning_rate": 8.241178979505795e-07, + "loss": 0.8079, + "step": 26780 + }, + { + "epoch": 1.0442417509598705, + "grad_norm": 16.076539993286133, + "learning_rate": 8.239451848204855e-07, + "loss": 0.7577, + "step": 26790 + }, + { + "epoch": 1.0446315461225126, + "grad_norm": 12.308966636657715, + "learning_rate": 8.237724050490716e-07, + "loss": 0.7585, + "step": 26800 + }, + { + "epoch": 1.0450213412851546, + "grad_norm": 15.163002014160156, + "learning_rate": 8.235995586718818e-07, + "loss": 0.7914, + "step": 26810 + }, + { + "epoch": 1.0454111364477967, + "grad_norm": 14.938960075378418, + "learning_rate": 8.234266457244739e-07, + "loss": 0.9044, + "step": 26820 + }, + { + "epoch": 1.0458009316104386, + "grad_norm": 12.543060302734375, + "learning_rate": 8.232536662424186e-07, + "loss": 0.7646, + "step": 26830 + }, + { + "epoch": 1.0461907267730808, + "grad_norm": 12.274667739868164, + "learning_rate": 8.230806202613012e-07, + "loss": 0.7323, + "step": 26840 + }, + { + "epoch": 1.0465805219357227, + "grad_norm": 14.465417861938477, + "learning_rate": 8.229075078167205e-07, + "loss": 0.8325, + "step": 26850 + }, + { + "epoch": 1.0469703170983649, + "grad_norm": 13.441286087036133, + "learning_rate": 8.227343289442887e-07, + "loss": 0.8143, + "step": 26860 + }, + { + "epoch": 1.0473601122610068, + "grad_norm": 11.35332202911377, + "learning_rate": 8.225610836796317e-07, + "loss": 0.8226, + "step": 26870 + }, + { + "epoch": 1.047749907423649, + "grad_norm": 12.011646270751953, + "learning_rate": 8.223877720583892e-07, + "loss": 0.8296, + "step": 26880 + }, + { + "epoch": 1.0481397025862909, + "grad_norm": 11.54375171661377, + "learning_rate": 8.222143941162148e-07, + "loss": 0.8178, + "step": 26890 + }, + { + "epoch": 1.048529497748933, + "grad_norm": 14.420514106750488, + "learning_rate": 8.220409498887749e-07, + "loss": 0.816, + "step": 26900 + }, + { + "epoch": 1.048919292911575, + "grad_norm": 13.92561149597168, + "learning_rate": 8.218674394117503e-07, + "loss": 0.8622, + "step": 26910 + }, + { + "epoch": 1.049309088074217, + "grad_norm": 10.262621879577637, + "learning_rate": 8.216938627208354e-07, + "loss": 0.7801, + "step": 26920 + }, + { + "epoch": 1.049698883236859, + "grad_norm": 15.776985168457031, + "learning_rate": 8.21520219851738e-07, + "loss": 0.8437, + "step": 26930 + }, + { + "epoch": 1.0500886783995012, + "grad_norm": 10.944559097290039, + "learning_rate": 8.213465108401791e-07, + "loss": 0.8029, + "step": 26940 + }, + { + "epoch": 1.050478473562143, + "grad_norm": 15.731622695922852, + "learning_rate": 8.211727357218944e-07, + "loss": 0.8122, + "step": 26950 + }, + { + "epoch": 1.0508682687247852, + "grad_norm": 15.617317199707031, + "learning_rate": 8.209988945326321e-07, + "loss": 0.7667, + "step": 26960 + }, + { + "epoch": 1.0512580638874272, + "grad_norm": 13.447525024414062, + "learning_rate": 8.208249873081548e-07, + "loss": 0.8487, + "step": 26970 + }, + { + "epoch": 1.0516478590500693, + "grad_norm": 12.386305809020996, + "learning_rate": 8.206510140842377e-07, + "loss": 0.8389, + "step": 26980 + }, + { + "epoch": 1.0520376542127112, + "grad_norm": 13.150272369384766, + "learning_rate": 8.204769748966709e-07, + "loss": 0.7905, + "step": 26990 + }, + { + "epoch": 1.0524274493753532, + "grad_norm": 13.261439323425293, + "learning_rate": 8.203028697812571e-07, + "loss": 0.812, + "step": 27000 + }, + { + "epoch": 1.0524274493753532, + "eval_loss": 0.9053342342376709, + "eval_runtime": 86.943, + "eval_samples_per_second": 47.698, + "eval_steps_per_second": 5.969, + "step": 27000 + }, + { + "epoch": 1.0528172445379953, + "grad_norm": 13.887731552124023, + "learning_rate": 8.201286987738128e-07, + "loss": 0.776, + "step": 27010 + }, + { + "epoch": 1.0532070397006372, + "grad_norm": 12.451720237731934, + "learning_rate": 8.199544619101679e-07, + "loss": 0.7812, + "step": 27020 + }, + { + "epoch": 1.0535968348632794, + "grad_norm": 15.174094200134277, + "learning_rate": 8.197801592261665e-07, + "loss": 0.8281, + "step": 27030 + }, + { + "epoch": 1.0539866300259213, + "grad_norm": 13.61098575592041, + "learning_rate": 8.196057907576656e-07, + "loss": 0.8165, + "step": 27040 + }, + { + "epoch": 1.0543764251885634, + "grad_norm": 10.957730293273926, + "learning_rate": 8.194313565405356e-07, + "loss": 0.8622, + "step": 27050 + }, + { + "epoch": 1.0547662203512054, + "grad_norm": 12.986677169799805, + "learning_rate": 8.192568566106612e-07, + "loss": 0.8619, + "step": 27060 + }, + { + "epoch": 1.0551560155138475, + "grad_norm": 11.946600914001465, + "learning_rate": 8.190822910039399e-07, + "loss": 0.7549, + "step": 27070 + }, + { + "epoch": 1.0555458106764894, + "grad_norm": 13.488011360168457, + "learning_rate": 8.189076597562828e-07, + "loss": 0.8185, + "step": 27080 + }, + { + "epoch": 1.0559356058391316, + "grad_norm": 14.743844985961914, + "learning_rate": 8.187329629036152e-07, + "loss": 0.7734, + "step": 27090 + }, + { + "epoch": 1.0563254010017735, + "grad_norm": 12.663313865661621, + "learning_rate": 8.18558200481875e-07, + "loss": 0.8223, + "step": 27100 + }, + { + "epoch": 1.0567151961644157, + "grad_norm": 11.354568481445312, + "learning_rate": 8.18383372527014e-07, + "loss": 0.8099, + "step": 27110 + }, + { + "epoch": 1.0571049913270576, + "grad_norm": 12.84670352935791, + "learning_rate": 8.182084790749973e-07, + "loss": 0.7548, + "step": 27120 + }, + { + "epoch": 1.0574947864896997, + "grad_norm": 12.930134773254395, + "learning_rate": 8.180335201618038e-07, + "loss": 0.8264, + "step": 27130 + }, + { + "epoch": 1.0578845816523417, + "grad_norm": 11.460614204406738, + "learning_rate": 8.178584958234259e-07, + "loss": 0.79, + "step": 27140 + }, + { + "epoch": 1.0582743768149838, + "grad_norm": 11.040495872497559, + "learning_rate": 8.176834060958686e-07, + "loss": 0.7852, + "step": 27150 + }, + { + "epoch": 1.0586641719776257, + "grad_norm": 13.990711212158203, + "learning_rate": 8.175082510151515e-07, + "loss": 0.858, + "step": 27160 + }, + { + "epoch": 1.059053967140268, + "grad_norm": 12.06353759765625, + "learning_rate": 8.173330306173071e-07, + "loss": 0.83, + "step": 27170 + }, + { + "epoch": 1.0594437623029098, + "grad_norm": 13.883865356445312, + "learning_rate": 8.171577449383813e-07, + "loss": 0.8401, + "step": 27180 + }, + { + "epoch": 1.0598335574655517, + "grad_norm": 12.454399108886719, + "learning_rate": 8.169823940144333e-07, + "loss": 0.8235, + "step": 27190 + }, + { + "epoch": 1.060223352628194, + "grad_norm": 12.524138450622559, + "learning_rate": 8.168069778815362e-07, + "loss": 0.777, + "step": 27200 + }, + { + "epoch": 1.0606131477908358, + "grad_norm": 12.797564506530762, + "learning_rate": 8.166314965757761e-07, + "loss": 0.8161, + "step": 27210 + }, + { + "epoch": 1.061002942953478, + "grad_norm": 13.436089515686035, + "learning_rate": 8.164559501332526e-07, + "loss": 0.7958, + "step": 27220 + }, + { + "epoch": 1.06139273811612, + "grad_norm": 13.03825569152832, + "learning_rate": 8.162803385900789e-07, + "loss": 0.8269, + "step": 27230 + }, + { + "epoch": 1.061782533278762, + "grad_norm": 10.898392677307129, + "learning_rate": 8.161046619823815e-07, + "loss": 0.8063, + "step": 27240 + }, + { + "epoch": 1.062172328441404, + "grad_norm": 12.551159858703613, + "learning_rate": 8.159289203462999e-07, + "loss": 0.7836, + "step": 27250 + }, + { + "epoch": 1.0625621236040461, + "grad_norm": 13.404977798461914, + "learning_rate": 8.157531137179874e-07, + "loss": 0.7558, + "step": 27260 + }, + { + "epoch": 1.062951918766688, + "grad_norm": 12.52367115020752, + "learning_rate": 8.155772421336109e-07, + "loss": 0.7982, + "step": 27270 + }, + { + "epoch": 1.0633417139293302, + "grad_norm": 13.959651947021484, + "learning_rate": 8.154013056293499e-07, + "loss": 0.8162, + "step": 27280 + }, + { + "epoch": 1.0637315090919721, + "grad_norm": 11.47115421295166, + "learning_rate": 8.152253042413979e-07, + "loss": 0.8067, + "step": 27290 + }, + { + "epoch": 1.0641213042546143, + "grad_norm": 12.519828796386719, + "learning_rate": 8.150492380059614e-07, + "loss": 0.7976, + "step": 27300 + }, + { + "epoch": 1.0645110994172562, + "grad_norm": 13.177289962768555, + "learning_rate": 8.148731069592606e-07, + "loss": 0.8153, + "step": 27310 + }, + { + "epoch": 1.0649008945798983, + "grad_norm": 12.059579849243164, + "learning_rate": 8.146969111375286e-07, + "loss": 0.7765, + "step": 27320 + }, + { + "epoch": 1.0652906897425403, + "grad_norm": 14.590814590454102, + "learning_rate": 8.145206505770121e-07, + "loss": 0.823, + "step": 27330 + }, + { + "epoch": 1.0656804849051824, + "grad_norm": 10.995859146118164, + "learning_rate": 8.143443253139711e-07, + "loss": 0.7824, + "step": 27340 + }, + { + "epoch": 1.0660702800678243, + "grad_norm": 12.843416213989258, + "learning_rate": 8.141679353846789e-07, + "loss": 0.7627, + "step": 27350 + }, + { + "epoch": 1.0664600752304665, + "grad_norm": 12.497159957885742, + "learning_rate": 8.139914808254219e-07, + "loss": 0.8263, + "step": 27360 + }, + { + "epoch": 1.0668498703931084, + "grad_norm": 12.867425918579102, + "learning_rate": 8.138149616725001e-07, + "loss": 0.784, + "step": 27370 + }, + { + "epoch": 1.0672396655557503, + "grad_norm": 14.112774848937988, + "learning_rate": 8.136383779622265e-07, + "loss": 0.8434, + "step": 27380 + }, + { + "epoch": 1.0676294607183925, + "grad_norm": 12.556109428405762, + "learning_rate": 8.134617297309276e-07, + "loss": 0.7726, + "step": 27390 + }, + { + "epoch": 1.0680192558810344, + "grad_norm": 11.861383438110352, + "learning_rate": 8.132850170149432e-07, + "loss": 0.7478, + "step": 27400 + }, + { + "epoch": 1.0684090510436766, + "grad_norm": 12.162129402160645, + "learning_rate": 8.131082398506262e-07, + "loss": 0.7926, + "step": 27410 + }, + { + "epoch": 1.0687988462063185, + "grad_norm": 13.376980781555176, + "learning_rate": 8.129313982743428e-07, + "loss": 0.8172, + "step": 27420 + }, + { + "epoch": 1.0691886413689606, + "grad_norm": 13.789857864379883, + "learning_rate": 8.127544923224725e-07, + "loss": 0.7945, + "step": 27430 + }, + { + "epoch": 1.0695784365316026, + "grad_norm": 12.148392677307129, + "learning_rate": 8.125775220314079e-07, + "loss": 0.8835, + "step": 27440 + }, + { + "epoch": 1.0699682316942447, + "grad_norm": 11.968158721923828, + "learning_rate": 8.124004874375552e-07, + "loss": 0.7736, + "step": 27450 + }, + { + "epoch": 1.0703580268568866, + "grad_norm": 12.269271850585938, + "learning_rate": 8.122233885773333e-07, + "loss": 0.797, + "step": 27460 + }, + { + "epoch": 1.0707478220195288, + "grad_norm": 12.464533805847168, + "learning_rate": 8.120462254871749e-07, + "loss": 0.771, + "step": 27470 + }, + { + "epoch": 1.0711376171821707, + "grad_norm": 16.566967010498047, + "learning_rate": 8.118689982035254e-07, + "loss": 0.8006, + "step": 27480 + }, + { + "epoch": 1.0715274123448129, + "grad_norm": 12.8241605758667, + "learning_rate": 8.116917067628437e-07, + "loss": 0.8142, + "step": 27490 + }, + { + "epoch": 1.0719172075074548, + "grad_norm": 12.9489107131958, + "learning_rate": 8.115143512016018e-07, + "loss": 0.741, + "step": 27500 + }, + { + "epoch": 1.0719172075074548, + "eval_loss": 0.9053013324737549, + "eval_runtime": 82.8654, + "eval_samples_per_second": 50.045, + "eval_steps_per_second": 6.263, + "step": 27500 + }, + { + "epoch": 1.072307002670097, + "grad_norm": 13.004432678222656, + "learning_rate": 8.113369315562849e-07, + "loss": 0.8276, + "step": 27510 + }, + { + "epoch": 1.0726967978327389, + "grad_norm": 12.650984764099121, + "learning_rate": 8.111594478633914e-07, + "loss": 0.8356, + "step": 27520 + }, + { + "epoch": 1.073086592995381, + "grad_norm": 13.477132797241211, + "learning_rate": 8.10981900159433e-07, + "loss": 0.8016, + "step": 27530 + }, + { + "epoch": 1.073476388158023, + "grad_norm": 13.738760948181152, + "learning_rate": 8.108042884809343e-07, + "loss": 0.8073, + "step": 27540 + }, + { + "epoch": 1.073866183320665, + "grad_norm": 14.808836936950684, + "learning_rate": 8.106266128644332e-07, + "loss": 0.8571, + "step": 27550 + }, + { + "epoch": 1.074255978483307, + "grad_norm": 13.267008781433105, + "learning_rate": 8.104488733464807e-07, + "loss": 0.8115, + "step": 27560 + }, + { + "epoch": 1.0746457736459492, + "grad_norm": 13.303712844848633, + "learning_rate": 8.102710699636413e-07, + "loss": 0.812, + "step": 27570 + }, + { + "epoch": 1.075035568808591, + "grad_norm": 10.556928634643555, + "learning_rate": 8.100932027524918e-07, + "loss": 0.7853, + "step": 27580 + }, + { + "epoch": 1.0754253639712332, + "grad_norm": 13.999531745910645, + "learning_rate": 8.099152717496232e-07, + "loss": 0.8336, + "step": 27590 + }, + { + "epoch": 1.0758151591338752, + "grad_norm": 13.299699783325195, + "learning_rate": 8.097372769916389e-07, + "loss": 0.8154, + "step": 27600 + }, + { + "epoch": 1.076204954296517, + "grad_norm": 12.60476303100586, + "learning_rate": 8.095592185151554e-07, + "loss": 0.7985, + "step": 27610 + }, + { + "epoch": 1.0765947494591592, + "grad_norm": 12.583000183105469, + "learning_rate": 8.093810963568029e-07, + "loss": 0.7568, + "step": 27620 + }, + { + "epoch": 1.0769845446218012, + "grad_norm": 13.485947608947754, + "learning_rate": 8.092029105532239e-07, + "loss": 0.8094, + "step": 27630 + }, + { + "epoch": 1.0773743397844433, + "grad_norm": 13.187661170959473, + "learning_rate": 8.090246611410747e-07, + "loss": 0.7481, + "step": 27640 + }, + { + "epoch": 1.0777641349470852, + "grad_norm": 12.163583755493164, + "learning_rate": 8.088463481570244e-07, + "loss": 0.7773, + "step": 27650 + }, + { + "epoch": 1.0781539301097274, + "grad_norm": 11.902276992797852, + "learning_rate": 8.086679716377549e-07, + "loss": 0.784, + "step": 27660 + }, + { + "epoch": 1.0785437252723693, + "grad_norm": 13.30802059173584, + "learning_rate": 8.084895316199618e-07, + "loss": 0.8006, + "step": 27670 + }, + { + "epoch": 1.0789335204350115, + "grad_norm": 14.507987022399902, + "learning_rate": 8.083110281403532e-07, + "loss": 0.7998, + "step": 27680 + }, + { + "epoch": 1.0793233155976534, + "grad_norm": 10.409671783447266, + "learning_rate": 8.081324612356503e-07, + "loss": 0.7818, + "step": 27690 + }, + { + "epoch": 1.0797131107602955, + "grad_norm": 13.713388442993164, + "learning_rate": 8.07953830942588e-07, + "loss": 0.842, + "step": 27700 + }, + { + "epoch": 1.0801029059229375, + "grad_norm": 11.162589073181152, + "learning_rate": 8.077751372979133e-07, + "loss": 0.8224, + "step": 27710 + }, + { + "epoch": 1.0804927010855796, + "grad_norm": 11.254040718078613, + "learning_rate": 8.075963803383867e-07, + "loss": 0.7765, + "step": 27720 + }, + { + "epoch": 1.0808824962482215, + "grad_norm": 13.396617889404297, + "learning_rate": 8.074175601007819e-07, + "loss": 0.7679, + "step": 27730 + }, + { + "epoch": 1.0812722914108637, + "grad_norm": 11.768601417541504, + "learning_rate": 8.072386766218853e-07, + "loss": 0.7609, + "step": 27740 + }, + { + "epoch": 1.0816620865735056, + "grad_norm": 13.934133529663086, + "learning_rate": 8.070597299384966e-07, + "loss": 0.7437, + "step": 27750 + }, + { + "epoch": 1.0820518817361477, + "grad_norm": 11.977148056030273, + "learning_rate": 8.068807200874281e-07, + "loss": 0.7737, + "step": 27760 + }, + { + "epoch": 1.0824416768987897, + "grad_norm": 12.239559173583984, + "learning_rate": 8.067016471055055e-07, + "loss": 0.8354, + "step": 27770 + }, + { + "epoch": 1.0828314720614318, + "grad_norm": 10.479872703552246, + "learning_rate": 8.065225110295672e-07, + "loss": 0.8267, + "step": 27780 + }, + { + "epoch": 1.0832212672240737, + "grad_norm": 11.449228286743164, + "learning_rate": 8.063433118964646e-07, + "loss": 0.7601, + "step": 27790 + }, + { + "epoch": 1.0836110623867157, + "grad_norm": 13.452764511108398, + "learning_rate": 8.061640497430627e-07, + "loss": 0.8573, + "step": 27800 + }, + { + "epoch": 1.0840008575493578, + "grad_norm": 12.841625213623047, + "learning_rate": 8.059847246062384e-07, + "loss": 0.8139, + "step": 27810 + }, + { + "epoch": 1.0843906527119997, + "grad_norm": 12.560860633850098, + "learning_rate": 8.058053365228821e-07, + "loss": 0.8324, + "step": 27820 + }, + { + "epoch": 1.084780447874642, + "grad_norm": 15.33007526397705, + "learning_rate": 8.056258855298973e-07, + "loss": 0.8302, + "step": 27830 + }, + { + "epoch": 1.0851702430372838, + "grad_norm": 14.899931907653809, + "learning_rate": 8.054463716642004e-07, + "loss": 0.8008, + "step": 27840 + }, + { + "epoch": 1.085560038199926, + "grad_norm": 14.829975128173828, + "learning_rate": 8.052667949627204e-07, + "loss": 0.7994, + "step": 27850 + }, + { + "epoch": 1.085949833362568, + "grad_norm": 16.214414596557617, + "learning_rate": 8.050871554623994e-07, + "loss": 0.8162, + "step": 27860 + }, + { + "epoch": 1.08633962852521, + "grad_norm": 14.20153522491455, + "learning_rate": 8.049074532001926e-07, + "loss": 0.8302, + "step": 27870 + }, + { + "epoch": 1.086729423687852, + "grad_norm": 10.92680835723877, + "learning_rate": 8.04727688213068e-07, + "loss": 0.8018, + "step": 27880 + }, + { + "epoch": 1.0871192188504941, + "grad_norm": 12.890881538391113, + "learning_rate": 8.045478605380065e-07, + "loss": 0.7759, + "step": 27890 + }, + { + "epoch": 1.087509014013136, + "grad_norm": 13.365934371948242, + "learning_rate": 8.043679702120017e-07, + "loss": 0.761, + "step": 27900 + }, + { + "epoch": 1.0878988091757782, + "grad_norm": 13.280560493469238, + "learning_rate": 8.041880172720603e-07, + "loss": 0.8242, + "step": 27910 + }, + { + "epoch": 1.0882886043384201, + "grad_norm": 15.511161804199219, + "learning_rate": 8.040080017552019e-07, + "loss": 0.8747, + "step": 27920 + }, + { + "epoch": 1.0886783995010623, + "grad_norm": 15.51421070098877, + "learning_rate": 8.038279236984588e-07, + "loss": 0.7076, + "step": 27930 + }, + { + "epoch": 1.0890681946637042, + "grad_norm": 13.820291519165039, + "learning_rate": 8.036477831388765e-07, + "loss": 0.7422, + "step": 27940 + }, + { + "epoch": 1.0894579898263463, + "grad_norm": 11.928253173828125, + "learning_rate": 8.03467580113513e-07, + "loss": 0.841, + "step": 27950 + }, + { + "epoch": 1.0898477849889883, + "grad_norm": 13.923212051391602, + "learning_rate": 8.032873146594392e-07, + "loss": 0.8223, + "step": 27960 + }, + { + "epoch": 1.0902375801516304, + "grad_norm": 13.978311538696289, + "learning_rate": 8.031069868137391e-07, + "loss": 0.7826, + "step": 27970 + }, + { + "epoch": 1.0906273753142723, + "grad_norm": 12.111215591430664, + "learning_rate": 8.029265966135092e-07, + "loss": 0.7967, + "step": 27980 + }, + { + "epoch": 1.0910171704769143, + "grad_norm": 15.658437728881836, + "learning_rate": 8.027461440958593e-07, + "loss": 0.7755, + "step": 27990 + }, + { + "epoch": 1.0914069656395564, + "grad_norm": 13.354958534240723, + "learning_rate": 8.025656292979112e-07, + "loss": 0.7819, + "step": 28000 + }, + { + "epoch": 1.0914069656395564, + "eval_loss": 0.9016230702400208, + "eval_runtime": 82.7021, + "eval_samples_per_second": 50.144, + "eval_steps_per_second": 6.276, + "step": 28000 + }, + { + "epoch": 1.0917967608021986, + "grad_norm": 12.706086158752441, + "learning_rate": 8.023850522568005e-07, + "loss": 0.7668, + "step": 28010 + }, + { + "epoch": 1.0921865559648405, + "grad_norm": 12.374022483825684, + "learning_rate": 8.022044130096749e-07, + "loss": 0.7616, + "step": 28020 + }, + { + "epoch": 1.0925763511274824, + "grad_norm": 15.056655883789062, + "learning_rate": 8.020237115936952e-07, + "loss": 0.7616, + "step": 28030 + }, + { + "epoch": 1.0929661462901246, + "grad_norm": 13.827685356140137, + "learning_rate": 8.018429480460347e-07, + "loss": 0.7995, + "step": 28040 + }, + { + "epoch": 1.0933559414527665, + "grad_norm": 13.510518074035645, + "learning_rate": 8.016621224038799e-07, + "loss": 0.7787, + "step": 28050 + }, + { + "epoch": 1.0937457366154086, + "grad_norm": 11.93447494506836, + "learning_rate": 8.0148123470443e-07, + "loss": 0.828, + "step": 28060 + }, + { + "epoch": 1.0941355317780506, + "grad_norm": 11.586976051330566, + "learning_rate": 8.013002849848967e-07, + "loss": 0.77, + "step": 28070 + }, + { + "epoch": 1.0945253269406927, + "grad_norm": 14.013508796691895, + "learning_rate": 8.011192732825044e-07, + "loss": 0.8101, + "step": 28080 + }, + { + "epoch": 1.0949151221033346, + "grad_norm": 11.864948272705078, + "learning_rate": 8.009381996344906e-07, + "loss": 0.8109, + "step": 28090 + }, + { + "epoch": 1.0953049172659768, + "grad_norm": 12.35537338256836, + "learning_rate": 8.007570640781056e-07, + "loss": 0.7089, + "step": 28100 + }, + { + "epoch": 1.0956947124286187, + "grad_norm": 14.167571067810059, + "learning_rate": 8.00575866650612e-07, + "loss": 0.8146, + "step": 28110 + }, + { + "epoch": 1.0960845075912609, + "grad_norm": 11.950983047485352, + "learning_rate": 8.003946073892852e-07, + "loss": 0.8504, + "step": 28120 + }, + { + "epoch": 1.0964743027539028, + "grad_norm": 12.764832496643066, + "learning_rate": 8.002132863314137e-07, + "loss": 0.8005, + "step": 28130 + }, + { + "epoch": 1.096864097916545, + "grad_norm": 13.662081718444824, + "learning_rate": 8.000319035142986e-07, + "loss": 0.8017, + "step": 28140 + }, + { + "epoch": 1.0972538930791869, + "grad_norm": 14.415201187133789, + "learning_rate": 7.998504589752532e-07, + "loss": 0.8137, + "step": 28150 + }, + { + "epoch": 1.097643688241829, + "grad_norm": 11.369823455810547, + "learning_rate": 7.996689527516044e-07, + "loss": 0.7757, + "step": 28160 + }, + { + "epoch": 1.098033483404471, + "grad_norm": 10.054341316223145, + "learning_rate": 7.994873848806909e-07, + "loss": 0.8149, + "step": 28170 + }, + { + "epoch": 1.098423278567113, + "grad_norm": 11.595555305480957, + "learning_rate": 7.993057553998645e-07, + "loss": 0.7874, + "step": 28180 + }, + { + "epoch": 1.098813073729755, + "grad_norm": 10.6121187210083, + "learning_rate": 7.991240643464896e-07, + "loss": 0.823, + "step": 28190 + }, + { + "epoch": 1.0992028688923972, + "grad_norm": 13.28363037109375, + "learning_rate": 7.989423117579434e-07, + "loss": 0.8045, + "step": 28200 + }, + { + "epoch": 1.099592664055039, + "grad_norm": 14.741850852966309, + "learning_rate": 7.987604976716159e-07, + "loss": 0.7846, + "step": 28210 + }, + { + "epoch": 1.099982459217681, + "grad_norm": 13.614274978637695, + "learning_rate": 7.985786221249091e-07, + "loss": 0.7895, + "step": 28220 + }, + { + "epoch": 1.1003722543803232, + "grad_norm": 11.489198684692383, + "learning_rate": 7.983966851552382e-07, + "loss": 0.7765, + "step": 28230 + }, + { + "epoch": 1.100762049542965, + "grad_norm": 11.574873924255371, + "learning_rate": 7.982146868000309e-07, + "loss": 0.8362, + "step": 28240 + }, + { + "epoch": 1.1011518447056072, + "grad_norm": 13.743842124938965, + "learning_rate": 7.980326270967275e-07, + "loss": 0.7808, + "step": 28250 + }, + { + "epoch": 1.1015416398682492, + "grad_norm": 12.682780265808105, + "learning_rate": 7.978505060827809e-07, + "loss": 0.8393, + "step": 28260 + }, + { + "epoch": 1.1019314350308913, + "grad_norm": 12.588360786437988, + "learning_rate": 7.976683237956569e-07, + "loss": 0.8107, + "step": 28270 + }, + { + "epoch": 1.1023212301935332, + "grad_norm": 14.450643539428711, + "learning_rate": 7.974860802728333e-07, + "loss": 0.7736, + "step": 28280 + }, + { + "epoch": 1.1027110253561754, + "grad_norm": 12.218282699584961, + "learning_rate": 7.973037755518011e-07, + "loss": 0.8129, + "step": 28290 + }, + { + "epoch": 1.1031008205188173, + "grad_norm": 12.179034233093262, + "learning_rate": 7.971214096700636e-07, + "loss": 0.803, + "step": 28300 + }, + { + "epoch": 1.1034906156814595, + "grad_norm": 11.605459213256836, + "learning_rate": 7.969389826651364e-07, + "loss": 0.8092, + "step": 28310 + }, + { + "epoch": 1.1038804108441014, + "grad_norm": 12.345074653625488, + "learning_rate": 7.967564945745487e-07, + "loss": 0.7793, + "step": 28320 + }, + { + "epoch": 1.1042702060067435, + "grad_norm": 12.991783142089844, + "learning_rate": 7.965739454358409e-07, + "loss": 0.71, + "step": 28330 + }, + { + "epoch": 1.1046600011693855, + "grad_norm": 12.648735046386719, + "learning_rate": 7.963913352865668e-07, + "loss": 0.8176, + "step": 28340 + }, + { + "epoch": 1.1050497963320276, + "grad_norm": 12.953163146972656, + "learning_rate": 7.962086641642929e-07, + "loss": 0.8344, + "step": 28350 + }, + { + "epoch": 1.1054395914946695, + "grad_norm": 14.095447540283203, + "learning_rate": 7.960259321065974e-07, + "loss": 0.8545, + "step": 28360 + }, + { + "epoch": 1.1058293866573117, + "grad_norm": 12.277088165283203, + "learning_rate": 7.958431391510719e-07, + "loss": 0.8347, + "step": 28370 + }, + { + "epoch": 1.1062191818199536, + "grad_norm": 13.299829483032227, + "learning_rate": 7.956602853353202e-07, + "loss": 0.8148, + "step": 28380 + }, + { + "epoch": 1.1066089769825957, + "grad_norm": 13.568120956420898, + "learning_rate": 7.954773706969584e-07, + "loss": 0.8028, + "step": 28390 + }, + { + "epoch": 1.1069987721452377, + "grad_norm": 12.079610824584961, + "learning_rate": 7.952943952736155e-07, + "loss": 0.8556, + "step": 28400 + }, + { + "epoch": 1.1073885673078796, + "grad_norm": 12.730743408203125, + "learning_rate": 7.951113591029325e-07, + "loss": 0.7034, + "step": 28410 + }, + { + "epoch": 1.1077783624705217, + "grad_norm": 12.168791770935059, + "learning_rate": 7.949282622225637e-07, + "loss": 0.752, + "step": 28420 + }, + { + "epoch": 1.1081681576331637, + "grad_norm": 11.914694786071777, + "learning_rate": 7.94745104670175e-07, + "loss": 0.7648, + "step": 28430 + }, + { + "epoch": 1.1085579527958058, + "grad_norm": 12.120577812194824, + "learning_rate": 7.945618864834454e-07, + "loss": 0.7829, + "step": 28440 + }, + { + "epoch": 1.1089477479584477, + "grad_norm": 11.039590835571289, + "learning_rate": 7.943786077000661e-07, + "loss": 0.802, + "step": 28450 + }, + { + "epoch": 1.10933754312109, + "grad_norm": 12.19485092163086, + "learning_rate": 7.941952683577409e-07, + "loss": 0.7884, + "step": 28460 + }, + { + "epoch": 1.1097273382837318, + "grad_norm": 13.17673110961914, + "learning_rate": 7.940118684941858e-07, + "loss": 0.8544, + "step": 28470 + }, + { + "epoch": 1.110117133446374, + "grad_norm": 12.946337699890137, + "learning_rate": 7.938284081471296e-07, + "loss": 0.8255, + "step": 28480 + }, + { + "epoch": 1.110506928609016, + "grad_norm": 11.45637321472168, + "learning_rate": 7.936448873543131e-07, + "loss": 0.8214, + "step": 28490 + }, + { + "epoch": 1.110896723771658, + "grad_norm": 11.550646781921387, + "learning_rate": 7.934613061534901e-07, + "loss": 0.7546, + "step": 28500 + }, + { + "epoch": 1.110896723771658, + "eval_loss": 0.8982346057891846, + "eval_runtime": 83.3703, + "eval_samples_per_second": 49.742, + "eval_steps_per_second": 6.225, + "step": 28500 + }, + { + "epoch": 1.1112865189343, + "grad_norm": 13.465926170349121, + "learning_rate": 7.932776645824266e-07, + "loss": 0.795, + "step": 28510 + }, + { + "epoch": 1.1116763140969421, + "grad_norm": 11.873233795166016, + "learning_rate": 7.930939626789008e-07, + "loss": 0.7854, + "step": 28520 + }, + { + "epoch": 1.112066109259584, + "grad_norm": 12.118992805480957, + "learning_rate": 7.929102004807034e-07, + "loss": 0.8092, + "step": 28530 + }, + { + "epoch": 1.1124559044222262, + "grad_norm": 14.147194862365723, + "learning_rate": 7.927263780256376e-07, + "loss": 0.8796, + "step": 28540 + }, + { + "epoch": 1.1128456995848681, + "grad_norm": 13.484308242797852, + "learning_rate": 7.92542495351519e-07, + "loss": 0.8033, + "step": 28550 + }, + { + "epoch": 1.1132354947475103, + "grad_norm": 11.869744300842285, + "learning_rate": 7.923585524961758e-07, + "loss": 0.7307, + "step": 28560 + }, + { + "epoch": 1.1136252899101522, + "grad_norm": 14.602240562438965, + "learning_rate": 7.921745494974479e-07, + "loss": 0.8187, + "step": 28570 + }, + { + "epoch": 1.1140150850727943, + "grad_norm": 12.614906311035156, + "learning_rate": 7.919904863931882e-07, + "loss": 0.8087, + "step": 28580 + }, + { + "epoch": 1.1144048802354363, + "grad_norm": 14.365131378173828, + "learning_rate": 7.918063632212619e-07, + "loss": 0.7967, + "step": 28590 + }, + { + "epoch": 1.1147946753980782, + "grad_norm": 13.164263725280762, + "learning_rate": 7.916221800195464e-07, + "loss": 0.7664, + "step": 28600 + }, + { + "epoch": 1.1151844705607203, + "grad_norm": 11.47569465637207, + "learning_rate": 7.914379368259312e-07, + "loss": 0.807, + "step": 28610 + }, + { + "epoch": 1.1155742657233625, + "grad_norm": 13.731983184814453, + "learning_rate": 7.912536336783187e-07, + "loss": 0.8535, + "step": 28620 + }, + { + "epoch": 1.1159640608860044, + "grad_norm": 11.534174919128418, + "learning_rate": 7.910692706146234e-07, + "loss": 0.7678, + "step": 28630 + }, + { + "epoch": 1.1163538560486463, + "grad_norm": 13.083117485046387, + "learning_rate": 7.908848476727719e-07, + "loss": 0.7529, + "step": 28640 + }, + { + "epoch": 1.1167436512112885, + "grad_norm": 13.10075855255127, + "learning_rate": 7.907003648907032e-07, + "loss": 0.7704, + "step": 28650 + }, + { + "epoch": 1.1171334463739304, + "grad_norm": 14.31292724609375, + "learning_rate": 7.905158223063691e-07, + "loss": 0.8193, + "step": 28660 + }, + { + "epoch": 1.1175232415365726, + "grad_norm": 13.96985149383545, + "learning_rate": 7.903312199577329e-07, + "loss": 0.8819, + "step": 28670 + }, + { + "epoch": 1.1179130366992145, + "grad_norm": 16.52273941040039, + "learning_rate": 7.901465578827707e-07, + "loss": 0.8304, + "step": 28680 + }, + { + "epoch": 1.1183028318618566, + "grad_norm": 11.734724998474121, + "learning_rate": 7.899618361194711e-07, + "loss": 0.7843, + "step": 28690 + }, + { + "epoch": 1.1186926270244986, + "grad_norm": 10.366153717041016, + "learning_rate": 7.897770547058345e-07, + "loss": 0.8437, + "step": 28700 + }, + { + "epoch": 1.1190824221871407, + "grad_norm": 14.169140815734863, + "learning_rate": 7.895922136798734e-07, + "loss": 0.8423, + "step": 28710 + }, + { + "epoch": 1.1194722173497826, + "grad_norm": 12.152835845947266, + "learning_rate": 7.894073130796132e-07, + "loss": 0.7791, + "step": 28720 + }, + { + "epoch": 1.1198620125124248, + "grad_norm": 11.646323204040527, + "learning_rate": 7.892223529430914e-07, + "loss": 0.7707, + "step": 28730 + }, + { + "epoch": 1.1202518076750667, + "grad_norm": 14.393430709838867, + "learning_rate": 7.890373333083577e-07, + "loss": 0.8172, + "step": 28740 + }, + { + "epoch": 1.1206416028377089, + "grad_norm": 13.921177864074707, + "learning_rate": 7.888522542134734e-07, + "loss": 0.766, + "step": 28750 + }, + { + "epoch": 1.1210313980003508, + "grad_norm": 12.956089973449707, + "learning_rate": 7.886671156965129e-07, + "loss": 0.7968, + "step": 28760 + }, + { + "epoch": 1.121421193162993, + "grad_norm": 11.328206062316895, + "learning_rate": 7.884819177955626e-07, + "loss": 0.7516, + "step": 28770 + }, + { + "epoch": 1.1218109883256349, + "grad_norm": 13.190888404846191, + "learning_rate": 7.882966605487209e-07, + "loss": 0.829, + "step": 28780 + }, + { + "epoch": 1.122200783488277, + "grad_norm": 12.017468452453613, + "learning_rate": 7.881113439940985e-07, + "loss": 0.7507, + "step": 28790 + }, + { + "epoch": 1.122590578650919, + "grad_norm": 12.675570487976074, + "learning_rate": 7.879259681698185e-07, + "loss": 0.8873, + "step": 28800 + }, + { + "epoch": 1.122980373813561, + "grad_norm": 12.14181900024414, + "learning_rate": 7.877405331140159e-07, + "loss": 0.8123, + "step": 28810 + }, + { + "epoch": 1.123370168976203, + "grad_norm": 11.03156566619873, + "learning_rate": 7.875550388648377e-07, + "loss": 0.7979, + "step": 28820 + }, + { + "epoch": 1.123759964138845, + "grad_norm": 11.81356143951416, + "learning_rate": 7.873694854604441e-07, + "loss": 0.8075, + "step": 28830 + }, + { + "epoch": 1.124149759301487, + "grad_norm": 11.741072654724121, + "learning_rate": 7.871838729390061e-07, + "loss": 0.8738, + "step": 28840 + }, + { + "epoch": 1.124539554464129, + "grad_norm": 13.357629776000977, + "learning_rate": 7.86998201338708e-07, + "loss": 0.7592, + "step": 28850 + }, + { + "epoch": 1.1249293496267712, + "grad_norm": 13.082197189331055, + "learning_rate": 7.868124706977452e-07, + "loss": 0.7785, + "step": 28860 + }, + { + "epoch": 1.125319144789413, + "grad_norm": 13.144143104553223, + "learning_rate": 7.866266810543265e-07, + "loss": 0.7587, + "step": 28870 + }, + { + "epoch": 1.1257089399520552, + "grad_norm": 12.02243709564209, + "learning_rate": 7.864408324466716e-07, + "loss": 0.7376, + "step": 28880 + }, + { + "epoch": 1.1260987351146972, + "grad_norm": 14.394508361816406, + "learning_rate": 7.862549249130133e-07, + "loss": 0.7165, + "step": 28890 + }, + { + "epoch": 1.1264885302773393, + "grad_norm": 14.148681640625, + "learning_rate": 7.860689584915956e-07, + "loss": 0.7613, + "step": 28900 + }, + { + "epoch": 1.1268783254399812, + "grad_norm": 13.458385467529297, + "learning_rate": 7.858829332206756e-07, + "loss": 0.7657, + "step": 28910 + }, + { + "epoch": 1.1272681206026234, + "grad_norm": 12.891824722290039, + "learning_rate": 7.856968491385218e-07, + "loss": 0.8456, + "step": 28920 + }, + { + "epoch": 1.1276579157652653, + "grad_norm": 11.17689037322998, + "learning_rate": 7.855107062834151e-07, + "loss": 0.7782, + "step": 28930 + }, + { + "epoch": 1.1280477109279075, + "grad_norm": 12.94887638092041, + "learning_rate": 7.853245046936483e-07, + "loss": 0.8039, + "step": 28940 + }, + { + "epoch": 1.1284375060905494, + "grad_norm": 10.716283798217773, + "learning_rate": 7.851382444075266e-07, + "loss": 0.7735, + "step": 28950 + }, + { + "epoch": 1.1288273012531915, + "grad_norm": 13.290563583374023, + "learning_rate": 7.849519254633669e-07, + "loss": 0.8443, + "step": 28960 + }, + { + "epoch": 1.1292170964158335, + "grad_norm": 12.26188850402832, + "learning_rate": 7.847655478994983e-07, + "loss": 0.7846, + "step": 28970 + }, + { + "epoch": 1.1296068915784756, + "grad_norm": 13.153176307678223, + "learning_rate": 7.845791117542623e-07, + "loss": 0.8222, + "step": 28980 + }, + { + "epoch": 1.1299966867411175, + "grad_norm": 11.604106903076172, + "learning_rate": 7.843926170660118e-07, + "loss": 0.7716, + "step": 28990 + }, + { + "epoch": 1.1303864819037597, + "grad_norm": 12.769079208374023, + "learning_rate": 7.842060638731123e-07, + "loss": 0.7506, + "step": 29000 + }, + { + "epoch": 1.1303864819037597, + "eval_loss": 0.9010828137397766, + "eval_runtime": 82.9884, + "eval_samples_per_second": 49.971, + "eval_steps_per_second": 6.254, + "step": 29000 + }, + { + "epoch": 1.1307762770664016, + "grad_norm": 14.361817359924316, + "learning_rate": 7.840194522139409e-07, + "loss": 0.8231, + "step": 29010 + }, + { + "epoch": 1.1311660722290435, + "grad_norm": 11.703439712524414, + "learning_rate": 7.838327821268873e-07, + "loss": 0.8074, + "step": 29020 + }, + { + "epoch": 1.1315558673916857, + "grad_norm": 12.691058158874512, + "learning_rate": 7.836460536503527e-07, + "loss": 0.8192, + "step": 29030 + }, + { + "epoch": 1.1319456625543278, + "grad_norm": 14.171671867370605, + "learning_rate": 7.834592668227505e-07, + "loss": 0.8583, + "step": 29040 + }, + { + "epoch": 1.1323354577169698, + "grad_norm": 11.508492469787598, + "learning_rate": 7.83272421682506e-07, + "loss": 0.7996, + "step": 29050 + }, + { + "epoch": 1.1327252528796117, + "grad_norm": 11.57665729522705, + "learning_rate": 7.830855182680567e-07, + "loss": 0.7707, + "step": 29060 + }, + { + "epoch": 1.1331150480422538, + "grad_norm": 11.781323432922363, + "learning_rate": 7.828985566178519e-07, + "loss": 0.8393, + "step": 29070 + }, + { + "epoch": 1.1335048432048958, + "grad_norm": 11.063751220703125, + "learning_rate": 7.827115367703529e-07, + "loss": 0.7654, + "step": 29080 + }, + { + "epoch": 1.133894638367538, + "grad_norm": 13.466583251953125, + "learning_rate": 7.825244587640334e-07, + "loss": 0.766, + "step": 29090 + }, + { + "epoch": 1.1342844335301798, + "grad_norm": 15.293683052062988, + "learning_rate": 7.823373226373783e-07, + "loss": 0.7472, + "step": 29100 + }, + { + "epoch": 1.134674228692822, + "grad_norm": 11.43502426147461, + "learning_rate": 7.821501284288849e-07, + "loss": 0.7894, + "step": 29110 + }, + { + "epoch": 1.135064023855464, + "grad_norm": 15.778165817260742, + "learning_rate": 7.819628761770626e-07, + "loss": 0.7807, + "step": 29120 + }, + { + "epoch": 1.135453819018106, + "grad_norm": 13.714278221130371, + "learning_rate": 7.817755659204323e-07, + "loss": 0.7718, + "step": 29130 + }, + { + "epoch": 1.135843614180748, + "grad_norm": 14.345159530639648, + "learning_rate": 7.815881976975271e-07, + "loss": 0.8572, + "step": 29140 + }, + { + "epoch": 1.1362334093433901, + "grad_norm": 13.235384941101074, + "learning_rate": 7.814007715468922e-07, + "loss": 0.8407, + "step": 29150 + }, + { + "epoch": 1.136623204506032, + "grad_norm": 13.398663520812988, + "learning_rate": 7.812132875070845e-07, + "loss": 0.7605, + "step": 29160 + }, + { + "epoch": 1.1370129996686742, + "grad_norm": 13.500993728637695, + "learning_rate": 7.810257456166725e-07, + "loss": 0.811, + "step": 29170 + }, + { + "epoch": 1.1374027948313161, + "grad_norm": 11.748517990112305, + "learning_rate": 7.808381459142373e-07, + "loss": 0.7511, + "step": 29180 + }, + { + "epoch": 1.1377925899939583, + "grad_norm": 14.273941993713379, + "learning_rate": 7.806504884383711e-07, + "loss": 0.8009, + "step": 29190 + }, + { + "epoch": 1.1381823851566002, + "grad_norm": 12.604330062866211, + "learning_rate": 7.804627732276788e-07, + "loss": 0.828, + "step": 29200 + }, + { + "epoch": 1.1385721803192421, + "grad_norm": 12.04293441772461, + "learning_rate": 7.802750003207766e-07, + "loss": 0.831, + "step": 29210 + }, + { + "epoch": 1.1389619754818843, + "grad_norm": 14.445537567138672, + "learning_rate": 7.800871697562927e-07, + "loss": 0.7466, + "step": 29220 + }, + { + "epoch": 1.1393517706445264, + "grad_norm": 14.378218650817871, + "learning_rate": 7.798992815728674e-07, + "loss": 0.7887, + "step": 29230 + }, + { + "epoch": 1.1397415658071683, + "grad_norm": 11.73642635345459, + "learning_rate": 7.797113358091525e-07, + "loss": 0.8283, + "step": 29240 + }, + { + "epoch": 1.1401313609698103, + "grad_norm": 12.116954803466797, + "learning_rate": 7.795233325038118e-07, + "loss": 0.7625, + "step": 29250 + }, + { + "epoch": 1.1405211561324524, + "grad_norm": 12.071039199829102, + "learning_rate": 7.79335271695521e-07, + "loss": 0.7537, + "step": 29260 + }, + { + "epoch": 1.1409109512950943, + "grad_norm": 12.283557891845703, + "learning_rate": 7.791471534229676e-07, + "loss": 0.8004, + "step": 29270 + }, + { + "epoch": 1.1413007464577365, + "grad_norm": 12.564621925354004, + "learning_rate": 7.789589777248509e-07, + "loss": 0.8094, + "step": 29280 + }, + { + "epoch": 1.1416905416203784, + "grad_norm": 12.00485897064209, + "learning_rate": 7.787707446398816e-07, + "loss": 0.7627, + "step": 29290 + }, + { + "epoch": 1.1420803367830206, + "grad_norm": 12.168889999389648, + "learning_rate": 7.785824542067833e-07, + "loss": 0.8163, + "step": 29300 + }, + { + "epoch": 1.1424701319456625, + "grad_norm": 12.260193824768066, + "learning_rate": 7.783941064642904e-07, + "loss": 0.7577, + "step": 29310 + }, + { + "epoch": 1.1428599271083046, + "grad_norm": 12.84726619720459, + "learning_rate": 7.782057014511492e-07, + "loss": 0.7747, + "step": 29320 + }, + { + "epoch": 1.1432497222709466, + "grad_norm": 12.008340835571289, + "learning_rate": 7.780172392061182e-07, + "loss": 0.7994, + "step": 29330 + }, + { + "epoch": 1.1436395174335887, + "grad_norm": 13.106820106506348, + "learning_rate": 7.778287197679676e-07, + "loss": 0.802, + "step": 29340 + }, + { + "epoch": 1.1440293125962306, + "grad_norm": 13.791325569152832, + "learning_rate": 7.776401431754789e-07, + "loss": 0.783, + "step": 29350 + }, + { + "epoch": 1.1444191077588728, + "grad_norm": 14.352214813232422, + "learning_rate": 7.774515094674459e-07, + "loss": 0.7919, + "step": 29360 + }, + { + "epoch": 1.1448089029215147, + "grad_norm": 13.447537422180176, + "learning_rate": 7.772628186826736e-07, + "loss": 0.7968, + "step": 29370 + }, + { + "epoch": 1.1451986980841569, + "grad_norm": 13.773513793945312, + "learning_rate": 7.770740708599795e-07, + "loss": 0.8069, + "step": 29380 + }, + { + "epoch": 1.1455884932467988, + "grad_norm": 11.33832836151123, + "learning_rate": 7.768852660381921e-07, + "loss": 0.7923, + "step": 29390 + }, + { + "epoch": 1.1459782884094407, + "grad_norm": 12.28536605834961, + "learning_rate": 7.766964042561524e-07, + "loss": 0.8604, + "step": 29400 + }, + { + "epoch": 1.1463680835720829, + "grad_norm": 11.981461524963379, + "learning_rate": 7.765074855527119e-07, + "loss": 0.8259, + "step": 29410 + }, + { + "epoch": 1.146757878734725, + "grad_norm": 14.30960750579834, + "learning_rate": 7.763185099667352e-07, + "loss": 0.8626, + "step": 29420 + }, + { + "epoch": 1.147147673897367, + "grad_norm": 12.49848461151123, + "learning_rate": 7.761294775370975e-07, + "loss": 0.7982, + "step": 29430 + }, + { + "epoch": 1.1475374690600089, + "grad_norm": 13.34499740600586, + "learning_rate": 7.759403883026865e-07, + "loss": 0.8197, + "step": 29440 + }, + { + "epoch": 1.147927264222651, + "grad_norm": 13.879630088806152, + "learning_rate": 7.75751242302401e-07, + "loss": 0.7655, + "step": 29450 + }, + { + "epoch": 1.148317059385293, + "grad_norm": 13.396138191223145, + "learning_rate": 7.755620395751521e-07, + "loss": 0.822, + "step": 29460 + }, + { + "epoch": 1.148706854547935, + "grad_norm": 13.903852462768555, + "learning_rate": 7.753727801598617e-07, + "loss": 0.8149, + "step": 29470 + }, + { + "epoch": 1.149096649710577, + "grad_norm": 14.990713119506836, + "learning_rate": 7.75183464095464e-07, + "loss": 0.9105, + "step": 29480 + }, + { + "epoch": 1.1494864448732192, + "grad_norm": 12.362504005432129, + "learning_rate": 7.749940914209049e-07, + "loss": 0.731, + "step": 29490 + }, + { + "epoch": 1.149876240035861, + "grad_norm": 12.535857200622559, + "learning_rate": 7.748046621751414e-07, + "loss": 0.7677, + "step": 29500 + }, + { + "epoch": 1.149876240035861, + "eval_loss": 0.8976953029632568, + "eval_runtime": 83.2291, + "eval_samples_per_second": 49.826, + "eval_steps_per_second": 6.236, + "step": 29500 + }, + { + "epoch": 1.1502660351985032, + "grad_norm": 13.507098197937012, + "learning_rate": 7.746151763971428e-07, + "loss": 0.8148, + "step": 29510 + }, + { + "epoch": 1.1506558303611452, + "grad_norm": 15.301827430725098, + "learning_rate": 7.744256341258897e-07, + "loss": 0.7788, + "step": 29520 + }, + { + "epoch": 1.1510456255237873, + "grad_norm": 12.410253524780273, + "learning_rate": 7.74236035400374e-07, + "loss": 0.8064, + "step": 29530 + }, + { + "epoch": 1.1514354206864292, + "grad_norm": 14.277114868164062, + "learning_rate": 7.740463802596001e-07, + "loss": 0.8227, + "step": 29540 + }, + { + "epoch": 1.1518252158490714, + "grad_norm": 13.844356536865234, + "learning_rate": 7.738566687425828e-07, + "loss": 0.7687, + "step": 29550 + }, + { + "epoch": 1.1522150110117133, + "grad_norm": 14.193256378173828, + "learning_rate": 7.736669008883497e-07, + "loss": 0.8008, + "step": 29560 + }, + { + "epoch": 1.1526048061743555, + "grad_norm": 11.465084075927734, + "learning_rate": 7.734770767359391e-07, + "loss": 0.8133, + "step": 29570 + }, + { + "epoch": 1.1529946013369974, + "grad_norm": 12.590509414672852, + "learning_rate": 7.732871963244016e-07, + "loss": 0.7098, + "step": 29580 + }, + { + "epoch": 1.1533843964996395, + "grad_norm": 14.244277000427246, + "learning_rate": 7.730972596927985e-07, + "loss": 0.796, + "step": 29590 + }, + { + "epoch": 1.1537741916622815, + "grad_norm": 15.2111177444458, + "learning_rate": 7.729072668802036e-07, + "loss": 0.8175, + "step": 29600 + }, + { + "epoch": 1.1541639868249236, + "grad_norm": 12.006050109863281, + "learning_rate": 7.727172179257016e-07, + "loss": 0.8188, + "step": 29610 + }, + { + "epoch": 1.1545537819875655, + "grad_norm": 14.53686237335205, + "learning_rate": 7.72527112868389e-07, + "loss": 0.8523, + "step": 29620 + }, + { + "epoch": 1.1549435771502075, + "grad_norm": 12.198389053344727, + "learning_rate": 7.723369517473738e-07, + "loss": 0.7883, + "step": 29630 + }, + { + "epoch": 1.1553333723128496, + "grad_norm": 12.871670722961426, + "learning_rate": 7.721467346017755e-07, + "loss": 0.8193, + "step": 29640 + }, + { + "epoch": 1.1557231674754918, + "grad_norm": 14.947281837463379, + "learning_rate": 7.719564614707253e-07, + "loss": 0.729, + "step": 29650 + }, + { + "epoch": 1.1561129626381337, + "grad_norm": 12.422595024108887, + "learning_rate": 7.717661323933657e-07, + "loss": 0.8567, + "step": 29660 + }, + { + "epoch": 1.1565027578007756, + "grad_norm": 12.203004837036133, + "learning_rate": 7.715757474088509e-07, + "loss": 0.7914, + "step": 29670 + }, + { + "epoch": 1.1568925529634178, + "grad_norm": 12.309530258178711, + "learning_rate": 7.713853065563464e-07, + "loss": 0.8302, + "step": 29680 + }, + { + "epoch": 1.1572823481260597, + "grad_norm": 14.704010963439941, + "learning_rate": 7.711948098750292e-07, + "loss": 0.7845, + "step": 29690 + }, + { + "epoch": 1.1576721432887018, + "grad_norm": 13.431638717651367, + "learning_rate": 7.710042574040883e-07, + "loss": 0.8068, + "step": 29700 + }, + { + "epoch": 1.1580619384513438, + "grad_norm": 11.632168769836426, + "learning_rate": 7.708136491827231e-07, + "loss": 0.7685, + "step": 29710 + }, + { + "epoch": 1.158451733613986, + "grad_norm": 13.139909744262695, + "learning_rate": 7.706229852501456e-07, + "loss": 0.8153, + "step": 29720 + }, + { + "epoch": 1.1588415287766278, + "grad_norm": 12.768489837646484, + "learning_rate": 7.704322656455786e-07, + "loss": 0.843, + "step": 29730 + }, + { + "epoch": 1.15923132393927, + "grad_norm": 15.098163604736328, + "learning_rate": 7.702414904082565e-07, + "loss": 0.7564, + "step": 29740 + }, + { + "epoch": 1.159621119101912, + "grad_norm": 13.74183177947998, + "learning_rate": 7.700506595774252e-07, + "loss": 0.8179, + "step": 29750 + }, + { + "epoch": 1.160010914264554, + "grad_norm": 12.652956008911133, + "learning_rate": 7.698597731923422e-07, + "loss": 0.8209, + "step": 29760 + }, + { + "epoch": 1.160400709427196, + "grad_norm": 12.253880500793457, + "learning_rate": 7.696688312922763e-07, + "loss": 0.807, + "step": 29770 + }, + { + "epoch": 1.1607905045898381, + "grad_norm": 12.988066673278809, + "learning_rate": 7.694778339165071e-07, + "loss": 0.7671, + "step": 29780 + }, + { + "epoch": 1.16118029975248, + "grad_norm": 10.315251350402832, + "learning_rate": 7.692867811043267e-07, + "loss": 0.7646, + "step": 29790 + }, + { + "epoch": 1.1615700949151222, + "grad_norm": 11.435159683227539, + "learning_rate": 7.690956728950381e-07, + "loss": 0.7574, + "step": 29800 + }, + { + "epoch": 1.1619598900777641, + "grad_norm": 11.66455078125, + "learning_rate": 7.689045093279554e-07, + "loss": 0.7768, + "step": 29810 + }, + { + "epoch": 1.162349685240406, + "grad_norm": 13.519287109375, + "learning_rate": 7.687132904424044e-07, + "loss": 0.8378, + "step": 29820 + }, + { + "epoch": 1.1627394804030482, + "grad_norm": 12.337118148803711, + "learning_rate": 7.685220162777224e-07, + "loss": 0.8206, + "step": 29830 + }, + { + "epoch": 1.1631292755656903, + "grad_norm": 13.620658874511719, + "learning_rate": 7.683306868732578e-07, + "loss": 0.8243, + "step": 29840 + }, + { + "epoch": 1.1635190707283323, + "grad_norm": 11.043423652648926, + "learning_rate": 7.681393022683706e-07, + "loss": 0.7838, + "step": 29850 + }, + { + "epoch": 1.1639088658909742, + "grad_norm": 11.704632759094238, + "learning_rate": 7.679478625024319e-07, + "loss": 0.809, + "step": 29860 + }, + { + "epoch": 1.1642986610536163, + "grad_norm": 12.972027778625488, + "learning_rate": 7.677563676148245e-07, + "loss": 0.8134, + "step": 29870 + }, + { + "epoch": 1.1646884562162583, + "grad_norm": 12.942703247070312, + "learning_rate": 7.675648176449418e-07, + "loss": 0.8012, + "step": 29880 + }, + { + "epoch": 1.1650782513789004, + "grad_norm": 14.840851783752441, + "learning_rate": 7.673732126321897e-07, + "loss": 0.826, + "step": 29890 + }, + { + "epoch": 1.1654680465415423, + "grad_norm": 11.833650588989258, + "learning_rate": 7.671815526159844e-07, + "loss": 0.7951, + "step": 29900 + }, + { + "epoch": 1.1658578417041845, + "grad_norm": 14.088912963867188, + "learning_rate": 7.66989837635754e-07, + "loss": 0.7863, + "step": 29910 + }, + { + "epoch": 1.1662476368668264, + "grad_norm": 13.074752807617188, + "learning_rate": 7.667980677309375e-07, + "loss": 0.7999, + "step": 29920 + }, + { + "epoch": 1.1666374320294686, + "grad_norm": 10.753192901611328, + "learning_rate": 7.666062429409856e-07, + "loss": 0.7644, + "step": 29930 + }, + { + "epoch": 1.1670272271921105, + "grad_norm": 12.945324897766113, + "learning_rate": 7.664143633053599e-07, + "loss": 0.7434, + "step": 29940 + }, + { + "epoch": 1.1674170223547526, + "grad_norm": 12.38203239440918, + "learning_rate": 7.662224288635337e-07, + "loss": 0.7791, + "step": 29950 + }, + { + "epoch": 1.1678068175173946, + "grad_norm": 15.810951232910156, + "learning_rate": 7.66030439654991e-07, + "loss": 0.7892, + "step": 29960 + }, + { + "epoch": 1.1681966126800367, + "grad_norm": 13.435308456420898, + "learning_rate": 7.658383957192277e-07, + "loss": 0.8473, + "step": 29970 + }, + { + "epoch": 1.1685864078426786, + "grad_norm": 11.759345054626465, + "learning_rate": 7.656462970957507e-07, + "loss": 0.7471, + "step": 29980 + }, + { + "epoch": 1.1689762030053208, + "grad_norm": 12.792527198791504, + "learning_rate": 7.65454143824078e-07, + "loss": 0.7935, + "step": 29990 + }, + { + "epoch": 1.1693659981679627, + "grad_norm": 14.981736183166504, + "learning_rate": 7.652619359437389e-07, + "loss": 0.8697, + "step": 30000 + }, + { + "epoch": 1.1693659981679627, + "eval_loss": 0.8986621499061584, + "eval_runtime": 82.8989, + "eval_samples_per_second": 50.025, + "eval_steps_per_second": 6.261, + "step": 30000 + } + ], + "logging_steps": 10, + "max_steps": 76962, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 15000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.1099801289599484e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}