| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.9981634527089072, | |
| "eval_steps": 500, | |
| "global_step": 1088, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.018365472910927456, | |
| "grad_norm": 0.27851340305035927, | |
| "learning_rate": 9.997915740506687e-05, | |
| "loss": 1.0445, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03673094582185491, | |
| "grad_norm": 0.061819145479423056, | |
| "learning_rate": 9.991664699681799e-05, | |
| "loss": 0.5539, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.05509641873278237, | |
| "grad_norm": 0.03760363481720962, | |
| "learning_rate": 9.981252089041809e-05, | |
| "loss": 0.4787, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.07346189164370982, | |
| "grad_norm": 0.0313950706656064, | |
| "learning_rate": 9.966686589619751e-05, | |
| "loss": 0.468, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.09182736455463728, | |
| "grad_norm": 0.022877630980033918, | |
| "learning_rate": 9.947980344727799e-05, | |
| "loss": 0.4451, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.11019283746556474, | |
| "grad_norm": 0.029933452784706518, | |
| "learning_rate": 9.925148949833355e-05, | |
| "loss": 0.4469, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.1285583103764922, | |
| "grad_norm": 0.023764995733481963, | |
| "learning_rate": 9.898211439557042e-05, | |
| "loss": 0.4424, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.14692378328741965, | |
| "grad_norm": 0.02327001477570512, | |
| "learning_rate": 9.867190271803465e-05, | |
| "loss": 0.4336, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.1652892561983471, | |
| "grad_norm": 0.024539265266611217, | |
| "learning_rate": 9.832111309037979e-05, | |
| "loss": 0.4321, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.18365472910927455, | |
| "grad_norm": 0.039657236107179715, | |
| "learning_rate": 9.793003796725048e-05, | |
| "loss": 0.4384, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.20202020202020202, | |
| "grad_norm": 0.027774934175057744, | |
| "learning_rate": 9.749900338946192e-05, | |
| "loss": 0.4283, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.22038567493112948, | |
| "grad_norm": 0.02480612425885802, | |
| "learning_rate": 9.702836871217839e-05, | |
| "loss": 0.4327, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.23875114784205692, | |
| "grad_norm": 0.025217057933479833, | |
| "learning_rate": 9.651852630531748e-05, | |
| "loss": 0.4191, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2571166207529844, | |
| "grad_norm": 0.042028378260899814, | |
| "learning_rate": 9.596990122642983e-05, | |
| "loss": 0.4171, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.27548209366391185, | |
| "grad_norm": 0.030731629233234352, | |
| "learning_rate": 9.538295086632703e-05, | |
| "loss": 0.4273, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.2938475665748393, | |
| "grad_norm": 0.06006724597666194, | |
| "learning_rate": 9.475816456775313e-05, | |
| "loss": 0.4266, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3122130394857668, | |
| "grad_norm": 0.02893865752550571, | |
| "learning_rate": 9.409606321741775e-05, | |
| "loss": 0.4196, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.3305785123966942, | |
| "grad_norm": 0.026458687435744248, | |
| "learning_rate": 9.339719881173093e-05, | |
| "loss": 0.423, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.34894398530762166, | |
| "grad_norm": 0.03576661132708211, | |
| "learning_rate": 9.266215399660146e-05, | |
| "loss": 0.4192, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.3673094582185491, | |
| "grad_norm": 0.02652249141257209, | |
| "learning_rate": 9.189154158168292e-05, | |
| "loss": 0.4164, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.3856749311294766, | |
| "grad_norm": 0.025216448544065753, | |
| "learning_rate": 9.108600402947192e-05, | |
| "loss": 0.4272, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.40404040404040403, | |
| "grad_norm": 0.026533959286346408, | |
| "learning_rate": 9.02462129196846e-05, | |
| "loss": 0.4189, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.42240587695133147, | |
| "grad_norm": 0.026029222217085504, | |
| "learning_rate": 8.93728683893582e-05, | |
| "loss": 0.4207, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.44077134986225897, | |
| "grad_norm": 0.028412018887179133, | |
| "learning_rate": 8.846669854914396e-05, | |
| "loss": 0.417, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.4591368227731864, | |
| "grad_norm": 0.026783824853837783, | |
| "learning_rate": 8.752845887627872e-05, | |
| "loss": 0.4147, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.47750229568411384, | |
| "grad_norm": 0.026947377894842927, | |
| "learning_rate": 8.655893158474055e-05, | |
| "loss": 0.4179, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.49586776859504134, | |
| "grad_norm": 0.03613738404815021, | |
| "learning_rate": 8.555892497311402e-05, | |
| "loss": 0.4068, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.5142332415059688, | |
| "grad_norm": 0.027745465710559532, | |
| "learning_rate": 8.452927275070858e-05, | |
| "loss": 0.4103, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.5325987144168962, | |
| "grad_norm": 0.025444027330206734, | |
| "learning_rate": 8.347083334249199e-05, | |
| "loss": 0.4009, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.5509641873278237, | |
| "grad_norm": 0.029718700126100683, | |
| "learning_rate": 8.23844891734181e-05, | |
| "loss": 0.4183, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.5693296602387512, | |
| "grad_norm": 0.029941432231859456, | |
| "learning_rate": 8.12711459327459e-05, | |
| "loss": 0.4199, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.5876951331496786, | |
| "grad_norm": 0.02728726308576057, | |
| "learning_rate": 8.013173181896283e-05, | |
| "loss": 0.4023, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.6060606060606061, | |
| "grad_norm": 0.02842461149504098, | |
| "learning_rate": 7.89671967659423e-05, | |
| "loss": 0.4185, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.6244260789715336, | |
| "grad_norm": 0.02557752553781187, | |
| "learning_rate": 7.777851165098012e-05, | |
| "loss": 0.4259, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.642791551882461, | |
| "grad_norm": 0.026972728188092137, | |
| "learning_rate": 7.656666748537045e-05, | |
| "loss": 0.4039, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.6611570247933884, | |
| "grad_norm": 0.024371307606132524, | |
| "learning_rate": 7.533267458819598e-05, | |
| "loss": 0.4216, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.6795224977043158, | |
| "grad_norm": 0.031096864607523835, | |
| "learning_rate": 7.407756174402089e-05, | |
| "loss": 0.4112, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.6978879706152433, | |
| "grad_norm": 0.027930614232647415, | |
| "learning_rate": 7.280237534518947e-05, | |
| "loss": 0.4154, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.7162534435261708, | |
| "grad_norm": 0.02788401233462263, | |
| "learning_rate": 7.150817851944472e-05, | |
| "loss": 0.4079, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7346189164370982, | |
| "grad_norm": 0.030760734784571572, | |
| "learning_rate": 7.019605024359474e-05, | |
| "loss": 0.411, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.7529843893480257, | |
| "grad_norm": 0.025453805261318345, | |
| "learning_rate": 6.886708444396573e-05, | |
| "loss": 0.4065, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.7713498622589532, | |
| "grad_norm": 0.028933030043804563, | |
| "learning_rate": 6.75223890843913e-05, | |
| "loss": 0.4171, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.7897153351698806, | |
| "grad_norm": 0.026727439486227408, | |
| "learning_rate": 6.6163085242499e-05, | |
| "loss": 0.408, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.8080808080808081, | |
| "grad_norm": 0.028829235828263126, | |
| "learning_rate": 6.479030617506353e-05, | |
| "loss": 0.4148, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.8264462809917356, | |
| "grad_norm": 0.02837397571927983, | |
| "learning_rate": 6.34051963732063e-05, | |
| "loss": 0.4142, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.8448117539026629, | |
| "grad_norm": 0.030025509075918014, | |
| "learning_rate": 6.200891060822883e-05, | |
| "loss": 0.4248, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.8631772268135904, | |
| "grad_norm": 0.02918663842326674, | |
| "learning_rate": 6.060261296887554e-05, | |
| "loss": 0.4106, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.8815426997245179, | |
| "grad_norm": 0.030657261653125826, | |
| "learning_rate": 5.918747589082853e-05, | |
| "loss": 0.4118, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.8999081726354453, | |
| "grad_norm": 0.029002395463699617, | |
| "learning_rate": 5.776467917924348e-05, | |
| "loss": 0.409, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.9182736455463728, | |
| "grad_norm": 0.02963746695292411, | |
| "learning_rate": 5.6335409025141694e-05, | |
| "loss": 0.406, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.9366391184573003, | |
| "grad_norm": 0.03089177834981146, | |
| "learning_rate": 5.490085701647805e-05, | |
| "loss": 0.4102, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.9550045913682277, | |
| "grad_norm": 0.032695907122492804, | |
| "learning_rate": 5.346221914470959e-05, | |
| "loss": 0.4025, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.9733700642791552, | |
| "grad_norm": 0.0283784844379621, | |
| "learning_rate": 5.2020694807693015e-05, | |
| "loss": 0.401, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.9917355371900827, | |
| "grad_norm": 0.02859989539113277, | |
| "learning_rate": 5.0577485809742044e-05, | |
| "loss": 0.4163, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.0101010101010102, | |
| "grad_norm": 0.03103627192686722, | |
| "learning_rate": 4.913379535967859e-05, | |
| "loss": 0.4039, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.0284664830119377, | |
| "grad_norm": 0.03322012900664279, | |
| "learning_rate": 4.7690827067713035e-05, | |
| "loss": 0.3914, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.046831955922865, | |
| "grad_norm": 0.03642666862930741, | |
| "learning_rate": 4.6249783941989785e-05, | |
| "loss": 0.3908, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.0651974288337924, | |
| "grad_norm": 0.0336389842141535, | |
| "learning_rate": 4.481186738563492e-05, | |
| "loss": 0.3916, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.08356290174472, | |
| "grad_norm": 0.03449706815272795, | |
| "learning_rate": 4.3378276195141665e-05, | |
| "loss": 0.3858, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.1019283746556474, | |
| "grad_norm": 0.036410754456203426, | |
| "learning_rate": 4.195020556092935e-05, | |
| "loss": 0.3837, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.120293847566575, | |
| "grad_norm": 0.03485127798044261, | |
| "learning_rate": 4.05288460709086e-05, | |
| "loss": 0.3904, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.1386593204775024, | |
| "grad_norm": 0.038047760695129655, | |
| "learning_rate": 3.911538271788358e-05, | |
| "loss": 0.3964, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.1570247933884297, | |
| "grad_norm": 0.038412283865629844, | |
| "learning_rate": 3.7710993911619094e-05, | |
| "loss": 0.3774, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.1753902662993572, | |
| "grad_norm": 0.036371859571502144, | |
| "learning_rate": 3.631685049639586e-05, | |
| "loss": 0.3907, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.1937557392102847, | |
| "grad_norm": 0.03976682097907334, | |
| "learning_rate": 3.493411477487315e-05, | |
| "loss": 0.3858, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.2121212121212122, | |
| "grad_norm": 0.037005097607228386, | |
| "learning_rate": 3.3563939539072705e-05, | |
| "loss": 0.3847, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.2304866850321396, | |
| "grad_norm": 0.03996197909827653, | |
| "learning_rate": 3.2207467109291586e-05, | |
| "loss": 0.3979, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.248852157943067, | |
| "grad_norm": 0.04158286829796218, | |
| "learning_rate": 3.086582838174551e-05, | |
| "loss": 0.39, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.2672176308539944, | |
| "grad_norm": 0.04139047858760264, | |
| "learning_rate": 2.9540141885736262e-05, | |
| "loss": 0.3874, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.285583103764922, | |
| "grad_norm": 0.041594762089601535, | |
| "learning_rate": 2.8231512851129593e-05, | |
| "loss": 0.3934, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.3039485766758494, | |
| "grad_norm": 0.0431336076994283, | |
| "learning_rate": 2.6941032286920985e-05, | |
| "loss": 0.3775, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.322314049586777, | |
| "grad_norm": 0.04035918987424071, | |
| "learning_rate": 2.5669776071657192e-05, | |
| "loss": 0.4017, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.3406795224977044, | |
| "grad_norm": 0.0417444272173153, | |
| "learning_rate": 2.4418804056472227e-05, | |
| "loss": 0.3799, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.3590449954086319, | |
| "grad_norm": 0.04538570278878064, | |
| "learning_rate": 2.3189159181485516e-05, | |
| "loss": 0.3921, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.3774104683195592, | |
| "grad_norm": 0.043013131185681716, | |
| "learning_rate": 2.1981866606298683e-05, | |
| "loss": 0.3857, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.3957759412304866, | |
| "grad_norm": 0.043552361092639345, | |
| "learning_rate": 2.079793285531618e-05, | |
| "loss": 0.3937, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.4141414141414141, | |
| "grad_norm": 0.041014272519928036, | |
| "learning_rate": 1.963834497860192e-05, | |
| "loss": 0.3801, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.4325068870523416, | |
| "grad_norm": 0.03865452597618009, | |
| "learning_rate": 1.8504069728972123e-05, | |
| "loss": 0.3834, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.4508723599632691, | |
| "grad_norm": 0.04173720509030274, | |
| "learning_rate": 1.7396052756009574e-05, | |
| "loss": 0.3829, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.4692378328741964, | |
| "grad_norm": 0.04256905712895654, | |
| "learning_rate": 1.631521781767214e-05, | |
| "loss": 0.3795, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.487603305785124, | |
| "grad_norm": 0.0473389022063686, | |
| "learning_rate": 1.52624660101522e-05, | |
| "loss": 0.3801, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.5059687786960514, | |
| "grad_norm": 0.0391194792265315, | |
| "learning_rate": 1.4238675016629338e-05, | |
| "loss": 0.3803, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.5243342516069789, | |
| "grad_norm": 0.046767033474025464, | |
| "learning_rate": 1.3244698375542491e-05, | |
| "loss": 0.3963, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.5426997245179064, | |
| "grad_norm": 0.042865074045739464, | |
| "learning_rate": 1.2281364768991804e-05, | |
| "loss": 0.3765, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.5610651974288339, | |
| "grad_norm": 0.04190703980014869, | |
| "learning_rate": 1.134947733186315e-05, | |
| "loss": 0.3761, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.5794306703397614, | |
| "grad_norm": 0.04183784386577142, | |
| "learning_rate": 1.0449812982251556e-05, | |
| "loss": 0.3852, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.5977961432506889, | |
| "grad_norm": 0.03879386200844887, | |
| "learning_rate": 9.58312177374157e-06, | |
| "loss": 0.3775, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.6161616161616161, | |
| "grad_norm": 0.0412465475003067, | |
| "learning_rate": 8.75012627008489e-06, | |
| "loss": 0.3818, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.6345270890725436, | |
| "grad_norm": 0.04486626572785468, | |
| "learning_rate": 7.951520942796025e-06, | |
| "loss": 0.3871, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.6528925619834711, | |
| "grad_norm": 0.040934222197984614, | |
| "learning_rate": 7.187971592168935e-06, | |
| "loss": 0.3739, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.6712580348943984, | |
| "grad_norm": 0.0440521821621867, | |
| "learning_rate": 6.460114792196642e-06, | |
| "loss": 0.4001, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.6896235078053259, | |
| "grad_norm": 0.04130542221023784, | |
| "learning_rate": 5.768557359857241e-06, | |
| "loss": 0.3833, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.7079889807162534, | |
| "grad_norm": 0.041016879516671184, | |
| "learning_rate": 5.113875849208099e-06, | |
| "loss": 0.3871, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.7263544536271809, | |
| "grad_norm": 0.04458130921484639, | |
| "learning_rate": 4.4966160707107076e-06, | |
| "loss": 0.3924, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.7447199265381084, | |
| "grad_norm": 0.04483909598982454, | |
| "learning_rate": 3.917292636186332e-06, | |
| "loss": 0.3868, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.7630853994490359, | |
| "grad_norm": 0.04374641202578716, | |
| "learning_rate": 3.376388529782215e-06, | |
| "loss": 0.3799, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.7814508723599634, | |
| "grad_norm": 0.04643290880252759, | |
| "learning_rate": 2.8743547053058427e-06, | |
| "loss": 0.3873, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.7998163452708908, | |
| "grad_norm": 0.04253533673018247, | |
| "learning_rate": 2.4116097102630907e-06, | |
| "loss": 0.3851, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.8181818181818183, | |
| "grad_norm": 0.039915967047357025, | |
| "learning_rate": 1.9885393369134976e-06, | |
| "loss": 0.3773, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.8365472910927456, | |
| "grad_norm": 0.04344253083810866, | |
| "learning_rate": 1.6054963006338742e-06, | |
| "loss": 0.3808, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.854912764003673, | |
| "grad_norm": 0.040517415128882496, | |
| "learning_rate": 1.2627999458580953e-06, | |
| "loss": 0.376, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.8732782369146006, | |
| "grad_norm": 0.040775411667955175, | |
| "learning_rate": 9.607359798384785e-07, | |
| "loss": 0.3824, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.8916437098255279, | |
| "grad_norm": 0.04533369281933566, | |
| "learning_rate": 6.995562344505214e-07, | |
| "loss": 0.3894, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.9100091827364554, | |
| "grad_norm": 0.0469790386931614, | |
| "learning_rate": 4.794784562397458e-07, | |
| "loss": 0.3746, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.9283746556473829, | |
| "grad_norm": 0.04179832849248068, | |
| "learning_rate": 3.006861248855408e-07, | |
| "loss": 0.3904, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.9467401285583104, | |
| "grad_norm": 0.04097669604600613, | |
| "learning_rate": 1.6332830023350064e-07, | |
| "loss": 0.3905, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.9651056014692379, | |
| "grad_norm": 0.043233457078553805, | |
| "learning_rate": 6.751949802362711e-08, | |
| "loss": 0.3777, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.9834710743801653, | |
| "grad_norm": 0.0432832073055713, | |
| "learning_rate": 1.3339594418138035e-08, | |
| "loss": 0.3719, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.9981634527089072, | |
| "step": 1088, | |
| "total_flos": 525248790528000.0, | |
| "train_loss": 0.4096706003388938, | |
| "train_runtime": 32183.0135, | |
| "train_samples_per_second": 4.058, | |
| "train_steps_per_second": 0.034 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1088, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 5000, | |
| "total_flos": 525248790528000.0, | |
| "train_batch_size": 10, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |