{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 5210, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009601536245799327, "grad_norm": 4.577141838896869, "learning_rate": 0.0, "loss": 1.026, "step": 1 }, { "epoch": 0.0019203072491598655, "grad_norm": 4.33984635778259, "learning_rate": 1.9193857965451055e-08, "loss": 0.9978, "step": 2 }, { "epoch": 0.0028804608737397984, "grad_norm": 4.238395010219315, "learning_rate": 3.838771593090211e-08, "loss": 0.972, "step": 3 }, { "epoch": 0.003840614498319731, "grad_norm": 3.861568907659275, "learning_rate": 5.758157389635317e-08, "loss": 0.798, "step": 4 }, { "epoch": 0.004800768122899664, "grad_norm": 4.154739873902508, "learning_rate": 7.677543186180422e-08, "loss": 1.0605, "step": 5 }, { "epoch": 0.005760921747479597, "grad_norm": 4.661923581887783, "learning_rate": 9.596928982725528e-08, "loss": 0.9341, "step": 6 }, { "epoch": 0.00672107537205953, "grad_norm": 4.535615329059657, "learning_rate": 1.1516314779270635e-07, "loss": 1.0143, "step": 7 }, { "epoch": 0.007681228996639462, "grad_norm": 3.9367669172969366, "learning_rate": 1.343570057581574e-07, "loss": 0.9396, "step": 8 }, { "epoch": 0.008641382621219395, "grad_norm": 3.939598752371097, "learning_rate": 1.5355086372360844e-07, "loss": 0.8451, "step": 9 }, { "epoch": 0.009601536245799328, "grad_norm": 3.712496280853144, "learning_rate": 1.7274472168905953e-07, "loss": 0.9018, "step": 10 }, { "epoch": 0.01056168987037926, "grad_norm": 4.1053756903052685, "learning_rate": 1.9193857965451055e-07, "loss": 0.916, "step": 11 }, { "epoch": 0.011521843494959194, "grad_norm": 4.3564642663005575, "learning_rate": 2.1113243761996164e-07, "loss": 0.9316, "step": 12 }, { "epoch": 0.012481997119539127, "grad_norm": 4.055364942831175, "learning_rate": 2.303262955854127e-07, "loss": 1.1022, "step": 13 }, { "epoch": 0.01344215074411906, "grad_norm": 5.008142380100818, "learning_rate": 2.4952015355086375e-07, "loss": 1.0371, "step": 14 }, { "epoch": 0.014402304368698993, "grad_norm": 4.262852332595414, "learning_rate": 2.687140115163148e-07, "loss": 0.9269, "step": 15 }, { "epoch": 0.015362457993278924, "grad_norm": 3.6823007289333525, "learning_rate": 2.8790786948176586e-07, "loss": 1.0006, "step": 16 }, { "epoch": 0.01632261161785886, "grad_norm": 3.6212713457229304, "learning_rate": 3.071017274472169e-07, "loss": 0.9386, "step": 17 }, { "epoch": 0.01728276524243879, "grad_norm": 3.6096643206155736, "learning_rate": 3.262955854126679e-07, "loss": 0.9339, "step": 18 }, { "epoch": 0.018242918867018725, "grad_norm": 3.409733479482879, "learning_rate": 3.4548944337811905e-07, "loss": 1.0225, "step": 19 }, { "epoch": 0.019203072491598656, "grad_norm": 3.2497717755470137, "learning_rate": 3.646833013435701e-07, "loss": 0.9405, "step": 20 }, { "epoch": 0.020163226116178587, "grad_norm": 3.650955685908668, "learning_rate": 3.838771593090211e-07, "loss": 0.9195, "step": 21 }, { "epoch": 0.02112337974075852, "grad_norm": 3.5003667751435876, "learning_rate": 4.0307101727447224e-07, "loss": 0.9831, "step": 22 }, { "epoch": 0.022083533365338453, "grad_norm": 3.4506557420558748, "learning_rate": 4.2226487523992327e-07, "loss": 1.0375, "step": 23 }, { "epoch": 0.023043686989918388, "grad_norm": 3.285126430330755, "learning_rate": 4.414587332053743e-07, "loss": 0.9756, "step": 24 }, { "epoch": 0.02400384061449832, "grad_norm": 2.9222572881725277, "learning_rate": 4.606525911708254e-07, "loss": 0.9868, "step": 25 }, { "epoch": 0.024963994239078253, "grad_norm": 3.459339083134459, "learning_rate": 4.798464491362765e-07, "loss": 0.9448, "step": 26 }, { "epoch": 0.025924147863658185, "grad_norm": 2.9732846541313522, "learning_rate": 4.990403071017275e-07, "loss": 0.9519, "step": 27 }, { "epoch": 0.02688430148823812, "grad_norm": 3.31554879406097, "learning_rate": 5.182341650671785e-07, "loss": 0.9141, "step": 28 }, { "epoch": 0.02784445511281805, "grad_norm": 3.3798887281294734, "learning_rate": 5.374280230326296e-07, "loss": 0.8705, "step": 29 }, { "epoch": 0.028804608737397985, "grad_norm": 2.742522537652992, "learning_rate": 5.566218809980807e-07, "loss": 0.8737, "step": 30 }, { "epoch": 0.029764762361977917, "grad_norm": 2.6529384154527045, "learning_rate": 5.758157389635317e-07, "loss": 0.9293, "step": 31 }, { "epoch": 0.030724915986557848, "grad_norm": 2.417535057087098, "learning_rate": 5.950095969289827e-07, "loss": 1.0313, "step": 32 }, { "epoch": 0.03168506961113778, "grad_norm": 2.160063288497681, "learning_rate": 6.142034548944338e-07, "loss": 0.9414, "step": 33 }, { "epoch": 0.03264522323571772, "grad_norm": 2.074020408221015, "learning_rate": 6.333973128598848e-07, "loss": 0.8629, "step": 34 }, { "epoch": 0.033605376860297645, "grad_norm": 1.8430847763516858, "learning_rate": 6.525911708253358e-07, "loss": 0.9071, "step": 35 }, { "epoch": 0.03456553048487758, "grad_norm": 1.9741468365769106, "learning_rate": 6.717850287907871e-07, "loss": 0.9168, "step": 36 }, { "epoch": 0.035525684109457514, "grad_norm": 1.9310913369096117, "learning_rate": 6.909788867562381e-07, "loss": 0.993, "step": 37 }, { "epoch": 0.03648583773403745, "grad_norm": 1.7692537692594104, "learning_rate": 7.101727447216891e-07, "loss": 0.8917, "step": 38 }, { "epoch": 0.03744599135861738, "grad_norm": 1.6857649916792743, "learning_rate": 7.293666026871402e-07, "loss": 0.8456, "step": 39 }, { "epoch": 0.03840614498319731, "grad_norm": 2.2613200931551534, "learning_rate": 7.485604606525912e-07, "loss": 0.9944, "step": 40 }, { "epoch": 0.039366298607777246, "grad_norm": 1.9438230294266092, "learning_rate": 7.677543186180422e-07, "loss": 0.8274, "step": 41 }, { "epoch": 0.040326452232357174, "grad_norm": 1.8043175525039512, "learning_rate": 7.869481765834934e-07, "loss": 0.8265, "step": 42 }, { "epoch": 0.04128660585693711, "grad_norm": 1.4149403723683052, "learning_rate": 8.061420345489445e-07, "loss": 0.8152, "step": 43 }, { "epoch": 0.04224675948151704, "grad_norm": 1.7426781210352167, "learning_rate": 8.253358925143955e-07, "loss": 0.855, "step": 44 }, { "epoch": 0.04320691310609698, "grad_norm": 1.6948390846540724, "learning_rate": 8.445297504798465e-07, "loss": 0.8459, "step": 45 }, { "epoch": 0.044167066730676906, "grad_norm": 1.9308735089919975, "learning_rate": 8.637236084452976e-07, "loss": 1.0274, "step": 46 }, { "epoch": 0.04512722035525684, "grad_norm": 1.7254665459901692, "learning_rate": 8.829174664107486e-07, "loss": 0.9015, "step": 47 }, { "epoch": 0.046087373979836775, "grad_norm": 1.6850734829034464, "learning_rate": 9.021113243761997e-07, "loss": 0.8491, "step": 48 }, { "epoch": 0.04704752760441671, "grad_norm": 1.880663174860765, "learning_rate": 9.213051823416508e-07, "loss": 0.945, "step": 49 }, { "epoch": 0.04800768122899664, "grad_norm": 1.519423360483066, "learning_rate": 9.404990403071018e-07, "loss": 0.8472, "step": 50 }, { "epoch": 0.04896783485357657, "grad_norm": 1.6015894794394687, "learning_rate": 9.59692898272553e-07, "loss": 0.8475, "step": 51 }, { "epoch": 0.04992798847815651, "grad_norm": 1.4916208624350973, "learning_rate": 9.788867562380039e-07, "loss": 0.8045, "step": 52 }, { "epoch": 0.050888142102736435, "grad_norm": 1.8298604039007258, "learning_rate": 9.98080614203455e-07, "loss": 1.0418, "step": 53 }, { "epoch": 0.05184829572731637, "grad_norm": 1.6099682751777635, "learning_rate": 1.0172744721689061e-06, "loss": 0.9138, "step": 54 }, { "epoch": 0.052808449351896304, "grad_norm": 1.6925226830530489, "learning_rate": 1.036468330134357e-06, "loss": 0.9055, "step": 55 }, { "epoch": 0.05376860297647624, "grad_norm": 1.2955809980411925, "learning_rate": 1.0556621880998082e-06, "loss": 0.9829, "step": 56 }, { "epoch": 0.05472875660105617, "grad_norm": 1.6078071748379568, "learning_rate": 1.074856046065259e-06, "loss": 0.9409, "step": 57 }, { "epoch": 0.0556889102256361, "grad_norm": 1.6716086522951294, "learning_rate": 1.0940499040307102e-06, "loss": 1.0491, "step": 58 }, { "epoch": 0.056649063850216036, "grad_norm": 1.7263216789007423, "learning_rate": 1.1132437619961614e-06, "loss": 1.0377, "step": 59 }, { "epoch": 0.05760921747479597, "grad_norm": 1.184805829043608, "learning_rate": 1.1324376199616125e-06, "loss": 0.842, "step": 60 }, { "epoch": 0.0585693710993759, "grad_norm": 1.1025041416558563, "learning_rate": 1.1516314779270634e-06, "loss": 0.8158, "step": 61 }, { "epoch": 0.05952952472395583, "grad_norm": 1.3821449512042823, "learning_rate": 1.1708253358925146e-06, "loss": 0.9313, "step": 62 }, { "epoch": 0.06048967834853577, "grad_norm": 1.0156180747346113, "learning_rate": 1.1900191938579655e-06, "loss": 0.7505, "step": 63 }, { "epoch": 0.061449831973115696, "grad_norm": 0.9255973361111742, "learning_rate": 1.2092130518234166e-06, "loss": 0.8504, "step": 64 }, { "epoch": 0.06240998559769563, "grad_norm": 1.0225622146422586, "learning_rate": 1.2284069097888675e-06, "loss": 0.7864, "step": 65 }, { "epoch": 0.06337013922227556, "grad_norm": 1.115144698795141, "learning_rate": 1.2476007677543187e-06, "loss": 0.8918, "step": 66 }, { "epoch": 0.0643302928468555, "grad_norm": 1.1256454687205975, "learning_rate": 1.2667946257197696e-06, "loss": 0.9486, "step": 67 }, { "epoch": 0.06529044647143543, "grad_norm": 1.0359375685106758, "learning_rate": 1.2859884836852207e-06, "loss": 0.7975, "step": 68 }, { "epoch": 0.06625060009601537, "grad_norm": 1.0475887376424418, "learning_rate": 1.3051823416506717e-06, "loss": 0.8228, "step": 69 }, { "epoch": 0.06721075372059529, "grad_norm": 1.1189059596016355, "learning_rate": 1.324376199616123e-06, "loss": 0.9463, "step": 70 }, { "epoch": 0.06817090734517522, "grad_norm": 1.04187594168728, "learning_rate": 1.3435700575815741e-06, "loss": 0.8424, "step": 71 }, { "epoch": 0.06913106096975516, "grad_norm": 1.0289225103982156, "learning_rate": 1.362763915547025e-06, "loss": 0.831, "step": 72 }, { "epoch": 0.0700912145943351, "grad_norm": 1.1117742328172833, "learning_rate": 1.3819577735124762e-06, "loss": 0.8985, "step": 73 }, { "epoch": 0.07105136821891503, "grad_norm": 0.9322019191432076, "learning_rate": 1.4011516314779271e-06, "loss": 0.7785, "step": 74 }, { "epoch": 0.07201152184349496, "grad_norm": 1.2474670198079632, "learning_rate": 1.4203454894433783e-06, "loss": 0.9439, "step": 75 }, { "epoch": 0.0729716754680749, "grad_norm": 0.8004689402193135, "learning_rate": 1.4395393474088292e-06, "loss": 0.7005, "step": 76 }, { "epoch": 0.07393182909265482, "grad_norm": 1.2653308951853899, "learning_rate": 1.4587332053742803e-06, "loss": 0.9875, "step": 77 }, { "epoch": 0.07489198271723475, "grad_norm": 1.2763542473058163, "learning_rate": 1.4779270633397312e-06, "loss": 1.0097, "step": 78 }, { "epoch": 0.07585213634181469, "grad_norm": 1.1183865864016693, "learning_rate": 1.4971209213051824e-06, "loss": 0.8761, "step": 79 }, { "epoch": 0.07681228996639462, "grad_norm": 0.9809086831651774, "learning_rate": 1.5163147792706335e-06, "loss": 0.8134, "step": 80 }, { "epoch": 0.07777244359097456, "grad_norm": 0.9182248069328082, "learning_rate": 1.5355086372360844e-06, "loss": 0.8311, "step": 81 }, { "epoch": 0.07873259721555449, "grad_norm": 0.9239941618382052, "learning_rate": 1.5547024952015358e-06, "loss": 0.883, "step": 82 }, { "epoch": 0.07969275084013443, "grad_norm": 0.8620749872248693, "learning_rate": 1.5738963531669867e-06, "loss": 0.7949, "step": 83 }, { "epoch": 0.08065290446471435, "grad_norm": 0.9647356359837779, "learning_rate": 1.5930902111324378e-06, "loss": 0.8291, "step": 84 }, { "epoch": 0.08161305808929428, "grad_norm": 0.9721463568898323, "learning_rate": 1.612284069097889e-06, "loss": 0.8164, "step": 85 }, { "epoch": 0.08257321171387422, "grad_norm": 0.9312032861088886, "learning_rate": 1.63147792706334e-06, "loss": 0.9157, "step": 86 }, { "epoch": 0.08353336533845415, "grad_norm": 0.8547144224759374, "learning_rate": 1.650671785028791e-06, "loss": 0.8582, "step": 87 }, { "epoch": 0.08449351896303409, "grad_norm": 0.8770736524961695, "learning_rate": 1.669865642994242e-06, "loss": 0.786, "step": 88 }, { "epoch": 0.08545367258761402, "grad_norm": 1.1046381600470017, "learning_rate": 1.689059500959693e-06, "loss": 0.9879, "step": 89 }, { "epoch": 0.08641382621219396, "grad_norm": 1.0369973370286998, "learning_rate": 1.708253358925144e-06, "loss": 0.7226, "step": 90 }, { "epoch": 0.08737397983677389, "grad_norm": 1.1281049949863124, "learning_rate": 1.7274472168905951e-06, "loss": 0.9568, "step": 91 }, { "epoch": 0.08833413346135381, "grad_norm": 1.4254943426120006, "learning_rate": 1.746641074856046e-06, "loss": 0.9576, "step": 92 }, { "epoch": 0.08929428708593375, "grad_norm": 1.3924097061049034, "learning_rate": 1.7658349328214972e-06, "loss": 0.9491, "step": 93 }, { "epoch": 0.09025444071051368, "grad_norm": 0.9759408830299551, "learning_rate": 1.7850287907869481e-06, "loss": 0.8288, "step": 94 }, { "epoch": 0.09121459433509362, "grad_norm": 2.009797042537888, "learning_rate": 1.8042226487523995e-06, "loss": 0.9751, "step": 95 }, { "epoch": 0.09217474795967355, "grad_norm": 1.2018425341550694, "learning_rate": 1.8234165067178506e-06, "loss": 0.897, "step": 96 }, { "epoch": 0.09313490158425348, "grad_norm": 0.8232170864029329, "learning_rate": 1.8426103646833015e-06, "loss": 0.759, "step": 97 }, { "epoch": 0.09409505520883342, "grad_norm": 0.8873931964391254, "learning_rate": 1.8618042226487527e-06, "loss": 0.8699, "step": 98 }, { "epoch": 0.09505520883341334, "grad_norm": 1.5059726994951654, "learning_rate": 1.8809980806142036e-06, "loss": 1.0718, "step": 99 }, { "epoch": 0.09601536245799328, "grad_norm": 1.1455129104016302, "learning_rate": 1.9001919385796547e-06, "loss": 0.9242, "step": 100 }, { "epoch": 0.09697551608257321, "grad_norm": 0.9792756606939649, "learning_rate": 1.919385796545106e-06, "loss": 0.8241, "step": 101 }, { "epoch": 0.09793566970715314, "grad_norm": 0.8338614782648381, "learning_rate": 1.9385796545105568e-06, "loss": 0.8217, "step": 102 }, { "epoch": 0.09889582333173308, "grad_norm": 1.116697001888954, "learning_rate": 1.9577735124760077e-06, "loss": 0.9195, "step": 103 }, { "epoch": 0.09985597695631301, "grad_norm": 1.1382902702177355, "learning_rate": 1.9769673704414586e-06, "loss": 0.8216, "step": 104 }, { "epoch": 0.10081613058089295, "grad_norm": 0.7875421921896231, "learning_rate": 1.99616122840691e-06, "loss": 0.7569, "step": 105 }, { "epoch": 0.10177628420547287, "grad_norm": 0.9462333923656521, "learning_rate": 2.015355086372361e-06, "loss": 0.8147, "step": 106 }, { "epoch": 0.1027364378300528, "grad_norm": 1.2471159392534685, "learning_rate": 2.0345489443378122e-06, "loss": 0.9496, "step": 107 }, { "epoch": 0.10369659145463274, "grad_norm": 1.0972978613025275, "learning_rate": 2.053742802303263e-06, "loss": 0.8787, "step": 108 }, { "epoch": 0.10465674507921267, "grad_norm": 0.9564358008503212, "learning_rate": 2.072936660268714e-06, "loss": 0.6505, "step": 109 }, { "epoch": 0.10561689870379261, "grad_norm": 0.8382354616537259, "learning_rate": 2.0921305182341654e-06, "loss": 0.7956, "step": 110 }, { "epoch": 0.10657705232837254, "grad_norm": 0.8915784275057668, "learning_rate": 2.1113243761996164e-06, "loss": 0.7803, "step": 111 }, { "epoch": 0.10753720595295248, "grad_norm": 1.142846540809407, "learning_rate": 2.1305182341650673e-06, "loss": 0.9204, "step": 112 }, { "epoch": 0.1084973595775324, "grad_norm": 0.8974039346501843, "learning_rate": 2.149712092130518e-06, "loss": 0.8523, "step": 113 }, { "epoch": 0.10945751320211233, "grad_norm": 0.9276605439304481, "learning_rate": 2.1689059500959696e-06, "loss": 0.8377, "step": 114 }, { "epoch": 0.11041766682669227, "grad_norm": 0.8822041742771103, "learning_rate": 2.1880998080614205e-06, "loss": 0.7516, "step": 115 }, { "epoch": 0.1113778204512722, "grad_norm": 1.4010891675131194, "learning_rate": 2.2072936660268714e-06, "loss": 0.9479, "step": 116 }, { "epoch": 0.11233797407585214, "grad_norm": 0.9826887792887448, "learning_rate": 2.2264875239923228e-06, "loss": 0.8469, "step": 117 }, { "epoch": 0.11329812770043207, "grad_norm": 1.336658129079735, "learning_rate": 2.2456813819577737e-06, "loss": 0.8057, "step": 118 }, { "epoch": 0.114258281325012, "grad_norm": 0.882482535329637, "learning_rate": 2.264875239923225e-06, "loss": 0.8773, "step": 119 }, { "epoch": 0.11521843494959194, "grad_norm": 1.0811394168808193, "learning_rate": 2.284069097888676e-06, "loss": 0.8309, "step": 120 }, { "epoch": 0.11617858857417186, "grad_norm": 0.8993855668700828, "learning_rate": 2.303262955854127e-06, "loss": 0.7727, "step": 121 }, { "epoch": 0.1171387421987518, "grad_norm": 0.7873978398789067, "learning_rate": 2.322456813819578e-06, "loss": 0.7331, "step": 122 }, { "epoch": 0.11809889582333173, "grad_norm": 1.1787412714073362, "learning_rate": 2.341650671785029e-06, "loss": 0.9466, "step": 123 }, { "epoch": 0.11905904944791167, "grad_norm": 0.9872507261181287, "learning_rate": 2.36084452975048e-06, "loss": 0.8958, "step": 124 }, { "epoch": 0.1200192030724916, "grad_norm": 1.1985408152722705, "learning_rate": 2.380038387715931e-06, "loss": 0.9142, "step": 125 }, { "epoch": 0.12097935669707154, "grad_norm": 0.9809678786360649, "learning_rate": 2.3992322456813823e-06, "loss": 0.8062, "step": 126 }, { "epoch": 0.12193951032165147, "grad_norm": 1.064765259377889, "learning_rate": 2.4184261036468333e-06, "loss": 0.8151, "step": 127 }, { "epoch": 0.12289966394623139, "grad_norm": 0.814350277256771, "learning_rate": 2.437619961612284e-06, "loss": 0.77, "step": 128 }, { "epoch": 0.12385981757081133, "grad_norm": 1.3447206060859422, "learning_rate": 2.456813819577735e-06, "loss": 0.8383, "step": 129 }, { "epoch": 0.12481997119539126, "grad_norm": 0.8405459708645042, "learning_rate": 2.4760076775431864e-06, "loss": 0.7069, "step": 130 }, { "epoch": 0.1257801248199712, "grad_norm": 0.9058649088192959, "learning_rate": 2.4952015355086374e-06, "loss": 0.8162, "step": 131 }, { "epoch": 0.12674027844455113, "grad_norm": 1.1155185616475223, "learning_rate": 2.5143953934740883e-06, "loss": 0.8553, "step": 132 }, { "epoch": 0.12770043206913106, "grad_norm": 0.8915853521268036, "learning_rate": 2.5335892514395392e-06, "loss": 0.8081, "step": 133 }, { "epoch": 0.128660585693711, "grad_norm": 0.8396890944339533, "learning_rate": 2.5527831094049906e-06, "loss": 0.7372, "step": 134 }, { "epoch": 0.12962073931829093, "grad_norm": 1.200008469124059, "learning_rate": 2.5719769673704415e-06, "loss": 0.8521, "step": 135 }, { "epoch": 0.13058089294287087, "grad_norm": 0.9525255053918179, "learning_rate": 2.5911708253358924e-06, "loss": 0.7797, "step": 136 }, { "epoch": 0.1315410465674508, "grad_norm": 1.1597511026225074, "learning_rate": 2.6103646833013433e-06, "loss": 0.9074, "step": 137 }, { "epoch": 0.13250120019203074, "grad_norm": 0.8839046873973051, "learning_rate": 2.629558541266795e-06, "loss": 0.733, "step": 138 }, { "epoch": 0.13346135381661065, "grad_norm": 1.015682098178237, "learning_rate": 2.648752399232246e-06, "loss": 0.8376, "step": 139 }, { "epoch": 0.13442150744119058, "grad_norm": 0.8385802239109873, "learning_rate": 2.6679462571976974e-06, "loss": 0.8148, "step": 140 }, { "epoch": 0.13538166106577051, "grad_norm": 1.0863453638320109, "learning_rate": 2.6871401151631483e-06, "loss": 0.8493, "step": 141 }, { "epoch": 0.13634181469035045, "grad_norm": 1.003803580124611, "learning_rate": 2.7063339731285992e-06, "loss": 0.8831, "step": 142 }, { "epoch": 0.13730196831493038, "grad_norm": 1.0478842238251738, "learning_rate": 2.72552783109405e-06, "loss": 0.7659, "step": 143 }, { "epoch": 0.13826212193951032, "grad_norm": 1.104387546380458, "learning_rate": 2.7447216890595015e-06, "loss": 0.7887, "step": 144 }, { "epoch": 0.13922227556409025, "grad_norm": 0.7309074641286217, "learning_rate": 2.7639155470249524e-06, "loss": 0.6578, "step": 145 }, { "epoch": 0.1401824291886702, "grad_norm": 1.0128217345749977, "learning_rate": 2.7831094049904033e-06, "loss": 0.8568, "step": 146 }, { "epoch": 0.14114258281325012, "grad_norm": 1.0577981551930282, "learning_rate": 2.8023032629558543e-06, "loss": 0.8375, "step": 147 }, { "epoch": 0.14210273643783006, "grad_norm": 0.8490365194042823, "learning_rate": 2.8214971209213056e-06, "loss": 0.7297, "step": 148 }, { "epoch": 0.14306289006241, "grad_norm": 0.8866110773593991, "learning_rate": 2.8406909788867565e-06, "loss": 0.8722, "step": 149 }, { "epoch": 0.14402304368698993, "grad_norm": 0.767067353988987, "learning_rate": 2.8598848368522074e-06, "loss": 0.7858, "step": 150 }, { "epoch": 0.14498319731156986, "grad_norm": 0.7856928727323835, "learning_rate": 2.8790786948176584e-06, "loss": 0.7862, "step": 151 }, { "epoch": 0.1459433509361498, "grad_norm": 0.790793584893058, "learning_rate": 2.8982725527831097e-06, "loss": 0.6779, "step": 152 }, { "epoch": 0.1469035045607297, "grad_norm": 0.7780817032213796, "learning_rate": 2.9174664107485606e-06, "loss": 0.7492, "step": 153 }, { "epoch": 0.14786365818530964, "grad_norm": 0.7952071858818339, "learning_rate": 2.9366602687140116e-06, "loss": 0.7407, "step": 154 }, { "epoch": 0.14882381180988957, "grad_norm": 0.9159480789420381, "learning_rate": 2.9558541266794625e-06, "loss": 0.7744, "step": 155 }, { "epoch": 0.1497839654344695, "grad_norm": 0.7973747144009165, "learning_rate": 2.975047984644914e-06, "loss": 0.7321, "step": 156 }, { "epoch": 0.15074411905904944, "grad_norm": 1.5745742190311103, "learning_rate": 2.9942418426103648e-06, "loss": 0.9273, "step": 157 }, { "epoch": 0.15170427268362938, "grad_norm": 1.2208467201775475, "learning_rate": 3.0134357005758157e-06, "loss": 0.8202, "step": 158 }, { "epoch": 0.1526644263082093, "grad_norm": 0.808498702793707, "learning_rate": 3.032629558541267e-06, "loss": 0.7722, "step": 159 }, { "epoch": 0.15362457993278925, "grad_norm": 1.1297771740469291, "learning_rate": 3.051823416506718e-06, "loss": 0.793, "step": 160 }, { "epoch": 0.15458473355736918, "grad_norm": 0.8920630230218092, "learning_rate": 3.071017274472169e-06, "loss": 0.7603, "step": 161 }, { "epoch": 0.15554488718194912, "grad_norm": 0.9911980773811446, "learning_rate": 3.09021113243762e-06, "loss": 0.8845, "step": 162 }, { "epoch": 0.15650504080652905, "grad_norm": 0.9637168974804745, "learning_rate": 3.1094049904030716e-06, "loss": 0.7493, "step": 163 }, { "epoch": 0.15746519443110898, "grad_norm": 1.008318644981146, "learning_rate": 3.1285988483685225e-06, "loss": 0.8444, "step": 164 }, { "epoch": 0.15842534805568892, "grad_norm": 0.8444810880090623, "learning_rate": 3.1477927063339734e-06, "loss": 0.731, "step": 165 }, { "epoch": 0.15938550168026885, "grad_norm": 0.7845005154103945, "learning_rate": 3.1669865642994248e-06, "loss": 0.7613, "step": 166 }, { "epoch": 0.1603456553048488, "grad_norm": 0.7585180664604924, "learning_rate": 3.1861804222648757e-06, "loss": 0.6808, "step": 167 }, { "epoch": 0.1613058089294287, "grad_norm": 0.708879804248452, "learning_rate": 3.2053742802303266e-06, "loss": 0.6368, "step": 168 }, { "epoch": 0.16226596255400863, "grad_norm": 0.8456065003931889, "learning_rate": 3.224568138195778e-06, "loss": 0.8338, "step": 169 }, { "epoch": 0.16322611617858857, "grad_norm": 0.8395891934311173, "learning_rate": 3.243761996161229e-06, "loss": 0.7225, "step": 170 }, { "epoch": 0.1641862698031685, "grad_norm": 0.9297060803605406, "learning_rate": 3.26295585412668e-06, "loss": 0.8321, "step": 171 }, { "epoch": 0.16514642342774843, "grad_norm": 0.8833795491136275, "learning_rate": 3.2821497120921307e-06, "loss": 0.8381, "step": 172 }, { "epoch": 0.16610657705232837, "grad_norm": 1.466746985306447, "learning_rate": 3.301343570057582e-06, "loss": 0.8577, "step": 173 }, { "epoch": 0.1670667306769083, "grad_norm": 1.4824534979935604, "learning_rate": 3.320537428023033e-06, "loss": 0.9089, "step": 174 }, { "epoch": 0.16802688430148824, "grad_norm": 0.8557828290379548, "learning_rate": 3.339731285988484e-06, "loss": 0.7944, "step": 175 }, { "epoch": 0.16898703792606817, "grad_norm": 0.9686708287892611, "learning_rate": 3.358925143953935e-06, "loss": 0.8574, "step": 176 }, { "epoch": 0.1699471915506481, "grad_norm": 0.8431136590487779, "learning_rate": 3.378119001919386e-06, "loss": 0.6604, "step": 177 }, { "epoch": 0.17090734517522804, "grad_norm": 1.4763785878877154, "learning_rate": 3.397312859884837e-06, "loss": 0.8402, "step": 178 }, { "epoch": 0.17186749879980798, "grad_norm": 1.1101229899925387, "learning_rate": 3.416506717850288e-06, "loss": 0.8633, "step": 179 }, { "epoch": 0.1728276524243879, "grad_norm": 0.8302601927585418, "learning_rate": 3.435700575815739e-06, "loss": 0.6392, "step": 180 }, { "epoch": 0.17378780604896785, "grad_norm": 0.8656759392779118, "learning_rate": 3.4548944337811903e-06, "loss": 0.8395, "step": 181 }, { "epoch": 0.17474795967354778, "grad_norm": 0.9547957252973762, "learning_rate": 3.4740882917466412e-06, "loss": 0.8163, "step": 182 }, { "epoch": 0.1757081132981277, "grad_norm": 0.8856150712098116, "learning_rate": 3.493282149712092e-06, "loss": 0.8307, "step": 183 }, { "epoch": 0.17666826692270762, "grad_norm": 0.9981943950309549, "learning_rate": 3.512476007677543e-06, "loss": 0.8446, "step": 184 }, { "epoch": 0.17762842054728756, "grad_norm": 0.9473648805992334, "learning_rate": 3.5316698656429944e-06, "loss": 0.7254, "step": 185 }, { "epoch": 0.1785885741718675, "grad_norm": 0.7920174010204585, "learning_rate": 3.5508637236084453e-06, "loss": 0.8296, "step": 186 }, { "epoch": 0.17954872779644743, "grad_norm": 0.6190134708281353, "learning_rate": 3.5700575815738963e-06, "loss": 0.6472, "step": 187 }, { "epoch": 0.18050888142102736, "grad_norm": 1.3526731510164427, "learning_rate": 3.589251439539348e-06, "loss": 0.8892, "step": 188 }, { "epoch": 0.1814690350456073, "grad_norm": 1.173948415985468, "learning_rate": 3.608445297504799e-06, "loss": 0.822, "step": 189 }, { "epoch": 0.18242918867018723, "grad_norm": 1.078900898862366, "learning_rate": 3.62763915547025e-06, "loss": 0.7581, "step": 190 }, { "epoch": 0.18338934229476717, "grad_norm": 0.7929795674519923, "learning_rate": 3.6468330134357012e-06, "loss": 0.793, "step": 191 }, { "epoch": 0.1843494959193471, "grad_norm": 0.8386451085029051, "learning_rate": 3.666026871401152e-06, "loss": 0.705, "step": 192 }, { "epoch": 0.18530964954392704, "grad_norm": 1.0735177595196022, "learning_rate": 3.685220729366603e-06, "loss": 0.732, "step": 193 }, { "epoch": 0.18626980316850697, "grad_norm": 0.9431902997077553, "learning_rate": 3.704414587332054e-06, "loss": 0.681, "step": 194 }, { "epoch": 0.1872299567930869, "grad_norm": 0.8926278710846778, "learning_rate": 3.7236084452975053e-06, "loss": 0.845, "step": 195 }, { "epoch": 0.18819011041766684, "grad_norm": 0.8991526201843326, "learning_rate": 3.7428023032629563e-06, "loss": 0.7584, "step": 196 }, { "epoch": 0.18915026404224675, "grad_norm": 0.7633942408146562, "learning_rate": 3.761996161228407e-06, "loss": 0.684, "step": 197 }, { "epoch": 0.19011041766682668, "grad_norm": 1.1455245589516796, "learning_rate": 3.781190019193858e-06, "loss": 0.7618, "step": 198 }, { "epoch": 0.19107057129140662, "grad_norm": 1.0821218403472503, "learning_rate": 3.8003838771593095e-06, "loss": 0.8418, "step": 199 }, { "epoch": 0.19203072491598655, "grad_norm": 1.3650458096237892, "learning_rate": 3.81957773512476e-06, "loss": 0.8632, "step": 200 }, { "epoch": 0.19299087854056649, "grad_norm": 0.8305562118590513, "learning_rate": 3.838771593090212e-06, "loss": 0.7579, "step": 201 }, { "epoch": 0.19395103216514642, "grad_norm": 1.1764982171600717, "learning_rate": 3.857965451055662e-06, "loss": 0.7343, "step": 202 }, { "epoch": 0.19491118578972635, "grad_norm": 0.8419631368807355, "learning_rate": 3.8771593090211136e-06, "loss": 0.728, "step": 203 }, { "epoch": 0.1958713394143063, "grad_norm": 1.412012697905847, "learning_rate": 3.896353166986565e-06, "loss": 0.9376, "step": 204 }, { "epoch": 0.19683149303888622, "grad_norm": 0.772276986102651, "learning_rate": 3.915547024952015e-06, "loss": 0.697, "step": 205 }, { "epoch": 0.19779164666346616, "grad_norm": 1.0576188701547733, "learning_rate": 3.934740882917467e-06, "loss": 0.9163, "step": 206 }, { "epoch": 0.1987518002880461, "grad_norm": 1.5918533652810105, "learning_rate": 3.953934740882917e-06, "loss": 0.7706, "step": 207 }, { "epoch": 0.19971195391262603, "grad_norm": 1.1943336552570978, "learning_rate": 3.973128598848369e-06, "loss": 0.6891, "step": 208 }, { "epoch": 0.20067210753720596, "grad_norm": 0.8781483957000793, "learning_rate": 3.99232245681382e-06, "loss": 0.861, "step": 209 }, { "epoch": 0.2016322611617859, "grad_norm": 0.8493715781408644, "learning_rate": 4.0115163147792705e-06, "loss": 0.7008, "step": 210 }, { "epoch": 0.20259241478636583, "grad_norm": 0.8940841063021919, "learning_rate": 4.030710172744722e-06, "loss": 0.7154, "step": 211 }, { "epoch": 0.20355256841094574, "grad_norm": 0.8246218388969546, "learning_rate": 4.049904030710173e-06, "loss": 0.6572, "step": 212 }, { "epoch": 0.20451272203552567, "grad_norm": 1.2380396503566105, "learning_rate": 4.0690978886756245e-06, "loss": 0.8725, "step": 213 }, { "epoch": 0.2054728756601056, "grad_norm": 0.9516245864123232, "learning_rate": 4.088291746641076e-06, "loss": 0.7383, "step": 214 }, { "epoch": 0.20643302928468554, "grad_norm": 0.804321194431788, "learning_rate": 4.107485604606526e-06, "loss": 0.6985, "step": 215 }, { "epoch": 0.20739318290926548, "grad_norm": 0.833858618146012, "learning_rate": 4.126679462571978e-06, "loss": 0.625, "step": 216 }, { "epoch": 0.2083533365338454, "grad_norm": 0.9078893019227422, "learning_rate": 4.145873320537428e-06, "loss": 0.6601, "step": 217 }, { "epoch": 0.20931349015842535, "grad_norm": 1.0437286841884523, "learning_rate": 4.1650671785028795e-06, "loss": 0.7764, "step": 218 }, { "epoch": 0.21027364378300528, "grad_norm": 0.9080260432854801, "learning_rate": 4.184261036468331e-06, "loss": 0.7048, "step": 219 }, { "epoch": 0.21123379740758522, "grad_norm": 0.9912616770447701, "learning_rate": 4.203454894433781e-06, "loss": 0.7016, "step": 220 }, { "epoch": 0.21219395103216515, "grad_norm": 0.9841767945506055, "learning_rate": 4.222648752399233e-06, "loss": 0.7472, "step": 221 }, { "epoch": 0.21315410465674509, "grad_norm": 0.8040295743930025, "learning_rate": 4.241842610364684e-06, "loss": 0.7741, "step": 222 }, { "epoch": 0.21411425828132502, "grad_norm": 1.029339529830975, "learning_rate": 4.2610364683301346e-06, "loss": 0.7542, "step": 223 }, { "epoch": 0.21507441190590496, "grad_norm": 0.8358337200626998, "learning_rate": 4.280230326295586e-06, "loss": 0.8003, "step": 224 }, { "epoch": 0.2160345655304849, "grad_norm": 0.7871074801617568, "learning_rate": 4.299424184261036e-06, "loss": 0.7555, "step": 225 }, { "epoch": 0.2169947191550648, "grad_norm": 0.779476971431696, "learning_rate": 4.318618042226488e-06, "loss": 0.6465, "step": 226 }, { "epoch": 0.21795487277964473, "grad_norm": 1.3645344064461826, "learning_rate": 4.337811900191939e-06, "loss": 0.8908, "step": 227 }, { "epoch": 0.21891502640422467, "grad_norm": 0.8369940377233636, "learning_rate": 4.35700575815739e-06, "loss": 0.7997, "step": 228 }, { "epoch": 0.2198751800288046, "grad_norm": 0.9145989263174269, "learning_rate": 4.376199616122841e-06, "loss": 0.6488, "step": 229 }, { "epoch": 0.22083533365338454, "grad_norm": 1.0331924469107403, "learning_rate": 4.395393474088292e-06, "loss": 0.8251, "step": 230 }, { "epoch": 0.22179548727796447, "grad_norm": 1.0223072806975653, "learning_rate": 4.414587332053743e-06, "loss": 0.7205, "step": 231 }, { "epoch": 0.2227556409025444, "grad_norm": 0.7940845329332421, "learning_rate": 4.433781190019194e-06, "loss": 0.6351, "step": 232 }, { "epoch": 0.22371579452712434, "grad_norm": 1.0194058049576815, "learning_rate": 4.4529750479846455e-06, "loss": 0.8269, "step": 233 }, { "epoch": 0.22467594815170427, "grad_norm": 0.7762787254756254, "learning_rate": 4.472168905950096e-06, "loss": 0.6229, "step": 234 }, { "epoch": 0.2256361017762842, "grad_norm": 0.7724891803670033, "learning_rate": 4.491362763915547e-06, "loss": 0.6554, "step": 235 }, { "epoch": 0.22659625540086414, "grad_norm": 0.9704788493195956, "learning_rate": 4.510556621880998e-06, "loss": 0.7888, "step": 236 }, { "epoch": 0.22755640902544408, "grad_norm": 0.7667671639268514, "learning_rate": 4.52975047984645e-06, "loss": 0.6245, "step": 237 }, { "epoch": 0.228516562650024, "grad_norm": 0.9433268933026349, "learning_rate": 4.5489443378119005e-06, "loss": 0.7965, "step": 238 }, { "epoch": 0.22947671627460395, "grad_norm": 0.8296551684508615, "learning_rate": 4.568138195777352e-06, "loss": 0.758, "step": 239 }, { "epoch": 0.23043686989918388, "grad_norm": 0.9595334836481333, "learning_rate": 4.587332053742803e-06, "loss": 0.8435, "step": 240 }, { "epoch": 0.2313970235237638, "grad_norm": 0.8971273475152683, "learning_rate": 4.606525911708254e-06, "loss": 0.865, "step": 241 }, { "epoch": 0.23235717714834372, "grad_norm": 1.010617686035143, "learning_rate": 4.625719769673705e-06, "loss": 0.8701, "step": 242 }, { "epoch": 0.23331733077292366, "grad_norm": 1.1676065764098063, "learning_rate": 4.644913627639156e-06, "loss": 0.7683, "step": 243 }, { "epoch": 0.2342774843975036, "grad_norm": 0.793638142701742, "learning_rate": 4.664107485604607e-06, "loss": 0.7304, "step": 244 }, { "epoch": 0.23523763802208353, "grad_norm": 1.1351477240401913, "learning_rate": 4.683301343570058e-06, "loss": 0.7883, "step": 245 }, { "epoch": 0.23619779164666346, "grad_norm": 0.9014252960603116, "learning_rate": 4.702495201535509e-06, "loss": 0.7882, "step": 246 }, { "epoch": 0.2371579452712434, "grad_norm": 1.1008716325649441, "learning_rate": 4.72168905950096e-06, "loss": 0.7722, "step": 247 }, { "epoch": 0.23811809889582333, "grad_norm": 0.8980719221578192, "learning_rate": 4.7408829174664115e-06, "loss": 0.6932, "step": 248 }, { "epoch": 0.23907825252040327, "grad_norm": 0.8633964545471303, "learning_rate": 4.760076775431862e-06, "loss": 0.724, "step": 249 }, { "epoch": 0.2400384061449832, "grad_norm": 0.924047032655447, "learning_rate": 4.779270633397313e-06, "loss": 0.8256, "step": 250 }, { "epoch": 0.24099855976956314, "grad_norm": 0.9117379272374883, "learning_rate": 4.798464491362765e-06, "loss": 0.8264, "step": 251 }, { "epoch": 0.24195871339414307, "grad_norm": 0.8673380153474116, "learning_rate": 4.817658349328215e-06, "loss": 0.8188, "step": 252 }, { "epoch": 0.242918867018723, "grad_norm": 0.7438781957499604, "learning_rate": 4.8368522072936665e-06, "loss": 0.6287, "step": 253 }, { "epoch": 0.24387902064330294, "grad_norm": 0.8623369374099497, "learning_rate": 4.856046065259117e-06, "loss": 0.7528, "step": 254 }, { "epoch": 0.24483917426788285, "grad_norm": 1.0761360899879107, "learning_rate": 4.875239923224568e-06, "loss": 0.8705, "step": 255 }, { "epoch": 0.24579932789246278, "grad_norm": 0.873485763638596, "learning_rate": 4.89443378119002e-06, "loss": 0.6785, "step": 256 }, { "epoch": 0.24675948151704272, "grad_norm": 0.9010971948972167, "learning_rate": 4.91362763915547e-06, "loss": 0.7482, "step": 257 }, { "epoch": 0.24771963514162265, "grad_norm": 0.8940763149292682, "learning_rate": 4.9328214971209215e-06, "loss": 0.7681, "step": 258 }, { "epoch": 0.2486797887662026, "grad_norm": 0.898185235000425, "learning_rate": 4.952015355086373e-06, "loss": 0.7237, "step": 259 }, { "epoch": 0.24963994239078252, "grad_norm": 1.161543197393246, "learning_rate": 4.971209213051823e-06, "loss": 0.7555, "step": 260 }, { "epoch": 0.2506000960153625, "grad_norm": 0.9649751862085733, "learning_rate": 4.990403071017275e-06, "loss": 0.7388, "step": 261 }, { "epoch": 0.2515602496399424, "grad_norm": 0.8610717852304463, "learning_rate": 5.009596928982726e-06, "loss": 0.7786, "step": 262 }, { "epoch": 0.2525204032645223, "grad_norm": 0.8345541448604458, "learning_rate": 5.028790786948177e-06, "loss": 0.7618, "step": 263 }, { "epoch": 0.25348055688910226, "grad_norm": 1.1675015090424423, "learning_rate": 5.047984644913628e-06, "loss": 0.8149, "step": 264 }, { "epoch": 0.25444071051368217, "grad_norm": 0.7596734943274147, "learning_rate": 5.0671785028790784e-06, "loss": 0.7057, "step": 265 }, { "epoch": 0.25540086413826213, "grad_norm": 0.9282163243692719, "learning_rate": 5.086372360844531e-06, "loss": 0.8051, "step": 266 }, { "epoch": 0.25636101776284204, "grad_norm": 0.7984156155023513, "learning_rate": 5.105566218809981e-06, "loss": 0.7145, "step": 267 }, { "epoch": 0.257321171387422, "grad_norm": 0.8603083731662003, "learning_rate": 5.1247600767754325e-06, "loss": 0.6682, "step": 268 }, { "epoch": 0.2582813250120019, "grad_norm": 1.0759438319412193, "learning_rate": 5.143953934740883e-06, "loss": 0.6945, "step": 269 }, { "epoch": 0.25924147863658187, "grad_norm": 0.8506307799583964, "learning_rate": 5.163147792706334e-06, "loss": 0.7533, "step": 270 }, { "epoch": 0.2602016322611618, "grad_norm": 1.048022855626229, "learning_rate": 5.182341650671785e-06, "loss": 0.7685, "step": 271 }, { "epoch": 0.26116178588574174, "grad_norm": 0.7783972958148241, "learning_rate": 5.201535508637236e-06, "loss": 0.6336, "step": 272 }, { "epoch": 0.26212193951032164, "grad_norm": 1.059911163502391, "learning_rate": 5.220729366602687e-06, "loss": 0.7675, "step": 273 }, { "epoch": 0.2630820931349016, "grad_norm": 0.845533453497823, "learning_rate": 5.239923224568139e-06, "loss": 0.6472, "step": 274 }, { "epoch": 0.2640422467594815, "grad_norm": 0.8532574653266674, "learning_rate": 5.25911708253359e-06, "loss": 0.7259, "step": 275 }, { "epoch": 0.2650024003840615, "grad_norm": 0.7902812643702479, "learning_rate": 5.278310940499041e-06, "loss": 0.6979, "step": 276 }, { "epoch": 0.2659625540086414, "grad_norm": 1.0324427635527533, "learning_rate": 5.297504798464492e-06, "loss": 0.7644, "step": 277 }, { "epoch": 0.2669227076332213, "grad_norm": 0.7227046291951138, "learning_rate": 5.3166986564299425e-06, "loss": 0.6526, "step": 278 }, { "epoch": 0.26788286125780125, "grad_norm": 0.9138118638895835, "learning_rate": 5.335892514395395e-06, "loss": 0.6938, "step": 279 }, { "epoch": 0.26884301488238116, "grad_norm": 0.9562573719490798, "learning_rate": 5.355086372360845e-06, "loss": 0.7261, "step": 280 }, { "epoch": 0.2698031685069611, "grad_norm": 1.0012369836801334, "learning_rate": 5.374280230326297e-06, "loss": 0.7157, "step": 281 }, { "epoch": 0.27076332213154103, "grad_norm": 1.1424962906070966, "learning_rate": 5.393474088291747e-06, "loss": 0.7556, "step": 282 }, { "epoch": 0.271723475756121, "grad_norm": 1.0073042304741078, "learning_rate": 5.4126679462571984e-06, "loss": 0.7152, "step": 283 }, { "epoch": 0.2726836293807009, "grad_norm": 0.8908765571268724, "learning_rate": 5.431861804222649e-06, "loss": 0.672, "step": 284 }, { "epoch": 0.27364378300528086, "grad_norm": 0.7885312908970684, "learning_rate": 5.4510556621881e-06, "loss": 0.6645, "step": 285 }, { "epoch": 0.27460393662986077, "grad_norm": 0.9118589180616191, "learning_rate": 5.470249520153551e-06, "loss": 0.6454, "step": 286 }, { "epoch": 0.27556409025444073, "grad_norm": 0.8036511042738933, "learning_rate": 5.489443378119003e-06, "loss": 0.7971, "step": 287 }, { "epoch": 0.27652424387902064, "grad_norm": 0.908848542317261, "learning_rate": 5.5086372360844535e-06, "loss": 0.7456, "step": 288 }, { "epoch": 0.2774843975036006, "grad_norm": 0.7888188117615401, "learning_rate": 5.527831094049905e-06, "loss": 0.7586, "step": 289 }, { "epoch": 0.2784445511281805, "grad_norm": 1.0740183382034862, "learning_rate": 5.547024952015355e-06, "loss": 0.715, "step": 290 }, { "epoch": 0.27940470475276047, "grad_norm": 0.7696355531181029, "learning_rate": 5.566218809980807e-06, "loss": 0.6671, "step": 291 }, { "epoch": 0.2803648583773404, "grad_norm": 0.9262838316202847, "learning_rate": 5.585412667946257e-06, "loss": 0.8105, "step": 292 }, { "epoch": 0.2813250120019203, "grad_norm": 0.8251592271265894, "learning_rate": 5.6046065259117085e-06, "loss": 0.7323, "step": 293 }, { "epoch": 0.28228516562650025, "grad_norm": 1.0410624894706555, "learning_rate": 5.623800383877159e-06, "loss": 0.6892, "step": 294 }, { "epoch": 0.28324531925108015, "grad_norm": 1.2878324863098023, "learning_rate": 5.642994241842611e-06, "loss": 0.807, "step": 295 }, { "epoch": 0.2842054728756601, "grad_norm": 0.9647133279427355, "learning_rate": 5.662188099808062e-06, "loss": 0.7326, "step": 296 }, { "epoch": 0.28516562650024, "grad_norm": 2.7371687843329298, "learning_rate": 5.681381957773513e-06, "loss": 0.721, "step": 297 }, { "epoch": 0.28612578012482, "grad_norm": 1.0246807185766358, "learning_rate": 5.7005758157389635e-06, "loss": 0.768, "step": 298 }, { "epoch": 0.2870859337493999, "grad_norm": 1.396773324506811, "learning_rate": 5.719769673704415e-06, "loss": 0.9106, "step": 299 }, { "epoch": 0.28804608737397985, "grad_norm": 0.996968569341027, "learning_rate": 5.738963531669866e-06, "loss": 0.7616, "step": 300 }, { "epoch": 0.28900624099855976, "grad_norm": 0.9793008359894099, "learning_rate": 5.758157389635317e-06, "loss": 0.9504, "step": 301 }, { "epoch": 0.2899663946231397, "grad_norm": 0.8742424413489684, "learning_rate": 5.777351247600769e-06, "loss": 0.7307, "step": 302 }, { "epoch": 0.29092654824771963, "grad_norm": 0.8954672809207785, "learning_rate": 5.7965451055662194e-06, "loss": 0.7001, "step": 303 }, { "epoch": 0.2918867018722996, "grad_norm": 1.072586377461691, "learning_rate": 5.815738963531671e-06, "loss": 0.7254, "step": 304 }, { "epoch": 0.2928468554968795, "grad_norm": 0.9170771315132297, "learning_rate": 5.834932821497121e-06, "loss": 0.7339, "step": 305 }, { "epoch": 0.2938070091214594, "grad_norm": 1.2961314183145254, "learning_rate": 5.854126679462573e-06, "loss": 0.7876, "step": 306 }, { "epoch": 0.29476716274603937, "grad_norm": 0.9529888788783099, "learning_rate": 5.873320537428023e-06, "loss": 0.7438, "step": 307 }, { "epoch": 0.2957273163706193, "grad_norm": 1.049261563046059, "learning_rate": 5.892514395393475e-06, "loss": 0.7846, "step": 308 }, { "epoch": 0.29668746999519924, "grad_norm": 0.889187432790426, "learning_rate": 5.911708253358925e-06, "loss": 0.6348, "step": 309 }, { "epoch": 0.29764762361977914, "grad_norm": 1.0836258652320905, "learning_rate": 5.930902111324377e-06, "loss": 0.6097, "step": 310 }, { "epoch": 0.2986077772443591, "grad_norm": 0.9668579895111062, "learning_rate": 5.950095969289828e-06, "loss": 0.7398, "step": 311 }, { "epoch": 0.299567930868939, "grad_norm": 0.9518328295296938, "learning_rate": 5.969289827255279e-06, "loss": 0.7836, "step": 312 }, { "epoch": 0.300528084493519, "grad_norm": 0.7800066286942839, "learning_rate": 5.9884836852207295e-06, "loss": 0.6945, "step": 313 }, { "epoch": 0.3014882381180989, "grad_norm": 0.8787425505431756, "learning_rate": 6.007677543186181e-06, "loss": 0.745, "step": 314 }, { "epoch": 0.30244839174267885, "grad_norm": 1.3426918709138653, "learning_rate": 6.026871401151631e-06, "loss": 0.9779, "step": 315 }, { "epoch": 0.30340854536725875, "grad_norm": 1.299941666888297, "learning_rate": 6.0460652591170836e-06, "loss": 0.9095, "step": 316 }, { "epoch": 0.3043686989918387, "grad_norm": 0.961293697561627, "learning_rate": 6.065259117082534e-06, "loss": 0.7947, "step": 317 }, { "epoch": 0.3053288526164186, "grad_norm": 0.9533634198240035, "learning_rate": 6.084452975047985e-06, "loss": 0.7173, "step": 318 }, { "epoch": 0.3062890062409986, "grad_norm": 0.7572427013425438, "learning_rate": 6.103646833013436e-06, "loss": 0.606, "step": 319 }, { "epoch": 0.3072491598655785, "grad_norm": 0.882483616545893, "learning_rate": 6.122840690978887e-06, "loss": 0.7647, "step": 320 }, { "epoch": 0.3082093134901584, "grad_norm": 1.06978173756966, "learning_rate": 6.142034548944338e-06, "loss": 0.8331, "step": 321 }, { "epoch": 0.30916946711473836, "grad_norm": 1.3703517602515691, "learning_rate": 6.161228406909789e-06, "loss": 0.7888, "step": 322 }, { "epoch": 0.31012962073931827, "grad_norm": 1.1826263897798015, "learning_rate": 6.18042226487524e-06, "loss": 0.7575, "step": 323 }, { "epoch": 0.31108977436389823, "grad_norm": 1.0945080266580793, "learning_rate": 6.199616122840692e-06, "loss": 0.7261, "step": 324 }, { "epoch": 0.31204992798847814, "grad_norm": 0.9012107985504348, "learning_rate": 6.218809980806143e-06, "loss": 0.8162, "step": 325 }, { "epoch": 0.3130100816130581, "grad_norm": 0.9995543247416003, "learning_rate": 6.238003838771594e-06, "loss": 0.7411, "step": 326 }, { "epoch": 0.313970235237638, "grad_norm": 0.7786967801573749, "learning_rate": 6.257197696737045e-06, "loss": 0.7166, "step": 327 }, { "epoch": 0.31493038886221797, "grad_norm": 0.8853004021773913, "learning_rate": 6.2763915547024955e-06, "loss": 0.8118, "step": 328 }, { "epoch": 0.3158905424867979, "grad_norm": 1.2951073638207742, "learning_rate": 6.295585412667947e-06, "loss": 0.6731, "step": 329 }, { "epoch": 0.31685069611137784, "grad_norm": 0.9880562373178796, "learning_rate": 6.314779270633397e-06, "loss": 0.6441, "step": 330 }, { "epoch": 0.31781084973595775, "grad_norm": 0.8577381760661504, "learning_rate": 6.3339731285988495e-06, "loss": 0.7221, "step": 331 }, { "epoch": 0.3187710033605377, "grad_norm": 0.7886802107419465, "learning_rate": 6.3531669865643e-06, "loss": 0.7036, "step": 332 }, { "epoch": 0.3197311569851176, "grad_norm": 0.8665790309959003, "learning_rate": 6.372360844529751e-06, "loss": 0.735, "step": 333 }, { "epoch": 0.3206913106096976, "grad_norm": 1.0794767268002694, "learning_rate": 6.391554702495202e-06, "loss": 0.8128, "step": 334 }, { "epoch": 0.3216514642342775, "grad_norm": 1.0866228599492807, "learning_rate": 6.410748560460653e-06, "loss": 0.7703, "step": 335 }, { "epoch": 0.3226116178588574, "grad_norm": 1.0418023056439853, "learning_rate": 6.429942418426104e-06, "loss": 0.7386, "step": 336 }, { "epoch": 0.32357177148343735, "grad_norm": 1.174499720924298, "learning_rate": 6.449136276391556e-06, "loss": 0.7899, "step": 337 }, { "epoch": 0.32453192510801726, "grad_norm": 0.7286823301909906, "learning_rate": 6.4683301343570056e-06, "loss": 0.5629, "step": 338 }, { "epoch": 0.3254920787325972, "grad_norm": 1.0584827642657384, "learning_rate": 6.487523992322458e-06, "loss": 0.7761, "step": 339 }, { "epoch": 0.32645223235717713, "grad_norm": 0.94495930134582, "learning_rate": 6.506717850287908e-06, "loss": 0.7505, "step": 340 }, { "epoch": 0.3274123859817571, "grad_norm": 0.8619043458231381, "learning_rate": 6.52591170825336e-06, "loss": 0.6527, "step": 341 }, { "epoch": 0.328372539606337, "grad_norm": 0.8180867311426304, "learning_rate": 6.54510556621881e-06, "loss": 0.7524, "step": 342 }, { "epoch": 0.32933269323091696, "grad_norm": 1.1605834329506106, "learning_rate": 6.5642994241842614e-06, "loss": 0.7572, "step": 343 }, { "epoch": 0.33029284685549687, "grad_norm": 0.8869058178225163, "learning_rate": 6.583493282149712e-06, "loss": 0.6303, "step": 344 }, { "epoch": 0.33125300048007683, "grad_norm": 1.0917961143110442, "learning_rate": 6.602687140115164e-06, "loss": 0.8181, "step": 345 }, { "epoch": 0.33221315410465674, "grad_norm": 0.7757532371426741, "learning_rate": 6.621880998080615e-06, "loss": 0.6531, "step": 346 }, { "epoch": 0.3331733077292367, "grad_norm": 1.0815351055429616, "learning_rate": 6.641074856046066e-06, "loss": 0.7101, "step": 347 }, { "epoch": 0.3341334613538166, "grad_norm": 0.9769891116066654, "learning_rate": 6.6602687140115165e-06, "loss": 0.72, "step": 348 }, { "epoch": 0.33509361497839657, "grad_norm": 0.9706165558184864, "learning_rate": 6.679462571976968e-06, "loss": 0.7017, "step": 349 }, { "epoch": 0.3360537686029765, "grad_norm": 1.2023504974077053, "learning_rate": 6.698656429942419e-06, "loss": 0.7904, "step": 350 }, { "epoch": 0.3370139222275564, "grad_norm": 0.9823047344707551, "learning_rate": 6.71785028790787e-06, "loss": 0.6535, "step": 351 }, { "epoch": 0.33797407585213635, "grad_norm": 1.2045397940307936, "learning_rate": 6.737044145873322e-06, "loss": 0.7977, "step": 352 }, { "epoch": 0.33893422947671625, "grad_norm": 0.8725735574186719, "learning_rate": 6.756238003838772e-06, "loss": 0.6739, "step": 353 }, { "epoch": 0.3398943831012962, "grad_norm": 0.9424441938795368, "learning_rate": 6.775431861804224e-06, "loss": 0.7596, "step": 354 }, { "epoch": 0.3408545367258761, "grad_norm": 0.7718425488729211, "learning_rate": 6.794625719769674e-06, "loss": 0.7186, "step": 355 }, { "epoch": 0.3418146903504561, "grad_norm": 1.1036993482534156, "learning_rate": 6.8138195777351256e-06, "loss": 0.7898, "step": 356 }, { "epoch": 0.342774843975036, "grad_norm": 1.5909401974325728, "learning_rate": 6.833013435700576e-06, "loss": 0.8943, "step": 357 }, { "epoch": 0.34373499759961595, "grad_norm": 0.9314489542524216, "learning_rate": 6.852207293666027e-06, "loss": 0.7374, "step": 358 }, { "epoch": 0.34469515122419586, "grad_norm": 0.8426759766762809, "learning_rate": 6.871401151631478e-06, "loss": 0.6889, "step": 359 }, { "epoch": 0.3456553048487758, "grad_norm": 0.7129722021603497, "learning_rate": 6.89059500959693e-06, "loss": 0.6388, "step": 360 }, { "epoch": 0.34661545847335573, "grad_norm": 0.7975394082026934, "learning_rate": 6.909788867562381e-06, "loss": 0.5806, "step": 361 }, { "epoch": 0.3475756120979357, "grad_norm": 1.4783745529402696, "learning_rate": 6.928982725527832e-06, "loss": 0.9238, "step": 362 }, { "epoch": 0.3485357657225156, "grad_norm": 0.8854102307091066, "learning_rate": 6.9481765834932824e-06, "loss": 0.7583, "step": 363 }, { "epoch": 0.34949591934709556, "grad_norm": 0.9825606768769671, "learning_rate": 6.967370441458734e-06, "loss": 0.8158, "step": 364 }, { "epoch": 0.35045607297167547, "grad_norm": 0.8840323142446338, "learning_rate": 6.986564299424184e-06, "loss": 0.781, "step": 365 }, { "epoch": 0.3514162265962554, "grad_norm": 0.7384945984152259, "learning_rate": 7.005758157389636e-06, "loss": 0.6453, "step": 366 }, { "epoch": 0.35237638022083534, "grad_norm": 0.902587466489343, "learning_rate": 7.024952015355086e-06, "loss": 0.6616, "step": 367 }, { "epoch": 0.35333653384541525, "grad_norm": 1.2114957952235266, "learning_rate": 7.044145873320538e-06, "loss": 0.9046, "step": 368 }, { "epoch": 0.3542966874699952, "grad_norm": 0.8558958611892297, "learning_rate": 7.063339731285989e-06, "loss": 0.7402, "step": 369 }, { "epoch": 0.3552568410945751, "grad_norm": 1.2795785169610148, "learning_rate": 7.08253358925144e-06, "loss": 0.7635, "step": 370 }, { "epoch": 0.3562169947191551, "grad_norm": 1.0046992054148385, "learning_rate": 7.101727447216891e-06, "loss": 0.8183, "step": 371 }, { "epoch": 0.357177148343735, "grad_norm": 0.9479551567936374, "learning_rate": 7.120921305182342e-06, "loss": 0.7862, "step": 372 }, { "epoch": 0.35813730196831495, "grad_norm": 0.8029812114649825, "learning_rate": 7.1401151631477925e-06, "loss": 0.6585, "step": 373 }, { "epoch": 0.35909745559289485, "grad_norm": 1.683090117241035, "learning_rate": 7.159309021113245e-06, "loss": 0.9663, "step": 374 }, { "epoch": 0.3600576092174748, "grad_norm": 0.8620985218199626, "learning_rate": 7.178502879078696e-06, "loss": 0.6748, "step": 375 }, { "epoch": 0.3610177628420547, "grad_norm": 1.098532977652713, "learning_rate": 7.1976967370441466e-06, "loss": 0.7628, "step": 376 }, { "epoch": 0.3619779164666347, "grad_norm": 1.121166416413462, "learning_rate": 7.216890595009598e-06, "loss": 0.7268, "step": 377 }, { "epoch": 0.3629380700912146, "grad_norm": 1.36961192743179, "learning_rate": 7.236084452975048e-06, "loss": 0.8127, "step": 378 }, { "epoch": 0.3638982237157945, "grad_norm": 1.1844005122326056, "learning_rate": 7.2552783109405e-06, "loss": 0.8257, "step": 379 }, { "epoch": 0.36485837734037446, "grad_norm": 0.9958938832198809, "learning_rate": 7.27447216890595e-06, "loss": 0.7103, "step": 380 }, { "epoch": 0.36581853096495437, "grad_norm": 1.1009158289719336, "learning_rate": 7.2936660268714024e-06, "loss": 0.8669, "step": 381 }, { "epoch": 0.36677868458953433, "grad_norm": 0.8349328223907851, "learning_rate": 7.312859884836853e-06, "loss": 0.7581, "step": 382 }, { "epoch": 0.36773883821411424, "grad_norm": 0.8233647000544146, "learning_rate": 7.332053742802304e-06, "loss": 0.6585, "step": 383 }, { "epoch": 0.3686989918386942, "grad_norm": 0.7785997082968157, "learning_rate": 7.351247600767755e-06, "loss": 0.6134, "step": 384 }, { "epoch": 0.3696591454632741, "grad_norm": 1.0795972922722161, "learning_rate": 7.370441458733206e-06, "loss": 0.7775, "step": 385 }, { "epoch": 0.37061929908785407, "grad_norm": 0.9827459840879991, "learning_rate": 7.389635316698657e-06, "loss": 0.6907, "step": 386 }, { "epoch": 0.371579452712434, "grad_norm": 1.1518044282226843, "learning_rate": 7.408829174664108e-06, "loss": 0.8148, "step": 387 }, { "epoch": 0.37253960633701394, "grad_norm": 0.8892270213317611, "learning_rate": 7.4280230326295585e-06, "loss": 0.714, "step": 388 }, { "epoch": 0.37349975996159385, "grad_norm": 0.8402787785526201, "learning_rate": 7.447216890595011e-06, "loss": 0.8013, "step": 389 }, { "epoch": 0.3744599135861738, "grad_norm": 0.7465357389659705, "learning_rate": 7.466410748560461e-06, "loss": 0.592, "step": 390 }, { "epoch": 0.3754200672107537, "grad_norm": 0.8943875406721344, "learning_rate": 7.4856046065259125e-06, "loss": 0.698, "step": 391 }, { "epoch": 0.3763802208353337, "grad_norm": 1.3035546863753054, "learning_rate": 7.504798464491363e-06, "loss": 0.8622, "step": 392 }, { "epoch": 0.3773403744599136, "grad_norm": 1.1141541301340707, "learning_rate": 7.523992322456814e-06, "loss": 0.764, "step": 393 }, { "epoch": 0.3783005280844935, "grad_norm": 0.8041583068170934, "learning_rate": 7.543186180422265e-06, "loss": 0.7432, "step": 394 }, { "epoch": 0.37926068170907346, "grad_norm": 1.056896179251162, "learning_rate": 7.562380038387716e-06, "loss": 0.5515, "step": 395 }, { "epoch": 0.38022083533365336, "grad_norm": 0.779917751913091, "learning_rate": 7.581573896353167e-06, "loss": 0.6253, "step": 396 }, { "epoch": 0.3811809889582333, "grad_norm": 0.9880246473175981, "learning_rate": 7.600767754318619e-06, "loss": 0.7614, "step": 397 }, { "epoch": 0.38214114258281323, "grad_norm": 1.0183208438084672, "learning_rate": 7.61996161228407e-06, "loss": 0.6934, "step": 398 }, { "epoch": 0.3831012962073932, "grad_norm": 0.9420061441791449, "learning_rate": 7.63915547024952e-06, "loss": 0.7486, "step": 399 }, { "epoch": 0.3840614498319731, "grad_norm": 1.2916648440151124, "learning_rate": 7.658349328214972e-06, "loss": 0.8062, "step": 400 }, { "epoch": 0.38502160345655306, "grad_norm": 0.9577597882182558, "learning_rate": 7.677543186180423e-06, "loss": 0.8577, "step": 401 }, { "epoch": 0.38598175708113297, "grad_norm": 0.9471769523102869, "learning_rate": 7.696737044145875e-06, "loss": 0.6681, "step": 402 }, { "epoch": 0.38694191070571293, "grad_norm": 0.78498206282339, "learning_rate": 7.715930902111324e-06, "loss": 0.7172, "step": 403 }, { "epoch": 0.38790206433029284, "grad_norm": 1.0077298023608472, "learning_rate": 7.735124760076776e-06, "loss": 0.7859, "step": 404 }, { "epoch": 0.3888622179548728, "grad_norm": 1.2951478396634468, "learning_rate": 7.754318618042227e-06, "loss": 0.782, "step": 405 }, { "epoch": 0.3898223715794527, "grad_norm": 0.9707550583344386, "learning_rate": 7.773512476007678e-06, "loss": 0.7049, "step": 406 }, { "epoch": 0.39078252520403267, "grad_norm": 1.0699366252835165, "learning_rate": 7.79270633397313e-06, "loss": 0.7962, "step": 407 }, { "epoch": 0.3917426788286126, "grad_norm": 0.9334634852532233, "learning_rate": 7.811900191938581e-06, "loss": 0.7302, "step": 408 }, { "epoch": 0.3927028324531925, "grad_norm": 0.9528340829487093, "learning_rate": 7.83109404990403e-06, "loss": 0.684, "step": 409 }, { "epoch": 0.39366298607777245, "grad_norm": 1.056330937697354, "learning_rate": 7.850287907869482e-06, "loss": 0.7689, "step": 410 }, { "epoch": 0.39462313970235235, "grad_norm": 0.9400269020958039, "learning_rate": 7.869481765834934e-06, "loss": 0.7945, "step": 411 }, { "epoch": 0.3955832933269323, "grad_norm": 0.8443889140301137, "learning_rate": 7.888675623800385e-06, "loss": 0.7389, "step": 412 }, { "epoch": 0.3965434469515122, "grad_norm": 1.2576795652780997, "learning_rate": 7.907869481765835e-06, "loss": 0.8591, "step": 413 }, { "epoch": 0.3975036005760922, "grad_norm": 0.9375464199274235, "learning_rate": 7.927063339731288e-06, "loss": 0.7015, "step": 414 }, { "epoch": 0.3984637542006721, "grad_norm": 0.7587116111552183, "learning_rate": 7.946257197696737e-06, "loss": 0.6345, "step": 415 }, { "epoch": 0.39942390782525206, "grad_norm": 0.8149000205465763, "learning_rate": 7.965451055662189e-06, "loss": 0.6723, "step": 416 }, { "epoch": 0.40038406144983196, "grad_norm": 0.8805777344629909, "learning_rate": 7.98464491362764e-06, "loss": 0.7578, "step": 417 }, { "epoch": 0.4013442150744119, "grad_norm": 1.3267050243855654, "learning_rate": 8.003838771593091e-06, "loss": 0.8162, "step": 418 }, { "epoch": 0.40230436869899183, "grad_norm": 1.15154809452485, "learning_rate": 8.023032629558541e-06, "loss": 0.6125, "step": 419 }, { "epoch": 0.4032645223235718, "grad_norm": 1.1271758708630868, "learning_rate": 8.042226487523992e-06, "loss": 0.7904, "step": 420 }, { "epoch": 0.4042246759481517, "grad_norm": 1.0157353934920994, "learning_rate": 8.061420345489444e-06, "loss": 0.7412, "step": 421 }, { "epoch": 0.40518482957273166, "grad_norm": 0.9938853797164021, "learning_rate": 8.080614203454895e-06, "loss": 0.7411, "step": 422 }, { "epoch": 0.40614498319731157, "grad_norm": 0.8593830852334721, "learning_rate": 8.099808061420346e-06, "loss": 0.6313, "step": 423 }, { "epoch": 0.4071051368218915, "grad_norm": 0.8807373078174986, "learning_rate": 8.119001919385798e-06, "loss": 0.6556, "step": 424 }, { "epoch": 0.40806529044647144, "grad_norm": 1.0015718371753883, "learning_rate": 8.138195777351249e-06, "loss": 0.7424, "step": 425 }, { "epoch": 0.40902544407105135, "grad_norm": 1.0337226680109526, "learning_rate": 8.157389635316699e-06, "loss": 0.6934, "step": 426 }, { "epoch": 0.4099855976956313, "grad_norm": 0.7937725114987125, "learning_rate": 8.176583493282152e-06, "loss": 0.6505, "step": 427 }, { "epoch": 0.4109457513202112, "grad_norm": 0.9299870157811758, "learning_rate": 8.195777351247601e-06, "loss": 0.7542, "step": 428 }, { "epoch": 0.4119059049447912, "grad_norm": 0.8791199207707324, "learning_rate": 8.214971209213053e-06, "loss": 0.7623, "step": 429 }, { "epoch": 0.4128660585693711, "grad_norm": 1.1426468623144077, "learning_rate": 8.234165067178504e-06, "loss": 0.789, "step": 430 }, { "epoch": 0.41382621219395105, "grad_norm": 0.8168542543188545, "learning_rate": 8.253358925143955e-06, "loss": 0.6747, "step": 431 }, { "epoch": 0.41478636581853096, "grad_norm": 1.0441968027761084, "learning_rate": 8.272552783109405e-06, "loss": 0.6908, "step": 432 }, { "epoch": 0.4157465194431109, "grad_norm": 0.7788632152667789, "learning_rate": 8.291746641074856e-06, "loss": 0.696, "step": 433 }, { "epoch": 0.4167066730676908, "grad_norm": 0.8974952524286895, "learning_rate": 8.310940499040308e-06, "loss": 0.6547, "step": 434 }, { "epoch": 0.4176668266922708, "grad_norm": 0.7968952100794404, "learning_rate": 8.330134357005759e-06, "loss": 0.6205, "step": 435 }, { "epoch": 0.4186269803168507, "grad_norm": 0.9724294213912004, "learning_rate": 8.34932821497121e-06, "loss": 0.7498, "step": 436 }, { "epoch": 0.4195871339414306, "grad_norm": 1.022082572417475, "learning_rate": 8.368522072936662e-06, "loss": 0.7414, "step": 437 }, { "epoch": 0.42054728756601056, "grad_norm": 0.959012781249683, "learning_rate": 8.387715930902111e-06, "loss": 0.7734, "step": 438 }, { "epoch": 0.42150744119059047, "grad_norm": 0.8707452353595925, "learning_rate": 8.406909788867563e-06, "loss": 0.6953, "step": 439 }, { "epoch": 0.42246759481517043, "grad_norm": 1.2526422559413455, "learning_rate": 8.426103646833014e-06, "loss": 0.9247, "step": 440 }, { "epoch": 0.42342774843975034, "grad_norm": 0.9984024389205554, "learning_rate": 8.445297504798465e-06, "loss": 0.7181, "step": 441 }, { "epoch": 0.4243879020643303, "grad_norm": 0.8710615567848037, "learning_rate": 8.464491362763915e-06, "loss": 0.7077, "step": 442 }, { "epoch": 0.4253480556889102, "grad_norm": 1.125311469315276, "learning_rate": 8.483685220729368e-06, "loss": 0.5924, "step": 443 }, { "epoch": 0.42630820931349017, "grad_norm": 0.9326542627880643, "learning_rate": 8.502879078694818e-06, "loss": 0.7247, "step": 444 }, { "epoch": 0.4272683629380701, "grad_norm": 0.773824757029302, "learning_rate": 8.522072936660269e-06, "loss": 0.667, "step": 445 }, { "epoch": 0.42822851656265004, "grad_norm": 1.1659356570563335, "learning_rate": 8.54126679462572e-06, "loss": 0.7784, "step": 446 }, { "epoch": 0.42918867018722995, "grad_norm": 1.3554461015381054, "learning_rate": 8.560460652591172e-06, "loss": 0.7618, "step": 447 }, { "epoch": 0.4301488238118099, "grad_norm": 0.9004778032227636, "learning_rate": 8.579654510556623e-06, "loss": 0.6726, "step": 448 }, { "epoch": 0.4311089774363898, "grad_norm": 0.8804816895336388, "learning_rate": 8.598848368522073e-06, "loss": 0.7555, "step": 449 }, { "epoch": 0.4320691310609698, "grad_norm": 1.0180498880285949, "learning_rate": 8.618042226487526e-06, "loss": 0.7426, "step": 450 }, { "epoch": 0.4330292846855497, "grad_norm": 0.9178279643781447, "learning_rate": 8.637236084452976e-06, "loss": 0.7887, "step": 451 }, { "epoch": 0.4339894383101296, "grad_norm": 1.0017678720320724, "learning_rate": 8.656429942418427e-06, "loss": 0.6795, "step": 452 }, { "epoch": 0.43494959193470956, "grad_norm": 0.7915978898225524, "learning_rate": 8.675623800383878e-06, "loss": 0.6427, "step": 453 }, { "epoch": 0.43590974555928946, "grad_norm": 0.9294552300992335, "learning_rate": 8.69481765834933e-06, "loss": 0.6818, "step": 454 }, { "epoch": 0.4368698991838694, "grad_norm": 1.716522641184676, "learning_rate": 8.71401151631478e-06, "loss": 0.9021, "step": 455 }, { "epoch": 0.43783005280844933, "grad_norm": 0.8336570255469592, "learning_rate": 8.73320537428023e-06, "loss": 0.6636, "step": 456 }, { "epoch": 0.4387902064330293, "grad_norm": 1.4400976413245044, "learning_rate": 8.752399232245682e-06, "loss": 0.8629, "step": 457 }, { "epoch": 0.4397503600576092, "grad_norm": 0.9377636278585642, "learning_rate": 8.771593090211133e-06, "loss": 0.6756, "step": 458 }, { "epoch": 0.44071051368218916, "grad_norm": 1.139205458011471, "learning_rate": 8.790786948176585e-06, "loss": 0.7958, "step": 459 }, { "epoch": 0.44167066730676907, "grad_norm": 0.9984708765391459, "learning_rate": 8.809980806142036e-06, "loss": 0.6759, "step": 460 }, { "epoch": 0.44263082093134903, "grad_norm": 1.3964495105488668, "learning_rate": 8.829174664107486e-06, "loss": 0.7354, "step": 461 }, { "epoch": 0.44359097455592894, "grad_norm": 1.0907869870804399, "learning_rate": 8.848368522072937e-06, "loss": 0.7507, "step": 462 }, { "epoch": 0.4445511281805089, "grad_norm": 1.0099897056128142, "learning_rate": 8.867562380038388e-06, "loss": 0.7477, "step": 463 }, { "epoch": 0.4455112818050888, "grad_norm": 1.0495564991335136, "learning_rate": 8.88675623800384e-06, "loss": 0.7837, "step": 464 }, { "epoch": 0.4464714354296688, "grad_norm": 1.048595445841977, "learning_rate": 8.905950095969291e-06, "loss": 0.7281, "step": 465 }, { "epoch": 0.4474315890542487, "grad_norm": 0.8548131245489575, "learning_rate": 8.925143953934742e-06, "loss": 0.8138, "step": 466 }, { "epoch": 0.4483917426788286, "grad_norm": 1.033284062429633, "learning_rate": 8.944337811900192e-06, "loss": 0.7175, "step": 467 }, { "epoch": 0.44935189630340855, "grad_norm": 1.1131511696561054, "learning_rate": 8.963531669865643e-06, "loss": 0.75, "step": 468 }, { "epoch": 0.45031204992798846, "grad_norm": 0.8798549127437901, "learning_rate": 8.982725527831095e-06, "loss": 0.6111, "step": 469 }, { "epoch": 0.4512722035525684, "grad_norm": 1.0670230523984496, "learning_rate": 9.001919385796546e-06, "loss": 0.7018, "step": 470 }, { "epoch": 0.4522323571771483, "grad_norm": 0.7977266252807448, "learning_rate": 9.021113243761996e-06, "loss": 0.6704, "step": 471 }, { "epoch": 0.4531925108017283, "grad_norm": 0.8303077539658585, "learning_rate": 9.040307101727449e-06, "loss": 0.7104, "step": 472 }, { "epoch": 0.4541526644263082, "grad_norm": 0.900809081224308, "learning_rate": 9.0595009596929e-06, "loss": 0.7467, "step": 473 }, { "epoch": 0.45511281805088816, "grad_norm": 0.9185208338703789, "learning_rate": 9.07869481765835e-06, "loss": 0.6357, "step": 474 }, { "epoch": 0.45607297167546806, "grad_norm": 0.935717932651676, "learning_rate": 9.097888675623801e-06, "loss": 0.7078, "step": 475 }, { "epoch": 0.457033125300048, "grad_norm": 1.363153880109318, "learning_rate": 9.117082533589252e-06, "loss": 0.8336, "step": 476 }, { "epoch": 0.45799327892462793, "grad_norm": 0.9398201823468145, "learning_rate": 9.136276391554704e-06, "loss": 0.7902, "step": 477 }, { "epoch": 0.4589534325492079, "grad_norm": 1.007249633490007, "learning_rate": 9.155470249520153e-06, "loss": 0.8324, "step": 478 }, { "epoch": 0.4599135861737878, "grad_norm": 0.922074000266469, "learning_rate": 9.174664107485606e-06, "loss": 0.6228, "step": 479 }, { "epoch": 0.46087373979836777, "grad_norm": 0.790500274226629, "learning_rate": 9.193857965451056e-06, "loss": 0.6233, "step": 480 }, { "epoch": 0.46183389342294767, "grad_norm": 1.1464149735182623, "learning_rate": 9.213051823416507e-06, "loss": 0.7877, "step": 481 }, { "epoch": 0.4627940470475276, "grad_norm": 0.8703218261259326, "learning_rate": 9.232245681381959e-06, "loss": 0.6669, "step": 482 }, { "epoch": 0.46375420067210754, "grad_norm": 1.0164033716576573, "learning_rate": 9.25143953934741e-06, "loss": 0.7762, "step": 483 }, { "epoch": 0.46471435429668745, "grad_norm": 0.8869171277007261, "learning_rate": 9.27063339731286e-06, "loss": 0.6907, "step": 484 }, { "epoch": 0.4656745079212674, "grad_norm": 1.0582506629713386, "learning_rate": 9.289827255278311e-06, "loss": 0.8009, "step": 485 }, { "epoch": 0.4666346615458473, "grad_norm": 1.0730518951482526, "learning_rate": 9.309021113243763e-06, "loss": 0.6907, "step": 486 }, { "epoch": 0.4675948151704273, "grad_norm": 0.7757250271658104, "learning_rate": 9.328214971209214e-06, "loss": 0.6545, "step": 487 }, { "epoch": 0.4685549687950072, "grad_norm": 0.6760914190301193, "learning_rate": 9.347408829174665e-06, "loss": 0.5518, "step": 488 }, { "epoch": 0.46951512241958715, "grad_norm": 0.8573952437495032, "learning_rate": 9.366602687140117e-06, "loss": 0.7337, "step": 489 }, { "epoch": 0.47047527604416706, "grad_norm": 0.9862002551432681, "learning_rate": 9.385796545105566e-06, "loss": 0.7595, "step": 490 }, { "epoch": 0.471435429668747, "grad_norm": 1.0278786897567849, "learning_rate": 9.404990403071018e-06, "loss": 0.5983, "step": 491 }, { "epoch": 0.4723955832933269, "grad_norm": 1.2101393913530925, "learning_rate": 9.424184261036469e-06, "loss": 0.8741, "step": 492 }, { "epoch": 0.4733557369179069, "grad_norm": 1.059531280827221, "learning_rate": 9.44337811900192e-06, "loss": 0.7624, "step": 493 }, { "epoch": 0.4743158905424868, "grad_norm": 1.0576120509385398, "learning_rate": 9.46257197696737e-06, "loss": 0.7433, "step": 494 }, { "epoch": 0.47527604416706676, "grad_norm": 0.8751161666493031, "learning_rate": 9.481765834932823e-06, "loss": 0.7384, "step": 495 }, { "epoch": 0.47623619779164666, "grad_norm": 0.9294963369121172, "learning_rate": 9.500959692898273e-06, "loss": 0.8079, "step": 496 }, { "epoch": 0.47719635141622657, "grad_norm": 0.9173082489414217, "learning_rate": 9.520153550863724e-06, "loss": 0.6774, "step": 497 }, { "epoch": 0.47815650504080653, "grad_norm": 1.2564899185306744, "learning_rate": 9.539347408829175e-06, "loss": 0.8885, "step": 498 }, { "epoch": 0.47911665866538644, "grad_norm": 0.8554570539207819, "learning_rate": 9.558541266794627e-06, "loss": 0.6752, "step": 499 }, { "epoch": 0.4800768122899664, "grad_norm": 0.8862232829744724, "learning_rate": 9.577735124760078e-06, "loss": 0.7101, "step": 500 }, { "epoch": 0.4810369659145463, "grad_norm": 1.0118218899071425, "learning_rate": 9.59692898272553e-06, "loss": 0.8252, "step": 501 }, { "epoch": 0.4819971195391263, "grad_norm": 0.9510071677433547, "learning_rate": 9.61612284069098e-06, "loss": 0.6858, "step": 502 }, { "epoch": 0.4829572731637062, "grad_norm": 0.983730217072968, "learning_rate": 9.63531669865643e-06, "loss": 0.6321, "step": 503 }, { "epoch": 0.48391742678828614, "grad_norm": 0.7941725714441769, "learning_rate": 9.654510556621882e-06, "loss": 0.5965, "step": 504 }, { "epoch": 0.48487758041286605, "grad_norm": 1.1292117660318046, "learning_rate": 9.673704414587333e-06, "loss": 0.6594, "step": 505 }, { "epoch": 0.485837734037446, "grad_norm": 1.2504275862262606, "learning_rate": 9.692898272552784e-06, "loss": 0.797, "step": 506 }, { "epoch": 0.4867978876620259, "grad_norm": 0.9027838696434731, "learning_rate": 9.712092130518234e-06, "loss": 0.6505, "step": 507 }, { "epoch": 0.4877580412866059, "grad_norm": 0.7390700284636669, "learning_rate": 9.731285988483687e-06, "loss": 0.6623, "step": 508 }, { "epoch": 0.4887181949111858, "grad_norm": 1.3546696766975825, "learning_rate": 9.750479846449137e-06, "loss": 0.8398, "step": 509 }, { "epoch": 0.4896783485357657, "grad_norm": 0.9308423377123827, "learning_rate": 9.769673704414588e-06, "loss": 0.7359, "step": 510 }, { "epoch": 0.49063850216034566, "grad_norm": 1.4164810668477907, "learning_rate": 9.78886756238004e-06, "loss": 0.8357, "step": 511 }, { "epoch": 0.49159865578492556, "grad_norm": 1.1678957755455932, "learning_rate": 9.80806142034549e-06, "loss": 0.8051, "step": 512 }, { "epoch": 0.4925588094095055, "grad_norm": 1.651690591039931, "learning_rate": 9.82725527831094e-06, "loss": 0.9281, "step": 513 }, { "epoch": 0.49351896303408543, "grad_norm": 0.7712655924568504, "learning_rate": 9.846449136276392e-06, "loss": 0.6199, "step": 514 }, { "epoch": 0.4944791166586654, "grad_norm": 0.8730812634245314, "learning_rate": 9.865642994241843e-06, "loss": 0.6054, "step": 515 }, { "epoch": 0.4954392702832453, "grad_norm": 1.2892294599499545, "learning_rate": 9.884836852207294e-06, "loss": 0.7117, "step": 516 }, { "epoch": 0.49639942390782527, "grad_norm": 0.8674583236092615, "learning_rate": 9.904030710172746e-06, "loss": 0.7137, "step": 517 }, { "epoch": 0.4973595775324052, "grad_norm": 1.0553488900307437, "learning_rate": 9.923224568138197e-06, "loss": 0.7206, "step": 518 }, { "epoch": 0.49831973115698514, "grad_norm": 1.1817240999316285, "learning_rate": 9.942418426103647e-06, "loss": 0.7629, "step": 519 }, { "epoch": 0.49927988478156504, "grad_norm": 0.980335079348206, "learning_rate": 9.961612284069098e-06, "loss": 0.7732, "step": 520 }, { "epoch": 0.500240038406145, "grad_norm": 1.0588741617720998, "learning_rate": 9.98080614203455e-06, "loss": 0.8913, "step": 521 }, { "epoch": 0.501200192030725, "grad_norm": 0.8974421759474529, "learning_rate": 1e-05, "loss": 0.6905, "step": 522 }, { "epoch": 0.5021603456553049, "grad_norm": 0.8158269666080086, "learning_rate": 9.999998877776725e-06, "loss": 0.6276, "step": 523 }, { "epoch": 0.5031204992798848, "grad_norm": 0.8853422435385098, "learning_rate": 9.999995511107399e-06, "loss": 0.6477, "step": 524 }, { "epoch": 0.5040806529044647, "grad_norm": 1.2285619791691047, "learning_rate": 9.999989899993536e-06, "loss": 0.8229, "step": 525 }, { "epoch": 0.5050408065290446, "grad_norm": 0.8694600709440518, "learning_rate": 9.999982044437652e-06, "loss": 0.7203, "step": 526 }, { "epoch": 0.5060009601536246, "grad_norm": 0.953040121864583, "learning_rate": 9.999971944443276e-06, "loss": 0.7572, "step": 527 }, { "epoch": 0.5069611137782045, "grad_norm": 1.3246484470019573, "learning_rate": 9.999959600014942e-06, "loss": 0.7749, "step": 528 }, { "epoch": 0.5079212674027844, "grad_norm": 0.8483217732652649, "learning_rate": 9.999945011158189e-06, "loss": 0.6593, "step": 529 }, { "epoch": 0.5088814210273643, "grad_norm": 1.0147646417211493, "learning_rate": 9.999928177879567e-06, "loss": 0.732, "step": 530 }, { "epoch": 0.5098415746519444, "grad_norm": 0.8722744341961544, "learning_rate": 9.999909100186634e-06, "loss": 0.6517, "step": 531 }, { "epoch": 0.5108017282765243, "grad_norm": 0.9064427704774831, "learning_rate": 9.999887778087948e-06, "loss": 0.7524, "step": 532 }, { "epoch": 0.5117618819011042, "grad_norm": 0.7885443931582331, "learning_rate": 9.999864211593088e-06, "loss": 0.6325, "step": 533 }, { "epoch": 0.5127220355256841, "grad_norm": 0.9376323379095572, "learning_rate": 9.999838400712627e-06, "loss": 0.7385, "step": 534 }, { "epoch": 0.5136821891502641, "grad_norm": 0.8159183187292293, "learning_rate": 9.999810345458153e-06, "loss": 0.6733, "step": 535 }, { "epoch": 0.514642342774844, "grad_norm": 0.9756134334934099, "learning_rate": 9.999780045842262e-06, "loss": 0.7396, "step": 536 }, { "epoch": 0.5156024963994239, "grad_norm": 0.9782658097677672, "learning_rate": 9.999747501878551e-06, "loss": 0.7325, "step": 537 }, { "epoch": 0.5165626500240038, "grad_norm": 0.9112399886746073, "learning_rate": 9.999712713581634e-06, "loss": 0.6487, "step": 538 }, { "epoch": 0.5175228036485837, "grad_norm": 0.7204106949046135, "learning_rate": 9.99967568096712e-06, "loss": 0.6185, "step": 539 }, { "epoch": 0.5184829572731637, "grad_norm": 0.7538538717939204, "learning_rate": 9.999636404051638e-06, "loss": 0.7175, "step": 540 }, { "epoch": 0.5194431108977436, "grad_norm": 0.8486480945638997, "learning_rate": 9.999594882852817e-06, "loss": 0.6473, "step": 541 }, { "epoch": 0.5204032645223235, "grad_norm": 1.60480335231693, "learning_rate": 9.999551117389298e-06, "loss": 0.8481, "step": 542 }, { "epoch": 0.5213634181469035, "grad_norm": 0.8915667388623162, "learning_rate": 9.999505107680722e-06, "loss": 0.6603, "step": 543 }, { "epoch": 0.5223235717714835, "grad_norm": 1.07346142794007, "learning_rate": 9.999456853747744e-06, "loss": 0.8322, "step": 544 }, { "epoch": 0.5232837253960634, "grad_norm": 1.1574562127163375, "learning_rate": 9.999406355612029e-06, "loss": 0.7647, "step": 545 }, { "epoch": 0.5242438790206433, "grad_norm": 1.0029199185116155, "learning_rate": 9.99935361329624e-06, "loss": 0.6991, "step": 546 }, { "epoch": 0.5252040326452232, "grad_norm": 0.8258547915816234, "learning_rate": 9.999298626824053e-06, "loss": 0.6539, "step": 547 }, { "epoch": 0.5261641862698032, "grad_norm": 0.9854532488709491, "learning_rate": 9.999241396220152e-06, "loss": 0.6705, "step": 548 }, { "epoch": 0.5271243398943831, "grad_norm": 1.1621509178033496, "learning_rate": 9.999181921510229e-06, "loss": 0.7312, "step": 549 }, { "epoch": 0.528084493518963, "grad_norm": 0.9313108714416162, "learning_rate": 9.999120202720978e-06, "loss": 0.7205, "step": 550 }, { "epoch": 0.5290446471435429, "grad_norm": 0.7784878580815145, "learning_rate": 9.999056239880107e-06, "loss": 0.6071, "step": 551 }, { "epoch": 0.530004800768123, "grad_norm": 1.0574272195823429, "learning_rate": 9.998990033016327e-06, "loss": 0.7161, "step": 552 }, { "epoch": 0.5309649543927029, "grad_norm": 1.0142877431907007, "learning_rate": 9.998921582159357e-06, "loss": 0.6893, "step": 553 }, { "epoch": 0.5319251080172828, "grad_norm": 0.9618624969877319, "learning_rate": 9.998850887339923e-06, "loss": 0.7194, "step": 554 }, { "epoch": 0.5328852616418627, "grad_norm": 0.8869046990555473, "learning_rate": 9.998777948589762e-06, "loss": 0.6927, "step": 555 }, { "epoch": 0.5338454152664426, "grad_norm": 1.2087985098064313, "learning_rate": 9.998702765941613e-06, "loss": 0.8321, "step": 556 }, { "epoch": 0.5348055688910226, "grad_norm": 0.7996681516217893, "learning_rate": 9.998625339429227e-06, "loss": 0.6178, "step": 557 }, { "epoch": 0.5357657225156025, "grad_norm": 0.9842437173142394, "learning_rate": 9.998545669087357e-06, "loss": 0.7869, "step": 558 }, { "epoch": 0.5367258761401824, "grad_norm": 1.1038370192395082, "learning_rate": 9.998463754951767e-06, "loss": 0.7929, "step": 559 }, { "epoch": 0.5376860297647623, "grad_norm": 0.8603358056037531, "learning_rate": 9.998379597059228e-06, "loss": 0.6458, "step": 560 }, { "epoch": 0.5386461833893423, "grad_norm": 0.9464807352785958, "learning_rate": 9.99829319544752e-06, "loss": 0.7622, "step": 561 }, { "epoch": 0.5396063370139222, "grad_norm": 1.022219497478962, "learning_rate": 9.998204550155424e-06, "loss": 0.7488, "step": 562 }, { "epoch": 0.5405664906385022, "grad_norm": 0.9661634238411974, "learning_rate": 9.998113661222732e-06, "loss": 0.6464, "step": 563 }, { "epoch": 0.5415266442630821, "grad_norm": 1.0639385161335813, "learning_rate": 9.998020528690247e-06, "loss": 0.8648, "step": 564 }, { "epoch": 0.5424867978876621, "grad_norm": 0.961132693140685, "learning_rate": 9.99792515259977e-06, "loss": 0.7142, "step": 565 }, { "epoch": 0.543446951512242, "grad_norm": 1.2407157454632114, "learning_rate": 9.997827532994118e-06, "loss": 0.7865, "step": 566 }, { "epoch": 0.5444071051368219, "grad_norm": 0.7996763278899846, "learning_rate": 9.99772766991711e-06, "loss": 0.5146, "step": 567 }, { "epoch": 0.5453672587614018, "grad_norm": 1.2009290522626825, "learning_rate": 9.997625563413574e-06, "loss": 0.8119, "step": 568 }, { "epoch": 0.5463274123859817, "grad_norm": 1.1706581459551841, "learning_rate": 9.997521213529345e-06, "loss": 0.826, "step": 569 }, { "epoch": 0.5472875660105617, "grad_norm": 1.1412143344040107, "learning_rate": 9.997414620311262e-06, "loss": 0.8276, "step": 570 }, { "epoch": 0.5482477196351416, "grad_norm": 0.8473326016697205, "learning_rate": 9.997305783807176e-06, "loss": 0.6663, "step": 571 }, { "epoch": 0.5492078732597215, "grad_norm": 0.7882035369218401, "learning_rate": 9.997194704065942e-06, "loss": 0.5954, "step": 572 }, { "epoch": 0.5501680268843014, "grad_norm": 0.7418262152383681, "learning_rate": 9.997081381137424e-06, "loss": 0.6863, "step": 573 }, { "epoch": 0.5511281805088815, "grad_norm": 0.8902881987186623, "learning_rate": 9.996965815072489e-06, "loss": 0.6168, "step": 574 }, { "epoch": 0.5520883341334614, "grad_norm": 0.9026956350055214, "learning_rate": 9.996848005923012e-06, "loss": 0.6171, "step": 575 }, { "epoch": 0.5530484877580413, "grad_norm": 0.7982859759425469, "learning_rate": 9.996727953741881e-06, "loss": 0.6739, "step": 576 }, { "epoch": 0.5540086413826212, "grad_norm": 1.2989972553857108, "learning_rate": 9.996605658582982e-06, "loss": 0.8252, "step": 577 }, { "epoch": 0.5549687950072012, "grad_norm": 1.0434312584552952, "learning_rate": 9.996481120501214e-06, "loss": 0.6939, "step": 578 }, { "epoch": 0.5559289486317811, "grad_norm": 0.8264764910208572, "learning_rate": 9.996354339552482e-06, "loss": 0.6598, "step": 579 }, { "epoch": 0.556889102256361, "grad_norm": 1.3115411679413957, "learning_rate": 9.996225315793692e-06, "loss": 0.6587, "step": 580 }, { "epoch": 0.5578492558809409, "grad_norm": 1.1324975752263866, "learning_rate": 9.996094049282768e-06, "loss": 0.6166, "step": 581 }, { "epoch": 0.5588094095055209, "grad_norm": 0.8630862856683241, "learning_rate": 9.995960540078627e-06, "loss": 0.6282, "step": 582 }, { "epoch": 0.5597695631301008, "grad_norm": 1.458659924159987, "learning_rate": 9.995824788241206e-06, "loss": 0.8426, "step": 583 }, { "epoch": 0.5607297167546808, "grad_norm": 0.7912246387482282, "learning_rate": 9.995686793831438e-06, "loss": 0.6011, "step": 584 }, { "epoch": 0.5616898703792607, "grad_norm": 0.7888802882899952, "learning_rate": 9.995546556911271e-06, "loss": 0.6968, "step": 585 }, { "epoch": 0.5626500240038406, "grad_norm": 1.1666763934468996, "learning_rate": 9.995404077543654e-06, "loss": 0.7349, "step": 586 }, { "epoch": 0.5636101776284206, "grad_norm": 1.6025839547244385, "learning_rate": 9.995259355792544e-06, "loss": 0.9321, "step": 587 }, { "epoch": 0.5645703312530005, "grad_norm": 1.798710423793645, "learning_rate": 9.995112391722905e-06, "loss": 0.8569, "step": 588 }, { "epoch": 0.5655304848775804, "grad_norm": 1.4011667823335725, "learning_rate": 9.994963185400708e-06, "loss": 0.7364, "step": 589 }, { "epoch": 0.5664906385021603, "grad_norm": 0.8749612709523675, "learning_rate": 9.994811736892931e-06, "loss": 0.6736, "step": 590 }, { "epoch": 0.5674507921267403, "grad_norm": 1.2800722045313515, "learning_rate": 9.994658046267556e-06, "loss": 0.9459, "step": 591 }, { "epoch": 0.5684109457513202, "grad_norm": 1.3267562805600717, "learning_rate": 9.994502113593575e-06, "loss": 0.8393, "step": 592 }, { "epoch": 0.5693710993759001, "grad_norm": 0.7855537661239357, "learning_rate": 9.994343938940983e-06, "loss": 0.6473, "step": 593 }, { "epoch": 0.57033125300048, "grad_norm": 0.8075005277083452, "learning_rate": 9.994183522380784e-06, "loss": 0.6449, "step": 594 }, { "epoch": 0.5712914066250601, "grad_norm": 0.8896907309530534, "learning_rate": 9.994020863984986e-06, "loss": 0.699, "step": 595 }, { "epoch": 0.57225156024964, "grad_norm": 0.8934011139457299, "learning_rate": 9.993855963826605e-06, "loss": 0.7624, "step": 596 }, { "epoch": 0.5732117138742199, "grad_norm": 0.8898104375705114, "learning_rate": 9.993688821979664e-06, "loss": 0.6694, "step": 597 }, { "epoch": 0.5741718674987998, "grad_norm": 0.987387424832426, "learning_rate": 9.993519438519191e-06, "loss": 0.741, "step": 598 }, { "epoch": 0.5751320211233797, "grad_norm": 0.8945606486133325, "learning_rate": 9.993347813521219e-06, "loss": 0.7427, "step": 599 }, { "epoch": 0.5760921747479597, "grad_norm": 1.4176390032602895, "learning_rate": 9.993173947062789e-06, "loss": 0.6867, "step": 600 }, { "epoch": 0.5770523283725396, "grad_norm": 1.1750060321207774, "learning_rate": 9.992997839221947e-06, "loss": 0.7227, "step": 601 }, { "epoch": 0.5780124819971195, "grad_norm": 0.8102107411039173, "learning_rate": 9.99281949007775e-06, "loss": 0.6431, "step": 602 }, { "epoch": 0.5789726356216994, "grad_norm": 0.9025304779685765, "learning_rate": 9.992638899710252e-06, "loss": 0.6817, "step": 603 }, { "epoch": 0.5799327892462794, "grad_norm": 1.0205983159610328, "learning_rate": 9.99245606820052e-06, "loss": 0.597, "step": 604 }, { "epoch": 0.5808929428708594, "grad_norm": 1.5466622002092316, "learning_rate": 9.992270995630628e-06, "loss": 0.8087, "step": 605 }, { "epoch": 0.5818530964954393, "grad_norm": 0.8797433439236639, "learning_rate": 9.992083682083649e-06, "loss": 0.7346, "step": 606 }, { "epoch": 0.5828132501200192, "grad_norm": 0.8166629612961828, "learning_rate": 9.991894127643666e-06, "loss": 0.5971, "step": 607 }, { "epoch": 0.5837734037445992, "grad_norm": 1.0714388938041954, "learning_rate": 9.991702332395771e-06, "loss": 0.8088, "step": 608 }, { "epoch": 0.5847335573691791, "grad_norm": 0.98642685487318, "learning_rate": 9.991508296426057e-06, "loss": 0.7302, "step": 609 }, { "epoch": 0.585693710993759, "grad_norm": 0.983924866582853, "learning_rate": 9.991312019821626e-06, "loss": 0.768, "step": 610 }, { "epoch": 0.5866538646183389, "grad_norm": 0.9168152531200842, "learning_rate": 9.991113502670583e-06, "loss": 0.7293, "step": 611 }, { "epoch": 0.5876140182429188, "grad_norm": 1.0651099187275237, "learning_rate": 9.99091274506204e-06, "loss": 0.6855, "step": 612 }, { "epoch": 0.5885741718674988, "grad_norm": 1.4726760989488568, "learning_rate": 9.990709747086118e-06, "loss": 0.7493, "step": 613 }, { "epoch": 0.5895343254920787, "grad_norm": 1.2843661398185344, "learning_rate": 9.990504508833936e-06, "loss": 0.7424, "step": 614 }, { "epoch": 0.5904944791166586, "grad_norm": 1.096241092094828, "learning_rate": 9.990297030397626e-06, "loss": 0.6876, "step": 615 }, { "epoch": 0.5914546327412386, "grad_norm": 0.838087466655316, "learning_rate": 9.990087311870322e-06, "loss": 0.6664, "step": 616 }, { "epoch": 0.5924147863658186, "grad_norm": 1.0358841105447545, "learning_rate": 9.989875353346164e-06, "loss": 0.7575, "step": 617 }, { "epoch": 0.5933749399903985, "grad_norm": 1.0685732965094747, "learning_rate": 9.9896611549203e-06, "loss": 0.8157, "step": 618 }, { "epoch": 0.5943350936149784, "grad_norm": 0.8973916511666883, "learning_rate": 9.98944471668888e-06, "loss": 0.6667, "step": 619 }, { "epoch": 0.5952952472395583, "grad_norm": 1.2713500455775724, "learning_rate": 9.98922603874906e-06, "loss": 0.626, "step": 620 }, { "epoch": 0.5962554008641383, "grad_norm": 1.0733413085887191, "learning_rate": 9.989005121199004e-06, "loss": 0.7723, "step": 621 }, { "epoch": 0.5972155544887182, "grad_norm": 0.9549881809411525, "learning_rate": 9.98878196413788e-06, "loss": 0.6427, "step": 622 }, { "epoch": 0.5981757081132981, "grad_norm": 0.771987088946783, "learning_rate": 9.988556567665858e-06, "loss": 0.7203, "step": 623 }, { "epoch": 0.599135861737878, "grad_norm": 0.8776127259205228, "learning_rate": 9.988328931884116e-06, "loss": 0.6224, "step": 624 }, { "epoch": 0.600096015362458, "grad_norm": 0.7283414359909332, "learning_rate": 9.98809905689484e-06, "loss": 0.5833, "step": 625 }, { "epoch": 0.601056168987038, "grad_norm": 0.9131307740100922, "learning_rate": 9.987866942801216e-06, "loss": 0.6774, "step": 626 }, { "epoch": 0.6020163226116179, "grad_norm": 1.3120938760386007, "learning_rate": 9.987632589707441e-06, "loss": 0.8937, "step": 627 }, { "epoch": 0.6029764762361978, "grad_norm": 1.3328538602196376, "learning_rate": 9.98739599771871e-06, "loss": 0.7257, "step": 628 }, { "epoch": 0.6039366298607777, "grad_norm": 1.2365612832258515, "learning_rate": 9.987157166941228e-06, "loss": 0.7599, "step": 629 }, { "epoch": 0.6048967834853577, "grad_norm": 0.8332792382216863, "learning_rate": 9.986916097482204e-06, "loss": 0.5862, "step": 630 }, { "epoch": 0.6058569371099376, "grad_norm": 0.9879194207090357, "learning_rate": 9.98667278944985e-06, "loss": 0.681, "step": 631 }, { "epoch": 0.6068170907345175, "grad_norm": 0.8464955362525541, "learning_rate": 9.986427242953387e-06, "loss": 0.6126, "step": 632 }, { "epoch": 0.6077772443590974, "grad_norm": 0.9792960849839638, "learning_rate": 9.986179458103036e-06, "loss": 0.7461, "step": 633 }, { "epoch": 0.6087373979836774, "grad_norm": 0.9435831540946961, "learning_rate": 9.985929435010024e-06, "loss": 0.7114, "step": 634 }, { "epoch": 0.6096975516082573, "grad_norm": 1.0566578633378716, "learning_rate": 9.985677173786587e-06, "loss": 0.6659, "step": 635 }, { "epoch": 0.6106577052328372, "grad_norm": 1.215304313152251, "learning_rate": 9.98542267454596e-06, "loss": 0.8623, "step": 636 }, { "epoch": 0.6116178588574172, "grad_norm": 0.6848191408167238, "learning_rate": 9.985165937402386e-06, "loss": 0.5351, "step": 637 }, { "epoch": 0.6125780124819972, "grad_norm": 0.8578491649464549, "learning_rate": 9.98490696247111e-06, "loss": 0.7246, "step": 638 }, { "epoch": 0.6135381661065771, "grad_norm": 0.7526315674793553, "learning_rate": 9.984645749868386e-06, "loss": 0.6217, "step": 639 }, { "epoch": 0.614498319731157, "grad_norm": 0.8642744699046789, "learning_rate": 9.984382299711466e-06, "loss": 0.7147, "step": 640 }, { "epoch": 0.6154584733557369, "grad_norm": 0.8974640256749795, "learning_rate": 9.984116612118612e-06, "loss": 0.7443, "step": 641 }, { "epoch": 0.6164186269803168, "grad_norm": 0.9219854325831495, "learning_rate": 9.98384868720909e-06, "loss": 0.6292, "step": 642 }, { "epoch": 0.6173787806048968, "grad_norm": 0.8120141197653961, "learning_rate": 9.983578525103166e-06, "loss": 0.6487, "step": 643 }, { "epoch": 0.6183389342294767, "grad_norm": 0.9391768132917029, "learning_rate": 9.983306125922111e-06, "loss": 0.5815, "step": 644 }, { "epoch": 0.6192990878540566, "grad_norm": 1.0025185518932291, "learning_rate": 9.983031489788208e-06, "loss": 0.745, "step": 645 }, { "epoch": 0.6202592414786365, "grad_norm": 0.9868369950892175, "learning_rate": 9.982754616824733e-06, "loss": 0.6369, "step": 646 }, { "epoch": 0.6212193951032166, "grad_norm": 1.0619459775088909, "learning_rate": 9.982475507155974e-06, "loss": 0.7368, "step": 647 }, { "epoch": 0.6221795487277965, "grad_norm": 1.0622349771799557, "learning_rate": 9.98219416090722e-06, "loss": 0.7361, "step": 648 }, { "epoch": 0.6231397023523764, "grad_norm": 1.2531773068999177, "learning_rate": 9.981910578204764e-06, "loss": 0.7721, "step": 649 }, { "epoch": 0.6240998559769563, "grad_norm": 0.864782341860092, "learning_rate": 9.981624759175902e-06, "loss": 0.5992, "step": 650 }, { "epoch": 0.6250600096015363, "grad_norm": 0.8683477425183821, "learning_rate": 9.981336703948936e-06, "loss": 0.6794, "step": 651 }, { "epoch": 0.6260201632261162, "grad_norm": 0.8988398019361346, "learning_rate": 9.98104641265317e-06, "loss": 0.6955, "step": 652 }, { "epoch": 0.6269803168506961, "grad_norm": 1.027446637799313, "learning_rate": 9.980753885418916e-06, "loss": 0.7974, "step": 653 }, { "epoch": 0.627940470475276, "grad_norm": 0.9445691444147462, "learning_rate": 9.980459122377484e-06, "loss": 0.7828, "step": 654 }, { "epoch": 0.628900624099856, "grad_norm": 1.0870348811683306, "learning_rate": 9.98016212366119e-06, "loss": 0.705, "step": 655 }, { "epoch": 0.6298607777244359, "grad_norm": 0.807370601980383, "learning_rate": 9.979862889403353e-06, "loss": 0.7469, "step": 656 }, { "epoch": 0.6308209313490158, "grad_norm": 1.2659344294906976, "learning_rate": 9.979561419738297e-06, "loss": 0.86, "step": 657 }, { "epoch": 0.6317810849735958, "grad_norm": 1.1850689437681503, "learning_rate": 9.979257714801348e-06, "loss": 0.7898, "step": 658 }, { "epoch": 0.6327412385981757, "grad_norm": 1.0533542239846692, "learning_rate": 9.978951774728837e-06, "loss": 0.7299, "step": 659 }, { "epoch": 0.6337013922227557, "grad_norm": 0.8853480714886022, "learning_rate": 9.978643599658095e-06, "loss": 0.6632, "step": 660 }, { "epoch": 0.6346615458473356, "grad_norm": 1.1871734772507836, "learning_rate": 9.97833318972746e-06, "loss": 0.7829, "step": 661 }, { "epoch": 0.6356216994719155, "grad_norm": 0.9508673098780193, "learning_rate": 9.978020545076274e-06, "loss": 0.725, "step": 662 }, { "epoch": 0.6365818530964954, "grad_norm": 0.7457623076392801, "learning_rate": 9.977705665844875e-06, "loss": 0.5596, "step": 663 }, { "epoch": 0.6375420067210754, "grad_norm": 1.337985442114, "learning_rate": 9.977388552174611e-06, "loss": 0.8824, "step": 664 }, { "epoch": 0.6385021603456553, "grad_norm": 0.8422382380385861, "learning_rate": 9.977069204207832e-06, "loss": 0.6956, "step": 665 }, { "epoch": 0.6394623139702352, "grad_norm": 1.0379832492504693, "learning_rate": 9.97674762208789e-06, "loss": 0.752, "step": 666 }, { "epoch": 0.6404224675948151, "grad_norm": 1.000732809816472, "learning_rate": 9.976423805959138e-06, "loss": 0.7414, "step": 667 }, { "epoch": 0.6413826212193952, "grad_norm": 1.0710443224255881, "learning_rate": 9.976097755966935e-06, "loss": 0.75, "step": 668 }, { "epoch": 0.6423427748439751, "grad_norm": 0.943052878689579, "learning_rate": 9.975769472257642e-06, "loss": 0.7453, "step": 669 }, { "epoch": 0.643302928468555, "grad_norm": 1.398220386073235, "learning_rate": 9.97543895497862e-06, "loss": 0.8438, "step": 670 }, { "epoch": 0.6442630820931349, "grad_norm": 1.306473888839786, "learning_rate": 9.975106204278235e-06, "loss": 0.8044, "step": 671 }, { "epoch": 0.6452232357177148, "grad_norm": 0.9379862762658544, "learning_rate": 9.974771220305855e-06, "loss": 0.7255, "step": 672 }, { "epoch": 0.6461833893422948, "grad_norm": 0.7811833131460245, "learning_rate": 9.974434003211853e-06, "loss": 0.6211, "step": 673 }, { "epoch": 0.6471435429668747, "grad_norm": 0.9406872178677893, "learning_rate": 9.974094553147599e-06, "loss": 0.7657, "step": 674 }, { "epoch": 0.6481036965914546, "grad_norm": 0.9730078122185979, "learning_rate": 9.973752870265473e-06, "loss": 0.6286, "step": 675 }, { "epoch": 0.6490638502160345, "grad_norm": 0.7970861090145978, "learning_rate": 9.973408954718848e-06, "loss": 0.6402, "step": 676 }, { "epoch": 0.6500240038406145, "grad_norm": 0.9576940717483807, "learning_rate": 9.973062806662106e-06, "loss": 0.6759, "step": 677 }, { "epoch": 0.6509841574651944, "grad_norm": 0.9843127470720656, "learning_rate": 9.97271442625063e-06, "loss": 0.7421, "step": 678 }, { "epoch": 0.6519443110897744, "grad_norm": 0.9181739068238992, "learning_rate": 9.972363813640804e-06, "loss": 0.7414, "step": 679 }, { "epoch": 0.6529044647143543, "grad_norm": 0.9649294090973161, "learning_rate": 9.972010968990012e-06, "loss": 0.7084, "step": 680 }, { "epoch": 0.6538646183389343, "grad_norm": 1.1338636162019948, "learning_rate": 9.971655892456646e-06, "loss": 0.788, "step": 681 }, { "epoch": 0.6548247719635142, "grad_norm": 0.9640765366682439, "learning_rate": 9.971298584200093e-06, "loss": 0.7256, "step": 682 }, { "epoch": 0.6557849255880941, "grad_norm": 0.7528074436817037, "learning_rate": 9.970939044380743e-06, "loss": 0.6326, "step": 683 }, { "epoch": 0.656745079212674, "grad_norm": 0.9217673288141575, "learning_rate": 9.970577273159995e-06, "loss": 0.7181, "step": 684 }, { "epoch": 0.6577052328372539, "grad_norm": 1.057024748949328, "learning_rate": 9.970213270700241e-06, "loss": 0.7826, "step": 685 }, { "epoch": 0.6586653864618339, "grad_norm": 0.9284475340492399, "learning_rate": 9.969847037164881e-06, "loss": 0.6094, "step": 686 }, { "epoch": 0.6596255400864138, "grad_norm": 0.954412074223108, "learning_rate": 9.969478572718307e-06, "loss": 0.6744, "step": 687 }, { "epoch": 0.6605856937109937, "grad_norm": 0.8472935540246108, "learning_rate": 9.969107877525925e-06, "loss": 0.6478, "step": 688 }, { "epoch": 0.6615458473355736, "grad_norm": 0.7681831113378361, "learning_rate": 9.968734951754134e-06, "loss": 0.6336, "step": 689 }, { "epoch": 0.6625060009601537, "grad_norm": 1.078217720701533, "learning_rate": 9.968359795570333e-06, "loss": 0.8075, "step": 690 }, { "epoch": 0.6634661545847336, "grad_norm": 0.8536505160834013, "learning_rate": 9.96798240914293e-06, "loss": 0.6535, "step": 691 }, { "epoch": 0.6644263082093135, "grad_norm": 0.9396078768585966, "learning_rate": 9.96760279264133e-06, "loss": 0.5677, "step": 692 }, { "epoch": 0.6653864618338934, "grad_norm": 0.8765175110726742, "learning_rate": 9.967220946235933e-06, "loss": 0.7104, "step": 693 }, { "epoch": 0.6663466154584734, "grad_norm": 1.342803083909664, "learning_rate": 9.966836870098153e-06, "loss": 0.7948, "step": 694 }, { "epoch": 0.6673067690830533, "grad_norm": 0.8962334113528844, "learning_rate": 9.966450564400396e-06, "loss": 0.7062, "step": 695 }, { "epoch": 0.6682669227076332, "grad_norm": 1.1668811838826516, "learning_rate": 9.966062029316067e-06, "loss": 0.8064, "step": 696 }, { "epoch": 0.6692270763322131, "grad_norm": 0.8605729676471373, "learning_rate": 9.965671265019576e-06, "loss": 0.6639, "step": 697 }, { "epoch": 0.6701872299567931, "grad_norm": 1.0065405580467948, "learning_rate": 9.965278271686335e-06, "loss": 0.6204, "step": 698 }, { "epoch": 0.671147383581373, "grad_norm": 0.8710649106767315, "learning_rate": 9.964883049492755e-06, "loss": 0.6561, "step": 699 }, { "epoch": 0.672107537205953, "grad_norm": 0.8379910071478627, "learning_rate": 9.964485598616244e-06, "loss": 0.6622, "step": 700 }, { "epoch": 0.6730676908305329, "grad_norm": 1.0273031648254707, "learning_rate": 9.964085919235217e-06, "loss": 0.7361, "step": 701 }, { "epoch": 0.6740278444551128, "grad_norm": 1.4822693715207664, "learning_rate": 9.963684011529084e-06, "loss": 0.7852, "step": 702 }, { "epoch": 0.6749879980796928, "grad_norm": 0.9438672170237147, "learning_rate": 9.963279875678256e-06, "loss": 0.722, "step": 703 }, { "epoch": 0.6759481517042727, "grad_norm": 1.0772850764800432, "learning_rate": 9.962873511864146e-06, "loss": 0.7197, "step": 704 }, { "epoch": 0.6769083053288526, "grad_norm": 1.2739525503600813, "learning_rate": 9.962464920269167e-06, "loss": 0.8162, "step": 705 }, { "epoch": 0.6778684589534325, "grad_norm": 0.968399084234287, "learning_rate": 9.962054101076733e-06, "loss": 0.7498, "step": 706 }, { "epoch": 0.6788286125780125, "grad_norm": 0.7015146717078158, "learning_rate": 9.961641054471252e-06, "loss": 0.6352, "step": 707 }, { "epoch": 0.6797887662025924, "grad_norm": 1.310024411817623, "learning_rate": 9.96122578063814e-06, "loss": 0.659, "step": 708 }, { "epoch": 0.6807489198271723, "grad_norm": 0.7713212422886758, "learning_rate": 9.960808279763807e-06, "loss": 0.649, "step": 709 }, { "epoch": 0.6817090734517522, "grad_norm": 1.1011926694234533, "learning_rate": 9.960388552035666e-06, "loss": 0.7521, "step": 710 }, { "epoch": 0.6826692270763323, "grad_norm": 0.8581748236140228, "learning_rate": 9.959966597642127e-06, "loss": 0.6925, "step": 711 }, { "epoch": 0.6836293807009122, "grad_norm": 0.7712883849847446, "learning_rate": 9.9595424167726e-06, "loss": 0.6454, "step": 712 }, { "epoch": 0.6845895343254921, "grad_norm": 0.914791412449119, "learning_rate": 9.9591160096175e-06, "loss": 0.6479, "step": 713 }, { "epoch": 0.685549687950072, "grad_norm": 1.2435527169543374, "learning_rate": 9.95868737636823e-06, "loss": 0.743, "step": 714 }, { "epoch": 0.6865098415746519, "grad_norm": 1.0135207937212491, "learning_rate": 9.958256517217204e-06, "loss": 0.8356, "step": 715 }, { "epoch": 0.6874699951992319, "grad_norm": 1.2089092087046092, "learning_rate": 9.95782343235783e-06, "loss": 0.8529, "step": 716 }, { "epoch": 0.6884301488238118, "grad_norm": 0.9824241106324622, "learning_rate": 9.95738812198451e-06, "loss": 0.7696, "step": 717 }, { "epoch": 0.6893903024483917, "grad_norm": 2.5701138816514058, "learning_rate": 9.956950586292655e-06, "loss": 0.7107, "step": 718 }, { "epoch": 0.6903504560729716, "grad_norm": 0.8761504487905307, "learning_rate": 9.956510825478669e-06, "loss": 0.7473, "step": 719 }, { "epoch": 0.6913106096975516, "grad_norm": 0.8750609124600874, "learning_rate": 9.956068839739955e-06, "loss": 0.758, "step": 720 }, { "epoch": 0.6922707633221316, "grad_norm": 1.3029199081023795, "learning_rate": 9.955624629274916e-06, "loss": 0.8352, "step": 721 }, { "epoch": 0.6932309169467115, "grad_norm": 0.9979950330696566, "learning_rate": 9.955178194282953e-06, "loss": 0.6331, "step": 722 }, { "epoch": 0.6941910705712914, "grad_norm": 0.9480533416480195, "learning_rate": 9.954729534964469e-06, "loss": 0.6919, "step": 723 }, { "epoch": 0.6951512241958714, "grad_norm": 0.7558938686625172, "learning_rate": 9.954278651520859e-06, "loss": 0.5586, "step": 724 }, { "epoch": 0.6961113778204513, "grad_norm": 0.887296337494471, "learning_rate": 9.953825544154519e-06, "loss": 0.7115, "step": 725 }, { "epoch": 0.6970715314450312, "grad_norm": 0.9756280380555864, "learning_rate": 9.953370213068847e-06, "loss": 0.716, "step": 726 }, { "epoch": 0.6980316850696111, "grad_norm": 1.20289491072228, "learning_rate": 9.952912658468235e-06, "loss": 0.9667, "step": 727 }, { "epoch": 0.6989918386941911, "grad_norm": 0.8647430145397105, "learning_rate": 9.952452880558073e-06, "loss": 0.7223, "step": 728 }, { "epoch": 0.699951992318771, "grad_norm": 0.844088474390158, "learning_rate": 9.951990879544753e-06, "loss": 0.6334, "step": 729 }, { "epoch": 0.7009121459433509, "grad_norm": 0.9563054388200516, "learning_rate": 9.95152665563566e-06, "loss": 0.6511, "step": 730 }, { "epoch": 0.7018722995679308, "grad_norm": 1.0654069792336476, "learning_rate": 9.951060209039182e-06, "loss": 0.6903, "step": 731 }, { "epoch": 0.7028324531925108, "grad_norm": 1.0065447982707603, "learning_rate": 9.950591539964698e-06, "loss": 0.7436, "step": 732 }, { "epoch": 0.7037926068170908, "grad_norm": 0.9021217906159897, "learning_rate": 9.950120648622592e-06, "loss": 0.7668, "step": 733 }, { "epoch": 0.7047527604416707, "grad_norm": 0.8695767205830083, "learning_rate": 9.94964753522424e-06, "loss": 0.6689, "step": 734 }, { "epoch": 0.7057129140662506, "grad_norm": 1.0925669867555596, "learning_rate": 9.949172199982019e-06, "loss": 0.653, "step": 735 }, { "epoch": 0.7066730676908305, "grad_norm": 1.1709910249844182, "learning_rate": 9.9486946431093e-06, "loss": 0.7551, "step": 736 }, { "epoch": 0.7076332213154105, "grad_norm": 0.7275696025758591, "learning_rate": 9.948214864820455e-06, "loss": 0.5922, "step": 737 }, { "epoch": 0.7085933749399904, "grad_norm": 1.3517674963096273, "learning_rate": 9.94773286533085e-06, "loss": 0.7935, "step": 738 }, { "epoch": 0.7095535285645703, "grad_norm": 0.7889304526279081, "learning_rate": 9.947248644856852e-06, "loss": 0.7072, "step": 739 }, { "epoch": 0.7105136821891502, "grad_norm": 0.8797653206686449, "learning_rate": 9.946762203615818e-06, "loss": 0.6852, "step": 740 }, { "epoch": 0.7114738358137302, "grad_norm": 1.0255799822888911, "learning_rate": 9.94627354182611e-06, "loss": 0.76, "step": 741 }, { "epoch": 0.7124339894383102, "grad_norm": 1.039117725966005, "learning_rate": 9.945782659707082e-06, "loss": 0.7375, "step": 742 }, { "epoch": 0.7133941430628901, "grad_norm": 1.0479684055692742, "learning_rate": 9.945289557479086e-06, "loss": 0.7642, "step": 743 }, { "epoch": 0.71435429668747, "grad_norm": 0.9510132711692371, "learning_rate": 9.944794235363468e-06, "loss": 0.7024, "step": 744 }, { "epoch": 0.7153144503120499, "grad_norm": 0.9760545592863202, "learning_rate": 9.944296693582576e-06, "loss": 0.7325, "step": 745 }, { "epoch": 0.7162746039366299, "grad_norm": 0.872222111699403, "learning_rate": 9.94379693235975e-06, "loss": 0.7083, "step": 746 }, { "epoch": 0.7172347575612098, "grad_norm": 0.9167189799984291, "learning_rate": 9.943294951919327e-06, "loss": 0.7014, "step": 747 }, { "epoch": 0.7181949111857897, "grad_norm": 0.7535461767467537, "learning_rate": 9.94279075248664e-06, "loss": 0.6243, "step": 748 }, { "epoch": 0.7191550648103696, "grad_norm": 0.9458025652258596, "learning_rate": 9.942284334288021e-06, "loss": 0.7394, "step": 749 }, { "epoch": 0.7201152184349496, "grad_norm": 0.9403355080002509, "learning_rate": 9.941775697550796e-06, "loss": 0.779, "step": 750 }, { "epoch": 0.7210753720595295, "grad_norm": 0.8658439730769567, "learning_rate": 9.941264842503282e-06, "loss": 0.7241, "step": 751 }, { "epoch": 0.7220355256841094, "grad_norm": 0.9122167972930422, "learning_rate": 9.9407517693748e-06, "loss": 0.6863, "step": 752 }, { "epoch": 0.7229956793086894, "grad_norm": 1.0253189637364748, "learning_rate": 9.940236478395662e-06, "loss": 0.779, "step": 753 }, { "epoch": 0.7239558329332694, "grad_norm": 0.6407168479663033, "learning_rate": 9.939718969797178e-06, "loss": 0.5078, "step": 754 }, { "epoch": 0.7249159865578493, "grad_norm": 1.0171655949343896, "learning_rate": 9.939199243811653e-06, "loss": 0.6718, "step": 755 }, { "epoch": 0.7258761401824292, "grad_norm": 0.8100811493970654, "learning_rate": 9.938677300672382e-06, "loss": 0.6425, "step": 756 }, { "epoch": 0.7268362938070091, "grad_norm": 1.0364023737815007, "learning_rate": 9.938153140613662e-06, "loss": 0.8269, "step": 757 }, { "epoch": 0.727796447431589, "grad_norm": 0.7724109954151283, "learning_rate": 9.937626763870784e-06, "loss": 0.5932, "step": 758 }, { "epoch": 0.728756601056169, "grad_norm": 0.9041135413893235, "learning_rate": 9.937098170680034e-06, "loss": 0.7555, "step": 759 }, { "epoch": 0.7297167546807489, "grad_norm": 1.0675031835777398, "learning_rate": 9.936567361278687e-06, "loss": 0.7445, "step": 760 }, { "epoch": 0.7306769083053288, "grad_norm": 0.9296058201785269, "learning_rate": 9.93603433590502e-06, "loss": 0.7226, "step": 761 }, { "epoch": 0.7316370619299087, "grad_norm": 0.8083744701378427, "learning_rate": 9.935499094798305e-06, "loss": 0.6561, "step": 762 }, { "epoch": 0.7325972155544888, "grad_norm": 0.878512891202631, "learning_rate": 9.934961638198804e-06, "loss": 0.7069, "step": 763 }, { "epoch": 0.7335573691790687, "grad_norm": 0.999154305652808, "learning_rate": 9.934421966347773e-06, "loss": 0.7021, "step": 764 }, { "epoch": 0.7345175228036486, "grad_norm": 0.9139644317688408, "learning_rate": 9.93388007948747e-06, "loss": 0.7189, "step": 765 }, { "epoch": 0.7354776764282285, "grad_norm": 1.0038416712292284, "learning_rate": 9.933335977861138e-06, "loss": 0.7294, "step": 766 }, { "epoch": 0.7364378300528085, "grad_norm": 1.036627888585655, "learning_rate": 9.93278966171302e-06, "loss": 0.6853, "step": 767 }, { "epoch": 0.7373979836773884, "grad_norm": 0.855701025900241, "learning_rate": 9.932241131288352e-06, "loss": 0.7076, "step": 768 }, { "epoch": 0.7383581373019683, "grad_norm": 0.8271554726320923, "learning_rate": 9.931690386833363e-06, "loss": 0.723, "step": 769 }, { "epoch": 0.7393182909265482, "grad_norm": 0.9347932377754757, "learning_rate": 9.931137428595276e-06, "loss": 0.6653, "step": 770 }, { "epoch": 0.7402784445511282, "grad_norm": 0.9476714461958836, "learning_rate": 9.930582256822307e-06, "loss": 0.7221, "step": 771 }, { "epoch": 0.7412385981757081, "grad_norm": 0.9089632393255248, "learning_rate": 9.93002487176367e-06, "loss": 0.7584, "step": 772 }, { "epoch": 0.742198751800288, "grad_norm": 0.8670073242639385, "learning_rate": 9.929465273669567e-06, "loss": 0.7277, "step": 773 }, { "epoch": 0.743158905424868, "grad_norm": 1.3368295561981998, "learning_rate": 9.928903462791195e-06, "loss": 0.7945, "step": 774 }, { "epoch": 0.7441190590494479, "grad_norm": 0.8063675410663966, "learning_rate": 9.928339439380745e-06, "loss": 0.6908, "step": 775 }, { "epoch": 0.7450792126740279, "grad_norm": 0.8815360239634885, "learning_rate": 9.927773203691403e-06, "loss": 0.6567, "step": 776 }, { "epoch": 0.7460393662986078, "grad_norm": 1.4050064012627475, "learning_rate": 9.927204755977344e-06, "loss": 0.7488, "step": 777 }, { "epoch": 0.7469995199231877, "grad_norm": 1.0228386376891416, "learning_rate": 9.926634096493738e-06, "loss": 0.6246, "step": 778 }, { "epoch": 0.7479596735477676, "grad_norm": 1.0437437295507945, "learning_rate": 9.92606122549675e-06, "loss": 0.5995, "step": 779 }, { "epoch": 0.7489198271723476, "grad_norm": 1.40571788159001, "learning_rate": 9.925486143243534e-06, "loss": 0.8701, "step": 780 }, { "epoch": 0.7498799807969275, "grad_norm": 0.9873203827580456, "learning_rate": 9.92490884999224e-06, "loss": 0.7831, "step": 781 }, { "epoch": 0.7508401344215074, "grad_norm": 0.9491754889114244, "learning_rate": 9.924329346002007e-06, "loss": 0.6775, "step": 782 }, { "epoch": 0.7518002880460873, "grad_norm": 1.2061352886494656, "learning_rate": 9.923747631532968e-06, "loss": 0.7006, "step": 783 }, { "epoch": 0.7527604416706674, "grad_norm": 0.9120336535677263, "learning_rate": 9.92316370684625e-06, "loss": 0.7195, "step": 784 }, { "epoch": 0.7537205952952473, "grad_norm": 0.9504338558646669, "learning_rate": 9.922577572203969e-06, "loss": 0.6941, "step": 785 }, { "epoch": 0.7546807489198272, "grad_norm": 0.9337759794159912, "learning_rate": 9.921989227869236e-06, "loss": 0.6846, "step": 786 }, { "epoch": 0.7556409025444071, "grad_norm": 0.9387479858347486, "learning_rate": 9.921398674106152e-06, "loss": 0.7023, "step": 787 }, { "epoch": 0.756601056168987, "grad_norm": 1.0831647019787463, "learning_rate": 9.92080591117981e-06, "loss": 0.8935, "step": 788 }, { "epoch": 0.757561209793567, "grad_norm": 0.809846367448999, "learning_rate": 9.920210939356294e-06, "loss": 0.6507, "step": 789 }, { "epoch": 0.7585213634181469, "grad_norm": 0.78094255122183, "learning_rate": 9.919613758902682e-06, "loss": 0.6527, "step": 790 }, { "epoch": 0.7594815170427268, "grad_norm": 0.8913631957384313, "learning_rate": 9.91901437008704e-06, "loss": 0.7031, "step": 791 }, { "epoch": 0.7604416706673067, "grad_norm": 0.7301484515971459, "learning_rate": 9.918412773178431e-06, "loss": 0.6363, "step": 792 }, { "epoch": 0.7614018242918867, "grad_norm": 0.9337727204820698, "learning_rate": 9.917808968446903e-06, "loss": 0.707, "step": 793 }, { "epoch": 0.7623619779164666, "grad_norm": 0.8551432928838548, "learning_rate": 9.917202956163496e-06, "loss": 0.6912, "step": 794 }, { "epoch": 0.7633221315410466, "grad_norm": 0.8363425945999091, "learning_rate": 9.916594736600244e-06, "loss": 0.7442, "step": 795 }, { "epoch": 0.7642822851656265, "grad_norm": 0.8770383358165755, "learning_rate": 9.915984310030172e-06, "loss": 0.7092, "step": 796 }, { "epoch": 0.7652424387902065, "grad_norm": 0.9316939792532943, "learning_rate": 9.91537167672729e-06, "loss": 0.6643, "step": 797 }, { "epoch": 0.7662025924147864, "grad_norm": 0.765563175887775, "learning_rate": 9.914756836966604e-06, "loss": 0.7014, "step": 798 }, { "epoch": 0.7671627460393663, "grad_norm": 0.9037370158874687, "learning_rate": 9.914139791024112e-06, "loss": 0.7337, "step": 799 }, { "epoch": 0.7681228996639462, "grad_norm": 0.8179278947915232, "learning_rate": 9.913520539176797e-06, "loss": 0.7249, "step": 800 }, { "epoch": 0.7690830532885262, "grad_norm": 0.8051401400764737, "learning_rate": 9.912899081702633e-06, "loss": 0.6315, "step": 801 }, { "epoch": 0.7700432069131061, "grad_norm": 1.012744351022652, "learning_rate": 9.912275418880588e-06, "loss": 0.7715, "step": 802 }, { "epoch": 0.771003360537686, "grad_norm": 1.0423051379668111, "learning_rate": 9.911649550990618e-06, "loss": 0.8197, "step": 803 }, { "epoch": 0.7719635141622659, "grad_norm": 0.9713769967527579, "learning_rate": 9.911021478313665e-06, "loss": 0.7934, "step": 804 }, { "epoch": 0.7729236677868458, "grad_norm": 1.1218933363990895, "learning_rate": 9.910391201131668e-06, "loss": 0.858, "step": 805 }, { "epoch": 0.7738838214114259, "grad_norm": 0.91878774275496, "learning_rate": 9.909758719727549e-06, "loss": 0.7228, "step": 806 }, { "epoch": 0.7748439750360058, "grad_norm": 1.0254321876070622, "learning_rate": 9.909124034385225e-06, "loss": 0.7539, "step": 807 }, { "epoch": 0.7758041286605857, "grad_norm": 0.8325192660974299, "learning_rate": 9.908487145389595e-06, "loss": 0.6815, "step": 808 }, { "epoch": 0.7767642822851656, "grad_norm": 0.9498225818470458, "learning_rate": 9.907848053026554e-06, "loss": 0.781, "step": 809 }, { "epoch": 0.7777244359097456, "grad_norm": 0.8593933253044965, "learning_rate": 9.907206757582987e-06, "loss": 0.7065, "step": 810 }, { "epoch": 0.7786845895343255, "grad_norm": 0.9307960052717634, "learning_rate": 9.90656325934676e-06, "loss": 0.7283, "step": 811 }, { "epoch": 0.7796447431589054, "grad_norm": 1.1118541450640134, "learning_rate": 9.905917558606734e-06, "loss": 0.6822, "step": 812 }, { "epoch": 0.7806048967834853, "grad_norm": 0.8395609426510988, "learning_rate": 9.905269655652758e-06, "loss": 0.5926, "step": 813 }, { "epoch": 0.7815650504080653, "grad_norm": 0.8026167428887037, "learning_rate": 9.904619550775668e-06, "loss": 0.6802, "step": 814 }, { "epoch": 0.7825252040326452, "grad_norm": 1.1512120584707919, "learning_rate": 9.903967244267287e-06, "loss": 0.7618, "step": 815 }, { "epoch": 0.7834853576572252, "grad_norm": 1.3867899320445112, "learning_rate": 9.903312736420433e-06, "loss": 0.8368, "step": 816 }, { "epoch": 0.7844455112818051, "grad_norm": 0.8741279789283188, "learning_rate": 9.902656027528903e-06, "loss": 0.6369, "step": 817 }, { "epoch": 0.785405664906385, "grad_norm": 0.998044190569138, "learning_rate": 9.90199711788749e-06, "loss": 0.6454, "step": 818 }, { "epoch": 0.786365818530965, "grad_norm": 0.7846748596942303, "learning_rate": 9.901336007791971e-06, "loss": 0.5394, "step": 819 }, { "epoch": 0.7873259721555449, "grad_norm": 1.0038990513854713, "learning_rate": 9.900672697539108e-06, "loss": 0.7827, "step": 820 }, { "epoch": 0.7882861257801248, "grad_norm": 0.8942582721648366, "learning_rate": 9.900007187426659e-06, "loss": 0.7091, "step": 821 }, { "epoch": 0.7892462794047047, "grad_norm": 0.7210794990991671, "learning_rate": 9.89933947775336e-06, "loss": 0.5573, "step": 822 }, { "epoch": 0.7902064330292847, "grad_norm": 0.9155235150799451, "learning_rate": 9.898669568818942e-06, "loss": 0.6508, "step": 823 }, { "epoch": 0.7911665866538646, "grad_norm": 0.8028335069537421, "learning_rate": 9.897997460924118e-06, "loss": 0.7453, "step": 824 }, { "epoch": 0.7921267402784445, "grad_norm": 0.8434626053221281, "learning_rate": 9.89732315437059e-06, "loss": 0.6408, "step": 825 }, { "epoch": 0.7930868939030244, "grad_norm": 0.8576802389221398, "learning_rate": 9.896646649461047e-06, "loss": 0.6028, "step": 826 }, { "epoch": 0.7940470475276045, "grad_norm": 0.941657251329196, "learning_rate": 9.895967946499165e-06, "loss": 0.6264, "step": 827 }, { "epoch": 0.7950072011521844, "grad_norm": 1.1240809102639933, "learning_rate": 9.895287045789608e-06, "loss": 0.6091, "step": 828 }, { "epoch": 0.7959673547767643, "grad_norm": 0.8608410755742362, "learning_rate": 9.894603947638023e-06, "loss": 0.6614, "step": 829 }, { "epoch": 0.7969275084013442, "grad_norm": 0.9217623002587092, "learning_rate": 9.893918652351048e-06, "loss": 0.6452, "step": 830 }, { "epoch": 0.7978876620259241, "grad_norm": 0.7256078100190181, "learning_rate": 9.893231160236303e-06, "loss": 0.5575, "step": 831 }, { "epoch": 0.7988478156505041, "grad_norm": 0.9243132053290668, "learning_rate": 9.892541471602395e-06, "loss": 0.6754, "step": 832 }, { "epoch": 0.799807969275084, "grad_norm": 1.2088154412879117, "learning_rate": 9.891849586758918e-06, "loss": 0.7448, "step": 833 }, { "epoch": 0.8007681228996639, "grad_norm": 0.8830438075820135, "learning_rate": 9.891155506016452e-06, "loss": 0.7247, "step": 834 }, { "epoch": 0.8017282765242438, "grad_norm": 0.8508724736309575, "learning_rate": 9.890459229686564e-06, "loss": 0.629, "step": 835 }, { "epoch": 0.8026884301488239, "grad_norm": 0.8024278647524528, "learning_rate": 9.889760758081803e-06, "loss": 0.6077, "step": 836 }, { "epoch": 0.8036485837734038, "grad_norm": 1.0030413253367623, "learning_rate": 9.889060091515707e-06, "loss": 0.6845, "step": 837 }, { "epoch": 0.8046087373979837, "grad_norm": 0.874541049428236, "learning_rate": 9.8883572303028e-06, "loss": 0.7206, "step": 838 }, { "epoch": 0.8055688910225636, "grad_norm": 1.1534786558487395, "learning_rate": 9.887652174758581e-06, "loss": 0.7731, "step": 839 }, { "epoch": 0.8065290446471436, "grad_norm": 1.3146878188491413, "learning_rate": 9.88694492519955e-06, "loss": 0.8833, "step": 840 }, { "epoch": 0.8074891982717235, "grad_norm": 1.0959727193682944, "learning_rate": 9.88623548194318e-06, "loss": 0.73, "step": 841 }, { "epoch": 0.8084493518963034, "grad_norm": 0.8281868442615177, "learning_rate": 9.88552384530793e-06, "loss": 0.6542, "step": 842 }, { "epoch": 0.8094095055208833, "grad_norm": 0.8952406397948682, "learning_rate": 9.884810015613255e-06, "loss": 0.7476, "step": 843 }, { "epoch": 0.8103696591454633, "grad_norm": 0.7918329688218624, "learning_rate": 9.884093993179575e-06, "loss": 0.5907, "step": 844 }, { "epoch": 0.8113298127700432, "grad_norm": 0.95622950904134, "learning_rate": 9.883375778328312e-06, "loss": 0.6889, "step": 845 }, { "epoch": 0.8122899663946231, "grad_norm": 1.2117761654311585, "learning_rate": 9.882655371381861e-06, "loss": 0.8179, "step": 846 }, { "epoch": 0.813250120019203, "grad_norm": 0.9740663523997564, "learning_rate": 9.881932772663609e-06, "loss": 0.7474, "step": 847 }, { "epoch": 0.814210273643783, "grad_norm": 1.082433837061907, "learning_rate": 9.881207982497917e-06, "loss": 0.8004, "step": 848 }, { "epoch": 0.815170427268363, "grad_norm": 0.7902891449226649, "learning_rate": 9.880481001210142e-06, "loss": 0.6336, "step": 849 }, { "epoch": 0.8161305808929429, "grad_norm": 1.048261276361648, "learning_rate": 9.879751829126612e-06, "loss": 0.6125, "step": 850 }, { "epoch": 0.8170907345175228, "grad_norm": 0.9811360851685772, "learning_rate": 9.87902046657465e-06, "loss": 0.7461, "step": 851 }, { "epoch": 0.8180508881421027, "grad_norm": 0.7392361651587824, "learning_rate": 9.878286913882553e-06, "loss": 0.6267, "step": 852 }, { "epoch": 0.8190110417666827, "grad_norm": 1.274716579557757, "learning_rate": 9.877551171379606e-06, "loss": 0.8107, "step": 853 }, { "epoch": 0.8199711953912626, "grad_norm": 1.3628896158110095, "learning_rate": 9.876813239396079e-06, "loss": 0.7651, "step": 854 }, { "epoch": 0.8209313490158425, "grad_norm": 1.1482517063536697, "learning_rate": 9.876073118263216e-06, "loss": 0.8026, "step": 855 }, { "epoch": 0.8218915026404224, "grad_norm": 0.835510282438675, "learning_rate": 9.875330808313254e-06, "loss": 0.6485, "step": 856 }, { "epoch": 0.8228516562650025, "grad_norm": 1.0422921843363226, "learning_rate": 9.874586309879406e-06, "loss": 0.6965, "step": 857 }, { "epoch": 0.8238118098895824, "grad_norm": 0.9484461312064201, "learning_rate": 9.87383962329587e-06, "loss": 0.6322, "step": 858 }, { "epoch": 0.8247719635141623, "grad_norm": 0.8380266256236384, "learning_rate": 9.873090748897826e-06, "loss": 0.6267, "step": 859 }, { "epoch": 0.8257321171387422, "grad_norm": 1.064093542459743, "learning_rate": 9.872339687021434e-06, "loss": 0.8146, "step": 860 }, { "epoch": 0.8266922707633221, "grad_norm": 0.9675532791714647, "learning_rate": 9.871586438003841e-06, "loss": 0.7391, "step": 861 }, { "epoch": 0.8276524243879021, "grad_norm": 1.1265448904523436, "learning_rate": 9.870831002183168e-06, "loss": 0.6595, "step": 862 }, { "epoch": 0.828612578012482, "grad_norm": 1.061263885878785, "learning_rate": 9.870073379898526e-06, "loss": 0.7424, "step": 863 }, { "epoch": 0.8295727316370619, "grad_norm": 0.9770752188789026, "learning_rate": 9.869313571490002e-06, "loss": 0.681, "step": 864 }, { "epoch": 0.8305328852616418, "grad_norm": 0.8963754875213547, "learning_rate": 9.868551577298665e-06, "loss": 0.7374, "step": 865 }, { "epoch": 0.8314930388862218, "grad_norm": 0.9341774231933359, "learning_rate": 9.867787397666568e-06, "loss": 0.7491, "step": 866 }, { "epoch": 0.8324531925108017, "grad_norm": 0.9607499708164543, "learning_rate": 9.86702103293674e-06, "loss": 0.7819, "step": 867 }, { "epoch": 0.8334133461353816, "grad_norm": 1.0084029791947833, "learning_rate": 9.866252483453199e-06, "loss": 0.7227, "step": 868 }, { "epoch": 0.8343734997599616, "grad_norm": 1.0239536589616365, "learning_rate": 9.865481749560933e-06, "loss": 0.7414, "step": 869 }, { "epoch": 0.8353336533845416, "grad_norm": 0.8673262030434952, "learning_rate": 9.864708831605918e-06, "loss": 0.687, "step": 870 }, { "epoch": 0.8362938070091215, "grad_norm": 1.0848034601574363, "learning_rate": 9.863933729935112e-06, "loss": 0.8258, "step": 871 }, { "epoch": 0.8372539606337014, "grad_norm": 0.844731708551887, "learning_rate": 9.863156444896446e-06, "loss": 0.7002, "step": 872 }, { "epoch": 0.8382141142582813, "grad_norm": 0.9158735678404796, "learning_rate": 9.862376976838835e-06, "loss": 0.7247, "step": 873 }, { "epoch": 0.8391742678828612, "grad_norm": 1.2531746241455983, "learning_rate": 9.861595326112176e-06, "loss": 0.7353, "step": 874 }, { "epoch": 0.8401344215074412, "grad_norm": 0.9716997561493464, "learning_rate": 9.860811493067344e-06, "loss": 0.7483, "step": 875 }, { "epoch": 0.8410945751320211, "grad_norm": 0.9986304238730049, "learning_rate": 9.86002547805619e-06, "loss": 0.7126, "step": 876 }, { "epoch": 0.842054728756601, "grad_norm": 0.7938433679180987, "learning_rate": 9.85923728143155e-06, "loss": 0.595, "step": 877 }, { "epoch": 0.8430148823811809, "grad_norm": 1.3106129980132002, "learning_rate": 9.858446903547239e-06, "loss": 0.7361, "step": 878 }, { "epoch": 0.843975036005761, "grad_norm": 0.8782891934412358, "learning_rate": 9.857654344758044e-06, "loss": 0.6779, "step": 879 }, { "epoch": 0.8449351896303409, "grad_norm": 0.9978597921359031, "learning_rate": 9.85685960541974e-06, "loss": 0.7363, "step": 880 }, { "epoch": 0.8458953432549208, "grad_norm": 1.1658691033239643, "learning_rate": 9.856062685889076e-06, "loss": 0.8015, "step": 881 }, { "epoch": 0.8468554968795007, "grad_norm": 0.9060160968841194, "learning_rate": 9.855263586523782e-06, "loss": 0.7059, "step": 882 }, { "epoch": 0.8478156505040807, "grad_norm": 1.2192197912112421, "learning_rate": 9.854462307682561e-06, "loss": 0.7882, "step": 883 }, { "epoch": 0.8487758041286606, "grad_norm": 0.9383775070893094, "learning_rate": 9.853658849725105e-06, "loss": 0.7185, "step": 884 }, { "epoch": 0.8497359577532405, "grad_norm": 1.129319623509329, "learning_rate": 9.852853213012072e-06, "loss": 0.8443, "step": 885 }, { "epoch": 0.8506961113778204, "grad_norm": 0.9130389021086519, "learning_rate": 9.852045397905106e-06, "loss": 0.7273, "step": 886 }, { "epoch": 0.8516562650024004, "grad_norm": 0.8149282139029422, "learning_rate": 9.851235404766827e-06, "loss": 0.6406, "step": 887 }, { "epoch": 0.8526164186269803, "grad_norm": 0.8464081813827931, "learning_rate": 9.850423233960831e-06, "loss": 0.6048, "step": 888 }, { "epoch": 0.8535765722515603, "grad_norm": 0.9833791383875089, "learning_rate": 9.849608885851693e-06, "loss": 0.7333, "step": 889 }, { "epoch": 0.8545367258761402, "grad_norm": 0.8980181685030804, "learning_rate": 9.848792360804966e-06, "loss": 0.6447, "step": 890 }, { "epoch": 0.8554968795007201, "grad_norm": 0.9644392880709245, "learning_rate": 9.84797365918718e-06, "loss": 0.6504, "step": 891 }, { "epoch": 0.8564570331253001, "grad_norm": 0.9111795810315034, "learning_rate": 9.847152781365839e-06, "loss": 0.6652, "step": 892 }, { "epoch": 0.85741718674988, "grad_norm": 0.8912622500119116, "learning_rate": 9.846329727709429e-06, "loss": 0.6965, "step": 893 }, { "epoch": 0.8583773403744599, "grad_norm": 0.9350961460489112, "learning_rate": 9.845504498587406e-06, "loss": 0.7688, "step": 894 }, { "epoch": 0.8593374939990398, "grad_norm": 0.7995232265895333, "learning_rate": 9.844677094370212e-06, "loss": 0.6449, "step": 895 }, { "epoch": 0.8602976476236198, "grad_norm": 0.743547346500059, "learning_rate": 9.843847515429256e-06, "loss": 0.6597, "step": 896 }, { "epoch": 0.8612578012481997, "grad_norm": 1.0323088447213449, "learning_rate": 9.843015762136927e-06, "loss": 0.7218, "step": 897 }, { "epoch": 0.8622179548727796, "grad_norm": 1.0403315943891926, "learning_rate": 9.842181834866593e-06, "loss": 0.672, "step": 898 }, { "epoch": 0.8631781084973595, "grad_norm": 1.2947492517817176, "learning_rate": 9.841345733992594e-06, "loss": 0.8347, "step": 899 }, { "epoch": 0.8641382621219396, "grad_norm": 1.3084559591862097, "learning_rate": 9.840507459890244e-06, "loss": 0.8375, "step": 900 }, { "epoch": 0.8650984157465195, "grad_norm": 0.7681518148331432, "learning_rate": 9.839667012935838e-06, "loss": 0.6563, "step": 901 }, { "epoch": 0.8660585693710994, "grad_norm": 0.891108318670506, "learning_rate": 9.838824393506644e-06, "loss": 0.6639, "step": 902 }, { "epoch": 0.8670187229956793, "grad_norm": 1.2819523414664935, "learning_rate": 9.837979601980902e-06, "loss": 0.8922, "step": 903 }, { "epoch": 0.8679788766202592, "grad_norm": 1.0699521408694517, "learning_rate": 9.837132638737834e-06, "loss": 0.7707, "step": 904 }, { "epoch": 0.8689390302448392, "grad_norm": 0.8341233125759041, "learning_rate": 9.83628350415763e-06, "loss": 0.706, "step": 905 }, { "epoch": 0.8698991838694191, "grad_norm": 0.7561115225679634, "learning_rate": 9.835432198621457e-06, "loss": 0.6734, "step": 906 }, { "epoch": 0.870859337493999, "grad_norm": 0.881283884400619, "learning_rate": 9.83457872251146e-06, "loss": 0.7738, "step": 907 }, { "epoch": 0.8718194911185789, "grad_norm": 1.1978984416476854, "learning_rate": 9.833723076210751e-06, "loss": 0.7505, "step": 908 }, { "epoch": 0.8727796447431589, "grad_norm": 1.1481270248278428, "learning_rate": 9.832865260103423e-06, "loss": 0.7334, "step": 909 }, { "epoch": 0.8737397983677389, "grad_norm": 0.9111430518719715, "learning_rate": 9.832005274574541e-06, "loss": 0.6842, "step": 910 }, { "epoch": 0.8746999519923188, "grad_norm": 1.0578083144965953, "learning_rate": 9.831143120010143e-06, "loss": 0.7227, "step": 911 }, { "epoch": 0.8756601056168987, "grad_norm": 0.9727605094485979, "learning_rate": 9.830278796797239e-06, "loss": 0.7765, "step": 912 }, { "epoch": 0.8766202592414787, "grad_norm": 0.8522041472682932, "learning_rate": 9.829412305323818e-06, "loss": 0.638, "step": 913 }, { "epoch": 0.8775804128660586, "grad_norm": 0.8361282547500545, "learning_rate": 9.828543645978835e-06, "loss": 0.6748, "step": 914 }, { "epoch": 0.8785405664906385, "grad_norm": 0.7472074445229582, "learning_rate": 9.827672819152223e-06, "loss": 0.7335, "step": 915 }, { "epoch": 0.8795007201152184, "grad_norm": 1.15834648106561, "learning_rate": 9.82679982523489e-06, "loss": 0.778, "step": 916 }, { "epoch": 0.8804608737397984, "grad_norm": 0.8885466749724152, "learning_rate": 9.82592466461871e-06, "loss": 0.7244, "step": 917 }, { "epoch": 0.8814210273643783, "grad_norm": 0.8669887658381694, "learning_rate": 9.825047337696532e-06, "loss": 0.6842, "step": 918 }, { "epoch": 0.8823811809889582, "grad_norm": 0.7781466505892219, "learning_rate": 9.824167844862184e-06, "loss": 0.6348, "step": 919 }, { "epoch": 0.8833413346135381, "grad_norm": 1.330334712353335, "learning_rate": 9.823286186510456e-06, "loss": 0.8418, "step": 920 }, { "epoch": 0.884301488238118, "grad_norm": 0.8570106334467112, "learning_rate": 9.822402363037118e-06, "loss": 0.5833, "step": 921 }, { "epoch": 0.8852616418626981, "grad_norm": 1.0851237782669956, "learning_rate": 9.82151637483891e-06, "loss": 0.5841, "step": 922 }, { "epoch": 0.886221795487278, "grad_norm": 0.8615215009596465, "learning_rate": 9.820628222313535e-06, "loss": 0.7113, "step": 923 }, { "epoch": 0.8871819491118579, "grad_norm": 1.333999168186296, "learning_rate": 9.819737905859685e-06, "loss": 0.8556, "step": 924 }, { "epoch": 0.8881421027364378, "grad_norm": 0.9437410839731489, "learning_rate": 9.818845425877008e-06, "loss": 0.695, "step": 925 }, { "epoch": 0.8891022563610178, "grad_norm": 0.8892979203133354, "learning_rate": 9.81795078276613e-06, "loss": 0.7289, "step": 926 }, { "epoch": 0.8900624099855977, "grad_norm": 0.7473130009268761, "learning_rate": 9.817053976928646e-06, "loss": 0.6251, "step": 927 }, { "epoch": 0.8910225636101776, "grad_norm": 0.9091577635447758, "learning_rate": 9.816155008767123e-06, "loss": 0.638, "step": 928 }, { "epoch": 0.8919827172347575, "grad_norm": 0.6899265570483941, "learning_rate": 9.8152538786851e-06, "loss": 0.6081, "step": 929 }, { "epoch": 0.8929428708593375, "grad_norm": 0.9457658417616214, "learning_rate": 9.814350587087083e-06, "loss": 0.7914, "step": 930 }, { "epoch": 0.8939030244839175, "grad_norm": 0.898977091056529, "learning_rate": 9.813445134378548e-06, "loss": 0.6827, "step": 931 }, { "epoch": 0.8948631781084974, "grad_norm": 1.324910867170117, "learning_rate": 9.812537520965946e-06, "loss": 0.8502, "step": 932 }, { "epoch": 0.8958233317330773, "grad_norm": 0.7129564222584683, "learning_rate": 9.811627747256694e-06, "loss": 0.6173, "step": 933 }, { "epoch": 0.8967834853576572, "grad_norm": 0.8266929472371213, "learning_rate": 9.81071581365918e-06, "loss": 0.656, "step": 934 }, { "epoch": 0.8977436389822372, "grad_norm": 1.2748221752001576, "learning_rate": 9.809801720582761e-06, "loss": 0.6919, "step": 935 }, { "epoch": 0.8987037926068171, "grad_norm": 1.242760662920845, "learning_rate": 9.808885468437764e-06, "loss": 0.8612, "step": 936 }, { "epoch": 0.899663946231397, "grad_norm": 0.9357911722104719, "learning_rate": 9.807967057635484e-06, "loss": 0.6793, "step": 937 }, { "epoch": 0.9006240998559769, "grad_norm": 0.9437700369822757, "learning_rate": 9.807046488588186e-06, "loss": 0.8019, "step": 938 }, { "epoch": 0.9015842534805569, "grad_norm": 0.8443105582798032, "learning_rate": 9.806123761709103e-06, "loss": 0.6051, "step": 939 }, { "epoch": 0.9025444071051368, "grad_norm": 1.0063940330624401, "learning_rate": 9.80519887741244e-06, "loss": 0.8243, "step": 940 }, { "epoch": 0.9035045607297167, "grad_norm": 0.9895940771508072, "learning_rate": 9.804271836113364e-06, "loss": 0.7985, "step": 941 }, { "epoch": 0.9044647143542967, "grad_norm": 0.7930947586233464, "learning_rate": 9.803342638228014e-06, "loss": 0.6574, "step": 942 }, { "epoch": 0.9054248679788767, "grad_norm": 0.8068731743312122, "learning_rate": 9.802411284173502e-06, "loss": 0.6914, "step": 943 }, { "epoch": 0.9063850216034566, "grad_norm": 0.83824063255733, "learning_rate": 9.801477774367898e-06, "loss": 0.6889, "step": 944 }, { "epoch": 0.9073451752280365, "grad_norm": 0.9141400618614127, "learning_rate": 9.800542109230247e-06, "loss": 0.67, "step": 945 }, { "epoch": 0.9083053288526164, "grad_norm": 0.9875803920368459, "learning_rate": 9.799604289180556e-06, "loss": 0.733, "step": 946 }, { "epoch": 0.9092654824771963, "grad_norm": 1.145911295155869, "learning_rate": 9.798664314639805e-06, "loss": 0.7601, "step": 947 }, { "epoch": 0.9102256361017763, "grad_norm": 0.9241634722586574, "learning_rate": 9.79772218602994e-06, "loss": 0.7563, "step": 948 }, { "epoch": 0.9111857897263562, "grad_norm": 0.7959551894032442, "learning_rate": 9.796777903773868e-06, "loss": 0.6112, "step": 949 }, { "epoch": 0.9121459433509361, "grad_norm": 0.7908473942024027, "learning_rate": 9.795831468295472e-06, "loss": 0.7024, "step": 950 }, { "epoch": 0.913106096975516, "grad_norm": 1.1388277745463495, "learning_rate": 9.794882880019594e-06, "loss": 0.6963, "step": 951 }, { "epoch": 0.914066250600096, "grad_norm": 0.7778229735008709, "learning_rate": 9.793932139372045e-06, "loss": 0.6709, "step": 952 }, { "epoch": 0.915026404224676, "grad_norm": 1.211908352901218, "learning_rate": 9.792979246779602e-06, "loss": 0.9628, "step": 953 }, { "epoch": 0.9159865578492559, "grad_norm": 1.0541973698199625, "learning_rate": 9.79202420267001e-06, "loss": 0.7445, "step": 954 }, { "epoch": 0.9169467114738358, "grad_norm": 0.7872434876171807, "learning_rate": 9.791067007471977e-06, "loss": 0.6549, "step": 955 }, { "epoch": 0.9179068650984158, "grad_norm": 0.927408132052613, "learning_rate": 9.790107661615179e-06, "loss": 0.6981, "step": 956 }, { "epoch": 0.9188670187229957, "grad_norm": 1.3538524520521609, "learning_rate": 9.789146165530255e-06, "loss": 0.8816, "step": 957 }, { "epoch": 0.9198271723475756, "grad_norm": 1.0974354541026947, "learning_rate": 9.78818251964881e-06, "loss": 0.6537, "step": 958 }, { "epoch": 0.9207873259721555, "grad_norm": 0.8068510308866969, "learning_rate": 9.787216724403414e-06, "loss": 0.7688, "step": 959 }, { "epoch": 0.9217474795967355, "grad_norm": 1.0555377461040094, "learning_rate": 9.786248780227604e-06, "loss": 0.7285, "step": 960 }, { "epoch": 0.9227076332213154, "grad_norm": 0.891265976606178, "learning_rate": 9.785278687555879e-06, "loss": 0.7087, "step": 961 }, { "epoch": 0.9236677868458953, "grad_norm": 0.8552240233328497, "learning_rate": 9.784306446823702e-06, "loss": 0.6345, "step": 962 }, { "epoch": 0.9246279404704753, "grad_norm": 0.9451060539884003, "learning_rate": 9.783332058467502e-06, "loss": 0.7707, "step": 963 }, { "epoch": 0.9255880940950552, "grad_norm": 0.9093858461922119, "learning_rate": 9.782355522924674e-06, "loss": 0.6896, "step": 964 }, { "epoch": 0.9265482477196352, "grad_norm": 1.1791715812342562, "learning_rate": 9.78137684063357e-06, "loss": 0.7016, "step": 965 }, { "epoch": 0.9275084013442151, "grad_norm": 1.0090480475115489, "learning_rate": 9.780396012033514e-06, "loss": 0.6965, "step": 966 }, { "epoch": 0.928468554968795, "grad_norm": 0.9367516260522296, "learning_rate": 9.779413037564787e-06, "loss": 0.6993, "step": 967 }, { "epoch": 0.9294287085933749, "grad_norm": 0.8185786107285561, "learning_rate": 9.778427917668636e-06, "loss": 0.6888, "step": 968 }, { "epoch": 0.9303888622179549, "grad_norm": 0.9545063594576916, "learning_rate": 9.777440652787273e-06, "loss": 0.727, "step": 969 }, { "epoch": 0.9313490158425348, "grad_norm": 0.8888044201749987, "learning_rate": 9.776451243363866e-06, "loss": 0.5569, "step": 970 }, { "epoch": 0.9323091694671147, "grad_norm": 1.2353083721685303, "learning_rate": 9.775459689842558e-06, "loss": 0.752, "step": 971 }, { "epoch": 0.9332693230916946, "grad_norm": 0.9131734550861986, "learning_rate": 9.774465992668438e-06, "loss": 0.7363, "step": 972 }, { "epoch": 0.9342294767162747, "grad_norm": 0.9289794365835782, "learning_rate": 9.773470152287573e-06, "loss": 0.7068, "step": 973 }, { "epoch": 0.9351896303408546, "grad_norm": 0.8548292159154727, "learning_rate": 9.772472169146977e-06, "loss": 0.6977, "step": 974 }, { "epoch": 0.9361497839654345, "grad_norm": 0.9404602856483443, "learning_rate": 9.771472043694645e-06, "loss": 0.6936, "step": 975 }, { "epoch": 0.9371099375900144, "grad_norm": 0.8234905275815235, "learning_rate": 9.770469776379514e-06, "loss": 0.6589, "step": 976 }, { "epoch": 0.9380700912145943, "grad_norm": 1.2112873410988239, "learning_rate": 9.769465367651494e-06, "loss": 0.7823, "step": 977 }, { "epoch": 0.9390302448391743, "grad_norm": 0.8924445921289379, "learning_rate": 9.768458817961455e-06, "loss": 0.6639, "step": 978 }, { "epoch": 0.9399903984637542, "grad_norm": 0.8435556725256517, "learning_rate": 9.767450127761223e-06, "loss": 0.7234, "step": 979 }, { "epoch": 0.9409505520883341, "grad_norm": 0.7952335337114745, "learning_rate": 9.766439297503589e-06, "loss": 0.7075, "step": 980 }, { "epoch": 0.941910705712914, "grad_norm": 0.9071538678978494, "learning_rate": 9.765426327642306e-06, "loss": 0.7117, "step": 981 }, { "epoch": 0.942870859337494, "grad_norm": 0.9574299670905045, "learning_rate": 9.764411218632085e-06, "loss": 0.7304, "step": 982 }, { "epoch": 0.943831012962074, "grad_norm": 1.1949279388909293, "learning_rate": 9.763393970928593e-06, "loss": 0.7824, "step": 983 }, { "epoch": 0.9447911665866539, "grad_norm": 0.9191504450890636, "learning_rate": 9.76237458498847e-06, "loss": 0.6658, "step": 984 }, { "epoch": 0.9457513202112338, "grad_norm": 0.8768484174055277, "learning_rate": 9.7613530612693e-06, "loss": 0.8204, "step": 985 }, { "epoch": 0.9467114738358138, "grad_norm": 0.8233103934263737, "learning_rate": 9.760329400229639e-06, "loss": 0.656, "step": 986 }, { "epoch": 0.9476716274603937, "grad_norm": 1.2754974710872697, "learning_rate": 9.759303602328993e-06, "loss": 0.8698, "step": 987 }, { "epoch": 0.9486317810849736, "grad_norm": 0.8947895536867726, "learning_rate": 9.758275668027834e-06, "loss": 0.7196, "step": 988 }, { "epoch": 0.9495919347095535, "grad_norm": 0.9030023627728169, "learning_rate": 9.757245597787592e-06, "loss": 0.6799, "step": 989 }, { "epoch": 0.9505520883341335, "grad_norm": 0.9365020686670895, "learning_rate": 9.756213392070654e-06, "loss": 0.6225, "step": 990 }, { "epoch": 0.9515122419587134, "grad_norm": 1.171309675590621, "learning_rate": 9.755179051340365e-06, "loss": 0.755, "step": 991 }, { "epoch": 0.9524723955832933, "grad_norm": 1.0047227725151537, "learning_rate": 9.75414257606103e-06, "loss": 0.6225, "step": 992 }, { "epoch": 0.9534325492078732, "grad_norm": 0.8800043826675522, "learning_rate": 9.75310396669791e-06, "loss": 0.7264, "step": 993 }, { "epoch": 0.9543927028324531, "grad_norm": 1.0292617201852627, "learning_rate": 9.752063223717227e-06, "loss": 0.7882, "step": 994 }, { "epoch": 0.9553528564570332, "grad_norm": 0.73731962941736, "learning_rate": 9.75102034758616e-06, "loss": 0.6364, "step": 995 }, { "epoch": 0.9563130100816131, "grad_norm": 0.8854364657213919, "learning_rate": 9.749975338772846e-06, "loss": 0.6928, "step": 996 }, { "epoch": 0.957273163706193, "grad_norm": 0.7973633950513039, "learning_rate": 9.748928197746376e-06, "loss": 0.6699, "step": 997 }, { "epoch": 0.9582333173307729, "grad_norm": 0.8490281259244137, "learning_rate": 9.7478789249768e-06, "loss": 0.6364, "step": 998 }, { "epoch": 0.9591934709553529, "grad_norm": 0.8259942851778136, "learning_rate": 9.746827520935129e-06, "loss": 0.6906, "step": 999 }, { "epoch": 0.9601536245799328, "grad_norm": 1.0017256053359092, "learning_rate": 9.745773986093325e-06, "loss": 0.6477, "step": 1000 }, { "epoch": 0.9611137782045127, "grad_norm": 0.8301049426037799, "learning_rate": 9.744718320924307e-06, "loss": 0.6669, "step": 1001 }, { "epoch": 0.9620739318290926, "grad_norm": 0.815871878274097, "learning_rate": 9.743660525901952e-06, "loss": 0.7006, "step": 1002 }, { "epoch": 0.9630340854536726, "grad_norm": 0.7971590413985893, "learning_rate": 9.742600601501095e-06, "loss": 0.6289, "step": 1003 }, { "epoch": 0.9639942390782525, "grad_norm": 0.9137063793667768, "learning_rate": 9.741538548197522e-06, "loss": 0.782, "step": 1004 }, { "epoch": 0.9649543927028325, "grad_norm": 0.7213276478208919, "learning_rate": 9.74047436646798e-06, "loss": 0.6145, "step": 1005 }, { "epoch": 0.9659145463274124, "grad_norm": 0.8981249883451546, "learning_rate": 9.739408056790168e-06, "loss": 0.6563, "step": 1006 }, { "epoch": 0.9668746999519923, "grad_norm": 0.8905259033957025, "learning_rate": 9.73833961964274e-06, "loss": 0.6252, "step": 1007 }, { "epoch": 0.9678348535765723, "grad_norm": 0.7697510003957507, "learning_rate": 9.737269055505308e-06, "loss": 0.6881, "step": 1008 }, { "epoch": 0.9687950072011522, "grad_norm": 0.9124705656250613, "learning_rate": 9.736196364858435e-06, "loss": 0.668, "step": 1009 }, { "epoch": 0.9697551608257321, "grad_norm": 0.8672004826416162, "learning_rate": 9.735121548183642e-06, "loss": 0.6759, "step": 1010 }, { "epoch": 0.970715314450312, "grad_norm": 0.8732930972989155, "learning_rate": 9.7340446059634e-06, "loss": 0.6472, "step": 1011 }, { "epoch": 0.971675468074892, "grad_norm": 1.3225327691741082, "learning_rate": 9.732965538681139e-06, "loss": 0.812, "step": 1012 }, { "epoch": 0.9726356216994719, "grad_norm": 0.8849487651127176, "learning_rate": 9.73188434682124e-06, "loss": 0.7202, "step": 1013 }, { "epoch": 0.9735957753240518, "grad_norm": 0.8499954021978494, "learning_rate": 9.73080103086904e-06, "loss": 0.7316, "step": 1014 }, { "epoch": 0.9745559289486317, "grad_norm": 0.9267510966104285, "learning_rate": 9.729715591310824e-06, "loss": 0.7044, "step": 1015 }, { "epoch": 0.9755160825732118, "grad_norm": 0.9216935703882152, "learning_rate": 9.728628028633839e-06, "loss": 0.7137, "step": 1016 }, { "epoch": 0.9764762361977917, "grad_norm": 0.7863947041900111, "learning_rate": 9.727538343326277e-06, "loss": 0.6401, "step": 1017 }, { "epoch": 0.9774363898223716, "grad_norm": 1.0243629267453345, "learning_rate": 9.72644653587729e-06, "loss": 0.6777, "step": 1018 }, { "epoch": 0.9783965434469515, "grad_norm": 0.9675710428170455, "learning_rate": 9.725352606776973e-06, "loss": 0.6736, "step": 1019 }, { "epoch": 0.9793566970715314, "grad_norm": 1.5938288328678791, "learning_rate": 9.724256556516383e-06, "loss": 0.889, "step": 1020 }, { "epoch": 0.9803168506961114, "grad_norm": 0.8433988341741272, "learning_rate": 9.723158385587525e-06, "loss": 0.6387, "step": 1021 }, { "epoch": 0.9812770043206913, "grad_norm": 0.8859466690885591, "learning_rate": 9.722058094483355e-06, "loss": 0.7276, "step": 1022 }, { "epoch": 0.9822371579452712, "grad_norm": 1.0912811495539692, "learning_rate": 9.72095568369778e-06, "loss": 0.7347, "step": 1023 }, { "epoch": 0.9831973115698511, "grad_norm": 0.8964981278290173, "learning_rate": 9.719851153725668e-06, "loss": 0.6458, "step": 1024 }, { "epoch": 0.9841574651944311, "grad_norm": 0.8457498802554669, "learning_rate": 9.718744505062822e-06, "loss": 0.6056, "step": 1025 }, { "epoch": 0.985117618819011, "grad_norm": 1.2301247933002828, "learning_rate": 9.717635738206007e-06, "loss": 0.8681, "step": 1026 }, { "epoch": 0.986077772443591, "grad_norm": 0.7958814531142709, "learning_rate": 9.716524853652942e-06, "loss": 0.6064, "step": 1027 }, { "epoch": 0.9870379260681709, "grad_norm": 1.0155696304953739, "learning_rate": 9.715411851902284e-06, "loss": 0.819, "step": 1028 }, { "epoch": 0.9879980796927509, "grad_norm": 0.7737576710910635, "learning_rate": 9.71429673345365e-06, "loss": 0.6494, "step": 1029 }, { "epoch": 0.9889582333173308, "grad_norm": 0.8940371555086873, "learning_rate": 9.713179498807607e-06, "loss": 0.768, "step": 1030 }, { "epoch": 0.9899183869419107, "grad_norm": 0.8787009888657692, "learning_rate": 9.712060148465669e-06, "loss": 0.6191, "step": 1031 }, { "epoch": 0.9908785405664906, "grad_norm": 0.8529293379247336, "learning_rate": 9.710938682930298e-06, "loss": 0.625, "step": 1032 }, { "epoch": 0.9918386941910706, "grad_norm": 1.176282009246556, "learning_rate": 9.709815102704909e-06, "loss": 0.7706, "step": 1033 }, { "epoch": 0.9927988478156505, "grad_norm": 1.1354427238851763, "learning_rate": 9.708689408293866e-06, "loss": 0.6876, "step": 1034 }, { "epoch": 0.9937590014402304, "grad_norm": 0.9730193616087658, "learning_rate": 9.707561600202481e-06, "loss": 0.7627, "step": 1035 }, { "epoch": 0.9947191550648103, "grad_norm": 0.7613666471327217, "learning_rate": 9.706431678937015e-06, "loss": 0.5528, "step": 1036 }, { "epoch": 0.9956793086893903, "grad_norm": 1.2305792639139255, "learning_rate": 9.705299645004679e-06, "loss": 0.8117, "step": 1037 }, { "epoch": 0.9966394623139703, "grad_norm": 0.7757653531466473, "learning_rate": 9.704165498913624e-06, "loss": 0.7053, "step": 1038 }, { "epoch": 0.9975996159385502, "grad_norm": 1.246968644865225, "learning_rate": 9.703029241172965e-06, "loss": 0.8161, "step": 1039 }, { "epoch": 0.9985597695631301, "grad_norm": 0.8702218871621071, "learning_rate": 9.701890872292753e-06, "loss": 0.7128, "step": 1040 }, { "epoch": 0.99951992318771, "grad_norm": 0.9381198480672964, "learning_rate": 9.700750392783986e-06, "loss": 0.7784, "step": 1041 }, { "epoch": 1.0, "grad_norm": 0.9381198480672964, "learning_rate": 9.699607803158618e-06, "loss": 0.8249, "step": 1042 }, { "epoch": 1.00096015362458, "grad_norm": 1.7125645713139614, "learning_rate": 9.698463103929542e-06, "loss": 0.6246, "step": 1043 }, { "epoch": 1.0019203072491598, "grad_norm": 1.057183572291711, "learning_rate": 9.697316295610604e-06, "loss": 0.5964, "step": 1044 }, { "epoch": 1.0028804608737398, "grad_norm": 0.6849097349351592, "learning_rate": 9.696167378716591e-06, "loss": 0.6223, "step": 1045 }, { "epoch": 1.0038406144983196, "grad_norm": 0.8263359817050435, "learning_rate": 9.695016353763241e-06, "loss": 0.6206, "step": 1046 }, { "epoch": 1.0048007681228996, "grad_norm": 0.9295062887639517, "learning_rate": 9.693863221267238e-06, "loss": 0.5531, "step": 1047 }, { "epoch": 1.0057609217474797, "grad_norm": 0.7142609587287906, "learning_rate": 9.692707981746208e-06, "loss": 0.5958, "step": 1048 }, { "epoch": 1.0067210753720595, "grad_norm": 0.9119764422251769, "learning_rate": 9.691550635718729e-06, "loss": 0.6614, "step": 1049 }, { "epoch": 1.0076812289966395, "grad_norm": 1.014922765903184, "learning_rate": 9.690391183704318e-06, "loss": 0.5559, "step": 1050 }, { "epoch": 1.0086413826212195, "grad_norm": 1.2633756530789586, "learning_rate": 9.689229626223441e-06, "loss": 0.6593, "step": 1051 }, { "epoch": 1.0096015362457993, "grad_norm": 1.3675344526789353, "learning_rate": 9.688065963797512e-06, "loss": 0.771, "step": 1052 }, { "epoch": 1.0105616898703793, "grad_norm": 0.8815956624440381, "learning_rate": 9.686900196948886e-06, "loss": 0.7449, "step": 1053 }, { "epoch": 1.011521843494959, "grad_norm": 1.4981762393494307, "learning_rate": 9.685732326200859e-06, "loss": 0.5799, "step": 1054 }, { "epoch": 1.0124819971195391, "grad_norm": 0.7890154422683266, "learning_rate": 9.684562352077679e-06, "loss": 0.6741, "step": 1055 }, { "epoch": 1.0134421507441191, "grad_norm": 1.001507132018874, "learning_rate": 9.683390275104533e-06, "loss": 0.5971, "step": 1056 }, { "epoch": 1.014402304368699, "grad_norm": 0.957391967481929, "learning_rate": 9.682216095807559e-06, "loss": 0.717, "step": 1057 }, { "epoch": 1.015362457993279, "grad_norm": 1.1223957893816978, "learning_rate": 9.681039814713827e-06, "loss": 0.6528, "step": 1058 }, { "epoch": 1.0163226116178588, "grad_norm": 1.066556277742611, "learning_rate": 9.67986143235136e-06, "loss": 0.66, "step": 1059 }, { "epoch": 1.0172827652424388, "grad_norm": 0.9625514897764756, "learning_rate": 9.67868094924912e-06, "loss": 0.6981, "step": 1060 }, { "epoch": 1.0182429188670188, "grad_norm": 1.0700126962675625, "learning_rate": 9.677498365937017e-06, "loss": 0.7225, "step": 1061 }, { "epoch": 1.0192030724915986, "grad_norm": 0.6188954380440629, "learning_rate": 9.676313682945896e-06, "loss": 0.4758, "step": 1062 }, { "epoch": 1.0201632261161786, "grad_norm": 0.9592469031116028, "learning_rate": 9.675126900807548e-06, "loss": 0.6623, "step": 1063 }, { "epoch": 1.0211233797407586, "grad_norm": 0.764797387874803, "learning_rate": 9.673938020054709e-06, "loss": 0.5526, "step": 1064 }, { "epoch": 1.0220835333653384, "grad_norm": 0.7557968419409261, "learning_rate": 9.672747041221055e-06, "loss": 0.565, "step": 1065 }, { "epoch": 1.0230436869899184, "grad_norm": 1.2616420544037212, "learning_rate": 9.671553964841206e-06, "loss": 0.7191, "step": 1066 }, { "epoch": 1.0240038406144982, "grad_norm": 1.0756958546722464, "learning_rate": 9.670358791450713e-06, "loss": 0.5347, "step": 1067 }, { "epoch": 1.0249639942390782, "grad_norm": 0.9081941496886377, "learning_rate": 9.669161521586086e-06, "loss": 0.6627, "step": 1068 }, { "epoch": 1.0259241478636583, "grad_norm": 1.0354732706351037, "learning_rate": 9.66796215578476e-06, "loss": 0.6754, "step": 1069 }, { "epoch": 1.026884301488238, "grad_norm": 1.088367846119722, "learning_rate": 9.666760694585122e-06, "loss": 0.6676, "step": 1070 }, { "epoch": 1.027844455112818, "grad_norm": 0.8650262402686005, "learning_rate": 9.665557138526492e-06, "loss": 0.6683, "step": 1071 }, { "epoch": 1.0288046087373979, "grad_norm": 1.001871641829671, "learning_rate": 9.664351488149136e-06, "loss": 0.7181, "step": 1072 }, { "epoch": 1.029764762361978, "grad_norm": 1.0126301038309395, "learning_rate": 9.663143743994254e-06, "loss": 0.7033, "step": 1073 }, { "epoch": 1.030724915986558, "grad_norm": 1.0790542370127933, "learning_rate": 9.661933906603994e-06, "loss": 0.5627, "step": 1074 }, { "epoch": 1.0316850696111377, "grad_norm": 0.6688326906984543, "learning_rate": 9.660721976521433e-06, "loss": 0.5844, "step": 1075 }, { "epoch": 1.0326452232357177, "grad_norm": 0.719234035612182, "learning_rate": 9.659507954290599e-06, "loss": 0.5796, "step": 1076 }, { "epoch": 1.0336053768602977, "grad_norm": 0.9195866983984744, "learning_rate": 9.658291840456454e-06, "loss": 0.6084, "step": 1077 }, { "epoch": 1.0345655304848775, "grad_norm": 0.9317838895820011, "learning_rate": 9.657073635564893e-06, "loss": 0.6217, "step": 1078 }, { "epoch": 1.0355256841094576, "grad_norm": 0.7165974509984959, "learning_rate": 9.65585334016276e-06, "loss": 0.7141, "step": 1079 }, { "epoch": 1.0364858377340374, "grad_norm": 1.0612295257340327, "learning_rate": 9.65463095479783e-06, "loss": 0.5758, "step": 1080 }, { "epoch": 1.0374459913586174, "grad_norm": 0.7663580911487703, "learning_rate": 9.653406480018819e-06, "loss": 0.6844, "step": 1081 }, { "epoch": 1.0384061449831974, "grad_norm": 1.1777745955640475, "learning_rate": 9.652179916375382e-06, "loss": 0.6787, "step": 1082 }, { "epoch": 1.0393662986077772, "grad_norm": 0.8036849468479822, "learning_rate": 9.65095126441811e-06, "loss": 0.6268, "step": 1083 }, { "epoch": 1.0403264522323572, "grad_norm": 0.907321842035561, "learning_rate": 9.649720524698532e-06, "loss": 0.7181, "step": 1084 }, { "epoch": 1.041286605856937, "grad_norm": 1.258368372044864, "learning_rate": 9.648487697769112e-06, "loss": 0.6943, "step": 1085 }, { "epoch": 1.042246759481517, "grad_norm": 1.1437258473419336, "learning_rate": 9.647252784183254e-06, "loss": 0.5924, "step": 1086 }, { "epoch": 1.043206913106097, "grad_norm": 0.8644551605082604, "learning_rate": 9.646015784495297e-06, "loss": 0.5629, "step": 1087 }, { "epoch": 1.0441670667306768, "grad_norm": 0.8764378087676976, "learning_rate": 9.64477669926052e-06, "loss": 0.6942, "step": 1088 }, { "epoch": 1.0451272203552568, "grad_norm": 1.0541260318055827, "learning_rate": 9.643535529035131e-06, "loss": 0.6281, "step": 1089 }, { "epoch": 1.0460873739798369, "grad_norm": 0.999237140365186, "learning_rate": 9.64229227437628e-06, "loss": 0.7314, "step": 1090 }, { "epoch": 1.0470475276044167, "grad_norm": 1.0121819658212057, "learning_rate": 9.641046935842048e-06, "loss": 0.712, "step": 1091 }, { "epoch": 1.0480076812289967, "grad_norm": 0.9146105448886497, "learning_rate": 9.63979951399146e-06, "loss": 0.6554, "step": 1092 }, { "epoch": 1.0489678348535765, "grad_norm": 0.9754090808119935, "learning_rate": 9.638550009384467e-06, "loss": 0.5202, "step": 1093 }, { "epoch": 1.0499279884781565, "grad_norm": 0.7249685967151078, "learning_rate": 9.637298422581955e-06, "loss": 0.7281, "step": 1094 }, { "epoch": 1.0508881421027365, "grad_norm": 1.0028668086789203, "learning_rate": 9.636044754145753e-06, "loss": 0.5796, "step": 1095 }, { "epoch": 1.0518482957273163, "grad_norm": 0.8394275846810937, "learning_rate": 9.634789004638618e-06, "loss": 0.6294, "step": 1096 }, { "epoch": 1.0528084493518963, "grad_norm": 0.9531848588238243, "learning_rate": 9.63353117462424e-06, "loss": 0.7111, "step": 1097 }, { "epoch": 1.0537686029764763, "grad_norm": 0.9988837369319868, "learning_rate": 9.632271264667249e-06, "loss": 0.5623, "step": 1098 }, { "epoch": 1.0547287566010561, "grad_norm": 0.9430224177273493, "learning_rate": 9.631009275333204e-06, "loss": 0.6943, "step": 1099 }, { "epoch": 1.0556889102256362, "grad_norm": 0.9936318119909381, "learning_rate": 9.629745207188596e-06, "loss": 0.7221, "step": 1100 }, { "epoch": 1.056649063850216, "grad_norm": 0.955824954631298, "learning_rate": 9.628479060800855e-06, "loss": 0.5994, "step": 1101 }, { "epoch": 1.057609217474796, "grad_norm": 0.9174852301821367, "learning_rate": 9.627210836738338e-06, "loss": 0.6419, "step": 1102 }, { "epoch": 1.058569371099376, "grad_norm": 1.109518753879785, "learning_rate": 9.62594053557034e-06, "loss": 0.7877, "step": 1103 }, { "epoch": 1.0595295247239558, "grad_norm": 1.1025335547547768, "learning_rate": 9.624668157867084e-06, "loss": 0.6789, "step": 1104 }, { "epoch": 1.0604896783485358, "grad_norm": 1.017053318327755, "learning_rate": 9.623393704199727e-06, "loss": 0.662, "step": 1105 }, { "epoch": 1.0614498319731156, "grad_norm": 0.923766902087191, "learning_rate": 9.622117175140357e-06, "loss": 0.6591, "step": 1106 }, { "epoch": 1.0624099855976956, "grad_norm": 1.0151942391554418, "learning_rate": 9.620838571261995e-06, "loss": 0.5951, "step": 1107 }, { "epoch": 1.0633701392222756, "grad_norm": 0.9122527548283507, "learning_rate": 9.619557893138593e-06, "loss": 0.6582, "step": 1108 }, { "epoch": 1.0643302928468554, "grad_norm": 0.9988851220710931, "learning_rate": 9.618275141345032e-06, "loss": 0.6439, "step": 1109 }, { "epoch": 1.0652904464714354, "grad_norm": 0.9262379798576478, "learning_rate": 9.616990316457127e-06, "loss": 0.6014, "step": 1110 }, { "epoch": 1.0662506000960155, "grad_norm": 0.7589102521103533, "learning_rate": 9.615703419051622e-06, "loss": 0.6684, "step": 1111 }, { "epoch": 1.0672107537205953, "grad_norm": 0.9180929445230135, "learning_rate": 9.614414449706191e-06, "loss": 0.6636, "step": 1112 }, { "epoch": 1.0681709073451753, "grad_norm": 0.7816364047058745, "learning_rate": 9.61312340899944e-06, "loss": 0.6369, "step": 1113 }, { "epoch": 1.069131060969755, "grad_norm": 1.073352960142353, "learning_rate": 9.6118302975109e-06, "loss": 0.6357, "step": 1114 }, { "epoch": 1.070091214594335, "grad_norm": 0.9397875131021031, "learning_rate": 9.61053511582104e-06, "loss": 0.6879, "step": 1115 }, { "epoch": 1.0710513682189151, "grad_norm": 0.8474892068057238, "learning_rate": 9.609237864511249e-06, "loss": 0.6922, "step": 1116 }, { "epoch": 1.072011521843495, "grad_norm": 0.9443474284777743, "learning_rate": 9.607938544163851e-06, "loss": 0.5977, "step": 1117 }, { "epoch": 1.072971675468075, "grad_norm": 0.8546547759789594, "learning_rate": 9.606637155362095e-06, "loss": 0.6344, "step": 1118 }, { "epoch": 1.0739318290926547, "grad_norm": 0.8501396936121727, "learning_rate": 9.605333698690164e-06, "loss": 0.5889, "step": 1119 }, { "epoch": 1.0748919827172347, "grad_norm": 0.8699332966018324, "learning_rate": 9.604028174733164e-06, "loss": 0.6902, "step": 1120 }, { "epoch": 1.0758521363418148, "grad_norm": 1.0162854744629004, "learning_rate": 9.60272058407713e-06, "loss": 0.6981, "step": 1121 }, { "epoch": 1.0768122899663946, "grad_norm": 0.7066746316882161, "learning_rate": 9.601410927309026e-06, "loss": 0.608, "step": 1122 }, { "epoch": 1.0777724435909746, "grad_norm": 1.1274593390371928, "learning_rate": 9.600099205016744e-06, "loss": 0.5615, "step": 1123 }, { "epoch": 1.0787325972155546, "grad_norm": 0.7728763319366568, "learning_rate": 9.5987854177891e-06, "loss": 0.6923, "step": 1124 }, { "epoch": 1.0796927508401344, "grad_norm": 0.7360881705887629, "learning_rate": 9.597469566215841e-06, "loss": 0.5241, "step": 1125 }, { "epoch": 1.0806529044647144, "grad_norm": 0.8942332787183556, "learning_rate": 9.596151650887637e-06, "loss": 0.5292, "step": 1126 }, { "epoch": 1.0816130580892942, "grad_norm": 0.9948069899043459, "learning_rate": 9.594831672396088e-06, "loss": 0.6241, "step": 1127 }, { "epoch": 1.0825732117138742, "grad_norm": 0.8386044158877731, "learning_rate": 9.593509631333717e-06, "loss": 0.6292, "step": 1128 }, { "epoch": 1.0835333653384542, "grad_norm": 0.8427578100127142, "learning_rate": 9.592185528293975e-06, "loss": 0.6235, "step": 1129 }, { "epoch": 1.084493518963034, "grad_norm": 0.9443420677278583, "learning_rate": 9.590859363871237e-06, "loss": 0.6377, "step": 1130 }, { "epoch": 1.085453672587614, "grad_norm": 0.8076349664403714, "learning_rate": 9.589531138660803e-06, "loss": 0.5056, "step": 1131 }, { "epoch": 1.086413826212194, "grad_norm": 1.11916391688949, "learning_rate": 9.588200853258901e-06, "loss": 0.6359, "step": 1132 }, { "epoch": 1.0873739798367739, "grad_norm": 1.2021340250903567, "learning_rate": 9.586868508262681e-06, "loss": 0.6441, "step": 1133 }, { "epoch": 1.0883341334613539, "grad_norm": 0.7077682066327982, "learning_rate": 9.58553410427022e-06, "loss": 0.608, "step": 1134 }, { "epoch": 1.0892942870859337, "grad_norm": 0.9724877644350576, "learning_rate": 9.584197641880513e-06, "loss": 0.7014, "step": 1135 }, { "epoch": 1.0902544407105137, "grad_norm": 1.0964798094861308, "learning_rate": 9.582859121693488e-06, "loss": 0.7918, "step": 1136 }, { "epoch": 1.0912145943350937, "grad_norm": 1.081321770591046, "learning_rate": 9.581518544309993e-06, "loss": 0.6268, "step": 1137 }, { "epoch": 1.0921747479596735, "grad_norm": 1.0606171196887748, "learning_rate": 9.580175910331797e-06, "loss": 0.5706, "step": 1138 }, { "epoch": 1.0931349015842535, "grad_norm": 0.8213800655112417, "learning_rate": 9.578831220361593e-06, "loss": 0.6721, "step": 1139 }, { "epoch": 1.0940950552088333, "grad_norm": 1.009351627624076, "learning_rate": 9.577484475003e-06, "loss": 0.6545, "step": 1140 }, { "epoch": 1.0950552088334133, "grad_norm": 0.8957907956813881, "learning_rate": 9.576135674860557e-06, "loss": 0.6228, "step": 1141 }, { "epoch": 1.0960153624579934, "grad_norm": 1.0005109864101902, "learning_rate": 9.574784820539725e-06, "loss": 0.6112, "step": 1142 }, { "epoch": 1.0969755160825732, "grad_norm": 0.9117693224107493, "learning_rate": 9.573431912646888e-06, "loss": 0.6049, "step": 1143 }, { "epoch": 1.0979356697071532, "grad_norm": 0.6541127321945297, "learning_rate": 9.572076951789353e-06, "loss": 0.5987, "step": 1144 }, { "epoch": 1.0988958233317332, "grad_norm": 1.045290344694907, "learning_rate": 9.570719938575349e-06, "loss": 0.5881, "step": 1145 }, { "epoch": 1.099855976956313, "grad_norm": 0.8707687021700571, "learning_rate": 9.56936087361402e-06, "loss": 0.7008, "step": 1146 }, { "epoch": 1.100816130580893, "grad_norm": 0.6677050353556723, "learning_rate": 9.56799975751544e-06, "loss": 0.6524, "step": 1147 }, { "epoch": 1.1017762842054728, "grad_norm": 0.9969684326782041, "learning_rate": 9.566636590890598e-06, "loss": 0.5002, "step": 1148 }, { "epoch": 1.1027364378300528, "grad_norm": 0.5872767904288713, "learning_rate": 9.565271374351405e-06, "loss": 0.5824, "step": 1149 }, { "epoch": 1.1036965914546328, "grad_norm": 1.054389458936764, "learning_rate": 9.563904108510693e-06, "loss": 0.5862, "step": 1150 }, { "epoch": 1.1046567450792126, "grad_norm": 0.7744288230577119, "learning_rate": 9.562534793982212e-06, "loss": 0.8205, "step": 1151 }, { "epoch": 1.1056168987037926, "grad_norm": 1.0220649454756339, "learning_rate": 9.56116343138063e-06, "loss": 0.6425, "step": 1152 }, { "epoch": 1.1065770523283724, "grad_norm": 0.9237996840315938, "learning_rate": 9.559790021321544e-06, "loss": 0.5967, "step": 1153 }, { "epoch": 1.1075372059529525, "grad_norm": 0.7671428999198082, "learning_rate": 9.558414564421456e-06, "loss": 0.621, "step": 1154 }, { "epoch": 1.1084973595775325, "grad_norm": 0.6324938985234146, "learning_rate": 9.557037061297799e-06, "loss": 0.4714, "step": 1155 }, { "epoch": 1.1094575132021123, "grad_norm": 0.7732047481906313, "learning_rate": 9.555657512568916e-06, "loss": 0.4852, "step": 1156 }, { "epoch": 1.1104176668266923, "grad_norm": 1.3279362825055938, "learning_rate": 9.554275918854073e-06, "loss": 0.6552, "step": 1157 }, { "epoch": 1.1113778204512723, "grad_norm": 0.8495529728545537, "learning_rate": 9.55289228077345e-06, "loss": 0.5642, "step": 1158 }, { "epoch": 1.112337974075852, "grad_norm": 1.2698548515852557, "learning_rate": 9.551506598948152e-06, "loss": 0.6373, "step": 1159 }, { "epoch": 1.1132981277004321, "grad_norm": 0.9410677173807662, "learning_rate": 9.550118874000196e-06, "loss": 0.5846, "step": 1160 }, { "epoch": 1.114258281325012, "grad_norm": 1.0491232539651245, "learning_rate": 9.548729106552515e-06, "loss": 0.7247, "step": 1161 }, { "epoch": 1.115218434949592, "grad_norm": 0.7497379574753003, "learning_rate": 9.54733729722896e-06, "loss": 0.6418, "step": 1162 }, { "epoch": 1.116178588574172, "grad_norm": 1.0743325628100975, "learning_rate": 9.5459434466543e-06, "loss": 0.6478, "step": 1163 }, { "epoch": 1.1171387421987518, "grad_norm": 1.0390924938011707, "learning_rate": 9.544547555454222e-06, "loss": 0.6315, "step": 1164 }, { "epoch": 1.1180988958233318, "grad_norm": 1.1529497269836626, "learning_rate": 9.543149624255326e-06, "loss": 0.7305, "step": 1165 }, { "epoch": 1.1190590494479116, "grad_norm": 0.9891519352467538, "learning_rate": 9.541749653685124e-06, "loss": 0.539, "step": 1166 }, { "epoch": 1.1200192030724916, "grad_norm": 0.6114350399104391, "learning_rate": 9.540347644372054e-06, "loss": 0.5723, "step": 1167 }, { "epoch": 1.1209793566970716, "grad_norm": 1.0577553936477444, "learning_rate": 9.538943596945459e-06, "loss": 0.5697, "step": 1168 }, { "epoch": 1.1219395103216514, "grad_norm": 0.9520410606414327, "learning_rate": 9.5375375120356e-06, "loss": 0.5998, "step": 1169 }, { "epoch": 1.1228996639462314, "grad_norm": 0.8748590984094271, "learning_rate": 9.536129390273659e-06, "loss": 0.5768, "step": 1170 }, { "epoch": 1.1238598175708114, "grad_norm": 0.7199519481868286, "learning_rate": 9.53471923229172e-06, "loss": 0.6895, "step": 1171 }, { "epoch": 1.1248199711953912, "grad_norm": 1.1949709000071098, "learning_rate": 9.533307038722792e-06, "loss": 0.6586, "step": 1172 }, { "epoch": 1.1257801248199713, "grad_norm": 0.7871397240556544, "learning_rate": 9.531892810200793e-06, "loss": 0.6499, "step": 1173 }, { "epoch": 1.126740278444551, "grad_norm": 1.1144631049687614, "learning_rate": 9.530476547360555e-06, "loss": 0.6563, "step": 1174 }, { "epoch": 1.127700432069131, "grad_norm": 0.9773788573764292, "learning_rate": 9.529058250837823e-06, "loss": 0.7166, "step": 1175 }, { "epoch": 1.128660585693711, "grad_norm": 1.1518946118850573, "learning_rate": 9.527637921269256e-06, "loss": 0.6336, "step": 1176 }, { "epoch": 1.1296207393182909, "grad_norm": 0.8294750724129425, "learning_rate": 9.526215559292423e-06, "loss": 0.6308, "step": 1177 }, { "epoch": 1.130580892942871, "grad_norm": 1.1288986508564527, "learning_rate": 9.524791165545808e-06, "loss": 0.7306, "step": 1178 }, { "epoch": 1.1315410465674507, "grad_norm": 0.9584797631052782, "learning_rate": 9.523364740668805e-06, "loss": 0.6369, "step": 1179 }, { "epoch": 1.1325012001920307, "grad_norm": 0.8711660980223555, "learning_rate": 9.521936285301723e-06, "loss": 0.6295, "step": 1180 }, { "epoch": 1.1334613538166107, "grad_norm": 1.0851248843747978, "learning_rate": 9.52050580008578e-06, "loss": 0.6461, "step": 1181 }, { "epoch": 1.1344215074411905, "grad_norm": 0.8720293729230344, "learning_rate": 9.519073285663103e-06, "loss": 0.6899, "step": 1182 }, { "epoch": 1.1353816610657705, "grad_norm": 0.6913901086911198, "learning_rate": 9.517638742676736e-06, "loss": 0.5984, "step": 1183 }, { "epoch": 1.1363418146903506, "grad_norm": 0.8341227662051727, "learning_rate": 9.516202171770627e-06, "loss": 0.555, "step": 1184 }, { "epoch": 1.1373019683149304, "grad_norm": 0.8655639955018779, "learning_rate": 9.51476357358964e-06, "loss": 0.5497, "step": 1185 }, { "epoch": 1.1382621219395104, "grad_norm": 1.0850439871709279, "learning_rate": 9.513322948779544e-06, "loss": 0.63, "step": 1186 }, { "epoch": 1.1392222755640902, "grad_norm": 1.1228550010491494, "learning_rate": 9.511880297987022e-06, "loss": 0.5552, "step": 1187 }, { "epoch": 1.1401824291886702, "grad_norm": 0.7921629742836642, "learning_rate": 9.510435621859662e-06, "loss": 0.6375, "step": 1188 }, { "epoch": 1.1411425828132502, "grad_norm": 1.148614185293995, "learning_rate": 9.508988921045967e-06, "loss": 0.7036, "step": 1189 }, { "epoch": 1.14210273643783, "grad_norm": 1.1561811459211908, "learning_rate": 9.507540196195342e-06, "loss": 0.5214, "step": 1190 }, { "epoch": 1.14306289006241, "grad_norm": 0.8040000505057628, "learning_rate": 9.506089447958108e-06, "loss": 0.7511, "step": 1191 }, { "epoch": 1.1440230436869898, "grad_norm": 0.9644690413066516, "learning_rate": 9.504636676985486e-06, "loss": 0.7235, "step": 1192 }, { "epoch": 1.1449831973115698, "grad_norm": 1.1894982974484598, "learning_rate": 9.503181883929613e-06, "loss": 0.626, "step": 1193 }, { "epoch": 1.1459433509361499, "grad_norm": 0.9329901934043924, "learning_rate": 9.501725069443528e-06, "loss": 0.669, "step": 1194 }, { "epoch": 1.1469035045607296, "grad_norm": 0.9237155675048903, "learning_rate": 9.50026623418118e-06, "loss": 0.601, "step": 1195 }, { "epoch": 1.1478636581853097, "grad_norm": 0.9715692627450506, "learning_rate": 9.498805378797427e-06, "loss": 0.7107, "step": 1196 }, { "epoch": 1.1488238118098897, "grad_norm": 0.9938491233348047, "learning_rate": 9.497342503948027e-06, "loss": 0.6365, "step": 1197 }, { "epoch": 1.1497839654344695, "grad_norm": 0.7588197753235485, "learning_rate": 9.49587761028965e-06, "loss": 0.5955, "step": 1198 }, { "epoch": 1.1507441190590495, "grad_norm": 1.0133692536057142, "learning_rate": 9.494410698479873e-06, "loss": 0.755, "step": 1199 }, { "epoch": 1.1517042726836293, "grad_norm": 0.8701973340578494, "learning_rate": 9.492941769177176e-06, "loss": 0.6845, "step": 1200 }, { "epoch": 1.1526644263082093, "grad_norm": 1.1337765659195391, "learning_rate": 9.491470823040944e-06, "loss": 0.7383, "step": 1201 }, { "epoch": 1.1536245799327893, "grad_norm": 1.0464531234767978, "learning_rate": 9.489997860731475e-06, "loss": 0.5811, "step": 1202 }, { "epoch": 1.1545847335573691, "grad_norm": 1.1282868580788523, "learning_rate": 9.488522882909959e-06, "loss": 0.6605, "step": 1203 }, { "epoch": 1.1555448871819491, "grad_norm": 1.0058103956479245, "learning_rate": 9.4870458902385e-06, "loss": 0.5538, "step": 1204 }, { "epoch": 1.156505040806529, "grad_norm": 0.8365145325866349, "learning_rate": 9.485566883380105e-06, "loss": 0.6131, "step": 1205 }, { "epoch": 1.157465194431109, "grad_norm": 0.7161485425648121, "learning_rate": 9.484085862998685e-06, "loss": 0.6198, "step": 1206 }, { "epoch": 1.158425348055689, "grad_norm": 1.211244116512397, "learning_rate": 9.482602829759052e-06, "loss": 0.6442, "step": 1207 }, { "epoch": 1.1593855016802688, "grad_norm": 0.9568728816934647, "learning_rate": 9.481117784326927e-06, "loss": 0.8419, "step": 1208 }, { "epoch": 1.1603456553048488, "grad_norm": 1.216531871506148, "learning_rate": 9.479630727368927e-06, "loss": 0.6568, "step": 1209 }, { "epoch": 1.1613058089294288, "grad_norm": 1.1786662405707418, "learning_rate": 9.47814165955258e-06, "loss": 0.6123, "step": 1210 }, { "epoch": 1.1622659625540086, "grad_norm": 0.952355551841717, "learning_rate": 9.47665058154631e-06, "loss": 0.6183, "step": 1211 }, { "epoch": 1.1632261161785886, "grad_norm": 0.6930943728988435, "learning_rate": 9.475157494019445e-06, "loss": 0.5412, "step": 1212 }, { "epoch": 1.1641862698031684, "grad_norm": 1.0625287702467052, "learning_rate": 9.473662397642219e-06, "loss": 0.5747, "step": 1213 }, { "epoch": 1.1651464234277484, "grad_norm": 0.7675165592979213, "learning_rate": 9.472165293085765e-06, "loss": 0.5416, "step": 1214 }, { "epoch": 1.1661065770523285, "grad_norm": 0.8091262728844713, "learning_rate": 9.470666181022114e-06, "loss": 0.6235, "step": 1215 }, { "epoch": 1.1670667306769082, "grad_norm": 0.9400548046884452, "learning_rate": 9.469165062124203e-06, "loss": 0.6277, "step": 1216 }, { "epoch": 1.1680268843014883, "grad_norm": 0.8889766497907815, "learning_rate": 9.467661937065869e-06, "loss": 0.6344, "step": 1217 }, { "epoch": 1.168987037926068, "grad_norm": 1.2303890871634446, "learning_rate": 9.466156806521847e-06, "loss": 0.6445, "step": 1218 }, { "epoch": 1.169947191550648, "grad_norm": 1.2954030759575375, "learning_rate": 9.464649671167777e-06, "loss": 0.6813, "step": 1219 }, { "epoch": 1.170907345175228, "grad_norm": 0.9067505364565195, "learning_rate": 9.463140531680193e-06, "loss": 0.5151, "step": 1220 }, { "epoch": 1.171867498799808, "grad_norm": 0.9474933236110304, "learning_rate": 9.461629388736534e-06, "loss": 0.5647, "step": 1221 }, { "epoch": 1.172827652424388, "grad_norm": 0.9486093383900981, "learning_rate": 9.460116243015131e-06, "loss": 0.7298, "step": 1222 }, { "epoch": 1.173787806048968, "grad_norm": 0.8197978337894188, "learning_rate": 9.458601095195225e-06, "loss": 0.5537, "step": 1223 }, { "epoch": 1.1747479596735477, "grad_norm": 0.7204772269693327, "learning_rate": 9.45708394595695e-06, "loss": 0.678, "step": 1224 }, { "epoch": 1.1757081132981277, "grad_norm": 1.033385177648348, "learning_rate": 9.455564795981331e-06, "loss": 0.5992, "step": 1225 }, { "epoch": 1.1766682669227075, "grad_norm": 0.8575159511680858, "learning_rate": 9.454043645950303e-06, "loss": 0.7304, "step": 1226 }, { "epoch": 1.1776284205472876, "grad_norm": 1.1799167009951452, "learning_rate": 9.452520496546693e-06, "loss": 0.7477, "step": 1227 }, { "epoch": 1.1785885741718676, "grad_norm": 0.8954288882594916, "learning_rate": 9.450995348454228e-06, "loss": 0.6244, "step": 1228 }, { "epoch": 1.1795487277964474, "grad_norm": 1.251034703997885, "learning_rate": 9.44946820235753e-06, "loss": 0.5839, "step": 1229 }, { "epoch": 1.1805088814210274, "grad_norm": 0.9292205500155021, "learning_rate": 9.447939058942119e-06, "loss": 0.4558, "step": 1230 }, { "epoch": 1.1814690350456072, "grad_norm": 0.973217644984768, "learning_rate": 9.446407918894408e-06, "loss": 0.6872, "step": 1231 }, { "epoch": 1.1824291886701872, "grad_norm": 1.090098265778865, "learning_rate": 9.444874782901713e-06, "loss": 0.6933, "step": 1232 }, { "epoch": 1.1833893422947672, "grad_norm": 0.89447287793134, "learning_rate": 9.443339651652241e-06, "loss": 0.5613, "step": 1233 }, { "epoch": 1.184349495919347, "grad_norm": 1.125103726432196, "learning_rate": 9.441802525835095e-06, "loss": 0.7413, "step": 1234 }, { "epoch": 1.185309649543927, "grad_norm": 0.724326219665278, "learning_rate": 9.440263406140276e-06, "loss": 0.5724, "step": 1235 }, { "epoch": 1.186269803168507, "grad_norm": 0.7344673980722954, "learning_rate": 9.438722293258678e-06, "loss": 0.5884, "step": 1236 }, { "epoch": 1.1872299567930868, "grad_norm": 0.9265320412850467, "learning_rate": 9.43717918788209e-06, "loss": 0.7694, "step": 1237 }, { "epoch": 1.1881901104176669, "grad_norm": 1.0233231376790144, "learning_rate": 9.435634090703195e-06, "loss": 0.6686, "step": 1238 }, { "epoch": 1.1891502640422467, "grad_norm": 1.1009753553750954, "learning_rate": 9.434087002415571e-06, "loss": 0.6657, "step": 1239 }, { "epoch": 1.1901104176668267, "grad_norm": 0.9601900328299103, "learning_rate": 9.432537923713689e-06, "loss": 0.647, "step": 1240 }, { "epoch": 1.1910705712914067, "grad_norm": 0.8709398628522105, "learning_rate": 9.430986855292916e-06, "loss": 0.5748, "step": 1241 }, { "epoch": 1.1920307249159865, "grad_norm": 1.0089072337961087, "learning_rate": 9.429433797849508e-06, "loss": 0.6799, "step": 1242 }, { "epoch": 1.1929908785405665, "grad_norm": 0.8086841391508848, "learning_rate": 9.427878752080613e-06, "loss": 0.5582, "step": 1243 }, { "epoch": 1.1939510321651463, "grad_norm": 0.8898880546837935, "learning_rate": 9.426321718684282e-06, "loss": 0.5707, "step": 1244 }, { "epoch": 1.1949111857897263, "grad_norm": 0.8203286994381711, "learning_rate": 9.424762698359442e-06, "loss": 0.7034, "step": 1245 }, { "epoch": 1.1958713394143063, "grad_norm": 0.9068614884501264, "learning_rate": 9.423201691805927e-06, "loss": 0.6349, "step": 1246 }, { "epoch": 1.1968314930388861, "grad_norm": 0.7917101635860333, "learning_rate": 9.421638699724453e-06, "loss": 0.552, "step": 1247 }, { "epoch": 1.1977916466634662, "grad_norm": 0.9591651699211111, "learning_rate": 9.420073722816634e-06, "loss": 0.6722, "step": 1248 }, { "epoch": 1.1987518002880462, "grad_norm": 0.94088672023118, "learning_rate": 9.418506761784967e-06, "loss": 0.7014, "step": 1249 }, { "epoch": 1.199711953912626, "grad_norm": 1.2036318456630022, "learning_rate": 9.416937817332846e-06, "loss": 0.7101, "step": 1250 }, { "epoch": 1.200672107537206, "grad_norm": 1.4650572940878548, "learning_rate": 9.415366890164553e-06, "loss": 0.7369, "step": 1251 }, { "epoch": 1.201632261161786, "grad_norm": 1.0352272780264828, "learning_rate": 9.41379398098526e-06, "loss": 0.5666, "step": 1252 }, { "epoch": 1.2025924147863658, "grad_norm": 1.1271950972267066, "learning_rate": 9.41221909050103e-06, "loss": 0.6535, "step": 1253 }, { "epoch": 1.2035525684109458, "grad_norm": 1.2464564636273787, "learning_rate": 9.410642219418816e-06, "loss": 0.5802, "step": 1254 }, { "epoch": 1.2045127220355256, "grad_norm": 1.0618678689122594, "learning_rate": 9.409063368446454e-06, "loss": 0.6645, "step": 1255 }, { "epoch": 1.2054728756601056, "grad_norm": 0.7256807862388882, "learning_rate": 9.407482538292679e-06, "loss": 0.602, "step": 1256 }, { "epoch": 1.2064330292846854, "grad_norm": 0.8082770105815767, "learning_rate": 9.405899729667105e-06, "loss": 0.6877, "step": 1257 }, { "epoch": 1.2073931829092654, "grad_norm": 1.0515964132430884, "learning_rate": 9.404314943280238e-06, "loss": 0.6523, "step": 1258 }, { "epoch": 1.2083533365338455, "grad_norm": 0.8112694384010882, "learning_rate": 9.402728179843472e-06, "loss": 0.6143, "step": 1259 }, { "epoch": 1.2093134901584253, "grad_norm": 0.6460406132784174, "learning_rate": 9.40113944006909e-06, "loss": 0.6402, "step": 1260 }, { "epoch": 1.2102736437830053, "grad_norm": 1.1816980022569024, "learning_rate": 9.39954872467026e-06, "loss": 0.7613, "step": 1261 }, { "epoch": 1.2112337974075853, "grad_norm": 1.1046019431932912, "learning_rate": 9.397956034361033e-06, "loss": 0.6029, "step": 1262 }, { "epoch": 1.212193951032165, "grad_norm": 0.9987259278154479, "learning_rate": 9.396361369856356e-06, "loss": 0.6138, "step": 1263 }, { "epoch": 1.2131541046567451, "grad_norm": 0.8040305219427624, "learning_rate": 9.394764731872054e-06, "loss": 0.6061, "step": 1264 }, { "epoch": 1.2141142582813251, "grad_norm": 1.0021404702279029, "learning_rate": 9.393166121124841e-06, "loss": 0.7061, "step": 1265 }, { "epoch": 1.215074411905905, "grad_norm": 0.906684841286482, "learning_rate": 9.391565538332317e-06, "loss": 0.5602, "step": 1266 }, { "epoch": 1.216034565530485, "grad_norm": 0.953457427308967, "learning_rate": 9.389962984212966e-06, "loss": 0.739, "step": 1267 }, { "epoch": 1.2169947191550647, "grad_norm": 0.8362999333522266, "learning_rate": 9.388358459486156e-06, "loss": 0.6118, "step": 1268 }, { "epoch": 1.2179548727796448, "grad_norm": 0.9406254998765409, "learning_rate": 9.386751964872144e-06, "loss": 0.6298, "step": 1269 }, { "epoch": 1.2189150264042246, "grad_norm": 0.9517253291678472, "learning_rate": 9.385143501092066e-06, "loss": 0.6803, "step": 1270 }, { "epoch": 1.2198751800288046, "grad_norm": 0.9589137349737066, "learning_rate": 9.383533068867947e-06, "loss": 0.602, "step": 1271 }, { "epoch": 1.2208353336533846, "grad_norm": 0.8429643304178753, "learning_rate": 9.381920668922688e-06, "loss": 0.6207, "step": 1272 }, { "epoch": 1.2217954872779644, "grad_norm": 0.7982396200792407, "learning_rate": 9.380306301980082e-06, "loss": 0.6522, "step": 1273 }, { "epoch": 1.2227556409025444, "grad_norm": 1.0422861399455476, "learning_rate": 9.378689968764801e-06, "loss": 0.6959, "step": 1274 }, { "epoch": 1.2237157945271244, "grad_norm": 0.9031021201358421, "learning_rate": 9.377071670002398e-06, "loss": 0.6387, "step": 1275 }, { "epoch": 1.2246759481517042, "grad_norm": 1.038000880630929, "learning_rate": 9.37545140641931e-06, "loss": 0.5319, "step": 1276 }, { "epoch": 1.2256361017762842, "grad_norm": 0.9644227792567388, "learning_rate": 9.373829178742857e-06, "loss": 0.631, "step": 1277 }, { "epoch": 1.2265962554008643, "grad_norm": 0.9578075448084754, "learning_rate": 9.37220498770124e-06, "loss": 0.6609, "step": 1278 }, { "epoch": 1.227556409025444, "grad_norm": 1.040453955733355, "learning_rate": 9.370578834023539e-06, "loss": 0.6307, "step": 1279 }, { "epoch": 1.228516562650024, "grad_norm": 0.7094371121556692, "learning_rate": 9.368950718439718e-06, "loss": 0.6609, "step": 1280 }, { "epoch": 1.2294767162746039, "grad_norm": 1.2701394292166805, "learning_rate": 9.367320641680621e-06, "loss": 0.6384, "step": 1281 }, { "epoch": 1.2304368698991839, "grad_norm": 0.904517082988849, "learning_rate": 9.365688604477974e-06, "loss": 0.6596, "step": 1282 }, { "epoch": 1.2313970235237637, "grad_norm": 0.8492487975840739, "learning_rate": 9.364054607564376e-06, "loss": 0.6466, "step": 1283 }, { "epoch": 1.2323571771483437, "grad_norm": 0.774844240423422, "learning_rate": 9.362418651673317e-06, "loss": 0.7352, "step": 1284 }, { "epoch": 1.2333173307729237, "grad_norm": 1.1183528366161186, "learning_rate": 9.360780737539155e-06, "loss": 0.6583, "step": 1285 }, { "epoch": 1.2342774843975035, "grad_norm": 1.199282502544458, "learning_rate": 9.359140865897136e-06, "loss": 0.6551, "step": 1286 }, { "epoch": 1.2352376380220835, "grad_norm": 0.8428486121336641, "learning_rate": 9.357499037483377e-06, "loss": 0.5954, "step": 1287 }, { "epoch": 1.2361977916466635, "grad_norm": 1.1570305344213019, "learning_rate": 9.35585525303488e-06, "loss": 0.5957, "step": 1288 }, { "epoch": 1.2371579452712433, "grad_norm": 0.914162725314509, "learning_rate": 9.35420951328952e-06, "loss": 0.614, "step": 1289 }, { "epoch": 1.2381180988958234, "grad_norm": 0.8522000082818546, "learning_rate": 9.352561818986056e-06, "loss": 0.6965, "step": 1290 }, { "epoch": 1.2390782525204034, "grad_norm": 0.8910582641835625, "learning_rate": 9.350912170864116e-06, "loss": 0.5836, "step": 1291 }, { "epoch": 1.2400384061449832, "grad_norm": 0.9565144298546774, "learning_rate": 9.349260569664212e-06, "loss": 0.5171, "step": 1292 }, { "epoch": 1.2409985597695632, "grad_norm": 0.868342784353127, "learning_rate": 9.347607016127728e-06, "loss": 0.6407, "step": 1293 }, { "epoch": 1.241958713394143, "grad_norm": 0.7753798046197758, "learning_rate": 9.34595151099693e-06, "loss": 0.5709, "step": 1294 }, { "epoch": 1.242918867018723, "grad_norm": 0.8779295402919994, "learning_rate": 9.344294055014952e-06, "loss": 0.5435, "step": 1295 }, { "epoch": 1.243879020643303, "grad_norm": 0.6701300609378331, "learning_rate": 9.342634648925813e-06, "loss": 0.6435, "step": 1296 }, { "epoch": 1.2448391742678828, "grad_norm": 0.8195024235932132, "learning_rate": 9.340973293474397e-06, "loss": 0.6905, "step": 1297 }, { "epoch": 1.2457993278924628, "grad_norm": 0.9803358785840908, "learning_rate": 9.339309989406474e-06, "loss": 0.6453, "step": 1298 }, { "epoch": 1.2467594815170426, "grad_norm": 0.981536203549909, "learning_rate": 9.337644737468682e-06, "loss": 0.5827, "step": 1299 }, { "epoch": 1.2477196351416227, "grad_norm": 1.2691353602994642, "learning_rate": 9.335977538408533e-06, "loss": 0.767, "step": 1300 }, { "epoch": 1.2486797887662027, "grad_norm": 1.053042218559518, "learning_rate": 9.334308392974417e-06, "loss": 0.6646, "step": 1301 }, { "epoch": 1.2496399423907825, "grad_norm": 1.034425302330388, "learning_rate": 9.332637301915594e-06, "loss": 0.6619, "step": 1302 }, { "epoch": 1.2506000960153625, "grad_norm": 1.4473585895653929, "learning_rate": 9.3309642659822e-06, "loss": 0.718, "step": 1303 }, { "epoch": 1.2515602496399425, "grad_norm": 1.2015993097700595, "learning_rate": 9.329289285925242e-06, "loss": 0.6585, "step": 1304 }, { "epoch": 1.2525204032645223, "grad_norm": 0.9276817685567383, "learning_rate": 9.327612362496601e-06, "loss": 0.5746, "step": 1305 }, { "epoch": 1.2534805568891023, "grad_norm": 0.9617117709737467, "learning_rate": 9.325933496449031e-06, "loss": 0.6339, "step": 1306 }, { "epoch": 1.2544407105136821, "grad_norm": 1.0821898760223272, "learning_rate": 9.324252688536156e-06, "loss": 0.6942, "step": 1307 }, { "epoch": 1.2554008641382621, "grad_norm": 0.792296875127517, "learning_rate": 9.322569939512472e-06, "loss": 0.5923, "step": 1308 }, { "epoch": 1.256361017762842, "grad_norm": 0.828905279910149, "learning_rate": 9.320885250133347e-06, "loss": 0.6835, "step": 1309 }, { "epoch": 1.257321171387422, "grad_norm": 1.249271151771068, "learning_rate": 9.319198621155022e-06, "loss": 0.6876, "step": 1310 }, { "epoch": 1.258281325012002, "grad_norm": 0.9483548212845954, "learning_rate": 9.317510053334604e-06, "loss": 0.5857, "step": 1311 }, { "epoch": 1.2592414786365818, "grad_norm": 1.0941520921260655, "learning_rate": 9.315819547430077e-06, "loss": 0.6045, "step": 1312 }, { "epoch": 1.2602016322611618, "grad_norm": 0.9808704503351134, "learning_rate": 9.314127104200287e-06, "loss": 0.6361, "step": 1313 }, { "epoch": 1.2611617858857418, "grad_norm": 0.9036133548267243, "learning_rate": 9.312432724404957e-06, "loss": 0.5976, "step": 1314 }, { "epoch": 1.2621219395103216, "grad_norm": 0.8830540296876271, "learning_rate": 9.310736408804674e-06, "loss": 0.5746, "step": 1315 }, { "epoch": 1.2630820931349016, "grad_norm": 0.7289134585582693, "learning_rate": 9.309038158160896e-06, "loss": 0.6067, "step": 1316 }, { "epoch": 1.2640422467594816, "grad_norm": 1.0328353644317936, "learning_rate": 9.30733797323595e-06, "loss": 0.6362, "step": 1317 }, { "epoch": 1.2650024003840614, "grad_norm": 0.9385310163317007, "learning_rate": 9.305635854793031e-06, "loss": 0.5555, "step": 1318 }, { "epoch": 1.2659625540086414, "grad_norm": 0.8152919298457398, "learning_rate": 9.3039318035962e-06, "loss": 0.5996, "step": 1319 }, { "epoch": 1.2669227076332212, "grad_norm": 0.731817495040664, "learning_rate": 9.30222582041039e-06, "loss": 0.6584, "step": 1320 }, { "epoch": 1.2678828612578013, "grad_norm": 1.1563383636613107, "learning_rate": 9.300517906001399e-06, "loss": 0.6217, "step": 1321 }, { "epoch": 1.268843014882381, "grad_norm": 1.1404617529356391, "learning_rate": 9.298808061135887e-06, "loss": 0.7582, "step": 1322 }, { "epoch": 1.269803168506961, "grad_norm": 0.8286602382678154, "learning_rate": 9.29709628658139e-06, "loss": 0.6436, "step": 1323 }, { "epoch": 1.270763322131541, "grad_norm": 0.8315776649856551, "learning_rate": 9.295382583106303e-06, "loss": 0.6211, "step": 1324 }, { "epoch": 1.2717234757561209, "grad_norm": 0.8575307218669823, "learning_rate": 9.29366695147989e-06, "loss": 0.529, "step": 1325 }, { "epoch": 1.272683629380701, "grad_norm": 1.0674676437720905, "learning_rate": 9.291949392472276e-06, "loss": 0.5605, "step": 1326 }, { "epoch": 1.273643783005281, "grad_norm": 1.0514222768108399, "learning_rate": 9.290229906854462e-06, "loss": 0.7326, "step": 1327 }, { "epoch": 1.2746039366298607, "grad_norm": 1.367273870048214, "learning_rate": 9.288508495398301e-06, "loss": 0.6779, "step": 1328 }, { "epoch": 1.2755640902544407, "grad_norm": 0.9840113974309325, "learning_rate": 9.28678515887652e-06, "loss": 0.688, "step": 1329 }, { "epoch": 1.2765242438790207, "grad_norm": 1.0473776856273445, "learning_rate": 9.285059898062701e-06, "loss": 0.6266, "step": 1330 }, { "epoch": 1.2774843975036005, "grad_norm": 0.9460657344690451, "learning_rate": 9.2833327137313e-06, "loss": 0.7224, "step": 1331 }, { "epoch": 1.2784445511281806, "grad_norm": 1.0875670300795142, "learning_rate": 9.281603606657632e-06, "loss": 0.6463, "step": 1332 }, { "epoch": 1.2794047047527606, "grad_norm": 1.0935833664042076, "learning_rate": 9.27987257761787e-06, "loss": 0.6915, "step": 1333 }, { "epoch": 1.2803648583773404, "grad_norm": 1.4130632172695041, "learning_rate": 9.27813962738906e-06, "loss": 0.7278, "step": 1334 }, { "epoch": 1.2813250120019202, "grad_norm": 1.2821098215853464, "learning_rate": 9.2764047567491e-06, "loss": 0.6803, "step": 1335 }, { "epoch": 1.2822851656265002, "grad_norm": 0.8007355438567274, "learning_rate": 9.274667966476757e-06, "loss": 0.5829, "step": 1336 }, { "epoch": 1.2832453192510802, "grad_norm": 1.0618502647336894, "learning_rate": 9.272929257351659e-06, "loss": 0.5955, "step": 1337 }, { "epoch": 1.28420547287566, "grad_norm": 0.7335032895481657, "learning_rate": 9.27118863015429e-06, "loss": 0.5658, "step": 1338 }, { "epoch": 1.28516562650024, "grad_norm": 0.8812622719896972, "learning_rate": 9.269446085666002e-06, "loss": 0.7627, "step": 1339 }, { "epoch": 1.28612578012482, "grad_norm": 1.1791477320179078, "learning_rate": 9.267701624669006e-06, "loss": 0.6364, "step": 1340 }, { "epoch": 1.2870859337493998, "grad_norm": 1.1909813263621865, "learning_rate": 9.265955247946369e-06, "loss": 0.6504, "step": 1341 }, { "epoch": 1.2880460873739799, "grad_norm": 0.6054452221262864, "learning_rate": 9.26420695628202e-06, "loss": 0.6241, "step": 1342 }, { "epoch": 1.2890062409985599, "grad_norm": 0.9296741355814061, "learning_rate": 9.262456750460754e-06, "loss": 0.6507, "step": 1343 }, { "epoch": 1.2899663946231397, "grad_norm": 0.7750390804493564, "learning_rate": 9.260704631268211e-06, "loss": 0.6636, "step": 1344 }, { "epoch": 1.2909265482477197, "grad_norm": 0.8528091640926342, "learning_rate": 9.258950599490906e-06, "loss": 0.6623, "step": 1345 }, { "epoch": 1.2918867018722997, "grad_norm": 1.1408902589755605, "learning_rate": 9.257194655916202e-06, "loss": 0.6239, "step": 1346 }, { "epoch": 1.2928468554968795, "grad_norm": 1.1898506727433453, "learning_rate": 9.255436801332324e-06, "loss": 0.7055, "step": 1347 }, { "epoch": 1.2938070091214593, "grad_norm": 1.2664469725478338, "learning_rate": 9.253677036528356e-06, "loss": 0.5658, "step": 1348 }, { "epoch": 1.2947671627460393, "grad_norm": 0.8250132731866192, "learning_rate": 9.251915362294233e-06, "loss": 0.6409, "step": 1349 }, { "epoch": 1.2957273163706193, "grad_norm": 0.9208331650229149, "learning_rate": 9.250151779420756e-06, "loss": 0.541, "step": 1350 }, { "epoch": 1.2966874699951991, "grad_norm": 0.7757577344481114, "learning_rate": 9.248386288699575e-06, "loss": 0.5985, "step": 1351 }, { "epoch": 1.2976476236197791, "grad_norm": 0.7571202014821614, "learning_rate": 9.246618890923203e-06, "loss": 0.6024, "step": 1352 }, { "epoch": 1.2986077772443592, "grad_norm": 0.8374972320528836, "learning_rate": 9.244849586885005e-06, "loss": 0.6374, "step": 1353 }, { "epoch": 1.299567930868939, "grad_norm": 0.7333364869914581, "learning_rate": 9.243078377379202e-06, "loss": 0.6555, "step": 1354 }, { "epoch": 1.300528084493519, "grad_norm": 0.794598447146553, "learning_rate": 9.24130526320087e-06, "loss": 0.6042, "step": 1355 }, { "epoch": 1.301488238118099, "grad_norm": 0.8390943627449102, "learning_rate": 9.239530245145945e-06, "loss": 0.5909, "step": 1356 }, { "epoch": 1.3024483917426788, "grad_norm": 0.6892954604292173, "learning_rate": 9.237753324011208e-06, "loss": 0.648, "step": 1357 }, { "epoch": 1.3034085453672588, "grad_norm": 1.0142756831259443, "learning_rate": 9.235974500594305e-06, "loss": 0.735, "step": 1358 }, { "epoch": 1.3043686989918388, "grad_norm": 1.0441391680135295, "learning_rate": 9.234193775693728e-06, "loss": 0.6194, "step": 1359 }, { "epoch": 1.3053288526164186, "grad_norm": 0.9842669581896375, "learning_rate": 9.232411150108826e-06, "loss": 0.5661, "step": 1360 }, { "epoch": 1.3062890062409986, "grad_norm": 0.8476790909647105, "learning_rate": 9.2306266246398e-06, "loss": 0.5939, "step": 1361 }, { "epoch": 1.3072491598655784, "grad_norm": 1.0052426009671198, "learning_rate": 9.228840200087708e-06, "loss": 0.6344, "step": 1362 }, { "epoch": 1.3082093134901585, "grad_norm": 0.8973270632848723, "learning_rate": 9.227051877254452e-06, "loss": 0.6379, "step": 1363 }, { "epoch": 1.3091694671147383, "grad_norm": 1.1511144902799122, "learning_rate": 9.225261656942792e-06, "loss": 0.665, "step": 1364 }, { "epoch": 1.3101296207393183, "grad_norm": 1.1881572962539777, "learning_rate": 9.22346953995634e-06, "loss": 0.6388, "step": 1365 }, { "epoch": 1.3110897743638983, "grad_norm": 0.9721378632755631, "learning_rate": 9.22167552709956e-06, "loss": 0.6249, "step": 1366 }, { "epoch": 1.312049927988478, "grad_norm": 1.0122712259057562, "learning_rate": 9.219879619177761e-06, "loss": 0.6951, "step": 1367 }, { "epoch": 1.313010081613058, "grad_norm": 0.9937419353627702, "learning_rate": 9.21808181699711e-06, "loss": 0.592, "step": 1368 }, { "epoch": 1.3139702352376381, "grad_norm": 0.7546057001005384, "learning_rate": 9.21628212136462e-06, "loss": 0.601, "step": 1369 }, { "epoch": 1.314930388862218, "grad_norm": 0.8329567194803621, "learning_rate": 9.214480533088154e-06, "loss": 0.7268, "step": 1370 }, { "epoch": 1.315890542486798, "grad_norm": 1.0480323495039077, "learning_rate": 9.212677052976428e-06, "loss": 0.6384, "step": 1371 }, { "epoch": 1.316850696111378, "grad_norm": 1.0151291641540212, "learning_rate": 9.210871681839006e-06, "loss": 0.7474, "step": 1372 }, { "epoch": 1.3178108497359577, "grad_norm": 1.1995580988098136, "learning_rate": 9.209064420486296e-06, "loss": 0.5595, "step": 1373 }, { "epoch": 1.3187710033605378, "grad_norm": 0.9806726383654879, "learning_rate": 9.207255269729559e-06, "loss": 0.6329, "step": 1374 }, { "epoch": 1.3197311569851176, "grad_norm": 0.8907078429290524, "learning_rate": 9.205444230380904e-06, "loss": 0.6806, "step": 1375 }, { "epoch": 1.3206913106096976, "grad_norm": 1.1654055001559578, "learning_rate": 9.203631303253292e-06, "loss": 0.6589, "step": 1376 }, { "epoch": 1.3216514642342774, "grad_norm": 0.8257781665432308, "learning_rate": 9.201816489160518e-06, "loss": 0.5637, "step": 1377 }, { "epoch": 1.3226116178588574, "grad_norm": 1.0372096632835965, "learning_rate": 9.199999788917238e-06, "loss": 0.666, "step": 1378 }, { "epoch": 1.3235717714834374, "grad_norm": 1.15265404548876, "learning_rate": 9.198181203338949e-06, "loss": 0.6254, "step": 1379 }, { "epoch": 1.3245319251080172, "grad_norm": 0.9128777275566708, "learning_rate": 9.196360733241992e-06, "loss": 0.6207, "step": 1380 }, { "epoch": 1.3254920787325972, "grad_norm": 0.80723607927469, "learning_rate": 9.194538379443561e-06, "loss": 0.6218, "step": 1381 }, { "epoch": 1.3264522323571772, "grad_norm": 1.1367990943481738, "learning_rate": 9.192714142761687e-06, "loss": 0.6644, "step": 1382 }, { "epoch": 1.327412385981757, "grad_norm": 1.059726619089504, "learning_rate": 9.190888024015252e-06, "loss": 0.7365, "step": 1383 }, { "epoch": 1.328372539606337, "grad_norm": 0.954503895062622, "learning_rate": 9.189060024023981e-06, "loss": 0.6336, "step": 1384 }, { "epoch": 1.329332693230917, "grad_norm": 1.2090363137789533, "learning_rate": 9.187230143608445e-06, "loss": 0.6822, "step": 1385 }, { "epoch": 1.3302928468554969, "grad_norm": 0.8205507838209184, "learning_rate": 9.185398383590056e-06, "loss": 0.646, "step": 1386 }, { "epoch": 1.3312530004800769, "grad_norm": 1.0332799972333997, "learning_rate": 9.18356474479107e-06, "loss": 0.5979, "step": 1387 }, { "epoch": 1.3322131541046567, "grad_norm": 0.8853141801527188, "learning_rate": 9.181729228034593e-06, "loss": 0.5973, "step": 1388 }, { "epoch": 1.3331733077292367, "grad_norm": 0.9931763691757564, "learning_rate": 9.179891834144565e-06, "loss": 0.5594, "step": 1389 }, { "epoch": 1.3341334613538165, "grad_norm": 1.0141444018678256, "learning_rate": 9.178052563945773e-06, "loss": 0.5852, "step": 1390 }, { "epoch": 1.3350936149783965, "grad_norm": 0.9382250617086866, "learning_rate": 9.176211418263845e-06, "loss": 0.573, "step": 1391 }, { "epoch": 1.3360537686029765, "grad_norm": 1.0577449536454064, "learning_rate": 9.174368397925254e-06, "loss": 0.6694, "step": 1392 }, { "epoch": 1.3370139222275563, "grad_norm": 1.1780959298737226, "learning_rate": 9.17252350375731e-06, "loss": 0.5794, "step": 1393 }, { "epoch": 1.3379740758521363, "grad_norm": 1.1133191933370012, "learning_rate": 9.170676736588167e-06, "loss": 0.5684, "step": 1394 }, { "epoch": 1.3389342294767164, "grad_norm": 1.2504545979140467, "learning_rate": 9.168828097246819e-06, "loss": 0.5671, "step": 1395 }, { "epoch": 1.3398943831012962, "grad_norm": 0.8216856310565852, "learning_rate": 9.1669775865631e-06, "loss": 0.7068, "step": 1396 }, { "epoch": 1.3408545367258762, "grad_norm": 1.006005338242595, "learning_rate": 9.165125205367685e-06, "loss": 0.7053, "step": 1397 }, { "epoch": 1.3418146903504562, "grad_norm": 1.0473157580813148, "learning_rate": 9.16327095449209e-06, "loss": 0.6148, "step": 1398 }, { "epoch": 1.342774843975036, "grad_norm": 1.0355353170104902, "learning_rate": 9.161414834768662e-06, "loss": 0.7913, "step": 1399 }, { "epoch": 1.343734997599616, "grad_norm": 1.045385236381784, "learning_rate": 9.159556847030602e-06, "loss": 0.6479, "step": 1400 }, { "epoch": 1.3446951512241958, "grad_norm": 1.2020664844787778, "learning_rate": 9.157696992111935e-06, "loss": 0.6217, "step": 1401 }, { "epoch": 1.3456553048487758, "grad_norm": 0.8989623828590095, "learning_rate": 9.15583527084753e-06, "loss": 0.6136, "step": 1402 }, { "epoch": 1.3466154584733556, "grad_norm": 0.8884164087119466, "learning_rate": 9.153971684073098e-06, "loss": 0.6385, "step": 1403 }, { "epoch": 1.3475756120979356, "grad_norm": 1.2005187667861983, "learning_rate": 9.15210623262518e-06, "loss": 0.6456, "step": 1404 }, { "epoch": 1.3485357657225157, "grad_norm": 0.9069057872577224, "learning_rate": 9.150238917341156e-06, "loss": 0.556, "step": 1405 }, { "epoch": 1.3494959193470955, "grad_norm": 0.901400172434011, "learning_rate": 9.148369739059247e-06, "loss": 0.6603, "step": 1406 }, { "epoch": 1.3504560729716755, "grad_norm": 1.0380031102346752, "learning_rate": 9.146498698618507e-06, "loss": 0.5941, "step": 1407 }, { "epoch": 1.3514162265962555, "grad_norm": 0.9205774573739466, "learning_rate": 9.14462579685882e-06, "loss": 0.5538, "step": 1408 }, { "epoch": 1.3523763802208353, "grad_norm": 0.7792268621586471, "learning_rate": 9.142751034620923e-06, "loss": 0.5899, "step": 1409 }, { "epoch": 1.3533365338454153, "grad_norm": 0.8126591166325283, "learning_rate": 9.140874412746365e-06, "loss": 0.6455, "step": 1410 }, { "epoch": 1.3542966874699953, "grad_norm": 0.6895843821665736, "learning_rate": 9.138995932077549e-06, "loss": 0.5935, "step": 1411 }, { "epoch": 1.3552568410945751, "grad_norm": 0.8810352515868705, "learning_rate": 9.137115593457702e-06, "loss": 0.6013, "step": 1412 }, { "epoch": 1.3562169947191551, "grad_norm": 0.8457435165010061, "learning_rate": 9.135233397730889e-06, "loss": 0.6886, "step": 1413 }, { "epoch": 1.357177148343735, "grad_norm": 1.068736292495337, "learning_rate": 9.133349345742007e-06, "loss": 0.5747, "step": 1414 }, { "epoch": 1.358137301968315, "grad_norm": 0.9469724558405831, "learning_rate": 9.131463438336785e-06, "loss": 0.638, "step": 1415 }, { "epoch": 1.3590974555928947, "grad_norm": 1.0519817408730467, "learning_rate": 9.12957567636179e-06, "loss": 0.6694, "step": 1416 }, { "epoch": 1.3600576092174748, "grad_norm": 0.9067467422574518, "learning_rate": 9.127686060664414e-06, "loss": 0.5847, "step": 1417 }, { "epoch": 1.3610177628420548, "grad_norm": 0.8814258645946218, "learning_rate": 9.125794592092891e-06, "loss": 0.6323, "step": 1418 }, { "epoch": 1.3619779164666346, "grad_norm": 0.8870532145594715, "learning_rate": 9.123901271496276e-06, "loss": 0.643, "step": 1419 }, { "epoch": 1.3629380700912146, "grad_norm": 0.9766259884210317, "learning_rate": 9.122006099724463e-06, "loss": 0.5925, "step": 1420 }, { "epoch": 1.3638982237157946, "grad_norm": 0.9383406312378868, "learning_rate": 9.120109077628175e-06, "loss": 0.6413, "step": 1421 }, { "epoch": 1.3648583773403744, "grad_norm": 1.0271936511187176, "learning_rate": 9.118210206058961e-06, "loss": 0.6881, "step": 1422 }, { "epoch": 1.3658185309649544, "grad_norm": 1.136451754590851, "learning_rate": 9.116309485869207e-06, "loss": 0.6601, "step": 1423 }, { "epoch": 1.3667786845895344, "grad_norm": 1.002403175255989, "learning_rate": 9.114406917912129e-06, "loss": 0.5782, "step": 1424 }, { "epoch": 1.3677388382141142, "grad_norm": 0.9987901437401137, "learning_rate": 9.112502503041763e-06, "loss": 0.6145, "step": 1425 }, { "epoch": 1.3686989918386943, "grad_norm": 0.7207819966435356, "learning_rate": 9.110596242112985e-06, "loss": 0.5453, "step": 1426 }, { "epoch": 1.369659145463274, "grad_norm": 1.1805516899547799, "learning_rate": 9.108688135981494e-06, "loss": 0.6795, "step": 1427 }, { "epoch": 1.370619299087854, "grad_norm": 1.0715538225099446, "learning_rate": 9.106778185503816e-06, "loss": 0.735, "step": 1428 }, { "epoch": 1.3715794527124339, "grad_norm": 1.3937996004752053, "learning_rate": 9.10486639153731e-06, "loss": 0.7152, "step": 1429 }, { "epoch": 1.3725396063370139, "grad_norm": 1.0237690635685353, "learning_rate": 9.102952754940162e-06, "loss": 0.592, "step": 1430 }, { "epoch": 1.373499759961594, "grad_norm": 0.8008443871579342, "learning_rate": 9.101037276571378e-06, "loss": 0.6194, "step": 1431 }, { "epoch": 1.3744599135861737, "grad_norm": 0.7835313156105745, "learning_rate": 9.099119957290798e-06, "loss": 0.6064, "step": 1432 }, { "epoch": 1.3754200672107537, "grad_norm": 0.7354649317904994, "learning_rate": 9.097200797959089e-06, "loss": 0.7287, "step": 1433 }, { "epoch": 1.3763802208353337, "grad_norm": 1.142771192758727, "learning_rate": 9.095279799437737e-06, "loss": 0.7437, "step": 1434 }, { "epoch": 1.3773403744599135, "grad_norm": 1.3047071232444747, "learning_rate": 9.09335696258906e-06, "loss": 0.5978, "step": 1435 }, { "epoch": 1.3783005280844935, "grad_norm": 0.8188529943710545, "learning_rate": 9.091432288276196e-06, "loss": 0.555, "step": 1436 }, { "epoch": 1.3792606817090736, "grad_norm": 0.8600707195931213, "learning_rate": 9.089505777363114e-06, "loss": 0.554, "step": 1437 }, { "epoch": 1.3802208353336534, "grad_norm": 0.7289063312206355, "learning_rate": 9.087577430714603e-06, "loss": 0.744, "step": 1438 }, { "epoch": 1.3811809889582334, "grad_norm": 1.0921082132800193, "learning_rate": 9.085647249196275e-06, "loss": 0.6392, "step": 1439 }, { "epoch": 1.3821411425828132, "grad_norm": 0.8184954324346673, "learning_rate": 9.083715233674572e-06, "loss": 0.5326, "step": 1440 }, { "epoch": 1.3831012962073932, "grad_norm": 0.8208166249051642, "learning_rate": 9.081781385016754e-06, "loss": 0.6762, "step": 1441 }, { "epoch": 1.384061449831973, "grad_norm": 0.8065528768148795, "learning_rate": 9.079845704090903e-06, "loss": 0.5346, "step": 1442 }, { "epoch": 1.385021603456553, "grad_norm": 0.7475627951963318, "learning_rate": 9.077908191765925e-06, "loss": 0.7119, "step": 1443 }, { "epoch": 1.385981757081133, "grad_norm": 1.0307973188196085, "learning_rate": 9.075968848911553e-06, "loss": 0.5785, "step": 1444 }, { "epoch": 1.3869419107057128, "grad_norm": 0.6680080220823404, "learning_rate": 9.074027676398333e-06, "loss": 0.6233, "step": 1445 }, { "epoch": 1.3879020643302928, "grad_norm": 0.857850012477854, "learning_rate": 9.07208467509764e-06, "loss": 0.5397, "step": 1446 }, { "epoch": 1.3888622179548729, "grad_norm": 0.9584778443460951, "learning_rate": 9.07013984588166e-06, "loss": 0.6262, "step": 1447 }, { "epoch": 1.3898223715794527, "grad_norm": 0.9642367081941637, "learning_rate": 9.068193189623412e-06, "loss": 0.7141, "step": 1448 }, { "epoch": 1.3907825252040327, "grad_norm": 1.0453633726935205, "learning_rate": 9.06624470719673e-06, "loss": 0.6107, "step": 1449 }, { "epoch": 1.3917426788286127, "grad_norm": 0.7639472923797955, "learning_rate": 9.064294399476265e-06, "loss": 0.5219, "step": 1450 }, { "epoch": 1.3927028324531925, "grad_norm": 1.1853409974320068, "learning_rate": 9.062342267337487e-06, "loss": 0.6751, "step": 1451 }, { "epoch": 1.3936629860777725, "grad_norm": 0.9446248120391193, "learning_rate": 9.060388311656689e-06, "loss": 0.5574, "step": 1452 }, { "epoch": 1.3946231397023523, "grad_norm": 0.6340149923871167, "learning_rate": 9.058432533310982e-06, "loss": 0.575, "step": 1453 }, { "epoch": 1.3955832933269323, "grad_norm": 1.1654738964718616, "learning_rate": 9.056474933178293e-06, "loss": 0.7492, "step": 1454 }, { "epoch": 1.3965434469515121, "grad_norm": 1.4239826522192331, "learning_rate": 9.054515512137367e-06, "loss": 0.6138, "step": 1455 }, { "epoch": 1.3975036005760921, "grad_norm": 1.1531531389503498, "learning_rate": 9.052554271067771e-06, "loss": 0.7146, "step": 1456 }, { "epoch": 1.3984637542006721, "grad_norm": 0.954332312121783, "learning_rate": 9.05059121084988e-06, "loss": 0.6574, "step": 1457 }, { "epoch": 1.399423907825252, "grad_norm": 0.9722937519679417, "learning_rate": 9.048626332364891e-06, "loss": 0.6338, "step": 1458 }, { "epoch": 1.400384061449832, "grad_norm": 0.7833890274001232, "learning_rate": 9.046659636494821e-06, "loss": 0.6248, "step": 1459 }, { "epoch": 1.401344215074412, "grad_norm": 0.8178548099088, "learning_rate": 9.044691124122496e-06, "loss": 0.5179, "step": 1460 }, { "epoch": 1.4023043686989918, "grad_norm": 0.8872623825271978, "learning_rate": 9.042720796131561e-06, "loss": 0.5995, "step": 1461 }, { "epoch": 1.4032645223235718, "grad_norm": 1.1050398071967948, "learning_rate": 9.040748653406475e-06, "loss": 0.7611, "step": 1462 }, { "epoch": 1.4042246759481518, "grad_norm": 1.2376504666813695, "learning_rate": 9.038774696832511e-06, "loss": 0.6661, "step": 1463 }, { "epoch": 1.4051848295727316, "grad_norm": 1.2584153711245483, "learning_rate": 9.036798927295757e-06, "loss": 0.6754, "step": 1464 }, { "epoch": 1.4061449831973116, "grad_norm": 1.0440784383189627, "learning_rate": 9.034821345683117e-06, "loss": 0.6131, "step": 1465 }, { "epoch": 1.4071051368218914, "grad_norm": 0.9149174875902278, "learning_rate": 9.032841952882306e-06, "loss": 0.583, "step": 1466 }, { "epoch": 1.4080652904464714, "grad_norm": 1.2008587408335687, "learning_rate": 9.030860749781848e-06, "loss": 0.6525, "step": 1467 }, { "epoch": 1.4090254440710512, "grad_norm": 0.6558486078081912, "learning_rate": 9.028877737271089e-06, "loss": 0.6168, "step": 1468 }, { "epoch": 1.4099855976956313, "grad_norm": 1.0276386458465896, "learning_rate": 9.02689291624018e-06, "loss": 0.6457, "step": 1469 }, { "epoch": 1.4109457513202113, "grad_norm": 1.0044822640912285, "learning_rate": 9.024906287580085e-06, "loss": 0.6459, "step": 1470 }, { "epoch": 1.411905904944791, "grad_norm": 0.696617706789436, "learning_rate": 9.022917852182582e-06, "loss": 0.5964, "step": 1471 }, { "epoch": 1.412866058569371, "grad_norm": 0.6405456733076281, "learning_rate": 9.020927610940256e-06, "loss": 0.5752, "step": 1472 }, { "epoch": 1.413826212193951, "grad_norm": 0.8515035515370114, "learning_rate": 9.018935564746509e-06, "loss": 0.6351, "step": 1473 }, { "epoch": 1.414786365818531, "grad_norm": 0.9901959764148056, "learning_rate": 9.016941714495546e-06, "loss": 0.5853, "step": 1474 }, { "epoch": 1.415746519443111, "grad_norm": 0.7176690439192978, "learning_rate": 9.014946061082385e-06, "loss": 0.6029, "step": 1475 }, { "epoch": 1.416706673067691, "grad_norm": 0.8343776234921784, "learning_rate": 9.012948605402856e-06, "loss": 0.6488, "step": 1476 }, { "epoch": 1.4176668266922707, "grad_norm": 0.9431049986213873, "learning_rate": 9.010949348353595e-06, "loss": 0.6526, "step": 1477 }, { "epoch": 1.4186269803168507, "grad_norm": 0.9869132162836172, "learning_rate": 9.008948290832046e-06, "loss": 0.6472, "step": 1478 }, { "epoch": 1.4195871339414305, "grad_norm": 0.6315220281277975, "learning_rate": 9.006945433736461e-06, "loss": 0.6623, "step": 1479 }, { "epoch": 1.4205472875660106, "grad_norm": 1.083500156437386, "learning_rate": 9.004940777965903e-06, "loss": 0.6943, "step": 1480 }, { "epoch": 1.4215074411905904, "grad_norm": 0.8799595545680973, "learning_rate": 9.002934324420242e-06, "loss": 0.605, "step": 1481 }, { "epoch": 1.4224675948151704, "grad_norm": 0.7790008972329887, "learning_rate": 9.000926074000149e-06, "loss": 0.6363, "step": 1482 }, { "epoch": 1.4234277484397504, "grad_norm": 0.8935383496731113, "learning_rate": 8.998916027607111e-06, "loss": 0.6597, "step": 1483 }, { "epoch": 1.4243879020643302, "grad_norm": 0.6807934214729915, "learning_rate": 8.996904186143414e-06, "loss": 0.5417, "step": 1484 }, { "epoch": 1.4253480556889102, "grad_norm": 1.007757324186503, "learning_rate": 8.994890550512152e-06, "loss": 0.691, "step": 1485 }, { "epoch": 1.4263082093134902, "grad_norm": 0.7821026858807716, "learning_rate": 8.992875121617224e-06, "loss": 0.688, "step": 1486 }, { "epoch": 1.42726836293807, "grad_norm": 1.0443981990586368, "learning_rate": 8.990857900363337e-06, "loss": 0.5489, "step": 1487 }, { "epoch": 1.42822851656265, "grad_norm": 0.8219507670836994, "learning_rate": 8.988838887655997e-06, "loss": 0.5789, "step": 1488 }, { "epoch": 1.42918867018723, "grad_norm": 0.7655362011879656, "learning_rate": 8.98681808440152e-06, "loss": 0.6662, "step": 1489 }, { "epoch": 1.4301488238118099, "grad_norm": 0.9531858687429643, "learning_rate": 8.984795491507021e-06, "loss": 0.5582, "step": 1490 }, { "epoch": 1.4311089774363899, "grad_norm": 0.6016365709377223, "learning_rate": 8.982771109880421e-06, "loss": 0.5289, "step": 1491 }, { "epoch": 1.43206913106097, "grad_norm": 0.8779220319870115, "learning_rate": 8.980744940430443e-06, "loss": 0.5987, "step": 1492 }, { "epoch": 1.4330292846855497, "grad_norm": 0.7406991747264607, "learning_rate": 8.978716984066613e-06, "loss": 0.7057, "step": 1493 }, { "epoch": 1.4339894383101295, "grad_norm": 0.8486552493878786, "learning_rate": 8.976687241699258e-06, "loss": 0.6996, "step": 1494 }, { "epoch": 1.4349495919347095, "grad_norm": 1.258298691352076, "learning_rate": 8.97465571423951e-06, "loss": 0.6579, "step": 1495 }, { "epoch": 1.4359097455592895, "grad_norm": 1.1168753305421455, "learning_rate": 8.972622402599298e-06, "loss": 0.5137, "step": 1496 }, { "epoch": 1.4368698991838693, "grad_norm": 0.8705581713705457, "learning_rate": 8.970587307691357e-06, "loss": 0.6282, "step": 1497 }, { "epoch": 1.4378300528084493, "grad_norm": 0.8996830383798771, "learning_rate": 8.968550430429212e-06, "loss": 0.6516, "step": 1498 }, { "epoch": 1.4387902064330294, "grad_norm": 1.5205341717211724, "learning_rate": 8.966511771727205e-06, "loss": 0.5685, "step": 1499 }, { "epoch": 1.4397503600576091, "grad_norm": 0.9227707052717918, "learning_rate": 8.96447133250046e-06, "loss": 0.6703, "step": 1500 }, { "epoch": 1.4407105136821892, "grad_norm": 0.9578166145064752, "learning_rate": 8.962429113664913e-06, "loss": 0.6366, "step": 1501 }, { "epoch": 1.4416706673067692, "grad_norm": 1.0802246607210446, "learning_rate": 8.960385116137293e-06, "loss": 0.7173, "step": 1502 }, { "epoch": 1.442630820931349, "grad_norm": 1.1441257062774577, "learning_rate": 8.958339340835128e-06, "loss": 0.6413, "step": 1503 }, { "epoch": 1.443590974555929, "grad_norm": 0.8890853575640166, "learning_rate": 8.956291788676746e-06, "loss": 0.6362, "step": 1504 }, { "epoch": 1.444551128180509, "grad_norm": 1.0789278937094944, "learning_rate": 8.95424246058127e-06, "loss": 0.6849, "step": 1505 }, { "epoch": 1.4455112818050888, "grad_norm": 0.8911349477731945, "learning_rate": 8.952191357468621e-06, "loss": 0.5714, "step": 1506 }, { "epoch": 1.4464714354296688, "grad_norm": 0.928480191549083, "learning_rate": 8.950138480259519e-06, "loss": 0.6621, "step": 1507 }, { "epoch": 1.4474315890542486, "grad_norm": 0.8227677309734666, "learning_rate": 8.948083829875478e-06, "loss": 0.5811, "step": 1508 }, { "epoch": 1.4483917426788286, "grad_norm": 1.0402500851274523, "learning_rate": 8.946027407238809e-06, "loss": 0.747, "step": 1509 }, { "epoch": 1.4493518963034084, "grad_norm": 1.246298317275389, "learning_rate": 8.943969213272616e-06, "loss": 0.7194, "step": 1510 }, { "epoch": 1.4503120499279885, "grad_norm": 0.9034150601588135, "learning_rate": 8.941909248900804e-06, "loss": 0.5435, "step": 1511 }, { "epoch": 1.4512722035525685, "grad_norm": 0.5458197336369576, "learning_rate": 8.939847515048066e-06, "loss": 0.6834, "step": 1512 }, { "epoch": 1.4522323571771483, "grad_norm": 1.1007416510853743, "learning_rate": 8.937784012639892e-06, "loss": 0.5753, "step": 1513 }, { "epoch": 1.4531925108017283, "grad_norm": 0.8969851309935141, "learning_rate": 8.935718742602568e-06, "loss": 0.6219, "step": 1514 }, { "epoch": 1.4541526644263083, "grad_norm": 0.8039487005218117, "learning_rate": 8.933651705863172e-06, "loss": 0.687, "step": 1515 }, { "epoch": 1.455112818050888, "grad_norm": 0.9217611468536476, "learning_rate": 8.931582903349571e-06, "loss": 0.6946, "step": 1516 }, { "epoch": 1.4560729716754681, "grad_norm": 0.9270316704038395, "learning_rate": 8.929512335990434e-06, "loss": 0.6029, "step": 1517 }, { "epoch": 1.4570331253000481, "grad_norm": 1.0744928869387713, "learning_rate": 8.92744000471521e-06, "loss": 0.6338, "step": 1518 }, { "epoch": 1.457993278924628, "grad_norm": 0.8834021674297483, "learning_rate": 8.925365910454153e-06, "loss": 0.7613, "step": 1519 }, { "epoch": 1.458953432549208, "grad_norm": 0.8953605907605613, "learning_rate": 8.923290054138298e-06, "loss": 0.5237, "step": 1520 }, { "epoch": 1.4599135861737877, "grad_norm": 0.9469411314290676, "learning_rate": 8.921212436699476e-06, "loss": 0.6431, "step": 1521 }, { "epoch": 1.4608737397983678, "grad_norm": 0.7364686162999547, "learning_rate": 8.919133059070307e-06, "loss": 0.575, "step": 1522 }, { "epoch": 1.4618338934229476, "grad_norm": 0.8767165640764638, "learning_rate": 8.9170519221842e-06, "loss": 0.5933, "step": 1523 }, { "epoch": 1.4627940470475276, "grad_norm": 1.2994847736525335, "learning_rate": 8.914969026975354e-06, "loss": 0.668, "step": 1524 }, { "epoch": 1.4637542006721076, "grad_norm": 0.7637249247395503, "learning_rate": 8.912884374378763e-06, "loss": 0.6212, "step": 1525 }, { "epoch": 1.4647143542966874, "grad_norm": 0.9080249321919854, "learning_rate": 8.910797965330202e-06, "loss": 0.6219, "step": 1526 }, { "epoch": 1.4656745079212674, "grad_norm": 0.99515408936552, "learning_rate": 8.908709800766237e-06, "loss": 0.7288, "step": 1527 }, { "epoch": 1.4666346615458474, "grad_norm": 1.0781081497172835, "learning_rate": 8.906619881624225e-06, "loss": 0.6379, "step": 1528 }, { "epoch": 1.4675948151704272, "grad_norm": 0.8403135519985371, "learning_rate": 8.90452820884231e-06, "loss": 0.617, "step": 1529 }, { "epoch": 1.4685549687950072, "grad_norm": 0.7714451680721026, "learning_rate": 8.902434783359417e-06, "loss": 0.6864, "step": 1530 }, { "epoch": 1.4695151224195873, "grad_norm": 1.0594991718367048, "learning_rate": 8.900339606115264e-06, "loss": 0.6854, "step": 1531 }, { "epoch": 1.470475276044167, "grad_norm": 0.9886179264798742, "learning_rate": 8.898242678050354e-06, "loss": 0.7258, "step": 1532 }, { "epoch": 1.471435429668747, "grad_norm": 1.0463485686807235, "learning_rate": 8.896144000105979e-06, "loss": 0.6727, "step": 1533 }, { "epoch": 1.4723955832933269, "grad_norm": 0.7929457773719131, "learning_rate": 8.894043573224207e-06, "loss": 0.6769, "step": 1534 }, { "epoch": 1.4733557369179069, "grad_norm": 1.1891572088914755, "learning_rate": 8.8919413983479e-06, "loss": 0.6483, "step": 1535 }, { "epoch": 1.4743158905424867, "grad_norm": 1.0050692403535442, "learning_rate": 8.889837476420703e-06, "loss": 0.4912, "step": 1536 }, { "epoch": 1.4752760441670667, "grad_norm": 0.7991468916988426, "learning_rate": 8.887731808387043e-06, "loss": 0.5311, "step": 1537 }, { "epoch": 1.4762361977916467, "grad_norm": 0.8328708607626627, "learning_rate": 8.885624395192131e-06, "loss": 0.5978, "step": 1538 }, { "epoch": 1.4771963514162265, "grad_norm": 0.8247309986352224, "learning_rate": 8.883515237781964e-06, "loss": 0.5652, "step": 1539 }, { "epoch": 1.4781565050408065, "grad_norm": 0.7136693135685164, "learning_rate": 8.881404337103316e-06, "loss": 0.7086, "step": 1540 }, { "epoch": 1.4791166586653866, "grad_norm": 1.2920557594883422, "learning_rate": 8.879291694103754e-06, "loss": 0.6611, "step": 1541 }, { "epoch": 1.4800768122899663, "grad_norm": 0.9764726793788544, "learning_rate": 8.877177309731618e-06, "loss": 0.6532, "step": 1542 }, { "epoch": 1.4810369659145464, "grad_norm": 1.228497063992622, "learning_rate": 8.875061184936031e-06, "loss": 0.6588, "step": 1543 }, { "epoch": 1.4819971195391264, "grad_norm": 0.6296761078247097, "learning_rate": 8.872943320666901e-06, "loss": 0.7106, "step": 1544 }, { "epoch": 1.4829572731637062, "grad_norm": 1.2780548781552477, "learning_rate": 8.870823717874912e-06, "loss": 0.681, "step": 1545 }, { "epoch": 1.4839174267882862, "grad_norm": 1.012456802685574, "learning_rate": 8.868702377511535e-06, "loss": 0.6492, "step": 1546 }, { "epoch": 1.484877580412866, "grad_norm": 0.9742705519016123, "learning_rate": 8.866579300529014e-06, "loss": 0.6671, "step": 1547 }, { "epoch": 1.485837734037446, "grad_norm": 0.8847052305716459, "learning_rate": 8.864454487880377e-06, "loss": 0.6589, "step": 1548 }, { "epoch": 1.4867978876620258, "grad_norm": 0.9759993887466291, "learning_rate": 8.862327940519429e-06, "loss": 0.7564, "step": 1549 }, { "epoch": 1.4877580412866058, "grad_norm": 0.759688503251527, "learning_rate": 8.860199659400754e-06, "loss": 0.6111, "step": 1550 }, { "epoch": 1.4887181949111858, "grad_norm": 1.2042592196678417, "learning_rate": 8.858069645479715e-06, "loss": 0.7262, "step": 1551 }, { "epoch": 1.4896783485357656, "grad_norm": 0.9313001850186893, "learning_rate": 8.855937899712454e-06, "loss": 0.6683, "step": 1552 }, { "epoch": 1.4906385021603457, "grad_norm": 0.833183431382319, "learning_rate": 8.853804423055887e-06, "loss": 0.5638, "step": 1553 }, { "epoch": 1.4915986557849257, "grad_norm": 0.8878442701742922, "learning_rate": 8.851669216467708e-06, "loss": 0.6066, "step": 1554 }, { "epoch": 1.4925588094095055, "grad_norm": 0.9513024626237748, "learning_rate": 8.849532280906391e-06, "loss": 0.6478, "step": 1555 }, { "epoch": 1.4935189630340855, "grad_norm": 0.973534450649038, "learning_rate": 8.847393617331182e-06, "loss": 0.6821, "step": 1556 }, { "epoch": 1.4944791166586655, "grad_norm": 0.8170646036900495, "learning_rate": 8.845253226702103e-06, "loss": 0.6528, "step": 1557 }, { "epoch": 1.4954392702832453, "grad_norm": 0.9997592114446631, "learning_rate": 8.843111109979957e-06, "loss": 0.6826, "step": 1558 }, { "epoch": 1.4963994239078253, "grad_norm": 0.8559330780341664, "learning_rate": 8.840967268126313e-06, "loss": 0.6091, "step": 1559 }, { "epoch": 1.4973595775324051, "grad_norm": 0.6685845808193616, "learning_rate": 8.83882170210352e-06, "loss": 0.5057, "step": 1560 }, { "epoch": 1.4983197311569851, "grad_norm": 0.8177236819292143, "learning_rate": 8.836674412874699e-06, "loss": 0.5945, "step": 1561 }, { "epoch": 1.499279884781565, "grad_norm": 1.1500981564705064, "learning_rate": 8.834525401403747e-06, "loss": 0.625, "step": 1562 }, { "epoch": 1.500240038406145, "grad_norm": 0.618863654899676, "learning_rate": 8.83237466865533e-06, "loss": 0.592, "step": 1563 }, { "epoch": 1.501200192030725, "grad_norm": 0.8382217168838498, "learning_rate": 8.83022221559489e-06, "loss": 0.6002, "step": 1564 }, { "epoch": 1.5021603456553048, "grad_norm": 0.8481890825202059, "learning_rate": 8.828068043188641e-06, "loss": 0.6245, "step": 1565 }, { "epoch": 1.5031204992798848, "grad_norm": 0.9271868770682927, "learning_rate": 8.825912152403568e-06, "loss": 0.5939, "step": 1566 }, { "epoch": 1.5040806529044648, "grad_norm": 0.9336134247042761, "learning_rate": 8.823754544207426e-06, "loss": 0.6325, "step": 1567 }, { "epoch": 1.5050408065290446, "grad_norm": 1.012380792661088, "learning_rate": 8.821595219568741e-06, "loss": 0.7471, "step": 1568 }, { "epoch": 1.5060009601536246, "grad_norm": 1.0616147948608838, "learning_rate": 8.819434179456814e-06, "loss": 0.6336, "step": 1569 }, { "epoch": 1.5069611137782046, "grad_norm": 0.6992877270669542, "learning_rate": 8.817271424841711e-06, "loss": 0.6673, "step": 1570 }, { "epoch": 1.5079212674027844, "grad_norm": 1.3433987875552045, "learning_rate": 8.81510695669427e-06, "loss": 0.6612, "step": 1571 }, { "epoch": 1.5088814210273642, "grad_norm": 0.8778045386191777, "learning_rate": 8.812940775986098e-06, "loss": 0.6372, "step": 1572 }, { "epoch": 1.5098415746519445, "grad_norm": 0.7288218390000476, "learning_rate": 8.810772883689567e-06, "loss": 0.7419, "step": 1573 }, { "epoch": 1.5108017282765243, "grad_norm": 1.2346650407455961, "learning_rate": 8.808603280777826e-06, "loss": 0.6145, "step": 1574 }, { "epoch": 1.511761881901104, "grad_norm": 0.8924791085643956, "learning_rate": 8.806431968224784e-06, "loss": 0.6146, "step": 1575 }, { "epoch": 1.512722035525684, "grad_norm": 1.018726486083731, "learning_rate": 8.80425894700512e-06, "loss": 0.6524, "step": 1576 }, { "epoch": 1.513682189150264, "grad_norm": 0.9192014100943581, "learning_rate": 8.802084218094278e-06, "loss": 0.7112, "step": 1577 }, { "epoch": 1.5146423427748439, "grad_norm": 0.8668683437471818, "learning_rate": 8.799907782468473e-06, "loss": 0.5638, "step": 1578 }, { "epoch": 1.515602496399424, "grad_norm": 0.97023658252574, "learning_rate": 8.797729641104687e-06, "loss": 0.6268, "step": 1579 }, { "epoch": 1.516562650024004, "grad_norm": 0.8747767413440061, "learning_rate": 8.795549794980655e-06, "loss": 0.6786, "step": 1580 }, { "epoch": 1.5175228036485837, "grad_norm": 0.8410420908526413, "learning_rate": 8.793368245074896e-06, "loss": 0.6297, "step": 1581 }, { "epoch": 1.5184829572731637, "grad_norm": 0.8633360455794895, "learning_rate": 8.791184992366678e-06, "loss": 0.6743, "step": 1582 }, { "epoch": 1.5194431108977438, "grad_norm": 0.9192118473660734, "learning_rate": 8.789000037836045e-06, "loss": 0.5667, "step": 1583 }, { "epoch": 1.5204032645223235, "grad_norm": 0.7015881186413507, "learning_rate": 8.786813382463796e-06, "loss": 0.64, "step": 1584 }, { "epoch": 1.5213634181469033, "grad_norm": 0.955572056981828, "learning_rate": 8.7846250272315e-06, "loss": 0.6924, "step": 1585 }, { "epoch": 1.5223235717714836, "grad_norm": 0.9165429411880639, "learning_rate": 8.782434973121481e-06, "loss": 0.5336, "step": 1586 }, { "epoch": 1.5232837253960634, "grad_norm": 0.7885352959975056, "learning_rate": 8.780243221116838e-06, "loss": 0.5799, "step": 1587 }, { "epoch": 1.5242438790206432, "grad_norm": 0.9352585056768101, "learning_rate": 8.77804977220142e-06, "loss": 0.6439, "step": 1588 }, { "epoch": 1.5252040326452232, "grad_norm": 0.8669104784873425, "learning_rate": 8.775854627359844e-06, "loss": 0.6806, "step": 1589 }, { "epoch": 1.5261641862698032, "grad_norm": 0.7850179442581622, "learning_rate": 8.773657787577488e-06, "loss": 0.6631, "step": 1590 }, { "epoch": 1.527124339894383, "grad_norm": 0.9197371548095034, "learning_rate": 8.771459253840489e-06, "loss": 0.6208, "step": 1591 }, { "epoch": 1.528084493518963, "grad_norm": 0.9926445508133437, "learning_rate": 8.769259027135746e-06, "loss": 0.5712, "step": 1592 }, { "epoch": 1.529044647143543, "grad_norm": 0.8807730455954347, "learning_rate": 8.767057108450919e-06, "loss": 0.6373, "step": 1593 }, { "epoch": 1.5300048007681228, "grad_norm": 1.4197040687645213, "learning_rate": 8.76485349877442e-06, "loss": 0.7018, "step": 1594 }, { "epoch": 1.5309649543927029, "grad_norm": 1.1984446425301636, "learning_rate": 8.76264819909543e-06, "loss": 0.6357, "step": 1595 }, { "epoch": 1.5319251080172829, "grad_norm": 0.9248115909133072, "learning_rate": 8.760441210403886e-06, "loss": 0.7186, "step": 1596 }, { "epoch": 1.5328852616418627, "grad_norm": 1.0236981619396788, "learning_rate": 8.758232533690479e-06, "loss": 0.5987, "step": 1597 }, { "epoch": 1.5338454152664425, "grad_norm": 1.412417901297909, "learning_rate": 8.75602216994666e-06, "loss": 0.7866, "step": 1598 }, { "epoch": 1.5348055688910227, "grad_norm": 1.2461327175519716, "learning_rate": 8.753810120164639e-06, "loss": 0.6178, "step": 1599 }, { "epoch": 1.5357657225156025, "grad_norm": 0.8245632868119092, "learning_rate": 8.75159638533738e-06, "loss": 0.6683, "step": 1600 }, { "epoch": 1.5367258761401823, "grad_norm": 1.080135282160629, "learning_rate": 8.749380966458608e-06, "loss": 0.6976, "step": 1601 }, { "epoch": 1.5376860297647623, "grad_norm": 1.2147002117891772, "learning_rate": 8.747163864522796e-06, "loss": 0.6414, "step": 1602 }, { "epoch": 1.5386461833893423, "grad_norm": 0.9873449238987781, "learning_rate": 8.744945080525182e-06, "loss": 0.7081, "step": 1603 }, { "epoch": 1.5396063370139221, "grad_norm": 1.0105527626311783, "learning_rate": 8.742724615461753e-06, "loss": 0.5994, "step": 1604 }, { "epoch": 1.5405664906385022, "grad_norm": 0.7832404596725351, "learning_rate": 8.740502470329251e-06, "loss": 0.6226, "step": 1605 }, { "epoch": 1.5415266442630822, "grad_norm": 0.8424191446673672, "learning_rate": 8.738278646125174e-06, "loss": 0.6953, "step": 1606 }, { "epoch": 1.542486797887662, "grad_norm": 1.0145291000742407, "learning_rate": 8.736053143847774e-06, "loss": 0.6591, "step": 1607 }, { "epoch": 1.543446951512242, "grad_norm": 0.8869815625159235, "learning_rate": 8.733825964496052e-06, "loss": 0.5908, "step": 1608 }, { "epoch": 1.544407105136822, "grad_norm": 0.9752400517495328, "learning_rate": 8.731597109069768e-06, "loss": 0.6697, "step": 1609 }, { "epoch": 1.5453672587614018, "grad_norm": 0.7085505610556894, "learning_rate": 8.72936657856943e-06, "loss": 0.6237, "step": 1610 }, { "epoch": 1.5463274123859816, "grad_norm": 0.8355663620558922, "learning_rate": 8.727134373996298e-06, "loss": 0.5473, "step": 1611 }, { "epoch": 1.5472875660105618, "grad_norm": 0.6511831667641756, "learning_rate": 8.724900496352388e-06, "loss": 0.68, "step": 1612 }, { "epoch": 1.5482477196351416, "grad_norm": 1.0574787239433237, "learning_rate": 8.722664946640463e-06, "loss": 0.5787, "step": 1613 }, { "epoch": 1.5492078732597214, "grad_norm": 0.7789518445270859, "learning_rate": 8.720427725864035e-06, "loss": 0.6077, "step": 1614 }, { "epoch": 1.5501680268843014, "grad_norm": 1.0630736789446655, "learning_rate": 8.71818883502737e-06, "loss": 0.6759, "step": 1615 }, { "epoch": 1.5511281805088815, "grad_norm": 0.7552948598573878, "learning_rate": 8.715948275135482e-06, "loss": 0.5438, "step": 1616 }, { "epoch": 1.5520883341334613, "grad_norm": 0.6690846516749812, "learning_rate": 8.713706047194135e-06, "loss": 0.6268, "step": 1617 }, { "epoch": 1.5530484877580413, "grad_norm": 0.833531325258287, "learning_rate": 8.711462152209843e-06, "loss": 0.6207, "step": 1618 }, { "epoch": 1.5540086413826213, "grad_norm": 0.7076354412335065, "learning_rate": 8.709216591189861e-06, "loss": 0.5824, "step": 1619 }, { "epoch": 1.554968795007201, "grad_norm": 0.8198370163722243, "learning_rate": 8.706969365142202e-06, "loss": 0.6218, "step": 1620 }, { "epoch": 1.555928948631781, "grad_norm": 0.9473693435723936, "learning_rate": 8.70472047507562e-06, "loss": 0.5882, "step": 1621 }, { "epoch": 1.5568891022563611, "grad_norm": 0.9545762828404588, "learning_rate": 8.702469921999617e-06, "loss": 0.6954, "step": 1622 }, { "epoch": 1.557849255880941, "grad_norm": 0.8459017402362939, "learning_rate": 8.700217706924445e-06, "loss": 0.6369, "step": 1623 }, { "epoch": 1.558809409505521, "grad_norm": 0.7452526917006129, "learning_rate": 8.697963830861095e-06, "loss": 0.6123, "step": 1624 }, { "epoch": 1.559769563130101, "grad_norm": 1.0908877902507779, "learning_rate": 8.695708294821314e-06, "loss": 0.6914, "step": 1625 }, { "epoch": 1.5607297167546808, "grad_norm": 1.0187320413088718, "learning_rate": 8.693451099817583e-06, "loss": 0.6487, "step": 1626 }, { "epoch": 1.5616898703792605, "grad_norm": 0.8776341911853939, "learning_rate": 8.691192246863133e-06, "loss": 0.5044, "step": 1627 }, { "epoch": 1.5626500240038406, "grad_norm": 0.657340079955985, "learning_rate": 8.688931736971941e-06, "loss": 0.6563, "step": 1628 }, { "epoch": 1.5636101776284206, "grad_norm": 0.7988956819908309, "learning_rate": 8.686669571158724e-06, "loss": 0.6388, "step": 1629 }, { "epoch": 1.5645703312530004, "grad_norm": 0.7830446234374473, "learning_rate": 8.684405750438944e-06, "loss": 0.5967, "step": 1630 }, { "epoch": 1.5655304848775804, "grad_norm": 0.7310560777476105, "learning_rate": 8.682140275828809e-06, "loss": 0.646, "step": 1631 }, { "epoch": 1.5664906385021604, "grad_norm": 0.8671157491214064, "learning_rate": 8.679873148345262e-06, "loss": 0.589, "step": 1632 }, { "epoch": 1.5674507921267402, "grad_norm": 0.7345826560695118, "learning_rate": 8.677604369005996e-06, "loss": 0.6657, "step": 1633 }, { "epoch": 1.5684109457513202, "grad_norm": 0.9072516683021853, "learning_rate": 8.67533393882944e-06, "loss": 0.5299, "step": 1634 }, { "epoch": 1.5693710993759002, "grad_norm": 0.7306213822986727, "learning_rate": 8.673061858834766e-06, "loss": 0.6811, "step": 1635 }, { "epoch": 1.57033125300048, "grad_norm": 0.7924804540471937, "learning_rate": 8.670788130041886e-06, "loss": 0.6552, "step": 1636 }, { "epoch": 1.57129140662506, "grad_norm": 1.2157651492134904, "learning_rate": 8.668512753471453e-06, "loss": 0.583, "step": 1637 }, { "epoch": 1.57225156024964, "grad_norm": 0.867597106965324, "learning_rate": 8.666235730144858e-06, "loss": 0.6814, "step": 1638 }, { "epoch": 1.5732117138742199, "grad_norm": 0.9159239897652129, "learning_rate": 8.663957061084234e-06, "loss": 0.5774, "step": 1639 }, { "epoch": 1.5741718674987997, "grad_norm": 0.8308415640246836, "learning_rate": 8.661676747312453e-06, "loss": 0.7869, "step": 1640 }, { "epoch": 1.5751320211233797, "grad_norm": 1.023787593621287, "learning_rate": 8.65939478985312e-06, "loss": 0.675, "step": 1641 }, { "epoch": 1.5760921747479597, "grad_norm": 0.8258767230245585, "learning_rate": 8.657111189730583e-06, "loss": 0.6292, "step": 1642 }, { "epoch": 1.5770523283725395, "grad_norm": 1.050780742066422, "learning_rate": 8.654825947969924e-06, "loss": 0.6899, "step": 1643 }, { "epoch": 1.5780124819971195, "grad_norm": 0.9709095017067964, "learning_rate": 8.652539065596966e-06, "loss": 0.5699, "step": 1644 }, { "epoch": 1.5789726356216995, "grad_norm": 0.8496019167724437, "learning_rate": 8.650250543638264e-06, "loss": 0.6464, "step": 1645 }, { "epoch": 1.5799327892462793, "grad_norm": 1.0216950201670862, "learning_rate": 8.647960383121113e-06, "loss": 0.6402, "step": 1646 }, { "epoch": 1.5808929428708594, "grad_norm": 0.6698143183891819, "learning_rate": 8.64566858507354e-06, "loss": 0.7502, "step": 1647 }, { "epoch": 1.5818530964954394, "grad_norm": 1.1691696441715864, "learning_rate": 8.643375150524308e-06, "loss": 0.7161, "step": 1648 }, { "epoch": 1.5828132501200192, "grad_norm": 0.8255801188976419, "learning_rate": 8.641080080502919e-06, "loss": 0.6416, "step": 1649 }, { "epoch": 1.5837734037445992, "grad_norm": 0.7885485147710118, "learning_rate": 8.6387833760396e-06, "loss": 0.7284, "step": 1650 }, { "epoch": 1.5847335573691792, "grad_norm": 0.8395066063613783, "learning_rate": 8.636485038165323e-06, "loss": 0.6063, "step": 1651 }, { "epoch": 1.585693710993759, "grad_norm": 1.0266440732154936, "learning_rate": 8.634185067911781e-06, "loss": 0.5884, "step": 1652 }, { "epoch": 1.5866538646183388, "grad_norm": 0.9456950895739955, "learning_rate": 8.631883466311412e-06, "loss": 0.624, "step": 1653 }, { "epoch": 1.5876140182429188, "grad_norm": 0.8246492072965308, "learning_rate": 8.629580234397377e-06, "loss": 0.5319, "step": 1654 }, { "epoch": 1.5885741718674988, "grad_norm": 0.7577778077416646, "learning_rate": 8.627275373203572e-06, "loss": 0.5504, "step": 1655 }, { "epoch": 1.5895343254920786, "grad_norm": 1.2123380596512194, "learning_rate": 8.624968883764626e-06, "loss": 0.7878, "step": 1656 }, { "epoch": 1.5904944791166586, "grad_norm": 0.9567815163692012, "learning_rate": 8.622660767115897e-06, "loss": 0.5716, "step": 1657 }, { "epoch": 1.5914546327412387, "grad_norm": 0.6955159862805869, "learning_rate": 8.620351024293475e-06, "loss": 0.6615, "step": 1658 }, { "epoch": 1.5924147863658185, "grad_norm": 0.9544101769328984, "learning_rate": 8.618039656334173e-06, "loss": 0.5284, "step": 1659 }, { "epoch": 1.5933749399903985, "grad_norm": 0.8069051871004468, "learning_rate": 8.615726664275547e-06, "loss": 0.5861, "step": 1660 }, { "epoch": 1.5943350936149785, "grad_norm": 0.7099097770449407, "learning_rate": 8.613412049155872e-06, "loss": 0.634, "step": 1661 }, { "epoch": 1.5952952472395583, "grad_norm": 0.6669083300402422, "learning_rate": 8.611095812014155e-06, "loss": 0.6185, "step": 1662 }, { "epoch": 1.5962554008641383, "grad_norm": 0.7287828035197083, "learning_rate": 8.608777953890125e-06, "loss": 0.6754, "step": 1663 }, { "epoch": 1.5972155544887183, "grad_norm": 1.0651963761806575, "learning_rate": 8.60645847582425e-06, "loss": 0.6145, "step": 1664 }, { "epoch": 1.5981757081132981, "grad_norm": 0.7448040220530598, "learning_rate": 8.604137378857713e-06, "loss": 0.6259, "step": 1665 }, { "epoch": 1.599135861737878, "grad_norm": 0.7653867797206239, "learning_rate": 8.601814664032434e-06, "loss": 0.6515, "step": 1666 }, { "epoch": 1.6000960153624582, "grad_norm": 0.8589353644986315, "learning_rate": 8.599490332391054e-06, "loss": 0.6479, "step": 1667 }, { "epoch": 1.601056168987038, "grad_norm": 0.8785099525840002, "learning_rate": 8.59716438497694e-06, "loss": 0.6181, "step": 1668 }, { "epoch": 1.6020163226116177, "grad_norm": 0.7464004089132802, "learning_rate": 8.594836822834185e-06, "loss": 0.6223, "step": 1669 }, { "epoch": 1.6029764762361978, "grad_norm": 0.910667615466374, "learning_rate": 8.592507647007606e-06, "loss": 0.594, "step": 1670 }, { "epoch": 1.6039366298607778, "grad_norm": 0.8602296741850002, "learning_rate": 8.590176858542748e-06, "loss": 0.6399, "step": 1671 }, { "epoch": 1.6048967834853576, "grad_norm": 0.8744214600610309, "learning_rate": 8.587844458485874e-06, "loss": 0.6141, "step": 1672 }, { "epoch": 1.6058569371099376, "grad_norm": 0.662853284206551, "learning_rate": 8.585510447883975e-06, "loss": 0.6203, "step": 1673 }, { "epoch": 1.6068170907345176, "grad_norm": 0.9792288798638955, "learning_rate": 8.583174827784762e-06, "loss": 0.7495, "step": 1674 }, { "epoch": 1.6077772443590974, "grad_norm": 0.9409128855644499, "learning_rate": 8.580837599236673e-06, "loss": 0.654, "step": 1675 }, { "epoch": 1.6087373979836774, "grad_norm": 0.9286073049032478, "learning_rate": 8.57849876328886e-06, "loss": 0.7328, "step": 1676 }, { "epoch": 1.6096975516082574, "grad_norm": 1.101091817340552, "learning_rate": 8.576158320991205e-06, "loss": 0.5005, "step": 1677 }, { "epoch": 1.6106577052328372, "grad_norm": 0.661884502839225, "learning_rate": 8.573816273394308e-06, "loss": 0.475, "step": 1678 }, { "epoch": 1.611617858857417, "grad_norm": 0.7970467324030691, "learning_rate": 8.571472621549488e-06, "loss": 0.5688, "step": 1679 }, { "epoch": 1.6125780124819973, "grad_norm": 0.954094043646711, "learning_rate": 8.569127366508782e-06, "loss": 0.5967, "step": 1680 }, { "epoch": 1.613538166106577, "grad_norm": 1.007477655548362, "learning_rate": 8.566780509324956e-06, "loss": 0.5615, "step": 1681 }, { "epoch": 1.6144983197311569, "grad_norm": 1.1664556577828813, "learning_rate": 8.564432051051485e-06, "loss": 0.701, "step": 1682 }, { "epoch": 1.615458473355737, "grad_norm": 0.8933818373435737, "learning_rate": 8.562081992742569e-06, "loss": 0.6414, "step": 1683 }, { "epoch": 1.616418626980317, "grad_norm": 0.8419703662508063, "learning_rate": 8.559730335453122e-06, "loss": 0.7242, "step": 1684 }, { "epoch": 1.6173787806048967, "grad_norm": 1.254509683695103, "learning_rate": 8.557377080238778e-06, "loss": 0.7802, "step": 1685 }, { "epoch": 1.6183389342294767, "grad_norm": 1.028343848747216, "learning_rate": 8.55502222815589e-06, "loss": 0.6072, "step": 1686 }, { "epoch": 1.6192990878540567, "grad_norm": 1.068724027540988, "learning_rate": 8.552665780261526e-06, "loss": 0.5744, "step": 1687 }, { "epoch": 1.6202592414786365, "grad_norm": 0.8130085399320413, "learning_rate": 8.550307737613468e-06, "loss": 0.491, "step": 1688 }, { "epoch": 1.6212193951032166, "grad_norm": 0.8189813821548464, "learning_rate": 8.547948101270215e-06, "loss": 0.5748, "step": 1689 }, { "epoch": 1.6221795487277966, "grad_norm": 1.005365984438001, "learning_rate": 8.545586872290987e-06, "loss": 0.6752, "step": 1690 }, { "epoch": 1.6231397023523764, "grad_norm": 1.1010744605730625, "learning_rate": 8.543224051735714e-06, "loss": 0.6649, "step": 1691 }, { "epoch": 1.6240998559769562, "grad_norm": 0.93841142956136, "learning_rate": 8.540859640665036e-06, "loss": 0.617, "step": 1692 }, { "epoch": 1.6250600096015364, "grad_norm": 0.9242743184031158, "learning_rate": 8.538493640140317e-06, "loss": 0.6485, "step": 1693 }, { "epoch": 1.6260201632261162, "grad_norm": 0.9164819217375217, "learning_rate": 8.536126051223627e-06, "loss": 0.529, "step": 1694 }, { "epoch": 1.626980316850696, "grad_norm": 0.8538570781698765, "learning_rate": 8.533756874977751e-06, "loss": 0.5425, "step": 1695 }, { "epoch": 1.627940470475276, "grad_norm": 1.1865311931027884, "learning_rate": 8.53138611246619e-06, "loss": 0.6645, "step": 1696 }, { "epoch": 1.628900624099856, "grad_norm": 0.8252003886493567, "learning_rate": 8.529013764753147e-06, "loss": 0.6737, "step": 1697 }, { "epoch": 1.6298607777244358, "grad_norm": 0.8708884416321419, "learning_rate": 8.526639832903552e-06, "loss": 0.6118, "step": 1698 }, { "epoch": 1.6308209313490158, "grad_norm": 0.8396782918201511, "learning_rate": 8.524264317983032e-06, "loss": 0.6115, "step": 1699 }, { "epoch": 1.6317810849735959, "grad_norm": 0.8070084486204835, "learning_rate": 8.521887221057932e-06, "loss": 0.6866, "step": 1700 }, { "epoch": 1.6327412385981757, "grad_norm": 0.8289682018263336, "learning_rate": 8.519508543195304e-06, "loss": 0.6481, "step": 1701 }, { "epoch": 1.6337013922227557, "grad_norm": 1.0807287294101995, "learning_rate": 8.517128285462914e-06, "loss": 0.6128, "step": 1702 }, { "epoch": 1.6346615458473357, "grad_norm": 0.7688699388753459, "learning_rate": 8.514746448929235e-06, "loss": 0.5924, "step": 1703 }, { "epoch": 1.6356216994719155, "grad_norm": 0.8031070471723358, "learning_rate": 8.512363034663441e-06, "loss": 0.6344, "step": 1704 }, { "epoch": 1.6365818530964953, "grad_norm": 0.9502808887022575, "learning_rate": 8.50997804373543e-06, "loss": 0.5883, "step": 1705 }, { "epoch": 1.6375420067210755, "grad_norm": 0.7459772227344592, "learning_rate": 8.507591477215793e-06, "loss": 0.5826, "step": 1706 }, { "epoch": 1.6385021603456553, "grad_norm": 1.16345517089271, "learning_rate": 8.505203336175836e-06, "loss": 0.5916, "step": 1707 }, { "epoch": 1.6394623139702351, "grad_norm": 0.9436009171949882, "learning_rate": 8.502813621687568e-06, "loss": 0.5918, "step": 1708 }, { "epoch": 1.6404224675948151, "grad_norm": 0.8941456657377294, "learning_rate": 8.50042233482371e-06, "loss": 0.5303, "step": 1709 }, { "epoch": 1.6413826212193952, "grad_norm": 0.7571933709892077, "learning_rate": 8.498029476657686e-06, "loss": 0.7042, "step": 1710 }, { "epoch": 1.642342774843975, "grad_norm": 1.1112968156853915, "learning_rate": 8.495635048263617e-06, "loss": 0.6855, "step": 1711 }, { "epoch": 1.643302928468555, "grad_norm": 0.8859366677977835, "learning_rate": 8.493239050716344e-06, "loss": 0.533, "step": 1712 }, { "epoch": 1.644263082093135, "grad_norm": 0.6994613499080129, "learning_rate": 8.4908414850914e-06, "loss": 0.5666, "step": 1713 }, { "epoch": 1.6452232357177148, "grad_norm": 0.974243199747327, "learning_rate": 8.488442352465029e-06, "loss": 0.6209, "step": 1714 }, { "epoch": 1.6461833893422948, "grad_norm": 0.9556499539283487, "learning_rate": 8.486041653914177e-06, "loss": 0.5956, "step": 1715 }, { "epoch": 1.6471435429668748, "grad_norm": 1.1687489923029992, "learning_rate": 8.483639390516488e-06, "loss": 0.6912, "step": 1716 }, { "epoch": 1.6481036965914546, "grad_norm": 1.0366291162608932, "learning_rate": 8.481235563350316e-06, "loss": 0.6332, "step": 1717 }, { "epoch": 1.6490638502160344, "grad_norm": 0.7206333367508821, "learning_rate": 8.478830173494712e-06, "loss": 0.574, "step": 1718 }, { "epoch": 1.6500240038406147, "grad_norm": 0.8443128902140005, "learning_rate": 8.47642322202943e-06, "loss": 0.7961, "step": 1719 }, { "epoch": 1.6509841574651944, "grad_norm": 1.0707738930921702, "learning_rate": 8.474014710034923e-06, "loss": 0.6037, "step": 1720 }, { "epoch": 1.6519443110897742, "grad_norm": 1.0205264560484588, "learning_rate": 8.471604638592348e-06, "loss": 0.6481, "step": 1721 }, { "epoch": 1.6529044647143543, "grad_norm": 1.0239069982776392, "learning_rate": 8.469193008783562e-06, "loss": 0.6865, "step": 1722 }, { "epoch": 1.6538646183389343, "grad_norm": 0.5958751457363809, "learning_rate": 8.466779821691117e-06, "loss": 0.613, "step": 1723 }, { "epoch": 1.654824771963514, "grad_norm": 0.7807036315706164, "learning_rate": 8.464365078398269e-06, "loss": 0.6087, "step": 1724 }, { "epoch": 1.655784925588094, "grad_norm": 0.8116541550606194, "learning_rate": 8.461948779988967e-06, "loss": 0.6971, "step": 1725 }, { "epoch": 1.656745079212674, "grad_norm": 0.823505089212548, "learning_rate": 8.459530927547864e-06, "loss": 0.63, "step": 1726 }, { "epoch": 1.657705232837254, "grad_norm": 1.0971531780222792, "learning_rate": 8.457111522160309e-06, "loss": 0.8003, "step": 1727 }, { "epoch": 1.658665386461834, "grad_norm": 1.2976753965396226, "learning_rate": 8.454690564912347e-06, "loss": 0.7746, "step": 1728 }, { "epoch": 1.659625540086414, "grad_norm": 1.098736237180015, "learning_rate": 8.452268056890717e-06, "loss": 0.7224, "step": 1729 }, { "epoch": 1.6605856937109937, "grad_norm": 0.8138650359510815, "learning_rate": 8.44984399918286e-06, "loss": 0.5711, "step": 1730 }, { "epoch": 1.6615458473355735, "grad_norm": 1.093292040005224, "learning_rate": 8.447418392876907e-06, "loss": 0.7268, "step": 1731 }, { "epoch": 1.6625060009601538, "grad_norm": 1.2264788433856098, "learning_rate": 8.44499123906169e-06, "loss": 0.6875, "step": 1732 }, { "epoch": 1.6634661545847336, "grad_norm": 0.8498869403338817, "learning_rate": 8.44256253882673e-06, "loss": 0.5861, "step": 1733 }, { "epoch": 1.6644263082093134, "grad_norm": 0.8122455699135925, "learning_rate": 8.440132293262246e-06, "loss": 0.6386, "step": 1734 }, { "epoch": 1.6653864618338934, "grad_norm": 0.7836964410559036, "learning_rate": 8.437700503459149e-06, "loss": 0.5384, "step": 1735 }, { "epoch": 1.6663466154584734, "grad_norm": 0.9905785264010473, "learning_rate": 8.435267170509044e-06, "loss": 0.6838, "step": 1736 }, { "epoch": 1.6673067690830532, "grad_norm": 1.0990297282146466, "learning_rate": 8.432832295504224e-06, "loss": 0.6673, "step": 1737 }, { "epoch": 1.6682669227076332, "grad_norm": 1.1317394784666022, "learning_rate": 8.430395879537684e-06, "loss": 0.6344, "step": 1738 }, { "epoch": 1.6692270763322132, "grad_norm": 0.7793434405378374, "learning_rate": 8.4279579237031e-06, "loss": 0.61, "step": 1739 }, { "epoch": 1.670187229956793, "grad_norm": 0.8674568448548475, "learning_rate": 8.425518429094848e-06, "loss": 0.5458, "step": 1740 }, { "epoch": 1.671147383581373, "grad_norm": 0.7375056425068617, "learning_rate": 8.423077396807991e-06, "loss": 0.6434, "step": 1741 }, { "epoch": 1.672107537205953, "grad_norm": 1.102040913043917, "learning_rate": 8.42063482793828e-06, "loss": 0.5531, "step": 1742 }, { "epoch": 1.6730676908305329, "grad_norm": 1.1233777552539144, "learning_rate": 8.41819072358216e-06, "loss": 0.7193, "step": 1743 }, { "epoch": 1.6740278444551127, "grad_norm": 0.8430058555562076, "learning_rate": 8.415745084836763e-06, "loss": 0.7292, "step": 1744 }, { "epoch": 1.674987998079693, "grad_norm": 0.9714046538033724, "learning_rate": 8.413297912799909e-06, "loss": 0.683, "step": 1745 }, { "epoch": 1.6759481517042727, "grad_norm": 1.0551198339092225, "learning_rate": 8.410849208570108e-06, "loss": 0.581, "step": 1746 }, { "epoch": 1.6769083053288525, "grad_norm": 0.9195301817671032, "learning_rate": 8.408398973246557e-06, "loss": 0.5141, "step": 1747 }, { "epoch": 1.6778684589534325, "grad_norm": 1.065852186405074, "learning_rate": 8.405947207929142e-06, "loss": 0.7392, "step": 1748 }, { "epoch": 1.6788286125780125, "grad_norm": 1.2405304850720786, "learning_rate": 8.40349391371843e-06, "loss": 0.6486, "step": 1749 }, { "epoch": 1.6797887662025923, "grad_norm": 0.6226374369862415, "learning_rate": 8.401039091715685e-06, "loss": 0.591, "step": 1750 }, { "epoch": 1.6807489198271723, "grad_norm": 0.8981972915393202, "learning_rate": 8.398582743022845e-06, "loss": 0.6378, "step": 1751 }, { "epoch": 1.6817090734517524, "grad_norm": 0.9514821928406277, "learning_rate": 8.39612486874254e-06, "loss": 0.658, "step": 1752 }, { "epoch": 1.6826692270763322, "grad_norm": 1.03636399739608, "learning_rate": 8.393665469978086e-06, "loss": 0.5538, "step": 1753 }, { "epoch": 1.6836293807009122, "grad_norm": 0.963630287541821, "learning_rate": 8.391204547833478e-06, "loss": 0.6433, "step": 1754 }, { "epoch": 1.6845895343254922, "grad_norm": 1.1805938476787494, "learning_rate": 8.388742103413397e-06, "loss": 0.8035, "step": 1755 }, { "epoch": 1.685549687950072, "grad_norm": 0.8478917922759251, "learning_rate": 8.38627813782321e-06, "loss": 0.6929, "step": 1756 }, { "epoch": 1.6865098415746518, "grad_norm": 1.0415136898918052, "learning_rate": 8.383812652168966e-06, "loss": 0.6862, "step": 1757 }, { "epoch": 1.687469995199232, "grad_norm": 1.1204187578913982, "learning_rate": 8.38134564755739e-06, "loss": 0.7281, "step": 1758 }, { "epoch": 1.6884301488238118, "grad_norm": 0.9185114965682186, "learning_rate": 8.378877125095901e-06, "loss": 0.5642, "step": 1759 }, { "epoch": 1.6893903024483916, "grad_norm": 1.0000242140902715, "learning_rate": 8.376407085892586e-06, "loss": 0.7041, "step": 1760 }, { "epoch": 1.6903504560729716, "grad_norm": 1.1560808327314147, "learning_rate": 8.373935531056222e-06, "loss": 0.4887, "step": 1761 }, { "epoch": 1.6913106096975516, "grad_norm": 0.8046750207667006, "learning_rate": 8.371462461696267e-06, "loss": 0.6296, "step": 1762 }, { "epoch": 1.6922707633221314, "grad_norm": 1.0959426102057963, "learning_rate": 8.36898787892285e-06, "loss": 0.6918, "step": 1763 }, { "epoch": 1.6932309169467115, "grad_norm": 0.7716778605825088, "learning_rate": 8.366511783846785e-06, "loss": 0.5644, "step": 1764 }, { "epoch": 1.6941910705712915, "grad_norm": 0.8352031926425314, "learning_rate": 8.364034177579568e-06, "loss": 0.5383, "step": 1765 }, { "epoch": 1.6951512241958713, "grad_norm": 0.7843151497670761, "learning_rate": 8.361555061233367e-06, "loss": 0.5996, "step": 1766 }, { "epoch": 1.6961113778204513, "grad_norm": 0.7805233830271293, "learning_rate": 8.359074435921032e-06, "loss": 0.5178, "step": 1767 }, { "epoch": 1.6970715314450313, "grad_norm": 0.9076298166931591, "learning_rate": 8.35659230275609e-06, "loss": 0.7428, "step": 1768 }, { "epoch": 1.698031685069611, "grad_norm": 0.790262982031244, "learning_rate": 8.354108662852743e-06, "loss": 0.6322, "step": 1769 }, { "epoch": 1.6989918386941911, "grad_norm": 0.7074380360483444, "learning_rate": 8.351623517325872e-06, "loss": 0.6715, "step": 1770 }, { "epoch": 1.6999519923187711, "grad_norm": 1.2023085673602094, "learning_rate": 8.349136867291029e-06, "loss": 0.6393, "step": 1771 }, { "epoch": 1.700912145943351, "grad_norm": 0.7453414313553465, "learning_rate": 8.346648713864447e-06, "loss": 0.5635, "step": 1772 }, { "epoch": 1.7018722995679307, "grad_norm": 1.0030687365897546, "learning_rate": 8.344159058163032e-06, "loss": 0.8021, "step": 1773 }, { "epoch": 1.7028324531925108, "grad_norm": 1.060946105124432, "learning_rate": 8.341667901304362e-06, "loss": 0.6451, "step": 1774 }, { "epoch": 1.7037926068170908, "grad_norm": 0.7120954916576836, "learning_rate": 8.339175244406693e-06, "loss": 0.7086, "step": 1775 }, { "epoch": 1.7047527604416706, "grad_norm": 1.1101859870432083, "learning_rate": 8.336681088588949e-06, "loss": 0.5524, "step": 1776 }, { "epoch": 1.7057129140662506, "grad_norm": 1.135667230010565, "learning_rate": 8.33418543497073e-06, "loss": 0.7581, "step": 1777 }, { "epoch": 1.7066730676908306, "grad_norm": 1.0508720188716456, "learning_rate": 8.331688284672313e-06, "loss": 0.6867, "step": 1778 }, { "epoch": 1.7076332213154104, "grad_norm": 0.9455686446961871, "learning_rate": 8.329189638814637e-06, "loss": 0.7351, "step": 1779 }, { "epoch": 1.7085933749399904, "grad_norm": 0.8062427606245879, "learning_rate": 8.32668949851932e-06, "loss": 0.674, "step": 1780 }, { "epoch": 1.7095535285645704, "grad_norm": 0.8167884202328937, "learning_rate": 8.324187864908646e-06, "loss": 0.6654, "step": 1781 }, { "epoch": 1.7105136821891502, "grad_norm": 0.9740431006360003, "learning_rate": 8.321684739105573e-06, "loss": 0.6858, "step": 1782 }, { "epoch": 1.7114738358137302, "grad_norm": 1.000256586364004, "learning_rate": 8.319180122233729e-06, "loss": 0.6532, "step": 1783 }, { "epoch": 1.7124339894383103, "grad_norm": 0.826129756378522, "learning_rate": 8.316674015417407e-06, "loss": 0.5965, "step": 1784 }, { "epoch": 1.71339414306289, "grad_norm": 0.995294558743917, "learning_rate": 8.314166419781574e-06, "loss": 0.6907, "step": 1785 }, { "epoch": 1.7143542966874699, "grad_norm": 1.078287686241976, "learning_rate": 8.311657336451862e-06, "loss": 0.6972, "step": 1786 }, { "epoch": 1.7153144503120499, "grad_norm": 1.0301861610063439, "learning_rate": 8.30914676655457e-06, "loss": 0.6603, "step": 1787 }, { "epoch": 1.71627460393663, "grad_norm": 0.9388895604984305, "learning_rate": 8.306634711216668e-06, "loss": 0.6598, "step": 1788 }, { "epoch": 1.7172347575612097, "grad_norm": 0.7198848533229457, "learning_rate": 8.304121171565788e-06, "loss": 0.6266, "step": 1789 }, { "epoch": 1.7181949111857897, "grad_norm": 0.7619300494945815, "learning_rate": 8.301606148730237e-06, "loss": 0.5771, "step": 1790 }, { "epoch": 1.7191550648103697, "grad_norm": 0.8553778362324546, "learning_rate": 8.299089643838976e-06, "loss": 0.5945, "step": 1791 }, { "epoch": 1.7201152184349495, "grad_norm": 0.8810958518353132, "learning_rate": 8.296571658021638e-06, "loss": 0.5985, "step": 1792 }, { "epoch": 1.7210753720595295, "grad_norm": 1.3317095235105623, "learning_rate": 8.294052192408522e-06, "loss": 0.6472, "step": 1793 }, { "epoch": 1.7220355256841096, "grad_norm": 0.8307601473549772, "learning_rate": 8.291531248130589e-06, "loss": 0.5802, "step": 1794 }, { "epoch": 1.7229956793086894, "grad_norm": 0.8500685752571073, "learning_rate": 8.289008826319463e-06, "loss": 0.5936, "step": 1795 }, { "epoch": 1.7239558329332694, "grad_norm": 1.130678992633, "learning_rate": 8.286484928107431e-06, "loss": 0.7867, "step": 1796 }, { "epoch": 1.7249159865578494, "grad_norm": 1.0139322422701196, "learning_rate": 8.283959554627448e-06, "loss": 0.5094, "step": 1797 }, { "epoch": 1.7258761401824292, "grad_norm": 0.8016336927430375, "learning_rate": 8.281432707013123e-06, "loss": 0.5748, "step": 1798 }, { "epoch": 1.726836293807009, "grad_norm": 0.9276916484783913, "learning_rate": 8.278904386398733e-06, "loss": 0.6664, "step": 1799 }, { "epoch": 1.727796447431589, "grad_norm": 0.9629166464822662, "learning_rate": 8.276374593919213e-06, "loss": 0.5988, "step": 1800 }, { "epoch": 1.728756601056169, "grad_norm": 0.985031028819971, "learning_rate": 8.27384333071016e-06, "loss": 0.6855, "step": 1801 }, { "epoch": 1.7297167546807488, "grad_norm": 0.7535715694122271, "learning_rate": 8.27131059790783e-06, "loss": 0.5803, "step": 1802 }, { "epoch": 1.7306769083053288, "grad_norm": 0.9209326675261907, "learning_rate": 8.268776396649145e-06, "loss": 0.7014, "step": 1803 }, { "epoch": 1.7316370619299088, "grad_norm": 0.9479593474910276, "learning_rate": 8.266240728071673e-06, "loss": 0.6471, "step": 1804 }, { "epoch": 1.7325972155544886, "grad_norm": 0.9130226740673932, "learning_rate": 8.263703593313654e-06, "loss": 0.6667, "step": 1805 }, { "epoch": 1.7335573691790687, "grad_norm": 1.1616810943501334, "learning_rate": 8.261164993513978e-06, "loss": 0.6518, "step": 1806 }, { "epoch": 1.7345175228036487, "grad_norm": 0.6808767406920158, "learning_rate": 8.258624929812198e-06, "loss": 0.604, "step": 1807 }, { "epoch": 1.7354776764282285, "grad_norm": 0.97315312168278, "learning_rate": 8.256083403348518e-06, "loss": 0.6773, "step": 1808 }, { "epoch": 1.7364378300528085, "grad_norm": 0.9095422712170884, "learning_rate": 8.253540415263805e-06, "loss": 0.5567, "step": 1809 }, { "epoch": 1.7373979836773885, "grad_norm": 0.7502187583867346, "learning_rate": 8.250995966699577e-06, "loss": 0.556, "step": 1810 }, { "epoch": 1.7383581373019683, "grad_norm": 0.8703781890510793, "learning_rate": 8.248450058798012e-06, "loss": 0.6912, "step": 1811 }, { "epoch": 1.739318290926548, "grad_norm": 0.836489102746461, "learning_rate": 8.245902692701939e-06, "loss": 0.5586, "step": 1812 }, { "epoch": 1.7402784445511283, "grad_norm": 0.9296393162212717, "learning_rate": 8.243353869554845e-06, "loss": 0.6693, "step": 1813 }, { "epoch": 1.7412385981757081, "grad_norm": 0.9179562592369949, "learning_rate": 8.240803590500865e-06, "loss": 0.6458, "step": 1814 }, { "epoch": 1.742198751800288, "grad_norm": 0.8997922793000442, "learning_rate": 8.238251856684801e-06, "loss": 0.6855, "step": 1815 }, { "epoch": 1.743158905424868, "grad_norm": 0.9430080262673844, "learning_rate": 8.23569866925209e-06, "loss": 0.6766, "step": 1816 }, { "epoch": 1.744119059049448, "grad_norm": 0.8139677735547685, "learning_rate": 8.233144029348834e-06, "loss": 0.6277, "step": 1817 }, { "epoch": 1.7450792126740278, "grad_norm": 0.9318490656946825, "learning_rate": 8.230587938121783e-06, "loss": 0.6579, "step": 1818 }, { "epoch": 1.7460393662986078, "grad_norm": 0.894082802215955, "learning_rate": 8.228030396718342e-06, "loss": 0.6577, "step": 1819 }, { "epoch": 1.7469995199231878, "grad_norm": 0.9662136385657485, "learning_rate": 8.22547140628656e-06, "loss": 0.6404, "step": 1820 }, { "epoch": 1.7479596735477676, "grad_norm": 0.8447927767332454, "learning_rate": 8.222910967975143e-06, "loss": 0.6364, "step": 1821 }, { "epoch": 1.7489198271723476, "grad_norm": 0.8415483763729134, "learning_rate": 8.220349082933444e-06, "loss": 0.5399, "step": 1822 }, { "epoch": 1.7498799807969276, "grad_norm": 0.9240329320740702, "learning_rate": 8.217785752311464e-06, "loss": 0.689, "step": 1823 }, { "epoch": 1.7508401344215074, "grad_norm": 0.9460069368310514, "learning_rate": 8.215220977259855e-06, "loss": 0.5613, "step": 1824 }, { "epoch": 1.7518002880460872, "grad_norm": 0.8600960041914113, "learning_rate": 8.21265475892992e-06, "loss": 0.693, "step": 1825 }, { "epoch": 1.7527604416706675, "grad_norm": 0.9644759840067197, "learning_rate": 8.210087098473605e-06, "loss": 0.6126, "step": 1826 }, { "epoch": 1.7537205952952473, "grad_norm": 1.1268192755908037, "learning_rate": 8.207517997043504e-06, "loss": 0.6751, "step": 1827 }, { "epoch": 1.754680748919827, "grad_norm": 1.0992756650418882, "learning_rate": 8.204947455792863e-06, "loss": 0.6884, "step": 1828 }, { "epoch": 1.755640902544407, "grad_norm": 0.9651369283959125, "learning_rate": 8.202375475875566e-06, "loss": 0.6377, "step": 1829 }, { "epoch": 1.756601056168987, "grad_norm": 1.1378718817096105, "learning_rate": 8.19980205844615e-06, "loss": 0.5727, "step": 1830 }, { "epoch": 1.757561209793567, "grad_norm": 0.9302749387546705, "learning_rate": 8.197227204659795e-06, "loss": 0.7435, "step": 1831 }, { "epoch": 1.758521363418147, "grad_norm": 1.3811748558321921, "learning_rate": 8.194650915672323e-06, "loss": 0.6798, "step": 1832 }, { "epoch": 1.759481517042727, "grad_norm": 0.9367759153733208, "learning_rate": 8.192073192640204e-06, "loss": 0.6709, "step": 1833 }, { "epoch": 1.7604416706673067, "grad_norm": 1.0272460481848618, "learning_rate": 8.18949403672055e-06, "loss": 0.4615, "step": 1834 }, { "epoch": 1.7614018242918867, "grad_norm": 0.6687862719098571, "learning_rate": 8.186913449071117e-06, "loss": 0.5922, "step": 1835 }, { "epoch": 1.7623619779164668, "grad_norm": 1.0871731901368777, "learning_rate": 8.184331430850304e-06, "loss": 0.73, "step": 1836 }, { "epoch": 1.7633221315410466, "grad_norm": 0.9161450271875266, "learning_rate": 8.181747983217148e-06, "loss": 0.6035, "step": 1837 }, { "epoch": 1.7642822851656264, "grad_norm": 1.4454043870822073, "learning_rate": 8.179163107331336e-06, "loss": 0.6771, "step": 1838 }, { "epoch": 1.7652424387902066, "grad_norm": 0.836440439239643, "learning_rate": 8.176576804353186e-06, "loss": 0.6471, "step": 1839 }, { "epoch": 1.7662025924147864, "grad_norm": 1.0535730264256495, "learning_rate": 8.173989075443666e-06, "loss": 0.6298, "step": 1840 }, { "epoch": 1.7671627460393662, "grad_norm": 1.2849561622011814, "learning_rate": 8.171399921764377e-06, "loss": 0.6806, "step": 1841 }, { "epoch": 1.7681228996639462, "grad_norm": 1.038202996351628, "learning_rate": 8.168809344477564e-06, "loss": 0.7502, "step": 1842 }, { "epoch": 1.7690830532885262, "grad_norm": 2.525971959843216, "learning_rate": 8.16621734474611e-06, "loss": 0.5708, "step": 1843 }, { "epoch": 1.770043206913106, "grad_norm": 1.0136705553829597, "learning_rate": 8.163623923733533e-06, "loss": 0.606, "step": 1844 }, { "epoch": 1.771003360537686, "grad_norm": 1.076410592077099, "learning_rate": 8.161029082603994e-06, "loss": 0.7613, "step": 1845 }, { "epoch": 1.771963514162266, "grad_norm": 1.1196986975398096, "learning_rate": 8.158432822522291e-06, "loss": 0.5411, "step": 1846 }, { "epoch": 1.7729236677868458, "grad_norm": 0.824027401743515, "learning_rate": 8.155835144653856e-06, "loss": 0.6939, "step": 1847 }, { "epoch": 1.7738838214114259, "grad_norm": 1.1066799608961153, "learning_rate": 8.153236050164757e-06, "loss": 0.6554, "step": 1848 }, { "epoch": 1.7748439750360059, "grad_norm": 0.7485857759384139, "learning_rate": 8.1506355402217e-06, "loss": 0.6374, "step": 1849 }, { "epoch": 1.7758041286605857, "grad_norm": 0.7928598966741474, "learning_rate": 8.14803361599203e-06, "loss": 0.5154, "step": 1850 }, { "epoch": 1.7767642822851655, "grad_norm": 1.065115125029083, "learning_rate": 8.14543027864372e-06, "loss": 0.6888, "step": 1851 }, { "epoch": 1.7777244359097457, "grad_norm": 0.9632568630957312, "learning_rate": 8.14282552934538e-06, "loss": 0.7608, "step": 1852 }, { "epoch": 1.7786845895343255, "grad_norm": 0.9685331911600413, "learning_rate": 8.140219369266255e-06, "loss": 0.5911, "step": 1853 }, { "epoch": 1.7796447431589053, "grad_norm": 0.7541378425434284, "learning_rate": 8.137611799576222e-06, "loss": 0.565, "step": 1854 }, { "epoch": 1.7806048967834853, "grad_norm": 0.8964093221096241, "learning_rate": 8.135002821445793e-06, "loss": 0.5549, "step": 1855 }, { "epoch": 1.7815650504080653, "grad_norm": 0.7642211271384314, "learning_rate": 8.132392436046107e-06, "loss": 0.6869, "step": 1856 }, { "epoch": 1.7825252040326451, "grad_norm": 1.3105231410170188, "learning_rate": 8.129780644548938e-06, "loss": 0.5671, "step": 1857 }, { "epoch": 1.7834853576572252, "grad_norm": 1.3184977217828324, "learning_rate": 8.127167448126698e-06, "loss": 0.6178, "step": 1858 }, { "epoch": 1.7844455112818052, "grad_norm": 0.9977859435042646, "learning_rate": 8.124552847952416e-06, "loss": 0.6752, "step": 1859 }, { "epoch": 1.785405664906385, "grad_norm": 0.715235139133677, "learning_rate": 8.121936845199759e-06, "loss": 0.6783, "step": 1860 }, { "epoch": 1.786365818530965, "grad_norm": 0.8741977521417024, "learning_rate": 8.119319441043024e-06, "loss": 0.5916, "step": 1861 }, { "epoch": 1.787325972155545, "grad_norm": 1.0352498400712868, "learning_rate": 8.116700636657137e-06, "loss": 0.5475, "step": 1862 }, { "epoch": 1.7882861257801248, "grad_norm": 1.2968265908761634, "learning_rate": 8.11408043321765e-06, "loss": 0.6652, "step": 1863 }, { "epoch": 1.7892462794047046, "grad_norm": 0.9227341954068595, "learning_rate": 8.111458831900743e-06, "loss": 0.6911, "step": 1864 }, { "epoch": 1.7902064330292848, "grad_norm": 0.964672586368304, "learning_rate": 8.108835833883227e-06, "loss": 0.7133, "step": 1865 }, { "epoch": 1.7911665866538646, "grad_norm": 0.8279944507502378, "learning_rate": 8.106211440342535e-06, "loss": 0.6549, "step": 1866 }, { "epoch": 1.7921267402784444, "grad_norm": 0.9400231801959814, "learning_rate": 8.103585652456732e-06, "loss": 0.6524, "step": 1867 }, { "epoch": 1.7930868939030244, "grad_norm": 0.8728981280869609, "learning_rate": 8.100958471404506e-06, "loss": 0.5947, "step": 1868 }, { "epoch": 1.7940470475276045, "grad_norm": 0.5960739026726652, "learning_rate": 8.098329898365168e-06, "loss": 0.635, "step": 1869 }, { "epoch": 1.7950072011521843, "grad_norm": 1.0252168027027986, "learning_rate": 8.095699934518659e-06, "loss": 0.563, "step": 1870 }, { "epoch": 1.7959673547767643, "grad_norm": 1.0280849061721453, "learning_rate": 8.093068581045538e-06, "loss": 0.5723, "step": 1871 }, { "epoch": 1.7969275084013443, "grad_norm": 0.711257699693533, "learning_rate": 8.090435839126998e-06, "loss": 0.5459, "step": 1872 }, { "epoch": 1.797887662025924, "grad_norm": 0.6813516793135845, "learning_rate": 8.087801709944842e-06, "loss": 0.6106, "step": 1873 }, { "epoch": 1.7988478156505041, "grad_norm": 0.8676402694622051, "learning_rate": 8.085166194681508e-06, "loss": 0.5863, "step": 1874 }, { "epoch": 1.7998079692750841, "grad_norm": 0.973067870868028, "learning_rate": 8.082529294520046e-06, "loss": 0.6973, "step": 1875 }, { "epoch": 1.800768122899664, "grad_norm": 0.9785784935410299, "learning_rate": 8.079891010644134e-06, "loss": 0.6984, "step": 1876 }, { "epoch": 1.8017282765242437, "grad_norm": 1.1092537387573012, "learning_rate": 8.077251344238068e-06, "loss": 0.6353, "step": 1877 }, { "epoch": 1.802688430148824, "grad_norm": 0.9496518514020257, "learning_rate": 8.074610296486771e-06, "loss": 0.5178, "step": 1878 }, { "epoch": 1.8036485837734038, "grad_norm": 0.9996544657295613, "learning_rate": 8.071967868575775e-06, "loss": 0.6676, "step": 1879 }, { "epoch": 1.8046087373979836, "grad_norm": 0.9208545810655526, "learning_rate": 8.069324061691242e-06, "loss": 0.6769, "step": 1880 }, { "epoch": 1.8055688910225636, "grad_norm": 0.8610504232139199, "learning_rate": 8.066678877019945e-06, "loss": 0.6389, "step": 1881 }, { "epoch": 1.8065290446471436, "grad_norm": 1.3134826751753745, "learning_rate": 8.064032315749283e-06, "loss": 0.6054, "step": 1882 }, { "epoch": 1.8074891982717234, "grad_norm": 0.9047762112835521, "learning_rate": 8.061384379067266e-06, "loss": 0.5767, "step": 1883 }, { "epoch": 1.8084493518963034, "grad_norm": 0.6673207258483606, "learning_rate": 8.058735068162527e-06, "loss": 0.5287, "step": 1884 }, { "epoch": 1.8094095055208834, "grad_norm": 0.772193364527559, "learning_rate": 8.05608438422431e-06, "loss": 0.5055, "step": 1885 }, { "epoch": 1.8103696591454632, "grad_norm": 0.7641990012136538, "learning_rate": 8.053432328442483e-06, "loss": 0.5489, "step": 1886 }, { "epoch": 1.8113298127700432, "grad_norm": 1.1113194097020602, "learning_rate": 8.05077890200752e-06, "loss": 0.6273, "step": 1887 }, { "epoch": 1.8122899663946233, "grad_norm": 0.8880017463620019, "learning_rate": 8.04812410611052e-06, "loss": 0.6144, "step": 1888 }, { "epoch": 1.813250120019203, "grad_norm": 0.8682629105731722, "learning_rate": 8.045467941943193e-06, "loss": 0.6296, "step": 1889 }, { "epoch": 1.8142102736437828, "grad_norm": 1.0173721245282328, "learning_rate": 8.042810410697861e-06, "loss": 0.6244, "step": 1890 }, { "epoch": 1.815170427268363, "grad_norm": 0.8331869149620711, "learning_rate": 8.04015151356746e-06, "loss": 0.5865, "step": 1891 }, { "epoch": 1.8161305808929429, "grad_norm": 0.7206455046929106, "learning_rate": 8.037491251745543e-06, "loss": 0.6525, "step": 1892 }, { "epoch": 1.8170907345175227, "grad_norm": 0.953351224641206, "learning_rate": 8.034829626426273e-06, "loss": 0.5604, "step": 1893 }, { "epoch": 1.8180508881421027, "grad_norm": 0.7211175479318865, "learning_rate": 8.032166638804422e-06, "loss": 0.6239, "step": 1894 }, { "epoch": 1.8190110417666827, "grad_norm": 0.8124945119128172, "learning_rate": 8.02950229007538e-06, "loss": 0.6377, "step": 1895 }, { "epoch": 1.8199711953912625, "grad_norm": 0.8167904969985839, "learning_rate": 8.026836581435142e-06, "loss": 0.6196, "step": 1896 }, { "epoch": 1.8209313490158425, "grad_norm": 1.0576799667522088, "learning_rate": 8.02416951408032e-06, "loss": 0.7303, "step": 1897 }, { "epoch": 1.8218915026404225, "grad_norm": 0.8550540311290219, "learning_rate": 8.02150108920813e-06, "loss": 0.6178, "step": 1898 }, { "epoch": 1.8228516562650023, "grad_norm": 0.7768541593621361, "learning_rate": 8.018831308016398e-06, "loss": 0.6265, "step": 1899 }, { "epoch": 1.8238118098895824, "grad_norm": 1.0508486877362824, "learning_rate": 8.01616017170356e-06, "loss": 0.6642, "step": 1900 }, { "epoch": 1.8247719635141624, "grad_norm": 0.8070004418867449, "learning_rate": 8.013487681468663e-06, "loss": 0.5749, "step": 1901 }, { "epoch": 1.8257321171387422, "grad_norm": 0.9115257639210748, "learning_rate": 8.010813838511357e-06, "loss": 0.5952, "step": 1902 }, { "epoch": 1.826692270763322, "grad_norm": 1.0083454662678875, "learning_rate": 8.008138644031906e-06, "loss": 0.6767, "step": 1903 }, { "epoch": 1.8276524243879022, "grad_norm": 0.9545105885097894, "learning_rate": 8.005462099231171e-06, "loss": 0.5091, "step": 1904 }, { "epoch": 1.828612578012482, "grad_norm": 0.8613874328136053, "learning_rate": 8.002784205310626e-06, "loss": 0.6417, "step": 1905 }, { "epoch": 1.8295727316370618, "grad_norm": 0.6274743007674297, "learning_rate": 8.00010496347235e-06, "loss": 0.6659, "step": 1906 }, { "epoch": 1.8305328852616418, "grad_norm": 1.0191180782249167, "learning_rate": 7.997424374919024e-06, "loss": 0.6484, "step": 1907 }, { "epoch": 1.8314930388862218, "grad_norm": 0.6690741720205372, "learning_rate": 7.994742440853937e-06, "loss": 0.6616, "step": 1908 }, { "epoch": 1.8324531925108016, "grad_norm": 0.8543977602524203, "learning_rate": 7.992059162480983e-06, "loss": 0.6567, "step": 1909 }, { "epoch": 1.8334133461353816, "grad_norm": 0.8552536798598672, "learning_rate": 7.98937454100465e-06, "loss": 0.6526, "step": 1910 }, { "epoch": 1.8343734997599617, "grad_norm": 0.6507995402633151, "learning_rate": 7.986688577630042e-06, "loss": 0.628, "step": 1911 }, { "epoch": 1.8353336533845415, "grad_norm": 0.8170030976976136, "learning_rate": 7.984001273562858e-06, "loss": 0.6267, "step": 1912 }, { "epoch": 1.8362938070091215, "grad_norm": 1.3053848351980555, "learning_rate": 7.981312630009398e-06, "loss": 0.7608, "step": 1913 }, { "epoch": 1.8372539606337015, "grad_norm": 1.210904717505064, "learning_rate": 7.97862264817657e-06, "loss": 0.5564, "step": 1914 }, { "epoch": 1.8382141142582813, "grad_norm": 0.9617384901288306, "learning_rate": 7.975931329271874e-06, "loss": 0.574, "step": 1915 }, { "epoch": 1.839174267882861, "grad_norm": 0.9692968533072517, "learning_rate": 7.973238674503412e-06, "loss": 0.6077, "step": 1916 }, { "epoch": 1.8401344215074413, "grad_norm": 1.1338048863083767, "learning_rate": 7.970544685079895e-06, "loss": 0.6843, "step": 1917 }, { "epoch": 1.8410945751320211, "grad_norm": 0.5451523744010679, "learning_rate": 7.96784936221062e-06, "loss": 0.6185, "step": 1918 }, { "epoch": 1.842054728756601, "grad_norm": 0.9226152426252683, "learning_rate": 7.965152707105491e-06, "loss": 0.579, "step": 1919 }, { "epoch": 1.843014882381181, "grad_norm": 0.9628452489787511, "learning_rate": 7.962454720975008e-06, "loss": 0.5414, "step": 1920 }, { "epoch": 1.843975036005761, "grad_norm": 1.0150578099513723, "learning_rate": 7.959755405030269e-06, "loss": 0.5742, "step": 1921 }, { "epoch": 1.8449351896303408, "grad_norm": 0.6343660162215211, "learning_rate": 7.957054760482964e-06, "loss": 0.5308, "step": 1922 }, { "epoch": 1.8458953432549208, "grad_norm": 0.8916524365186375, "learning_rate": 7.954352788545388e-06, "loss": 0.6555, "step": 1923 }, { "epoch": 1.8468554968795008, "grad_norm": 0.7540701874596591, "learning_rate": 7.951649490430424e-06, "loss": 0.5805, "step": 1924 }, { "epoch": 1.8478156505040806, "grad_norm": 1.1462886242377603, "learning_rate": 7.948944867351556e-06, "loss": 0.5869, "step": 1925 }, { "epoch": 1.8487758041286606, "grad_norm": 1.1924749362761624, "learning_rate": 7.946238920522862e-06, "loss": 0.7007, "step": 1926 }, { "epoch": 1.8497359577532406, "grad_norm": 0.929335754803921, "learning_rate": 7.943531651159007e-06, "loss": 0.5868, "step": 1927 }, { "epoch": 1.8506961113778204, "grad_norm": 1.0260627318956923, "learning_rate": 7.940823060475257e-06, "loss": 0.626, "step": 1928 }, { "epoch": 1.8516562650024004, "grad_norm": 0.8369868062264989, "learning_rate": 7.938113149687474e-06, "loss": 0.676, "step": 1929 }, { "epoch": 1.8526164186269805, "grad_norm": 1.166192668128317, "learning_rate": 7.935401920012105e-06, "loss": 0.6383, "step": 1930 }, { "epoch": 1.8535765722515603, "grad_norm": 0.9803809462364839, "learning_rate": 7.932689372666192e-06, "loss": 0.6612, "step": 1931 }, { "epoch": 1.85453672587614, "grad_norm": 0.995937822370504, "learning_rate": 7.929975508867364e-06, "loss": 0.6974, "step": 1932 }, { "epoch": 1.85549687950072, "grad_norm": 0.7790416829511575, "learning_rate": 7.927260329833854e-06, "loss": 0.5569, "step": 1933 }, { "epoch": 1.8564570331253, "grad_norm": 0.8983196881158549, "learning_rate": 7.924543836784471e-06, "loss": 0.6736, "step": 1934 }, { "epoch": 1.8574171867498799, "grad_norm": 1.001190651510258, "learning_rate": 7.921826030938623e-06, "loss": 0.596, "step": 1935 }, { "epoch": 1.85837734037446, "grad_norm": 1.2401998634118503, "learning_rate": 7.919106913516299e-06, "loss": 0.5965, "step": 1936 }, { "epoch": 1.85933749399904, "grad_norm": 1.1432582517076924, "learning_rate": 7.916386485738089e-06, "loss": 0.7684, "step": 1937 }, { "epoch": 1.8602976476236197, "grad_norm": 1.038102567947624, "learning_rate": 7.913664748825156e-06, "loss": 0.6517, "step": 1938 }, { "epoch": 1.8612578012481997, "grad_norm": 1.127833622527371, "learning_rate": 7.910941703999265e-06, "loss": 0.6987, "step": 1939 }, { "epoch": 1.8622179548727797, "grad_norm": 0.9546239488138775, "learning_rate": 7.908217352482759e-06, "loss": 0.5502, "step": 1940 }, { "epoch": 1.8631781084973595, "grad_norm": 0.8421076780871938, "learning_rate": 7.905491695498571e-06, "loss": 0.516, "step": 1941 }, { "epoch": 1.8641382621219396, "grad_norm": 0.8617972234475988, "learning_rate": 7.902764734270219e-06, "loss": 0.6196, "step": 1942 }, { "epoch": 1.8650984157465196, "grad_norm": 0.8877788996506157, "learning_rate": 7.900036470021805e-06, "loss": 0.5935, "step": 1943 }, { "epoch": 1.8660585693710994, "grad_norm": 0.7155851128005509, "learning_rate": 7.89730690397802e-06, "loss": 0.54, "step": 1944 }, { "epoch": 1.8670187229956792, "grad_norm": 0.7768358234083267, "learning_rate": 7.894576037364136e-06, "loss": 0.5708, "step": 1945 }, { "epoch": 1.8679788766202592, "grad_norm": 1.0943032311191803, "learning_rate": 7.89184387140601e-06, "loss": 0.6259, "step": 1946 }, { "epoch": 1.8689390302448392, "grad_norm": 0.8120826850279467, "learning_rate": 7.889110407330083e-06, "loss": 0.6853, "step": 1947 }, { "epoch": 1.869899183869419, "grad_norm": 1.2077321057308066, "learning_rate": 7.886375646363378e-06, "loss": 0.7487, "step": 1948 }, { "epoch": 1.870859337493999, "grad_norm": 1.3588621005683725, "learning_rate": 7.883639589733498e-06, "loss": 0.681, "step": 1949 }, { "epoch": 1.871819491118579, "grad_norm": 1.0921909950445614, "learning_rate": 7.880902238668631e-06, "loss": 0.6644, "step": 1950 }, { "epoch": 1.8727796447431588, "grad_norm": 0.8788628608306976, "learning_rate": 7.878163594397543e-06, "loss": 0.5389, "step": 1951 }, { "epoch": 1.8737397983677389, "grad_norm": 0.8927504751676186, "learning_rate": 7.875423658149583e-06, "loss": 0.7776, "step": 1952 }, { "epoch": 1.8746999519923189, "grad_norm": 0.8712767572295782, "learning_rate": 7.87268243115468e-06, "loss": 0.5927, "step": 1953 }, { "epoch": 1.8756601056168987, "grad_norm": 0.8321068817677691, "learning_rate": 7.869939914643343e-06, "loss": 0.7287, "step": 1954 }, { "epoch": 1.8766202592414787, "grad_norm": 0.9411183027501857, "learning_rate": 7.867196109846653e-06, "loss": 0.5747, "step": 1955 }, { "epoch": 1.8775804128660587, "grad_norm": 0.7186427431575353, "learning_rate": 7.86445101799628e-06, "loss": 0.5847, "step": 1956 }, { "epoch": 1.8785405664906385, "grad_norm": 0.862079889131208, "learning_rate": 7.861704640324466e-06, "loss": 0.6688, "step": 1957 }, { "epoch": 1.8795007201152183, "grad_norm": 1.2058582401837974, "learning_rate": 7.858956978064025e-06, "loss": 0.7939, "step": 1958 }, { "epoch": 1.8804608737397985, "grad_norm": 0.9167834309877442, "learning_rate": 7.85620803244836e-06, "loss": 0.646, "step": 1959 }, { "epoch": 1.8814210273643783, "grad_norm": 0.9049180507115854, "learning_rate": 7.85345780471144e-06, "loss": 0.5838, "step": 1960 }, { "epoch": 1.8823811809889581, "grad_norm": 0.8825621119403076, "learning_rate": 7.850706296087811e-06, "loss": 0.5738, "step": 1961 }, { "epoch": 1.8833413346135381, "grad_norm": 0.7822401296935878, "learning_rate": 7.8479535078126e-06, "loss": 0.6312, "step": 1962 }, { "epoch": 1.8843014882381182, "grad_norm": 1.0508084878507158, "learning_rate": 7.8451994411215e-06, "loss": 0.593, "step": 1963 }, { "epoch": 1.885261641862698, "grad_norm": 0.8592355163718516, "learning_rate": 7.842444097250787e-06, "loss": 0.6083, "step": 1964 }, { "epoch": 1.886221795487278, "grad_norm": 0.8092310890490754, "learning_rate": 7.839687477437304e-06, "loss": 0.5776, "step": 1965 }, { "epoch": 1.887181949111858, "grad_norm": 0.9808215080257987, "learning_rate": 7.836929582918464e-06, "loss": 0.6498, "step": 1966 }, { "epoch": 1.8881421027364378, "grad_norm": 0.9509854633863327, "learning_rate": 7.834170414932259e-06, "loss": 0.4972, "step": 1967 }, { "epoch": 1.8891022563610178, "grad_norm": 0.9236989238521253, "learning_rate": 7.831409974717252e-06, "loss": 0.5969, "step": 1968 }, { "epoch": 1.8900624099855978, "grad_norm": 0.9635468812392368, "learning_rate": 7.828648263512573e-06, "loss": 0.6135, "step": 1969 }, { "epoch": 1.8910225636101776, "grad_norm": 1.015913845498851, "learning_rate": 7.825885282557926e-06, "loss": 0.6639, "step": 1970 }, { "epoch": 1.8919827172347574, "grad_norm": 0.9088176721665507, "learning_rate": 7.823121033093581e-06, "loss": 0.5623, "step": 1971 }, { "epoch": 1.8929428708593377, "grad_norm": 0.8821601286282678, "learning_rate": 7.820355516360383e-06, "loss": 0.6314, "step": 1972 }, { "epoch": 1.8939030244839175, "grad_norm": 0.9949704447465338, "learning_rate": 7.817588733599742e-06, "loss": 0.6589, "step": 1973 }, { "epoch": 1.8948631781084972, "grad_norm": 0.9173934571111674, "learning_rate": 7.814820686053634e-06, "loss": 0.7043, "step": 1974 }, { "epoch": 1.8958233317330773, "grad_norm": 0.8224460751427073, "learning_rate": 7.812051374964611e-06, "loss": 0.6051, "step": 1975 }, { "epoch": 1.8967834853576573, "grad_norm": 0.9862247697383825, "learning_rate": 7.809280801575784e-06, "loss": 0.6787, "step": 1976 }, { "epoch": 1.897743638982237, "grad_norm": 0.6941269932251598, "learning_rate": 7.806508967130838e-06, "loss": 0.7478, "step": 1977 }, { "epoch": 1.898703792606817, "grad_norm": 0.7874487120932511, "learning_rate": 7.803735872874012e-06, "loss": 0.5921, "step": 1978 }, { "epoch": 1.8996639462313971, "grad_norm": 0.9593184186937603, "learning_rate": 7.800961520050125e-06, "loss": 0.655, "step": 1979 }, { "epoch": 1.900624099855977, "grad_norm": 0.9099673431290162, "learning_rate": 7.798185909904552e-06, "loss": 0.7259, "step": 1980 }, { "epoch": 1.901584253480557, "grad_norm": 0.8858398240818072, "learning_rate": 7.795409043683237e-06, "loss": 0.6003, "step": 1981 }, { "epoch": 1.902544407105137, "grad_norm": 1.1769693317074694, "learning_rate": 7.792630922632682e-06, "loss": 0.5932, "step": 1982 }, { "epoch": 1.9035045607297167, "grad_norm": 0.9300028472682665, "learning_rate": 7.789851547999957e-06, "loss": 0.4654, "step": 1983 }, { "epoch": 1.9044647143542965, "grad_norm": 0.6584156321813891, "learning_rate": 7.787070921032693e-06, "loss": 0.5807, "step": 1984 }, { "epoch": 1.9054248679788768, "grad_norm": 1.0419509897686148, "learning_rate": 7.784289042979085e-06, "loss": 0.692, "step": 1985 }, { "epoch": 1.9063850216034566, "grad_norm": 0.8525830001425926, "learning_rate": 7.781505915087891e-06, "loss": 0.6657, "step": 1986 }, { "epoch": 1.9073451752280364, "grad_norm": 0.8970703623176188, "learning_rate": 7.778721538608422e-06, "loss": 0.7161, "step": 1987 }, { "epoch": 1.9083053288526164, "grad_norm": 0.9726696896046253, "learning_rate": 7.775935914790557e-06, "loss": 0.5935, "step": 1988 }, { "epoch": 1.9092654824771964, "grad_norm": 0.8326676110073934, "learning_rate": 7.773149044884732e-06, "loss": 0.7149, "step": 1989 }, { "epoch": 1.9102256361017762, "grad_norm": 1.1652210787647013, "learning_rate": 7.770360930141946e-06, "loss": 0.6701, "step": 1990 }, { "epoch": 1.9111857897263562, "grad_norm": 0.986141048397407, "learning_rate": 7.767571571813752e-06, "loss": 0.6278, "step": 1991 }, { "epoch": 1.9121459433509362, "grad_norm": 0.6463897652151377, "learning_rate": 7.764780971152262e-06, "loss": 0.5156, "step": 1992 }, { "epoch": 1.913106096975516, "grad_norm": 1.0207173018849236, "learning_rate": 7.761989129410149e-06, "loss": 0.6799, "step": 1993 }, { "epoch": 1.914066250600096, "grad_norm": 0.8744673399609925, "learning_rate": 7.759196047840639e-06, "loss": 0.5516, "step": 1994 }, { "epoch": 1.915026404224676, "grad_norm": 0.8835770152678148, "learning_rate": 7.756401727697518e-06, "loss": 0.6347, "step": 1995 }, { "epoch": 1.9159865578492559, "grad_norm": 0.9085317010118922, "learning_rate": 7.753606170235124e-06, "loss": 0.6569, "step": 1996 }, { "epoch": 1.9169467114738357, "grad_norm": 1.1432888248533526, "learning_rate": 7.750809376708357e-06, "loss": 0.6359, "step": 1997 }, { "epoch": 1.917906865098416, "grad_norm": 0.7231217733470161, "learning_rate": 7.748011348372663e-06, "loss": 0.6102, "step": 1998 }, { "epoch": 1.9188670187229957, "grad_norm": 0.660531747776664, "learning_rate": 7.74521208648405e-06, "loss": 0.6881, "step": 1999 }, { "epoch": 1.9198271723475755, "grad_norm": 1.3057933739221608, "learning_rate": 7.742411592299077e-06, "loss": 0.6124, "step": 2000 }, { "epoch": 1.9207873259721555, "grad_norm": 0.7493070277054039, "learning_rate": 7.739609867074856e-06, "loss": 0.6677, "step": 2001 }, { "epoch": 1.9217474795967355, "grad_norm": 0.9814391204134008, "learning_rate": 7.736806912069048e-06, "loss": 0.6635, "step": 2002 }, { "epoch": 1.9227076332213153, "grad_norm": 0.8549823217736486, "learning_rate": 7.734002728539873e-06, "loss": 0.611, "step": 2003 }, { "epoch": 1.9236677868458953, "grad_norm": 0.8811659393472916, "learning_rate": 7.731197317746099e-06, "loss": 0.6484, "step": 2004 }, { "epoch": 1.9246279404704754, "grad_norm": 0.9278052402415213, "learning_rate": 7.728390680947043e-06, "loss": 0.6125, "step": 2005 }, { "epoch": 1.9255880940950552, "grad_norm": 0.7508633719960072, "learning_rate": 7.725582819402575e-06, "loss": 0.6966, "step": 2006 }, { "epoch": 1.9265482477196352, "grad_norm": 0.7828767537016417, "learning_rate": 7.722773734373114e-06, "loss": 0.5919, "step": 2007 }, { "epoch": 1.9275084013442152, "grad_norm": 1.0285242591513555, "learning_rate": 7.71996342711963e-06, "loss": 0.6015, "step": 2008 }, { "epoch": 1.928468554968795, "grad_norm": 0.6708537686970968, "learning_rate": 7.717151898903636e-06, "loss": 0.5813, "step": 2009 }, { "epoch": 1.9294287085933748, "grad_norm": 0.929554336690013, "learning_rate": 7.714339150987202e-06, "loss": 0.6617, "step": 2010 }, { "epoch": 1.930388862217955, "grad_norm": 1.0790732356261517, "learning_rate": 7.711525184632936e-06, "loss": 0.7555, "step": 2011 }, { "epoch": 1.9313490158425348, "grad_norm": 0.9889792336004267, "learning_rate": 7.708710001103998e-06, "loss": 0.5573, "step": 2012 }, { "epoch": 1.9323091694671146, "grad_norm": 0.7439096259670189, "learning_rate": 7.705893601664099e-06, "loss": 0.6654, "step": 2013 }, { "epoch": 1.9332693230916946, "grad_norm": 1.0459699459112652, "learning_rate": 7.703075987577483e-06, "loss": 0.71, "step": 2014 }, { "epoch": 1.9342294767162747, "grad_norm": 1.5309420939577785, "learning_rate": 7.700257160108953e-06, "loss": 0.6374, "step": 2015 }, { "epoch": 1.9351896303408544, "grad_norm": 1.087191818408777, "learning_rate": 7.697437120523845e-06, "loss": 0.6608, "step": 2016 }, { "epoch": 1.9361497839654345, "grad_norm": 0.7596383446651823, "learning_rate": 7.694615870088051e-06, "loss": 0.7033, "step": 2017 }, { "epoch": 1.9371099375900145, "grad_norm": 1.4335138465256476, "learning_rate": 7.691793410067993e-06, "loss": 0.6816, "step": 2018 }, { "epoch": 1.9380700912145943, "grad_norm": 1.3366943314980546, "learning_rate": 7.68896974173065e-06, "loss": 0.6753, "step": 2019 }, { "epoch": 1.9390302448391743, "grad_norm": 1.0976121426216767, "learning_rate": 7.686144866343533e-06, "loss": 0.5043, "step": 2020 }, { "epoch": 1.9399903984637543, "grad_norm": 0.7444643553829531, "learning_rate": 7.683318785174698e-06, "loss": 0.6333, "step": 2021 }, { "epoch": 1.9409505520883341, "grad_norm": 1.004129579216168, "learning_rate": 7.680491499492743e-06, "loss": 0.5213, "step": 2022 }, { "epoch": 1.941910705712914, "grad_norm": 0.6332859563607501, "learning_rate": 7.677663010566807e-06, "loss": 0.5522, "step": 2023 }, { "epoch": 1.9428708593374941, "grad_norm": 0.9216162469281932, "learning_rate": 7.674833319666568e-06, "loss": 0.6607, "step": 2024 }, { "epoch": 1.943831012962074, "grad_norm": 0.8530290895544359, "learning_rate": 7.672002428062245e-06, "loss": 0.6011, "step": 2025 }, { "epoch": 1.9447911665866537, "grad_norm": 0.8171244040123942, "learning_rate": 7.669170337024593e-06, "loss": 0.6783, "step": 2026 }, { "epoch": 1.9457513202112338, "grad_norm": 0.7617586901042728, "learning_rate": 7.66633704782491e-06, "loss": 0.6099, "step": 2027 }, { "epoch": 1.9467114738358138, "grad_norm": 1.0619657573764127, "learning_rate": 7.663502561735028e-06, "loss": 0.7876, "step": 2028 }, { "epoch": 1.9476716274603936, "grad_norm": 1.0230982711513525, "learning_rate": 7.660666880027316e-06, "loss": 0.6329, "step": 2029 }, { "epoch": 1.9486317810849736, "grad_norm": 0.7852172354373707, "learning_rate": 7.657830003974684e-06, "loss": 0.5559, "step": 2030 }, { "epoch": 1.9495919347095536, "grad_norm": 0.9269256683602936, "learning_rate": 7.654991934850573e-06, "loss": 0.5673, "step": 2031 }, { "epoch": 1.9505520883341334, "grad_norm": 0.843650232371707, "learning_rate": 7.652152673928964e-06, "loss": 0.6354, "step": 2032 }, { "epoch": 1.9515122419587134, "grad_norm": 1.1612508001346127, "learning_rate": 7.649312222484367e-06, "loss": 0.7032, "step": 2033 }, { "epoch": 1.9524723955832934, "grad_norm": 1.24201474352753, "learning_rate": 7.646470581791834e-06, "loss": 0.6985, "step": 2034 }, { "epoch": 1.9534325492078732, "grad_norm": 1.3147079498099148, "learning_rate": 7.643627753126946e-06, "loss": 0.6139, "step": 2035 }, { "epoch": 1.954392702832453, "grad_norm": 0.8248393338163675, "learning_rate": 7.640783737765817e-06, "loss": 0.6101, "step": 2036 }, { "epoch": 1.9553528564570333, "grad_norm": 0.8867457028992051, "learning_rate": 7.6379385369851e-06, "loss": 0.6586, "step": 2037 }, { "epoch": 1.956313010081613, "grad_norm": 1.0257296124012056, "learning_rate": 7.63509215206197e-06, "loss": 0.6088, "step": 2038 }, { "epoch": 1.9572731637061929, "grad_norm": 1.0214509376297933, "learning_rate": 7.63224458427414e-06, "loss": 0.6562, "step": 2039 }, { "epoch": 1.9582333173307729, "grad_norm": 0.9431014047348625, "learning_rate": 7.629395834899852e-06, "loss": 0.6551, "step": 2040 }, { "epoch": 1.959193470955353, "grad_norm": 0.8129492651144443, "learning_rate": 7.626545905217884e-06, "loss": 0.6636, "step": 2041 }, { "epoch": 1.9601536245799327, "grad_norm": 0.9428832010577324, "learning_rate": 7.6236947965075325e-06, "loss": 0.6631, "step": 2042 }, { "epoch": 1.9611137782045127, "grad_norm": 0.8899440361392587, "learning_rate": 7.6208425100486334e-06, "loss": 0.4846, "step": 2043 }, { "epoch": 1.9620739318290927, "grad_norm": 0.9530451753576852, "learning_rate": 7.617989047121548e-06, "loss": 0.6187, "step": 2044 }, { "epoch": 1.9630340854536725, "grad_norm": 0.7524458697819769, "learning_rate": 7.6151344090071624e-06, "loss": 0.5431, "step": 2045 }, { "epoch": 1.9639942390782525, "grad_norm": 0.7315972863465067, "learning_rate": 7.612278596986895e-06, "loss": 0.5519, "step": 2046 }, { "epoch": 1.9649543927028326, "grad_norm": 1.0393541734956016, "learning_rate": 7.609421612342689e-06, "loss": 0.5928, "step": 2047 }, { "epoch": 1.9659145463274124, "grad_norm": 1.017146338482968, "learning_rate": 7.606563456357016e-06, "loss": 0.6977, "step": 2048 }, { "epoch": 1.9668746999519922, "grad_norm": 0.9558613875666172, "learning_rate": 7.603704130312868e-06, "loss": 0.7088, "step": 2049 }, { "epoch": 1.9678348535765724, "grad_norm": 1.0419164698092818, "learning_rate": 7.6008436354937695e-06, "loss": 0.6781, "step": 2050 }, { "epoch": 1.9687950072011522, "grad_norm": 0.8263691601521463, "learning_rate": 7.5979819731837635e-06, "loss": 0.6303, "step": 2051 }, { "epoch": 1.969755160825732, "grad_norm": 0.8880967981996022, "learning_rate": 7.595119144667421e-06, "loss": 0.6627, "step": 2052 }, { "epoch": 1.970715314450312, "grad_norm": 0.8138022056222668, "learning_rate": 7.5922551512298345e-06, "loss": 0.5556, "step": 2053 }, { "epoch": 1.971675468074892, "grad_norm": 0.7852022764945256, "learning_rate": 7.589389994156622e-06, "loss": 0.7877, "step": 2054 }, { "epoch": 1.9726356216994718, "grad_norm": 0.8994337684627339, "learning_rate": 7.586523674733918e-06, "loss": 0.606, "step": 2055 }, { "epoch": 1.9735957753240518, "grad_norm": 0.8323716191561732, "learning_rate": 7.583656194248386e-06, "loss": 0.6238, "step": 2056 }, { "epoch": 1.9745559289486319, "grad_norm": 1.2244033399320309, "learning_rate": 7.5807875539872054e-06, "loss": 0.6958, "step": 2057 }, { "epoch": 1.9755160825732117, "grad_norm": 1.2862333536465442, "learning_rate": 7.577917755238079e-06, "loss": 0.7279, "step": 2058 }, { "epoch": 1.9764762361977917, "grad_norm": 0.8390603142713938, "learning_rate": 7.57504679928923e-06, "loss": 0.7119, "step": 2059 }, { "epoch": 1.9774363898223717, "grad_norm": 0.8828609096341173, "learning_rate": 7.572174687429395e-06, "loss": 0.6371, "step": 2060 }, { "epoch": 1.9783965434469515, "grad_norm": 0.7066800922035822, "learning_rate": 7.569301420947841e-06, "loss": 0.7256, "step": 2061 }, { "epoch": 1.9793566970715313, "grad_norm": 0.7545224637464156, "learning_rate": 7.566427001134342e-06, "loss": 0.6563, "step": 2062 }, { "epoch": 1.9803168506961115, "grad_norm": 0.9322557623401021, "learning_rate": 7.563551429279198e-06, "loss": 0.61, "step": 2063 }, { "epoch": 1.9812770043206913, "grad_norm": 0.950739376561709, "learning_rate": 7.560674706673219e-06, "loss": 0.6508, "step": 2064 }, { "epoch": 1.982237157945271, "grad_norm": 1.2877424283027523, "learning_rate": 7.557796834607736e-06, "loss": 0.886, "step": 2065 }, { "epoch": 1.9831973115698511, "grad_norm": 0.9761205332583609, "learning_rate": 7.554917814374595e-06, "loss": 0.5689, "step": 2066 }, { "epoch": 1.9841574651944311, "grad_norm": 0.870833821805864, "learning_rate": 7.552037647266157e-06, "loss": 0.78, "step": 2067 }, { "epoch": 1.985117618819011, "grad_norm": 1.0836240017815606, "learning_rate": 7.5491563345753e-06, "loss": 0.6305, "step": 2068 }, { "epoch": 1.986077772443591, "grad_norm": 0.8216282900489478, "learning_rate": 7.546273877595413e-06, "loss": 0.5955, "step": 2069 }, { "epoch": 1.987037926068171, "grad_norm": 0.7883366881722336, "learning_rate": 7.5433902776204015e-06, "loss": 0.6055, "step": 2070 }, { "epoch": 1.9879980796927508, "grad_norm": 0.6999008964429867, "learning_rate": 7.54050553594468e-06, "loss": 0.6111, "step": 2071 }, { "epoch": 1.9889582333173308, "grad_norm": 1.0159234714359926, "learning_rate": 7.537619653863182e-06, "loss": 0.6221, "step": 2072 }, { "epoch": 1.9899183869419108, "grad_norm": 0.5482281958820241, "learning_rate": 7.5347326326713454e-06, "loss": 0.6192, "step": 2073 }, { "epoch": 1.9908785405664906, "grad_norm": 1.0541225922860435, "learning_rate": 7.531844473665125e-06, "loss": 0.6774, "step": 2074 }, { "epoch": 1.9918386941910706, "grad_norm": 1.125709703168444, "learning_rate": 7.528955178140985e-06, "loss": 0.5405, "step": 2075 }, { "epoch": 1.9927988478156506, "grad_norm": 0.7889115070185686, "learning_rate": 7.526064747395898e-06, "loss": 0.6675, "step": 2076 }, { "epoch": 1.9937590014402304, "grad_norm": 0.9286942062413462, "learning_rate": 7.523173182727348e-06, "loss": 0.5757, "step": 2077 }, { "epoch": 1.9947191550648102, "grad_norm": 0.6637918948394058, "learning_rate": 7.520280485433327e-06, "loss": 0.5832, "step": 2078 }, { "epoch": 1.9956793086893903, "grad_norm": 0.7849877300670812, "learning_rate": 7.517386656812338e-06, "loss": 0.727, "step": 2079 }, { "epoch": 1.9966394623139703, "grad_norm": 1.1146165558987766, "learning_rate": 7.514491698163387e-06, "loss": 0.7178, "step": 2080 }, { "epoch": 1.99759961593855, "grad_norm": 0.8888348404077672, "learning_rate": 7.511595610785991e-06, "loss": 0.6351, "step": 2081 }, { "epoch": 1.99855976956313, "grad_norm": 1.1976041884694015, "learning_rate": 7.508698395980173e-06, "loss": 0.5697, "step": 2082 }, { "epoch": 1.99951992318771, "grad_norm": 1.2628319129533099, "learning_rate": 7.5058000550464615e-06, "loss": 0.8398, "step": 2083 }, { "epoch": 2.0, "grad_norm": 1.931543366467905, "learning_rate": 7.502900589285891e-06, "loss": 0.6745, "step": 2084 }, { "epoch": 2.00096015362458, "grad_norm": 1.1180056154934426, "learning_rate": 7.500000000000001e-06, "loss": 0.5387, "step": 2085 }, { "epoch": 2.00192030724916, "grad_norm": 0.7730065479119468, "learning_rate": 7.497098288490834e-06, "loss": 0.5526, "step": 2086 }, { "epoch": 2.00288046087374, "grad_norm": 1.0866516313919004, "learning_rate": 7.494195456060938e-06, "loss": 0.5759, "step": 2087 }, { "epoch": 2.0038406144983196, "grad_norm": 1.8397315190715802, "learning_rate": 7.491291504013363e-06, "loss": 0.5983, "step": 2088 }, { "epoch": 2.0048007681229, "grad_norm": 0.776105850272823, "learning_rate": 7.488386433651662e-06, "loss": 0.5189, "step": 2089 }, { "epoch": 2.0057609217474797, "grad_norm": 0.7126494004031685, "learning_rate": 7.4854802462798905e-06, "loss": 0.4941, "step": 2090 }, { "epoch": 2.0067210753720595, "grad_norm": 1.1397895783571348, "learning_rate": 7.482572943202604e-06, "loss": 0.5848, "step": 2091 }, { "epoch": 2.0076812289966393, "grad_norm": 0.8321450397090203, "learning_rate": 7.4796645257248625e-06, "loss": 0.5532, "step": 2092 }, { "epoch": 2.0086413826212195, "grad_norm": 0.8435558682422999, "learning_rate": 7.47675499515222e-06, "loss": 0.6158, "step": 2093 }, { "epoch": 2.0096015362457993, "grad_norm": 1.2215648192265838, "learning_rate": 7.473844352790735e-06, "loss": 0.6214, "step": 2094 }, { "epoch": 2.010561689870379, "grad_norm": 0.9424757524520742, "learning_rate": 7.470932599946965e-06, "loss": 0.5582, "step": 2095 }, { "epoch": 2.0115218434949593, "grad_norm": 0.871410681023748, "learning_rate": 7.468019737927963e-06, "loss": 0.5603, "step": 2096 }, { "epoch": 2.012481997119539, "grad_norm": 0.8562787581640613, "learning_rate": 7.465105768041283e-06, "loss": 0.5108, "step": 2097 }, { "epoch": 2.013442150744119, "grad_norm": 0.8046341380987055, "learning_rate": 7.462190691594973e-06, "loss": 0.59, "step": 2098 }, { "epoch": 2.014402304368699, "grad_norm": 1.017525274323804, "learning_rate": 7.459274509897583e-06, "loss": 0.5555, "step": 2099 }, { "epoch": 2.015362457993279, "grad_norm": 1.1326547210109956, "learning_rate": 7.45635722425815e-06, "loss": 0.5437, "step": 2100 }, { "epoch": 2.0163226116178588, "grad_norm": 0.7221752238092628, "learning_rate": 7.45343883598622e-06, "loss": 0.5284, "step": 2101 }, { "epoch": 2.017282765242439, "grad_norm": 0.8799212226885517, "learning_rate": 7.45051934639182e-06, "loss": 0.4749, "step": 2102 }, { "epoch": 2.018242918867019, "grad_norm": 0.8133772576516266, "learning_rate": 7.447598756785482e-06, "loss": 0.5414, "step": 2103 }, { "epoch": 2.0192030724915986, "grad_norm": 1.390902349323953, "learning_rate": 7.444677068478223e-06, "loss": 0.6227, "step": 2104 }, { "epoch": 2.0201632261161784, "grad_norm": 0.828238810882178, "learning_rate": 7.441754282781561e-06, "loss": 0.6078, "step": 2105 }, { "epoch": 2.0211233797407586, "grad_norm": 0.8553585379339558, "learning_rate": 7.438830401007504e-06, "loss": 0.5715, "step": 2106 }, { "epoch": 2.0220835333653384, "grad_norm": 1.0236484837693596, "learning_rate": 7.435905424468548e-06, "loss": 0.5513, "step": 2107 }, { "epoch": 2.023043686989918, "grad_norm": 0.826953754989398, "learning_rate": 7.432979354477686e-06, "loss": 0.5076, "step": 2108 }, { "epoch": 2.0240038406144985, "grad_norm": 0.7584933626443712, "learning_rate": 7.430052192348398e-06, "loss": 0.4633, "step": 2109 }, { "epoch": 2.0249639942390782, "grad_norm": 0.8235456457771038, "learning_rate": 7.427123939394658e-06, "loss": 0.4515, "step": 2110 }, { "epoch": 2.025924147863658, "grad_norm": 1.0486352220687156, "learning_rate": 7.424194596930925e-06, "loss": 0.5914, "step": 2111 }, { "epoch": 2.0268843014882383, "grad_norm": 0.8172127796335653, "learning_rate": 7.421264166272152e-06, "loss": 0.4839, "step": 2112 }, { "epoch": 2.027844455112818, "grad_norm": 0.9334083534606835, "learning_rate": 7.418332648733774e-06, "loss": 0.6246, "step": 2113 }, { "epoch": 2.028804608737398, "grad_norm": 0.8576019838037999, "learning_rate": 7.415400045631723e-06, "loss": 0.6102, "step": 2114 }, { "epoch": 2.029764762361978, "grad_norm": 1.0106516311732836, "learning_rate": 7.41246635828241e-06, "loss": 0.5129, "step": 2115 }, { "epoch": 2.030724915986558, "grad_norm": 0.9572200517165217, "learning_rate": 7.4095315880027365e-06, "loss": 0.5725, "step": 2116 }, { "epoch": 2.0316850696111377, "grad_norm": 0.6751717002830564, "learning_rate": 7.406595736110089e-06, "loss": 0.5697, "step": 2117 }, { "epoch": 2.0326452232357175, "grad_norm": 0.972391323393412, "learning_rate": 7.4036588039223415e-06, "loss": 0.5948, "step": 2118 }, { "epoch": 2.0336053768602977, "grad_norm": 0.9885599902424669, "learning_rate": 7.400720792757851e-06, "loss": 0.6134, "step": 2119 }, { "epoch": 2.0345655304848775, "grad_norm": 0.853448051645855, "learning_rate": 7.397781703935458e-06, "loss": 0.4918, "step": 2120 }, { "epoch": 2.0355256841094573, "grad_norm": 0.9017476127067426, "learning_rate": 7.39484153877449e-06, "loss": 0.5985, "step": 2121 }, { "epoch": 2.0364858377340376, "grad_norm": 0.8217422026672089, "learning_rate": 7.391900298594752e-06, "loss": 0.4737, "step": 2122 }, { "epoch": 2.0374459913586174, "grad_norm": 1.100503685019264, "learning_rate": 7.388957984716541e-06, "loss": 0.5475, "step": 2123 }, { "epoch": 2.038406144983197, "grad_norm": 0.9721742321479334, "learning_rate": 7.386014598460626e-06, "loss": 0.5214, "step": 2124 }, { "epoch": 2.0393662986077774, "grad_norm": 0.8775807010172804, "learning_rate": 7.383070141148263e-06, "loss": 0.5399, "step": 2125 }, { "epoch": 2.040326452232357, "grad_norm": 0.7233458505911536, "learning_rate": 7.380124614101185e-06, "loss": 0.537, "step": 2126 }, { "epoch": 2.041286605856937, "grad_norm": 0.8249810382094585, "learning_rate": 7.377178018641613e-06, "loss": 0.5188, "step": 2127 }, { "epoch": 2.0422467594815172, "grad_norm": 1.0822061855292786, "learning_rate": 7.374230356092238e-06, "loss": 0.6193, "step": 2128 }, { "epoch": 2.043206913106097, "grad_norm": 0.6665403682733809, "learning_rate": 7.371281627776237e-06, "loss": 0.5361, "step": 2129 }, { "epoch": 2.044167066730677, "grad_norm": 0.8842234144809232, "learning_rate": 7.368331835017257e-06, "loss": 0.5326, "step": 2130 }, { "epoch": 2.0451272203552566, "grad_norm": 0.8506742321393448, "learning_rate": 7.365380979139435e-06, "loss": 0.6244, "step": 2131 }, { "epoch": 2.046087373979837, "grad_norm": 0.8315239843452709, "learning_rate": 7.362429061467377e-06, "loss": 0.5684, "step": 2132 }, { "epoch": 2.0470475276044167, "grad_norm": 1.2476725993360396, "learning_rate": 7.359476083326164e-06, "loss": 0.599, "step": 2133 }, { "epoch": 2.0480076812289965, "grad_norm": 0.9469317731976715, "learning_rate": 7.35652204604136e-06, "loss": 0.6144, "step": 2134 }, { "epoch": 2.0489678348535767, "grad_norm": 0.7895077279439779, "learning_rate": 7.353566950938997e-06, "loss": 0.5313, "step": 2135 }, { "epoch": 2.0499279884781565, "grad_norm": 0.6685640243490947, "learning_rate": 7.35061079934559e-06, "loss": 0.4747, "step": 2136 }, { "epoch": 2.0508881421027363, "grad_norm": 1.2320836652429985, "learning_rate": 7.347653592588122e-06, "loss": 0.6027, "step": 2137 }, { "epoch": 2.0518482957273165, "grad_norm": 0.8211402351100694, "learning_rate": 7.344695331994052e-06, "loss": 0.5506, "step": 2138 }, { "epoch": 2.0528084493518963, "grad_norm": 0.87088219177073, "learning_rate": 7.341736018891308e-06, "loss": 0.5468, "step": 2139 }, { "epoch": 2.053768602976476, "grad_norm": 1.1838467546696474, "learning_rate": 7.338775654608299e-06, "loss": 0.5508, "step": 2140 }, { "epoch": 2.0547287566010564, "grad_norm": 1.090295562775228, "learning_rate": 7.335814240473899e-06, "loss": 0.5286, "step": 2141 }, { "epoch": 2.055688910225636, "grad_norm": 0.7837543162884755, "learning_rate": 7.332851777817453e-06, "loss": 0.5513, "step": 2142 }, { "epoch": 2.056649063850216, "grad_norm": 0.7885777088231247, "learning_rate": 7.3298882679687836e-06, "loss": 0.5063, "step": 2143 }, { "epoch": 2.0576092174747957, "grad_norm": 0.8004586077403608, "learning_rate": 7.326923712258173e-06, "loss": 0.5312, "step": 2144 }, { "epoch": 2.058569371099376, "grad_norm": 0.6944460534744284, "learning_rate": 7.323958112016383e-06, "loss": 0.5174, "step": 2145 }, { "epoch": 2.059529524723956, "grad_norm": 0.6927286853585791, "learning_rate": 7.320991468574637e-06, "loss": 0.5108, "step": 2146 }, { "epoch": 2.0604896783485356, "grad_norm": 1.1153670784129026, "learning_rate": 7.318023783264633e-06, "loss": 0.5214, "step": 2147 }, { "epoch": 2.061449831973116, "grad_norm": 0.8718335709700077, "learning_rate": 7.315055057418528e-06, "loss": 0.6212, "step": 2148 }, { "epoch": 2.0624099855976956, "grad_norm": 0.9903339631434328, "learning_rate": 7.312085292368957e-06, "loss": 0.5141, "step": 2149 }, { "epoch": 2.0633701392222754, "grad_norm": 0.9775331280459144, "learning_rate": 7.309114489449012e-06, "loss": 0.5642, "step": 2150 }, { "epoch": 2.0643302928468557, "grad_norm": 0.8443920599454572, "learning_rate": 7.306142649992256e-06, "loss": 0.5461, "step": 2151 }, { "epoch": 2.0652904464714354, "grad_norm": 1.1356011135244748, "learning_rate": 7.3031697753327145e-06, "loss": 0.6355, "step": 2152 }, { "epoch": 2.0662506000960152, "grad_norm": 1.045201426995347, "learning_rate": 7.300195866804881e-06, "loss": 0.6073, "step": 2153 }, { "epoch": 2.0672107537205955, "grad_norm": 1.1772263424640241, "learning_rate": 7.2972209257437125e-06, "loss": 0.4705, "step": 2154 }, { "epoch": 2.0681709073451753, "grad_norm": 0.9696129870637803, "learning_rate": 7.2942449534846234e-06, "loss": 0.5448, "step": 2155 }, { "epoch": 2.069131060969755, "grad_norm": 1.1866941658313037, "learning_rate": 7.291267951363501e-06, "loss": 0.5278, "step": 2156 }, { "epoch": 2.070091214594335, "grad_norm": 0.7673378188367882, "learning_rate": 7.288289920716686e-06, "loss": 0.5484, "step": 2157 }, { "epoch": 2.071051368218915, "grad_norm": 1.1095276224196864, "learning_rate": 7.285310862880987e-06, "loss": 0.5054, "step": 2158 }, { "epoch": 2.072011521843495, "grad_norm": 0.9374105385272088, "learning_rate": 7.28233077919367e-06, "loss": 0.5213, "step": 2159 }, { "epoch": 2.0729716754680747, "grad_norm": 0.8890832333754094, "learning_rate": 7.279349670992464e-06, "loss": 0.5707, "step": 2160 }, { "epoch": 2.073931829092655, "grad_norm": 0.7556041343489539, "learning_rate": 7.2763675396155545e-06, "loss": 0.5557, "step": 2161 }, { "epoch": 2.0748919827172347, "grad_norm": 0.8214569852773665, "learning_rate": 7.27338438640159e-06, "loss": 0.5433, "step": 2162 }, { "epoch": 2.0758521363418145, "grad_norm": 0.7901768174798454, "learning_rate": 7.270400212689675e-06, "loss": 0.5736, "step": 2163 }, { "epoch": 2.0768122899663948, "grad_norm": 0.8074844274470517, "learning_rate": 7.2674150198193735e-06, "loss": 0.5904, "step": 2164 }, { "epoch": 2.0777724435909746, "grad_norm": 1.2272722968069443, "learning_rate": 7.264428809130708e-06, "loss": 0.6196, "step": 2165 }, { "epoch": 2.0787325972155544, "grad_norm": 1.0062856666124138, "learning_rate": 7.2614415819641535e-06, "loss": 0.5718, "step": 2166 }, { "epoch": 2.0796927508401346, "grad_norm": 0.8103789009331849, "learning_rate": 7.258453339660648e-06, "loss": 0.563, "step": 2167 }, { "epoch": 2.0806529044647144, "grad_norm": 0.9340702743192837, "learning_rate": 7.25546408356158e-06, "loss": 0.5879, "step": 2168 }, { "epoch": 2.081613058089294, "grad_norm": 0.8582435390962432, "learning_rate": 7.252473815008794e-06, "loss": 0.5373, "step": 2169 }, { "epoch": 2.082573211713874, "grad_norm": 0.9345531594935731, "learning_rate": 7.24948253534459e-06, "loss": 0.5439, "step": 2170 }, { "epoch": 2.0835333653384542, "grad_norm": 0.7694924977873263, "learning_rate": 7.246490245911722e-06, "loss": 0.5418, "step": 2171 }, { "epoch": 2.084493518963034, "grad_norm": 1.2722834496407283, "learning_rate": 7.243496948053394e-06, "loss": 0.633, "step": 2172 }, { "epoch": 2.085453672587614, "grad_norm": 0.8267984668992502, "learning_rate": 7.240502643113271e-06, "loss": 0.4939, "step": 2173 }, { "epoch": 2.086413826212194, "grad_norm": 0.8195833482628507, "learning_rate": 7.237507332435459e-06, "loss": 0.5076, "step": 2174 }, { "epoch": 2.087373979836774, "grad_norm": 1.040409217785711, "learning_rate": 7.234511017364524e-06, "loss": 0.6581, "step": 2175 }, { "epoch": 2.0883341334613537, "grad_norm": 0.8861415047181347, "learning_rate": 7.2315136992454794e-06, "loss": 0.6015, "step": 2176 }, { "epoch": 2.089294287085934, "grad_norm": 1.0444328966905378, "learning_rate": 7.228515379423787e-06, "loss": 0.5343, "step": 2177 }, { "epoch": 2.0902544407105137, "grad_norm": 0.9050227750870652, "learning_rate": 7.225516059245365e-06, "loss": 0.6025, "step": 2178 }, { "epoch": 2.0912145943350935, "grad_norm": 0.9793683787428816, "learning_rate": 7.222515740056571e-06, "loss": 0.5322, "step": 2179 }, { "epoch": 2.0921747479596737, "grad_norm": 0.9736765075627355, "learning_rate": 7.219514423204219e-06, "loss": 0.5518, "step": 2180 }, { "epoch": 2.0931349015842535, "grad_norm": 0.8552388037570061, "learning_rate": 7.216512110035567e-06, "loss": 0.5171, "step": 2181 }, { "epoch": 2.0940950552088333, "grad_norm": 0.8858809554354977, "learning_rate": 7.213508801898325e-06, "loss": 0.518, "step": 2182 }, { "epoch": 2.095055208833413, "grad_norm": 0.9467693628928557, "learning_rate": 7.21050450014064e-06, "loss": 0.596, "step": 2183 }, { "epoch": 2.0960153624579934, "grad_norm": 1.133831005367814, "learning_rate": 7.2074992061111145e-06, "loss": 0.5638, "step": 2184 }, { "epoch": 2.096975516082573, "grad_norm": 1.0068135718385338, "learning_rate": 7.204492921158794e-06, "loss": 0.6315, "step": 2185 }, { "epoch": 2.097935669707153, "grad_norm": 0.8236412939332182, "learning_rate": 7.201485646633164e-06, "loss": 0.5714, "step": 2186 }, { "epoch": 2.098895823331733, "grad_norm": 0.9118405628769543, "learning_rate": 7.198477383884161e-06, "loss": 0.5822, "step": 2187 }, { "epoch": 2.099855976956313, "grad_norm": 0.7294127716306451, "learning_rate": 7.195468134262159e-06, "loss": 0.5324, "step": 2188 }, { "epoch": 2.100816130580893, "grad_norm": 1.2554681531351808, "learning_rate": 7.1924578991179815e-06, "loss": 0.4759, "step": 2189 }, { "epoch": 2.101776284205473, "grad_norm": 0.6966842302226403, "learning_rate": 7.189446679802889e-06, "loss": 0.4412, "step": 2190 }, { "epoch": 2.102736437830053, "grad_norm": 0.7063745000096698, "learning_rate": 7.186434477668586e-06, "loss": 0.5245, "step": 2191 }, { "epoch": 2.1036965914546326, "grad_norm": 1.040788739636212, "learning_rate": 7.1834212940672155e-06, "loss": 0.5443, "step": 2192 }, { "epoch": 2.104656745079213, "grad_norm": 1.4116018703653066, "learning_rate": 7.180407130351368e-06, "loss": 0.5781, "step": 2193 }, { "epoch": 2.1056168987037926, "grad_norm": 0.9749124807276462, "learning_rate": 7.177391987874065e-06, "loss": 0.6301, "step": 2194 }, { "epoch": 2.1065770523283724, "grad_norm": 0.7595222832472861, "learning_rate": 7.174375867988776e-06, "loss": 0.4806, "step": 2195 }, { "epoch": 2.1075372059529527, "grad_norm": 0.715272870202692, "learning_rate": 7.171358772049402e-06, "loss": 0.4749, "step": 2196 }, { "epoch": 2.1084973595775325, "grad_norm": 1.0505004749821678, "learning_rate": 7.168340701410284e-06, "loss": 0.5521, "step": 2197 }, { "epoch": 2.1094575132021123, "grad_norm": 0.9767950296446208, "learning_rate": 7.165321657426205e-06, "loss": 0.4539, "step": 2198 }, { "epoch": 2.110417666826692, "grad_norm": 1.3722868882754091, "learning_rate": 7.1623016414523795e-06, "loss": 0.6327, "step": 2199 }, { "epoch": 2.1113778204512723, "grad_norm": 1.0696007028352503, "learning_rate": 7.159280654844462e-06, "loss": 0.6627, "step": 2200 }, { "epoch": 2.112337974075852, "grad_norm": 0.8267290787885438, "learning_rate": 7.156258698958537e-06, "loss": 0.6042, "step": 2201 }, { "epoch": 2.113298127700432, "grad_norm": 1.414431349468318, "learning_rate": 7.153235775151134e-06, "loss": 0.5633, "step": 2202 }, { "epoch": 2.114258281325012, "grad_norm": 1.2550142917084253, "learning_rate": 7.150211884779208e-06, "loss": 0.5708, "step": 2203 }, { "epoch": 2.115218434949592, "grad_norm": 0.9226235775619207, "learning_rate": 7.147187029200151e-06, "loss": 0.5352, "step": 2204 }, { "epoch": 2.1161785885741717, "grad_norm": 1.1236135516836152, "learning_rate": 7.144161209771788e-06, "loss": 0.5115, "step": 2205 }, { "epoch": 2.117138742198752, "grad_norm": 1.539184601170347, "learning_rate": 7.141134427852378e-06, "loss": 0.5127, "step": 2206 }, { "epoch": 2.1180988958233318, "grad_norm": 1.078767338492532, "learning_rate": 7.138106684800612e-06, "loss": 0.6234, "step": 2207 }, { "epoch": 2.1190590494479116, "grad_norm": 0.8532453201860914, "learning_rate": 7.1350779819756085e-06, "loss": 0.5012, "step": 2208 }, { "epoch": 2.120019203072492, "grad_norm": 0.9011164553039275, "learning_rate": 7.132048320736924e-06, "loss": 0.5972, "step": 2209 }, { "epoch": 2.1209793566970716, "grad_norm": 0.8499778542983218, "learning_rate": 7.129017702444536e-06, "loss": 0.5756, "step": 2210 }, { "epoch": 2.1219395103216514, "grad_norm": 1.3217188768774837, "learning_rate": 7.12598612845886e-06, "loss": 0.5925, "step": 2211 }, { "epoch": 2.122899663946231, "grad_norm": 0.850933227542242, "learning_rate": 7.1229536001407376e-06, "loss": 0.5874, "step": 2212 }, { "epoch": 2.1238598175708114, "grad_norm": 0.8006336417142614, "learning_rate": 7.119920118851438e-06, "loss": 0.5571, "step": 2213 }, { "epoch": 2.1248199711953912, "grad_norm": 0.7957143814125812, "learning_rate": 7.116885685952655e-06, "loss": 0.6198, "step": 2214 }, { "epoch": 2.125780124819971, "grad_norm": 0.9446868842582681, "learning_rate": 7.113850302806518e-06, "loss": 0.5845, "step": 2215 }, { "epoch": 2.1267402784445513, "grad_norm": 0.8723748700684102, "learning_rate": 7.110813970775574e-06, "loss": 0.5938, "step": 2216 }, { "epoch": 2.127700432069131, "grad_norm": 0.9805893461253402, "learning_rate": 7.107776691222803e-06, "loss": 0.634, "step": 2217 }, { "epoch": 2.128660585693711, "grad_norm": 0.8799936401014986, "learning_rate": 7.104738465511606e-06, "loss": 0.57, "step": 2218 }, { "epoch": 2.129620739318291, "grad_norm": 0.7986385110798393, "learning_rate": 7.1016992950058104e-06, "loss": 0.559, "step": 2219 }, { "epoch": 2.130580892942871, "grad_norm": 1.2115378566836874, "learning_rate": 7.098659181069667e-06, "loss": 0.5761, "step": 2220 }, { "epoch": 2.1315410465674507, "grad_norm": 1.0142402126614256, "learning_rate": 7.0956181250678514e-06, "loss": 0.547, "step": 2221 }, { "epoch": 2.132501200192031, "grad_norm": 0.7658521288613331, "learning_rate": 7.09257612836546e-06, "loss": 0.5293, "step": 2222 }, { "epoch": 2.1334613538166107, "grad_norm": 0.8294845270158665, "learning_rate": 7.089533192328011e-06, "loss": 0.5169, "step": 2223 }, { "epoch": 2.1344215074411905, "grad_norm": 0.8024683130638101, "learning_rate": 7.08648931832145e-06, "loss": 0.61, "step": 2224 }, { "epoch": 2.1353816610657703, "grad_norm": 0.9026826615751471, "learning_rate": 7.0834445077121374e-06, "loss": 0.5954, "step": 2225 }, { "epoch": 2.1363418146903506, "grad_norm": 0.761598215348195, "learning_rate": 7.080398761866856e-06, "loss": 0.5211, "step": 2226 }, { "epoch": 2.1373019683149304, "grad_norm": 0.979006382775271, "learning_rate": 7.077352082152809e-06, "loss": 0.5699, "step": 2227 }, { "epoch": 2.13826212193951, "grad_norm": 0.9252619772749834, "learning_rate": 7.0743044699376165e-06, "loss": 0.5165, "step": 2228 }, { "epoch": 2.1392222755640904, "grad_norm": 0.7668035098364651, "learning_rate": 7.0712559265893214e-06, "loss": 0.5491, "step": 2229 }, { "epoch": 2.14018242918867, "grad_norm": 0.7414287423743836, "learning_rate": 7.068206453476381e-06, "loss": 0.5036, "step": 2230 }, { "epoch": 2.14114258281325, "grad_norm": 0.8178240021551951, "learning_rate": 7.065156051967672e-06, "loss": 0.5368, "step": 2231 }, { "epoch": 2.1421027364378302, "grad_norm": 0.7388627592597732, "learning_rate": 7.062104723432485e-06, "loss": 0.4954, "step": 2232 }, { "epoch": 2.14306289006241, "grad_norm": 1.164769360920896, "learning_rate": 7.059052469240533e-06, "loss": 0.5178, "step": 2233 }, { "epoch": 2.14402304368699, "grad_norm": 0.8772811007079424, "learning_rate": 7.0559992907619346e-06, "loss": 0.5855, "step": 2234 }, { "epoch": 2.14498319731157, "grad_norm": 0.6906935653270964, "learning_rate": 7.0529451893672356e-06, "loss": 0.5476, "step": 2235 }, { "epoch": 2.14594335093615, "grad_norm": 1.1363754671988087, "learning_rate": 7.0498901664273825e-06, "loss": 0.5681, "step": 2236 }, { "epoch": 2.1469035045607296, "grad_norm": 0.7756208798244094, "learning_rate": 7.046834223313746e-06, "loss": 0.4298, "step": 2237 }, { "epoch": 2.1478636581853094, "grad_norm": 0.7114868910310154, "learning_rate": 7.043777361398108e-06, "loss": 0.5178, "step": 2238 }, { "epoch": 2.1488238118098897, "grad_norm": 0.9026957415063291, "learning_rate": 7.040719582052659e-06, "loss": 0.6007, "step": 2239 }, { "epoch": 2.1497839654344695, "grad_norm": 1.0942635993690641, "learning_rate": 7.037660886650003e-06, "loss": 0.6136, "step": 2240 }, { "epoch": 2.1507441190590493, "grad_norm": 1.2573894316596292, "learning_rate": 7.034601276563155e-06, "loss": 0.6018, "step": 2241 }, { "epoch": 2.1517042726836295, "grad_norm": 0.9770680987086197, "learning_rate": 7.031540753165545e-06, "loss": 0.4367, "step": 2242 }, { "epoch": 2.1526644263082093, "grad_norm": 0.9892381660706268, "learning_rate": 7.028479317831006e-06, "loss": 0.6132, "step": 2243 }, { "epoch": 2.153624579932789, "grad_norm": 0.7112835946643007, "learning_rate": 7.0254169719337864e-06, "loss": 0.5142, "step": 2244 }, { "epoch": 2.1545847335573693, "grad_norm": 1.065777786696222, "learning_rate": 7.022353716848535e-06, "loss": 0.5956, "step": 2245 }, { "epoch": 2.155544887181949, "grad_norm": 1.5539690439648255, "learning_rate": 7.0192895539503205e-06, "loss": 0.6245, "step": 2246 }, { "epoch": 2.156505040806529, "grad_norm": 0.8599626031646135, "learning_rate": 7.016224484614609e-06, "loss": 0.5631, "step": 2247 }, { "epoch": 2.157465194431109, "grad_norm": 0.6719318705655244, "learning_rate": 7.01315851021728e-06, "loss": 0.4895, "step": 2248 }, { "epoch": 2.158425348055689, "grad_norm": 1.3558382554935693, "learning_rate": 7.010091632134612e-06, "loss": 0.5787, "step": 2249 }, { "epoch": 2.1593855016802688, "grad_norm": 0.97600255215016, "learning_rate": 7.0070238517433e-06, "loss": 0.5959, "step": 2250 }, { "epoch": 2.160345655304849, "grad_norm": 1.2991139416399855, "learning_rate": 7.003955170420434e-06, "loss": 0.6888, "step": 2251 }, { "epoch": 2.161305808929429, "grad_norm": 0.7198982154950009, "learning_rate": 7.000885589543513e-06, "loss": 0.6094, "step": 2252 }, { "epoch": 2.1622659625540086, "grad_norm": 1.1673018256504468, "learning_rate": 6.99781511049044e-06, "loss": 0.637, "step": 2253 }, { "epoch": 2.1632261161785884, "grad_norm": 0.9534705925232878, "learning_rate": 6.994743734639517e-06, "loss": 0.6197, "step": 2254 }, { "epoch": 2.1641862698031686, "grad_norm": 1.0002855698401616, "learning_rate": 6.991671463369457e-06, "loss": 0.5925, "step": 2255 }, { "epoch": 2.1651464234277484, "grad_norm": 0.763911863914701, "learning_rate": 6.988598298059366e-06, "loss": 0.5177, "step": 2256 }, { "epoch": 2.1661065770523282, "grad_norm": 0.7865107655764889, "learning_rate": 6.985524240088756e-06, "loss": 0.5229, "step": 2257 }, { "epoch": 2.1670667306769085, "grad_norm": 0.8654886244499703, "learning_rate": 6.982449290837538e-06, "loss": 0.5776, "step": 2258 }, { "epoch": 2.1680268843014883, "grad_norm": 0.6858240484062622, "learning_rate": 6.979373451686025e-06, "loss": 0.5609, "step": 2259 }, { "epoch": 2.168987037926068, "grad_norm": 0.8249988281529682, "learning_rate": 6.976296724014927e-06, "loss": 0.6297, "step": 2260 }, { "epoch": 2.1699471915506483, "grad_norm": 0.8343555119067906, "learning_rate": 6.973219109205356e-06, "loss": 0.5801, "step": 2261 }, { "epoch": 2.170907345175228, "grad_norm": 1.3255750313555195, "learning_rate": 6.970140608638818e-06, "loss": 0.5841, "step": 2262 }, { "epoch": 2.171867498799808, "grad_norm": 0.810420911846056, "learning_rate": 6.967061223697223e-06, "loss": 0.566, "step": 2263 }, { "epoch": 2.172827652424388, "grad_norm": 0.6577554875572367, "learning_rate": 6.963980955762869e-06, "loss": 0.3834, "step": 2264 }, { "epoch": 2.173787806048968, "grad_norm": 0.9803999857215404, "learning_rate": 6.960899806218458e-06, "loss": 0.5313, "step": 2265 }, { "epoch": 2.1747479596735477, "grad_norm": 0.7647434810511963, "learning_rate": 6.957817776447087e-06, "loss": 0.5202, "step": 2266 }, { "epoch": 2.1757081132981275, "grad_norm": 0.9908443655571111, "learning_rate": 6.954734867832241e-06, "loss": 0.5405, "step": 2267 }, { "epoch": 2.1766682669227078, "grad_norm": 1.2638869240941317, "learning_rate": 6.9516510817578105e-06, "loss": 0.5437, "step": 2268 }, { "epoch": 2.1776284205472876, "grad_norm": 1.0039906754725876, "learning_rate": 6.948566419608069e-06, "loss": 0.6595, "step": 2269 }, { "epoch": 2.1785885741718674, "grad_norm": 0.777383483919325, "learning_rate": 6.945480882767694e-06, "loss": 0.5789, "step": 2270 }, { "epoch": 2.1795487277964476, "grad_norm": 1.0473632642305972, "learning_rate": 6.942394472621744e-06, "loss": 0.4763, "step": 2271 }, { "epoch": 2.1805088814210274, "grad_norm": 0.9702453120501578, "learning_rate": 6.939307190555679e-06, "loss": 0.5878, "step": 2272 }, { "epoch": 2.181469035045607, "grad_norm": 0.8758949961513824, "learning_rate": 6.936219037955347e-06, "loss": 0.613, "step": 2273 }, { "epoch": 2.1824291886701874, "grad_norm": 0.9102154498740379, "learning_rate": 6.9331300162069846e-06, "loss": 0.5359, "step": 2274 }, { "epoch": 2.183389342294767, "grad_norm": 0.767100547523797, "learning_rate": 6.930040126697224e-06, "loss": 0.5204, "step": 2275 }, { "epoch": 2.184349495919347, "grad_norm": 1.0742745837620984, "learning_rate": 6.926949370813079e-06, "loss": 0.6586, "step": 2276 }, { "epoch": 2.1853096495439273, "grad_norm": 0.778206194282406, "learning_rate": 6.92385774994196e-06, "loss": 0.5792, "step": 2277 }, { "epoch": 2.186269803168507, "grad_norm": 0.7267575332969071, "learning_rate": 6.920765265471663e-06, "loss": 0.5829, "step": 2278 }, { "epoch": 2.187229956793087, "grad_norm": 1.215780724669486, "learning_rate": 6.91767191879037e-06, "loss": 0.4964, "step": 2279 }, { "epoch": 2.1881901104176666, "grad_norm": 0.8954544868849986, "learning_rate": 6.9145777112866495e-06, "loss": 0.5537, "step": 2280 }, { "epoch": 2.189150264042247, "grad_norm": 1.1920330317568804, "learning_rate": 6.911482644349462e-06, "loss": 0.5238, "step": 2281 }, { "epoch": 2.1901104176668267, "grad_norm": 0.7822228685891184, "learning_rate": 6.908386719368147e-06, "loss": 0.6381, "step": 2282 }, { "epoch": 2.1910705712914065, "grad_norm": 0.6973971159414132, "learning_rate": 6.9052899377324345e-06, "loss": 0.5336, "step": 2283 }, { "epoch": 2.1920307249159867, "grad_norm": 0.8104438243699706, "learning_rate": 6.902192300832435e-06, "loss": 0.6069, "step": 2284 }, { "epoch": 2.1929908785405665, "grad_norm": 1.0017800399528134, "learning_rate": 6.899093810058643e-06, "loss": 0.4497, "step": 2285 }, { "epoch": 2.1939510321651463, "grad_norm": 1.4197944574410453, "learning_rate": 6.895994466801943e-06, "loss": 0.545, "step": 2286 }, { "epoch": 2.1949111857897265, "grad_norm": 0.8935501033526516, "learning_rate": 6.892894272453592e-06, "loss": 0.5228, "step": 2287 }, { "epoch": 2.1958713394143063, "grad_norm": 1.0883689086111068, "learning_rate": 6.889793228405237e-06, "loss": 0.6184, "step": 2288 }, { "epoch": 2.196831493038886, "grad_norm": 1.1326471858009117, "learning_rate": 6.8866913360489005e-06, "loss": 0.5031, "step": 2289 }, { "epoch": 2.1977916466634664, "grad_norm": 0.6972885404272577, "learning_rate": 6.883588596776994e-06, "loss": 0.563, "step": 2290 }, { "epoch": 2.198751800288046, "grad_norm": 1.2716807501519334, "learning_rate": 6.8804850119822986e-06, "loss": 0.6434, "step": 2291 }, { "epoch": 2.199711953912626, "grad_norm": 0.7774604563769336, "learning_rate": 6.8773805830579845e-06, "loss": 0.6302, "step": 2292 }, { "epoch": 2.2006721075372058, "grad_norm": 1.1664101212879447, "learning_rate": 6.8742753113975925e-06, "loss": 0.6466, "step": 2293 }, { "epoch": 2.201632261161786, "grad_norm": 0.8546969752206524, "learning_rate": 6.871169198395049e-06, "loss": 0.5545, "step": 2294 }, { "epoch": 2.202592414786366, "grad_norm": 0.7106716203100409, "learning_rate": 6.868062245444655e-06, "loss": 0.5895, "step": 2295 }, { "epoch": 2.2035525684109456, "grad_norm": 0.8638131203453104, "learning_rate": 6.864954453941086e-06, "loss": 0.5925, "step": 2296 }, { "epoch": 2.204512722035526, "grad_norm": 0.8759755721839326, "learning_rate": 6.861845825279399e-06, "loss": 0.5027, "step": 2297 }, { "epoch": 2.2054728756601056, "grad_norm": 0.7554023217795944, "learning_rate": 6.85873636085502e-06, "loss": 0.5781, "step": 2298 }, { "epoch": 2.2064330292846854, "grad_norm": 0.9368870403320614, "learning_rate": 6.855626062063761e-06, "loss": 0.6287, "step": 2299 }, { "epoch": 2.2073931829092657, "grad_norm": 1.6181387943044583, "learning_rate": 6.852514930301797e-06, "loss": 0.727, "step": 2300 }, { "epoch": 2.2083533365338455, "grad_norm": 1.3959774822893378, "learning_rate": 6.8494029669656845e-06, "loss": 0.6687, "step": 2301 }, { "epoch": 2.2093134901584253, "grad_norm": 1.121430573797831, "learning_rate": 6.846290173452348e-06, "loss": 0.5816, "step": 2302 }, { "epoch": 2.2102736437830055, "grad_norm": 0.7342880662474258, "learning_rate": 6.843176551159087e-06, "loss": 0.5523, "step": 2303 }, { "epoch": 2.2112337974075853, "grad_norm": 1.0401505202042727, "learning_rate": 6.8400621014835775e-06, "loss": 0.685, "step": 2304 }, { "epoch": 2.212193951032165, "grad_norm": 0.8706936365952261, "learning_rate": 6.836946825823858e-06, "loss": 0.572, "step": 2305 }, { "epoch": 2.213154104656745, "grad_norm": 0.951134051796077, "learning_rate": 6.833830725578346e-06, "loss": 0.6174, "step": 2306 }, { "epoch": 2.214114258281325, "grad_norm": 1.0742464836933634, "learning_rate": 6.83071380214582e-06, "loss": 0.5727, "step": 2307 }, { "epoch": 2.215074411905905, "grad_norm": 1.0253390130066518, "learning_rate": 6.827596056925441e-06, "loss": 0.6745, "step": 2308 }, { "epoch": 2.2160345655304847, "grad_norm": 0.9753973639686732, "learning_rate": 6.824477491316725e-06, "loss": 0.5455, "step": 2309 }, { "epoch": 2.216994719155065, "grad_norm": 1.6641997872582994, "learning_rate": 6.821358106719569e-06, "loss": 0.5112, "step": 2310 }, { "epoch": 2.2179548727796448, "grad_norm": 1.334630302636793, "learning_rate": 6.818237904534226e-06, "loss": 0.6364, "step": 2311 }, { "epoch": 2.2189150264042246, "grad_norm": 0.8706829266439885, "learning_rate": 6.815116886161324e-06, "loss": 0.6125, "step": 2312 }, { "epoch": 2.219875180028805, "grad_norm": 1.1280844361609608, "learning_rate": 6.811995053001854e-06, "loss": 0.664, "step": 2313 }, { "epoch": 2.2208353336533846, "grad_norm": 1.0897465102219406, "learning_rate": 6.808872406457175e-06, "loss": 0.5094, "step": 2314 }, { "epoch": 2.2217954872779644, "grad_norm": 0.9808832474780659, "learning_rate": 6.8057489479290085e-06, "loss": 0.578, "step": 2315 }, { "epoch": 2.2227556409025446, "grad_norm": 0.7771621013840303, "learning_rate": 6.802624678819441e-06, "loss": 0.5118, "step": 2316 }, { "epoch": 2.2237157945271244, "grad_norm": 0.8526791800990299, "learning_rate": 6.799499600530925e-06, "loss": 0.6069, "step": 2317 }, { "epoch": 2.224675948151704, "grad_norm": 0.8584544172957495, "learning_rate": 6.796373714466272e-06, "loss": 0.6003, "step": 2318 }, { "epoch": 2.225636101776284, "grad_norm": 0.7136462429716051, "learning_rate": 6.793247022028663e-06, "loss": 0.5084, "step": 2319 }, { "epoch": 2.2265962554008643, "grad_norm": 1.1930138979695286, "learning_rate": 6.790119524621632e-06, "loss": 0.6307, "step": 2320 }, { "epoch": 2.227556409025444, "grad_norm": 1.1750224668952614, "learning_rate": 6.786991223649084e-06, "loss": 0.6143, "step": 2321 }, { "epoch": 2.228516562650024, "grad_norm": 1.1669679935198085, "learning_rate": 6.783862120515277e-06, "loss": 0.5525, "step": 2322 }, { "epoch": 2.229476716274604, "grad_norm": 0.7647258656596809, "learning_rate": 6.780732216624833e-06, "loss": 0.5381, "step": 2323 }, { "epoch": 2.230436869899184, "grad_norm": 1.0238811047404157, "learning_rate": 6.77760151338273e-06, "loss": 0.5239, "step": 2324 }, { "epoch": 2.2313970235237637, "grad_norm": 0.9681798393162807, "learning_rate": 6.7744700121943095e-06, "loss": 0.6768, "step": 2325 }, { "epoch": 2.232357177148344, "grad_norm": 0.8436996268226329, "learning_rate": 6.77133771446527e-06, "loss": 0.5706, "step": 2326 }, { "epoch": 2.2333173307729237, "grad_norm": 0.7943711905250763, "learning_rate": 6.768204621601665e-06, "loss": 0.5326, "step": 2327 }, { "epoch": 2.2342774843975035, "grad_norm": 0.9629485127123245, "learning_rate": 6.765070735009906e-06, "loss": 0.5508, "step": 2328 }, { "epoch": 2.2352376380220838, "grad_norm": 1.3459133727734214, "learning_rate": 6.761936056096761e-06, "loss": 0.4556, "step": 2329 }, { "epoch": 2.2361977916466635, "grad_norm": 1.1383538845400374, "learning_rate": 6.758800586269355e-06, "loss": 0.5237, "step": 2330 }, { "epoch": 2.2371579452712433, "grad_norm": 0.732888679483021, "learning_rate": 6.755664326935167e-06, "loss": 0.4607, "step": 2331 }, { "epoch": 2.238118098895823, "grad_norm": 0.9865410999907581, "learning_rate": 6.752527279502029e-06, "loss": 0.5826, "step": 2332 }, { "epoch": 2.2390782525204034, "grad_norm": 0.951160087685012, "learning_rate": 6.749389445378127e-06, "loss": 0.5468, "step": 2333 }, { "epoch": 2.240038406144983, "grad_norm": 1.3136937883921078, "learning_rate": 6.746250825972005e-06, "loss": 0.5868, "step": 2334 }, { "epoch": 2.240998559769563, "grad_norm": 0.8122534155481174, "learning_rate": 6.743111422692553e-06, "loss": 0.5349, "step": 2335 }, { "epoch": 2.241958713394143, "grad_norm": 1.1089356976879372, "learning_rate": 6.739971236949017e-06, "loss": 0.6057, "step": 2336 }, { "epoch": 2.242918867018723, "grad_norm": 0.9357134705059069, "learning_rate": 6.736830270150992e-06, "loss": 0.6111, "step": 2337 }, { "epoch": 2.243879020643303, "grad_norm": 0.7384225502866528, "learning_rate": 6.733688523708422e-06, "loss": 0.5593, "step": 2338 }, { "epoch": 2.244839174267883, "grad_norm": 0.8639186408254549, "learning_rate": 6.7305459990316094e-06, "loss": 0.5489, "step": 2339 }, { "epoch": 2.245799327892463, "grad_norm": 0.7730199458933303, "learning_rate": 6.727402697531193e-06, "loss": 0.4998, "step": 2340 }, { "epoch": 2.2467594815170426, "grad_norm": 1.0038578364232367, "learning_rate": 6.7242586206181735e-06, "loss": 0.5552, "step": 2341 }, { "epoch": 2.247719635141623, "grad_norm": 1.3620602388258491, "learning_rate": 6.721113769703887e-06, "loss": 0.4578, "step": 2342 }, { "epoch": 2.2486797887662027, "grad_norm": 0.7861287708700152, "learning_rate": 6.71796814620003e-06, "loss": 0.5493, "step": 2343 }, { "epoch": 2.2496399423907825, "grad_norm": 0.8295176803252249, "learning_rate": 6.714821751518635e-06, "loss": 0.6012, "step": 2344 }, { "epoch": 2.2506000960153623, "grad_norm": 0.7489999962206016, "learning_rate": 6.7116745870720875e-06, "loss": 0.4831, "step": 2345 }, { "epoch": 2.2515602496399425, "grad_norm": 1.0590681070165233, "learning_rate": 6.708526654273113e-06, "loss": 0.5847, "step": 2346 }, { "epoch": 2.2525204032645223, "grad_norm": 0.7873516251106734, "learning_rate": 6.7053779545347874e-06, "loss": 0.5958, "step": 2347 }, { "epoch": 2.253480556889102, "grad_norm": 1.043915253647952, "learning_rate": 6.702228489270528e-06, "loss": 0.5759, "step": 2348 }, { "epoch": 2.2544407105136823, "grad_norm": 1.3124866185031892, "learning_rate": 6.699078259894094e-06, "loss": 0.5687, "step": 2349 }, { "epoch": 2.255400864138262, "grad_norm": 1.0673403606710374, "learning_rate": 6.695927267819593e-06, "loss": 0.5282, "step": 2350 }, { "epoch": 2.256361017762842, "grad_norm": 0.9821335951348141, "learning_rate": 6.692775514461468e-06, "loss": 0.5468, "step": 2351 }, { "epoch": 2.257321171387422, "grad_norm": 0.7765230338279765, "learning_rate": 6.689623001234511e-06, "loss": 0.473, "step": 2352 }, { "epoch": 2.258281325012002, "grad_norm": 0.9877080896749931, "learning_rate": 6.686469729553849e-06, "loss": 0.6309, "step": 2353 }, { "epoch": 2.2592414786365818, "grad_norm": 0.8024710970715238, "learning_rate": 6.683315700834953e-06, "loss": 0.59, "step": 2354 }, { "epoch": 2.260201632261162, "grad_norm": 0.7685511186569934, "learning_rate": 6.680160916493631e-06, "loss": 0.5714, "step": 2355 }, { "epoch": 2.261161785885742, "grad_norm": 0.8910370837797735, "learning_rate": 6.677005377946037e-06, "loss": 0.4756, "step": 2356 }, { "epoch": 2.2621219395103216, "grad_norm": 0.7988130099362978, "learning_rate": 6.6738490866086515e-06, "loss": 0.52, "step": 2357 }, { "epoch": 2.2630820931349014, "grad_norm": 1.2512007210073648, "learning_rate": 6.670692043898306e-06, "loss": 0.5519, "step": 2358 }, { "epoch": 2.2640422467594816, "grad_norm": 0.9928481135872863, "learning_rate": 6.66753425123216e-06, "loss": 0.4711, "step": 2359 }, { "epoch": 2.2650024003840614, "grad_norm": 1.0752150528291098, "learning_rate": 6.664375710027713e-06, "loss": 0.5437, "step": 2360 }, { "epoch": 2.265962554008641, "grad_norm": 0.9790731678286968, "learning_rate": 6.661216421702801e-06, "loss": 0.5305, "step": 2361 }, { "epoch": 2.2669227076332215, "grad_norm": 0.9401671154771365, "learning_rate": 6.6580563876755945e-06, "loss": 0.5668, "step": 2362 }, { "epoch": 2.2678828612578013, "grad_norm": 1.1201848037414552, "learning_rate": 6.6548956093646e-06, "loss": 0.6381, "step": 2363 }, { "epoch": 2.268843014882381, "grad_norm": 0.8390992789153865, "learning_rate": 6.651734088188655e-06, "loss": 0.5656, "step": 2364 }, { "epoch": 2.2698031685069613, "grad_norm": 0.8835462208604588, "learning_rate": 6.648571825566936e-06, "loss": 0.5634, "step": 2365 }, { "epoch": 2.270763322131541, "grad_norm": 0.7741225520687126, "learning_rate": 6.645408822918945e-06, "loss": 0.5722, "step": 2366 }, { "epoch": 2.271723475756121, "grad_norm": 0.8133902795470719, "learning_rate": 6.642245081664524e-06, "loss": 0.5169, "step": 2367 }, { "epoch": 2.272683629380701, "grad_norm": 1.0092800717641697, "learning_rate": 6.6390806032238375e-06, "loss": 0.5223, "step": 2368 }, { "epoch": 2.273643783005281, "grad_norm": 1.2183770876810358, "learning_rate": 6.63591538901739e-06, "loss": 0.4758, "step": 2369 }, { "epoch": 2.2746039366298607, "grad_norm": 0.9699182134234299, "learning_rate": 6.632749440466013e-06, "loss": 0.4907, "step": 2370 }, { "epoch": 2.2755640902544405, "grad_norm": 0.8621209530250445, "learning_rate": 6.629582758990864e-06, "loss": 0.5537, "step": 2371 }, { "epoch": 2.2765242438790207, "grad_norm": 1.2245036523430783, "learning_rate": 6.626415346013434e-06, "loss": 0.6427, "step": 2372 }, { "epoch": 2.2774843975036005, "grad_norm": 1.2809266146911322, "learning_rate": 6.623247202955538e-06, "loss": 0.6078, "step": 2373 }, { "epoch": 2.2784445511281803, "grad_norm": 0.7928894778485797, "learning_rate": 6.620078331239328e-06, "loss": 0.5551, "step": 2374 }, { "epoch": 2.2794047047527606, "grad_norm": 0.9592387241749792, "learning_rate": 6.61690873228727e-06, "loss": 0.5605, "step": 2375 }, { "epoch": 2.2803648583773404, "grad_norm": 0.7824093705440623, "learning_rate": 6.613738407522169e-06, "loss": 0.4686, "step": 2376 }, { "epoch": 2.28132501200192, "grad_norm": 0.9724173798772967, "learning_rate": 6.610567358367143e-06, "loss": 0.582, "step": 2377 }, { "epoch": 2.2822851656265004, "grad_norm": 0.7986369717944548, "learning_rate": 6.607395586245648e-06, "loss": 0.4701, "step": 2378 }, { "epoch": 2.28324531925108, "grad_norm": 1.1603692403514454, "learning_rate": 6.6042230925814565e-06, "loss": 0.5343, "step": 2379 }, { "epoch": 2.28420547287566, "grad_norm": 0.9351265268720285, "learning_rate": 6.601049878798667e-06, "loss": 0.5114, "step": 2380 }, { "epoch": 2.2851656265002402, "grad_norm": 0.9226927097733172, "learning_rate": 6.597875946321701e-06, "loss": 0.4153, "step": 2381 }, { "epoch": 2.28612578012482, "grad_norm": 0.8011323153957083, "learning_rate": 6.594701296575303e-06, "loss": 0.549, "step": 2382 }, { "epoch": 2.2870859337494, "grad_norm": 0.7707617081985462, "learning_rate": 6.5915259309845404e-06, "loss": 0.4796, "step": 2383 }, { "epoch": 2.2880460873739796, "grad_norm": 0.9325805221668414, "learning_rate": 6.588349850974798e-06, "loss": 0.6414, "step": 2384 }, { "epoch": 2.28900624099856, "grad_norm": 0.6572331056324962, "learning_rate": 6.585173057971787e-06, "loss": 0.5431, "step": 2385 }, { "epoch": 2.2899663946231397, "grad_norm": 0.9496403807366189, "learning_rate": 6.5819955534015335e-06, "loss": 0.5388, "step": 2386 }, { "epoch": 2.2909265482477195, "grad_norm": 0.9332839960241, "learning_rate": 6.578817338690389e-06, "loss": 0.5196, "step": 2387 }, { "epoch": 2.2918867018722997, "grad_norm": 0.8872038349682433, "learning_rate": 6.575638415265016e-06, "loss": 0.611, "step": 2388 }, { "epoch": 2.2928468554968795, "grad_norm": 1.095860683225861, "learning_rate": 6.5724587845524024e-06, "loss": 0.452, "step": 2389 }, { "epoch": 2.2938070091214593, "grad_norm": 0.7866685885471305, "learning_rate": 6.569278447979848e-06, "loss": 0.4459, "step": 2390 }, { "epoch": 2.2947671627460395, "grad_norm": 0.978503021487999, "learning_rate": 6.566097406974972e-06, "loss": 0.7169, "step": 2391 }, { "epoch": 2.2957273163706193, "grad_norm": 0.8128803043876026, "learning_rate": 6.562915662965713e-06, "loss": 0.5249, "step": 2392 }, { "epoch": 2.296687469995199, "grad_norm": 0.7612125394015503, "learning_rate": 6.5597332173803175e-06, "loss": 0.5594, "step": 2393 }, { "epoch": 2.2976476236197794, "grad_norm": 1.4126483029004413, "learning_rate": 6.556550071647355e-06, "loss": 0.5007, "step": 2394 }, { "epoch": 2.298607777244359, "grad_norm": 1.2741787590669489, "learning_rate": 6.553366227195702e-06, "loss": 0.64, "step": 2395 }, { "epoch": 2.299567930868939, "grad_norm": 0.8095186933297978, "learning_rate": 6.550181685454556e-06, "loss": 0.5044, "step": 2396 }, { "epoch": 2.3005280844935188, "grad_norm": 0.8226823342970352, "learning_rate": 6.546996447853421e-06, "loss": 0.607, "step": 2397 }, { "epoch": 2.301488238118099, "grad_norm": 0.8715421720307045, "learning_rate": 6.5438105158221175e-06, "loss": 0.656, "step": 2398 }, { "epoch": 2.302448391742679, "grad_norm": 1.101523216835279, "learning_rate": 6.5406238907907745e-06, "loss": 0.5678, "step": 2399 }, { "epoch": 2.3034085453672586, "grad_norm": 0.7704240982743515, "learning_rate": 6.537436574189837e-06, "loss": 0.4556, "step": 2400 }, { "epoch": 2.304368698991839, "grad_norm": 0.8652780790595904, "learning_rate": 6.5342485674500554e-06, "loss": 0.555, "step": 2401 }, { "epoch": 2.3053288526164186, "grad_norm": 0.8223093427849086, "learning_rate": 6.531059872002492e-06, "loss": 0.6208, "step": 2402 }, { "epoch": 2.3062890062409984, "grad_norm": 0.7804148510032975, "learning_rate": 6.527870489278519e-06, "loss": 0.4957, "step": 2403 }, { "epoch": 2.3072491598655787, "grad_norm": 1.0162376854521107, "learning_rate": 6.524680420709813e-06, "loss": 0.6554, "step": 2404 }, { "epoch": 2.3082093134901585, "grad_norm": 1.127049629454465, "learning_rate": 6.521489667728365e-06, "loss": 0.6514, "step": 2405 }, { "epoch": 2.3091694671147383, "grad_norm": 1.0337902888295218, "learning_rate": 6.51829823176647e-06, "loss": 0.5556, "step": 2406 }, { "epoch": 2.3101296207393185, "grad_norm": 1.5368027072115629, "learning_rate": 6.5151061142567285e-06, "loss": 0.7153, "step": 2407 }, { "epoch": 2.3110897743638983, "grad_norm": 1.2735974926243365, "learning_rate": 6.511913316632046e-06, "loss": 0.6018, "step": 2408 }, { "epoch": 2.312049927988478, "grad_norm": 1.0713024100411563, "learning_rate": 6.508719840325639e-06, "loss": 0.5701, "step": 2409 }, { "epoch": 2.313010081613058, "grad_norm": 1.1720074944375385, "learning_rate": 6.505525686771021e-06, "loss": 0.5361, "step": 2410 }, { "epoch": 2.313970235237638, "grad_norm": 0.8740805188987985, "learning_rate": 6.502330857402018e-06, "loss": 0.4821, "step": 2411 }, { "epoch": 2.314930388862218, "grad_norm": 1.365982632179892, "learning_rate": 6.49913535365275e-06, "loss": 0.5042, "step": 2412 }, { "epoch": 2.3158905424867977, "grad_norm": 0.9936505671135138, "learning_rate": 6.495939176957648e-06, "loss": 0.5361, "step": 2413 }, { "epoch": 2.316850696111378, "grad_norm": 1.1147677332627948, "learning_rate": 6.49274232875144e-06, "loss": 0.6195, "step": 2414 }, { "epoch": 2.3178108497359577, "grad_norm": 0.8316161499661874, "learning_rate": 6.489544810469157e-06, "loss": 0.5517, "step": 2415 }, { "epoch": 2.3187710033605375, "grad_norm": 1.0740806960093239, "learning_rate": 6.4863466235461316e-06, "loss": 0.5923, "step": 2416 }, { "epoch": 2.319731156985118, "grad_norm": 0.9476388826500379, "learning_rate": 6.483147769417992e-06, "loss": 0.638, "step": 2417 }, { "epoch": 2.3206913106096976, "grad_norm": 0.856570508314201, "learning_rate": 6.479948249520676e-06, "loss": 0.5319, "step": 2418 }, { "epoch": 2.3216514642342774, "grad_norm": 0.9952770962681086, "learning_rate": 6.476748065290409e-06, "loss": 0.5742, "step": 2419 }, { "epoch": 2.3226116178588576, "grad_norm": 0.8505254098783909, "learning_rate": 6.473547218163723e-06, "loss": 0.5583, "step": 2420 }, { "epoch": 2.3235717714834374, "grad_norm": 1.3576772631509149, "learning_rate": 6.4703457095774395e-06, "loss": 0.6635, "step": 2421 }, { "epoch": 2.324531925108017, "grad_norm": 1.5917294486410767, "learning_rate": 6.467143540968684e-06, "loss": 0.503, "step": 2422 }, { "epoch": 2.325492078732597, "grad_norm": 1.3246665425153055, "learning_rate": 6.4639407137748774e-06, "loss": 0.5936, "step": 2423 }, { "epoch": 2.3264522323571772, "grad_norm": 0.9339596885616838, "learning_rate": 6.4607372294337335e-06, "loss": 0.5399, "step": 2424 }, { "epoch": 2.327412385981757, "grad_norm": 0.8177555876561722, "learning_rate": 6.457533089383261e-06, "loss": 0.4806, "step": 2425 }, { "epoch": 2.328372539606337, "grad_norm": 0.7588436046599981, "learning_rate": 6.454328295061764e-06, "loss": 0.5089, "step": 2426 }, { "epoch": 2.329332693230917, "grad_norm": 0.9211308121686369, "learning_rate": 6.451122847907843e-06, "loss": 0.5871, "step": 2427 }, { "epoch": 2.330292846855497, "grad_norm": 0.8258185144679521, "learning_rate": 6.447916749360386e-06, "loss": 0.4748, "step": 2428 }, { "epoch": 2.3312530004800767, "grad_norm": 0.6898369953331842, "learning_rate": 6.444710000858578e-06, "loss": 0.5427, "step": 2429 }, { "epoch": 2.332213154104657, "grad_norm": 0.9947492183760548, "learning_rate": 6.441502603841892e-06, "loss": 0.539, "step": 2430 }, { "epoch": 2.3331733077292367, "grad_norm": 0.8827734737760261, "learning_rate": 6.438294559750098e-06, "loss": 0.4801, "step": 2431 }, { "epoch": 2.3341334613538165, "grad_norm": 0.7846317356112407, "learning_rate": 6.43508587002325e-06, "loss": 0.4433, "step": 2432 }, { "epoch": 2.3350936149783967, "grad_norm": 1.1191951854872464, "learning_rate": 6.431876536101695e-06, "loss": 0.5674, "step": 2433 }, { "epoch": 2.3360537686029765, "grad_norm": 0.7097853258500872, "learning_rate": 6.42866655942607e-06, "loss": 0.4894, "step": 2434 }, { "epoch": 2.3370139222275563, "grad_norm": 0.837806444193277, "learning_rate": 6.425455941437297e-06, "loss": 0.5178, "step": 2435 }, { "epoch": 2.337974075852136, "grad_norm": 0.8657577841711145, "learning_rate": 6.422244683576589e-06, "loss": 0.6231, "step": 2436 }, { "epoch": 2.3389342294767164, "grad_norm": 0.8181209535021943, "learning_rate": 6.419032787285446e-06, "loss": 0.5201, "step": 2437 }, { "epoch": 2.339894383101296, "grad_norm": 0.9921109623933675, "learning_rate": 6.4158202540056536e-06, "loss": 0.5179, "step": 2438 }, { "epoch": 2.340854536725876, "grad_norm": 0.6649322971466218, "learning_rate": 6.412607085179282e-06, "loss": 0.4719, "step": 2439 }, { "epoch": 2.341814690350456, "grad_norm": 0.9852969707488567, "learning_rate": 6.409393282248692e-06, "loss": 0.5692, "step": 2440 }, { "epoch": 2.342774843975036, "grad_norm": 1.1233013689558224, "learning_rate": 6.406178846656521e-06, "loss": 0.5793, "step": 2441 }, { "epoch": 2.343734997599616, "grad_norm": 1.37952102639272, "learning_rate": 6.402963779845699e-06, "loss": 0.5499, "step": 2442 }, { "epoch": 2.344695151224196, "grad_norm": 0.7376306575442765, "learning_rate": 6.399748083259432e-06, "loss": 0.5611, "step": 2443 }, { "epoch": 2.345655304848776, "grad_norm": 1.0137573272707938, "learning_rate": 6.3965317583412135e-06, "loss": 0.5197, "step": 2444 }, { "epoch": 2.3466154584733556, "grad_norm": 1.0886001252606252, "learning_rate": 6.393314806534815e-06, "loss": 0.6471, "step": 2445 }, { "epoch": 2.347575612097936, "grad_norm": 1.0490385137939633, "learning_rate": 6.390097229284296e-06, "loss": 0.5345, "step": 2446 }, { "epoch": 2.3485357657225157, "grad_norm": 0.7487415961539117, "learning_rate": 6.386879028033988e-06, "loss": 0.509, "step": 2447 }, { "epoch": 2.3494959193470955, "grad_norm": 0.934736531458572, "learning_rate": 6.3836602042285105e-06, "loss": 0.5224, "step": 2448 }, { "epoch": 2.3504560729716752, "grad_norm": 0.7904575332692185, "learning_rate": 6.3804407593127585e-06, "loss": 0.6055, "step": 2449 }, { "epoch": 2.3514162265962555, "grad_norm": 0.9664013802691584, "learning_rate": 6.377220694731904e-06, "loss": 0.5395, "step": 2450 }, { "epoch": 2.3523763802208353, "grad_norm": 1.225826846258607, "learning_rate": 6.374000011931403e-06, "loss": 0.5096, "step": 2451 }, { "epoch": 2.353336533845415, "grad_norm": 0.772428388492896, "learning_rate": 6.370778712356981e-06, "loss": 0.5621, "step": 2452 }, { "epoch": 2.3542966874699953, "grad_norm": 0.9052249121862281, "learning_rate": 6.3675567974546506e-06, "loss": 0.517, "step": 2453 }, { "epoch": 2.355256841094575, "grad_norm": 0.9233270644654981, "learning_rate": 6.364334268670691e-06, "loss": 0.6037, "step": 2454 }, { "epoch": 2.356216994719155, "grad_norm": 0.7753383660795188, "learning_rate": 6.361111127451664e-06, "loss": 0.5784, "step": 2455 }, { "epoch": 2.357177148343735, "grad_norm": 0.671127375300393, "learning_rate": 6.357887375244398e-06, "loss": 0.5535, "step": 2456 }, { "epoch": 2.358137301968315, "grad_norm": 0.9416894937612309, "learning_rate": 6.354663013496006e-06, "loss": 0.526, "step": 2457 }, { "epoch": 2.3590974555928947, "grad_norm": 0.7616480594031287, "learning_rate": 6.3514380436538695e-06, "loss": 0.5698, "step": 2458 }, { "epoch": 2.360057609217475, "grad_norm": 0.8209662430591267, "learning_rate": 6.348212467165639e-06, "loss": 0.5857, "step": 2459 }, { "epoch": 2.361017762842055, "grad_norm": 0.7941866590537507, "learning_rate": 6.344986285479244e-06, "loss": 0.4297, "step": 2460 }, { "epoch": 2.3619779164666346, "grad_norm": 0.9182698646287325, "learning_rate": 6.341759500042882e-06, "loss": 0.5319, "step": 2461 }, { "epoch": 2.3629380700912144, "grad_norm": 0.695721262667141, "learning_rate": 6.338532112305025e-06, "loss": 0.5207, "step": 2462 }, { "epoch": 2.3638982237157946, "grad_norm": 0.8286413082556289, "learning_rate": 6.3353041237144096e-06, "loss": 0.6152, "step": 2463 }, { "epoch": 2.3648583773403744, "grad_norm": 0.7291567543035669, "learning_rate": 6.332075535720047e-06, "loss": 0.4518, "step": 2464 }, { "epoch": 2.365818530964954, "grad_norm": 0.853051426604433, "learning_rate": 6.328846349771215e-06, "loss": 0.4506, "step": 2465 }, { "epoch": 2.3667786845895344, "grad_norm": 1.0738469238991144, "learning_rate": 6.325616567317461e-06, "loss": 0.6481, "step": 2466 }, { "epoch": 2.3677388382141142, "grad_norm": 0.7962807731142384, "learning_rate": 6.3223861898086e-06, "loss": 0.5375, "step": 2467 }, { "epoch": 2.368698991838694, "grad_norm": 0.7000380230503076, "learning_rate": 6.319155218694716e-06, "loss": 0.5056, "step": 2468 }, { "epoch": 2.3696591454632743, "grad_norm": 0.8519805808057961, "learning_rate": 6.315923655426152e-06, "loss": 0.5445, "step": 2469 }, { "epoch": 2.370619299087854, "grad_norm": 0.8403082707183263, "learning_rate": 6.312691501453527e-06, "loss": 0.4714, "step": 2470 }, { "epoch": 2.371579452712434, "grad_norm": 0.8245300889511389, "learning_rate": 6.30945875822772e-06, "loss": 0.5908, "step": 2471 }, { "epoch": 2.372539606337014, "grad_norm": 0.7960992419057993, "learning_rate": 6.306225427199873e-06, "loss": 0.532, "step": 2472 }, { "epoch": 2.373499759961594, "grad_norm": 0.7010826798128924, "learning_rate": 6.302991509821396e-06, "loss": 0.5752, "step": 2473 }, { "epoch": 2.3744599135861737, "grad_norm": 0.8767652832272371, "learning_rate": 6.299757007543954e-06, "loss": 0.6182, "step": 2474 }, { "epoch": 2.3754200672107535, "grad_norm": 0.7931231099255094, "learning_rate": 6.296521921819489e-06, "loss": 0.5304, "step": 2475 }, { "epoch": 2.3763802208353337, "grad_norm": 0.7801694222664188, "learning_rate": 6.29328625410019e-06, "loss": 0.5923, "step": 2476 }, { "epoch": 2.3773403744599135, "grad_norm": 0.7313064367290673, "learning_rate": 6.290050005838519e-06, "loss": 0.5369, "step": 2477 }, { "epoch": 2.3783005280844933, "grad_norm": 1.0568361596817375, "learning_rate": 6.286813178487188e-06, "loss": 0.6592, "step": 2478 }, { "epoch": 2.3792606817090736, "grad_norm": 1.0850130662757338, "learning_rate": 6.283575773499176e-06, "loss": 0.6231, "step": 2479 }, { "epoch": 2.3802208353336534, "grad_norm": 0.8159694780633017, "learning_rate": 6.280337792327722e-06, "loss": 0.5752, "step": 2480 }, { "epoch": 2.381180988958233, "grad_norm": 1.1630419337007967, "learning_rate": 6.277099236426317e-06, "loss": 0.5964, "step": 2481 }, { "epoch": 2.3821411425828134, "grad_norm": 1.0654026310852518, "learning_rate": 6.273860107248718e-06, "loss": 0.5348, "step": 2482 }, { "epoch": 2.383101296207393, "grad_norm": 0.9338942066493532, "learning_rate": 6.270620406248931e-06, "loss": 0.5775, "step": 2483 }, { "epoch": 2.384061449831973, "grad_norm": 0.9846103676052312, "learning_rate": 6.267380134881229e-06, "loss": 0.5793, "step": 2484 }, { "epoch": 2.3850216034565532, "grad_norm": 0.8950992050740468, "learning_rate": 6.26413929460013e-06, "loss": 0.5484, "step": 2485 }, { "epoch": 2.385981757081133, "grad_norm": 1.0231346447472465, "learning_rate": 6.260897886860415e-06, "loss": 0.6427, "step": 2486 }, { "epoch": 2.386941910705713, "grad_norm": 0.9103755200643812, "learning_rate": 6.2576559131171174e-06, "loss": 0.5332, "step": 2487 }, { "epoch": 2.3879020643302926, "grad_norm": 1.0385592409584714, "learning_rate": 6.254413374825523e-06, "loss": 0.5936, "step": 2488 }, { "epoch": 2.388862217954873, "grad_norm": 0.8511355949801235, "learning_rate": 6.251170273441174e-06, "loss": 0.6136, "step": 2489 }, { "epoch": 2.3898223715794527, "grad_norm": 1.6255217972351936, "learning_rate": 6.247926610419865e-06, "loss": 0.5253, "step": 2490 }, { "epoch": 2.390782525204033, "grad_norm": 0.8621126588442573, "learning_rate": 6.244682387217638e-06, "loss": 0.608, "step": 2491 }, { "epoch": 2.3917426788286127, "grad_norm": 0.9102643394903416, "learning_rate": 6.241437605290792e-06, "loss": 0.561, "step": 2492 }, { "epoch": 2.3927028324531925, "grad_norm": 0.718517224449862, "learning_rate": 6.238192266095877e-06, "loss": 0.5079, "step": 2493 }, { "epoch": 2.3936629860777723, "grad_norm": 0.8539126464658889, "learning_rate": 6.234946371089688e-06, "loss": 0.6101, "step": 2494 }, { "epoch": 2.3946231397023525, "grad_norm": 1.0150250357111408, "learning_rate": 6.231699921729275e-06, "loss": 0.669, "step": 2495 }, { "epoch": 2.3955832933269323, "grad_norm": 0.7973364255556628, "learning_rate": 6.228452919471933e-06, "loss": 0.5133, "step": 2496 }, { "epoch": 2.396543446951512, "grad_norm": 0.6775272495463913, "learning_rate": 6.225205365775206e-06, "loss": 0.5478, "step": 2497 }, { "epoch": 2.3975036005760924, "grad_norm": 0.7268722453301552, "learning_rate": 6.221957262096887e-06, "loss": 0.5443, "step": 2498 }, { "epoch": 2.398463754200672, "grad_norm": 0.6320228249184098, "learning_rate": 6.218708609895017e-06, "loss": 0.4648, "step": 2499 }, { "epoch": 2.399423907825252, "grad_norm": 0.8599499360805127, "learning_rate": 6.2154594106278775e-06, "loss": 0.626, "step": 2500 }, { "epoch": 2.4003840614498317, "grad_norm": 1.1867711149291895, "learning_rate": 6.212209665754001e-06, "loss": 0.5687, "step": 2501 }, { "epoch": 2.401344215074412, "grad_norm": 1.0312144455734815, "learning_rate": 6.2089593767321645e-06, "loss": 0.5691, "step": 2502 }, { "epoch": 2.4023043686989918, "grad_norm": 0.6945306805643403, "learning_rate": 6.2057085450213864e-06, "loss": 0.4948, "step": 2503 }, { "epoch": 2.403264522323572, "grad_norm": 0.8814450198559802, "learning_rate": 6.202457172080931e-06, "loss": 0.5091, "step": 2504 }, { "epoch": 2.404224675948152, "grad_norm": 0.8151043308464335, "learning_rate": 6.199205259370304e-06, "loss": 0.5326, "step": 2505 }, { "epoch": 2.4051848295727316, "grad_norm": 0.8106367569477382, "learning_rate": 6.195952808349255e-06, "loss": 0.548, "step": 2506 }, { "epoch": 2.4061449831973114, "grad_norm": 0.7240442322885156, "learning_rate": 6.192699820477775e-06, "loss": 0.446, "step": 2507 }, { "epoch": 2.4071051368218916, "grad_norm": 0.8468881766842207, "learning_rate": 6.189446297216094e-06, "loss": 0.499, "step": 2508 }, { "epoch": 2.4080652904464714, "grad_norm": 0.7916866548270733, "learning_rate": 6.186192240024684e-06, "loss": 0.5276, "step": 2509 }, { "epoch": 2.4090254440710512, "grad_norm": 1.0785379665584012, "learning_rate": 6.182937650364258e-06, "loss": 0.5873, "step": 2510 }, { "epoch": 2.4099855976956315, "grad_norm": 0.8669622864258372, "learning_rate": 6.179682529695766e-06, "loss": 0.452, "step": 2511 }, { "epoch": 2.4109457513202113, "grad_norm": 0.6291905586594797, "learning_rate": 6.176426879480396e-06, "loss": 0.4964, "step": 2512 }, { "epoch": 2.411905904944791, "grad_norm": 1.0081488350066787, "learning_rate": 6.173170701179576e-06, "loss": 0.6294, "step": 2513 }, { "epoch": 2.412866058569371, "grad_norm": 0.8467977670598451, "learning_rate": 6.169913996254967e-06, "loss": 0.5398, "step": 2514 }, { "epoch": 2.413826212193951, "grad_norm": 0.9660318559455972, "learning_rate": 6.166656766168473e-06, "loss": 0.6377, "step": 2515 }, { "epoch": 2.414786365818531, "grad_norm": 1.1855666698060003, "learning_rate": 6.1633990123822265e-06, "loss": 0.5671, "step": 2516 }, { "epoch": 2.415746519443111, "grad_norm": 0.9114746246111138, "learning_rate": 6.1601407363586e-06, "loss": 0.6051, "step": 2517 }, { "epoch": 2.416706673067691, "grad_norm": 1.1117619773902394, "learning_rate": 6.156881939560196e-06, "loss": 0.4939, "step": 2518 }, { "epoch": 2.4176668266922707, "grad_norm": 1.0851973216042656, "learning_rate": 6.153622623449858e-06, "loss": 0.6459, "step": 2519 }, { "epoch": 2.4186269803168505, "grad_norm": 1.094801315584149, "learning_rate": 6.150362789490654e-06, "loss": 0.5101, "step": 2520 }, { "epoch": 2.4195871339414308, "grad_norm": 1.1872767670770754, "learning_rate": 6.147102439145892e-06, "loss": 0.4894, "step": 2521 }, { "epoch": 2.4205472875660106, "grad_norm": 0.8342998603639464, "learning_rate": 6.143841573879107e-06, "loss": 0.6031, "step": 2522 }, { "epoch": 2.4215074411905904, "grad_norm": 0.9269967350467448, "learning_rate": 6.140580195154065e-06, "loss": 0.5567, "step": 2523 }, { "epoch": 2.4224675948151706, "grad_norm": 0.7881215340500689, "learning_rate": 6.137318304434768e-06, "loss": 0.4211, "step": 2524 }, { "epoch": 2.4234277484397504, "grad_norm": 0.9054412068404949, "learning_rate": 6.1340559031854385e-06, "loss": 0.5865, "step": 2525 }, { "epoch": 2.42438790206433, "grad_norm": 0.7678615469923623, "learning_rate": 6.130792992870539e-06, "loss": 0.4593, "step": 2526 }, { "epoch": 2.42534805568891, "grad_norm": 1.38526633381621, "learning_rate": 6.127529574954749e-06, "loss": 0.6489, "step": 2527 }, { "epoch": 2.4263082093134902, "grad_norm": 0.8002741701409554, "learning_rate": 6.124265650902989e-06, "loss": 0.5325, "step": 2528 }, { "epoch": 2.42726836293807, "grad_norm": 1.0439454953549643, "learning_rate": 6.121001222180394e-06, "loss": 0.5914, "step": 2529 }, { "epoch": 2.4282285165626503, "grad_norm": 1.136308977329623, "learning_rate": 6.117736290252335e-06, "loss": 0.534, "step": 2530 }, { "epoch": 2.42918867018723, "grad_norm": 1.2907476156378515, "learning_rate": 6.1144708565844026e-06, "loss": 0.5692, "step": 2531 }, { "epoch": 2.43014882381181, "grad_norm": 0.9229416758707362, "learning_rate": 6.111204922642414e-06, "loss": 0.57, "step": 2532 }, { "epoch": 2.4311089774363897, "grad_norm": 0.8946857688706149, "learning_rate": 6.107938489892417e-06, "loss": 0.5151, "step": 2533 }, { "epoch": 2.43206913106097, "grad_norm": 0.934673631422437, "learning_rate": 6.104671559800672e-06, "loss": 0.6107, "step": 2534 }, { "epoch": 2.4330292846855497, "grad_norm": 0.6521504234555383, "learning_rate": 6.101404133833674e-06, "loss": 0.4653, "step": 2535 }, { "epoch": 2.4339894383101295, "grad_norm": 0.7190371642225567, "learning_rate": 6.098136213458131e-06, "loss": 0.5193, "step": 2536 }, { "epoch": 2.4349495919347097, "grad_norm": 0.81716834203505, "learning_rate": 6.0948678001409835e-06, "loss": 0.6028, "step": 2537 }, { "epoch": 2.4359097455592895, "grad_norm": 0.8223490585415987, "learning_rate": 6.091598895349382e-06, "loss": 0.5116, "step": 2538 }, { "epoch": 2.4368698991838693, "grad_norm": 0.9327330636146485, "learning_rate": 6.088329500550706e-06, "loss": 0.5232, "step": 2539 }, { "epoch": 2.437830052808449, "grad_norm": 0.9036509285453073, "learning_rate": 6.08505961721255e-06, "loss": 0.6361, "step": 2540 }, { "epoch": 2.4387902064330294, "grad_norm": 0.9281376235550863, "learning_rate": 6.081789246802731e-06, "loss": 0.5389, "step": 2541 }, { "epoch": 2.439750360057609, "grad_norm": 0.837035661612115, "learning_rate": 6.078518390789282e-06, "loss": 0.5544, "step": 2542 }, { "epoch": 2.4407105136821894, "grad_norm": 0.9419539604952558, "learning_rate": 6.075247050640458e-06, "loss": 0.5887, "step": 2543 }, { "epoch": 2.441670667306769, "grad_norm": 0.7006552953140716, "learning_rate": 6.071975227824724e-06, "loss": 0.4197, "step": 2544 }, { "epoch": 2.442630820931349, "grad_norm": 1.1036403262202916, "learning_rate": 6.068702923810771e-06, "loss": 0.5927, "step": 2545 }, { "epoch": 2.4435909745559288, "grad_norm": 1.0369776663094283, "learning_rate": 6.0654301400675e-06, "loss": 0.5834, "step": 2546 }, { "epoch": 2.444551128180509, "grad_norm": 1.0670725190330697, "learning_rate": 6.062156878064026e-06, "loss": 0.5631, "step": 2547 }, { "epoch": 2.445511281805089, "grad_norm": 0.7244979837313604, "learning_rate": 6.058883139269685e-06, "loss": 0.5169, "step": 2548 }, { "epoch": 2.4464714354296686, "grad_norm": 0.7279948668054826, "learning_rate": 6.055608925154019e-06, "loss": 0.5487, "step": 2549 }, { "epoch": 2.447431589054249, "grad_norm": 0.7894902126684348, "learning_rate": 6.052334237186792e-06, "loss": 0.4844, "step": 2550 }, { "epoch": 2.4483917426788286, "grad_norm": 0.9028041771275973, "learning_rate": 6.049059076837974e-06, "loss": 0.5776, "step": 2551 }, { "epoch": 2.4493518963034084, "grad_norm": 0.8602142649151429, "learning_rate": 6.045783445577752e-06, "loss": 0.6746, "step": 2552 }, { "epoch": 2.4503120499279882, "grad_norm": 0.6221760764346755, "learning_rate": 6.042507344876518e-06, "loss": 0.432, "step": 2553 }, { "epoch": 2.4512722035525685, "grad_norm": 1.3994045359709253, "learning_rate": 6.0392307762048816e-06, "loss": 0.5736, "step": 2554 }, { "epoch": 2.4522323571771483, "grad_norm": 0.7919210716842101, "learning_rate": 6.03595374103366e-06, "loss": 0.5527, "step": 2555 }, { "epoch": 2.4531925108017285, "grad_norm": 1.268337782412299, "learning_rate": 6.032676240833876e-06, "loss": 0.5697, "step": 2556 }, { "epoch": 2.4541526644263083, "grad_norm": 1.1668017098395764, "learning_rate": 6.029398277076766e-06, "loss": 0.5218, "step": 2557 }, { "epoch": 2.455112818050888, "grad_norm": 0.7976536922741492, "learning_rate": 6.0261198512337725e-06, "loss": 0.5845, "step": 2558 }, { "epoch": 2.456072971675468, "grad_norm": 1.4019875372343662, "learning_rate": 6.022840964776548e-06, "loss": 0.6341, "step": 2559 }, { "epoch": 2.457033125300048, "grad_norm": 0.8710181222304418, "learning_rate": 6.019561619176948e-06, "loss": 0.59, "step": 2560 }, { "epoch": 2.457993278924628, "grad_norm": 1.541788861632369, "learning_rate": 6.016281815907036e-06, "loss": 0.5645, "step": 2561 }, { "epoch": 2.4589534325492077, "grad_norm": 0.8699506962665938, "learning_rate": 6.013001556439079e-06, "loss": 0.4908, "step": 2562 }, { "epoch": 2.459913586173788, "grad_norm": 0.8342701063892572, "learning_rate": 6.0097208422455526e-06, "loss": 0.5588, "step": 2563 }, { "epoch": 2.4608737397983678, "grad_norm": 1.1289484753469146, "learning_rate": 6.006439674799132e-06, "loss": 0.5776, "step": 2564 }, { "epoch": 2.4618338934229476, "grad_norm": 0.7900733026656773, "learning_rate": 6.0031580555727005e-06, "loss": 0.5226, "step": 2565 }, { "epoch": 2.4627940470475274, "grad_norm": 1.028822003256681, "learning_rate": 5.999875986039341e-06, "loss": 0.5302, "step": 2566 }, { "epoch": 2.4637542006721076, "grad_norm": 0.7147457955602052, "learning_rate": 5.996593467672339e-06, "loss": 0.4601, "step": 2567 }, { "epoch": 2.4647143542966874, "grad_norm": 0.7999468549687858, "learning_rate": 5.993310501945184e-06, "loss": 0.5522, "step": 2568 }, { "epoch": 2.4656745079212676, "grad_norm": 0.7916369469848131, "learning_rate": 5.990027090331562e-06, "loss": 0.4862, "step": 2569 }, { "epoch": 2.4666346615458474, "grad_norm": 0.8621250227939997, "learning_rate": 5.986743234305361e-06, "loss": 0.5293, "step": 2570 }, { "epoch": 2.4675948151704272, "grad_norm": 1.5152833660085934, "learning_rate": 5.983458935340669e-06, "loss": 0.5714, "step": 2571 }, { "epoch": 2.468554968795007, "grad_norm": 0.9816276557530337, "learning_rate": 5.980174194911776e-06, "loss": 0.551, "step": 2572 }, { "epoch": 2.4695151224195873, "grad_norm": 0.8311067377126226, "learning_rate": 5.9768890144931626e-06, "loss": 0.572, "step": 2573 }, { "epoch": 2.470475276044167, "grad_norm": 0.8139869998718569, "learning_rate": 5.973603395559514e-06, "loss": 0.6212, "step": 2574 }, { "epoch": 2.471435429668747, "grad_norm": 0.8009013791732073, "learning_rate": 5.970317339585707e-06, "loss": 0.5873, "step": 2575 }, { "epoch": 2.472395583293327, "grad_norm": 0.7741463250953557, "learning_rate": 5.96703084804682e-06, "loss": 0.4747, "step": 2576 }, { "epoch": 2.473355736917907, "grad_norm": 0.8634808960804815, "learning_rate": 5.963743922418122e-06, "loss": 0.5962, "step": 2577 }, { "epoch": 2.4743158905424867, "grad_norm": 1.0551041437421191, "learning_rate": 5.960456564175079e-06, "loss": 0.4885, "step": 2578 }, { "epoch": 2.475276044167067, "grad_norm": 0.7633703215595228, "learning_rate": 5.9571687747933505e-06, "loss": 0.5321, "step": 2579 }, { "epoch": 2.4762361977916467, "grad_norm": 0.8846760381567773, "learning_rate": 5.9538805557487896e-06, "loss": 0.5782, "step": 2580 }, { "epoch": 2.4771963514162265, "grad_norm": 1.1069558920762408, "learning_rate": 5.950591908517446e-06, "loss": 0.659, "step": 2581 }, { "epoch": 2.4781565050408068, "grad_norm": 0.9887189151512948, "learning_rate": 5.9473028345755535e-06, "loss": 0.6149, "step": 2582 }, { "epoch": 2.4791166586653866, "grad_norm": 0.7407597490688271, "learning_rate": 5.944013335399546e-06, "loss": 0.5388, "step": 2583 }, { "epoch": 2.4800768122899663, "grad_norm": 0.721015806178182, "learning_rate": 5.940723412466044e-06, "loss": 0.5588, "step": 2584 }, { "epoch": 2.481036965914546, "grad_norm": 1.1569650155916376, "learning_rate": 5.937433067251857e-06, "loss": 0.618, "step": 2585 }, { "epoch": 2.4819971195391264, "grad_norm": 0.8452246998184166, "learning_rate": 5.934142301233986e-06, "loss": 0.6016, "step": 2586 }, { "epoch": 2.482957273163706, "grad_norm": 0.7626939764109372, "learning_rate": 5.9308511158896235e-06, "loss": 0.5162, "step": 2587 }, { "epoch": 2.483917426788286, "grad_norm": 0.6678690436111278, "learning_rate": 5.927559512696143e-06, "loss": 0.5555, "step": 2588 }, { "epoch": 2.484877580412866, "grad_norm": 1.1940539022434866, "learning_rate": 5.924267493131113e-06, "loss": 0.5045, "step": 2589 }, { "epoch": 2.485837734037446, "grad_norm": 0.8289472087525059, "learning_rate": 5.9209750586722855e-06, "loss": 0.709, "step": 2590 }, { "epoch": 2.486797887662026, "grad_norm": 0.8942456956420332, "learning_rate": 5.917682210797598e-06, "loss": 0.5476, "step": 2591 }, { "epoch": 2.487758041286606, "grad_norm": 0.8190611328252395, "learning_rate": 5.9143889509851785e-06, "loss": 0.5357, "step": 2592 }, { "epoch": 2.488718194911186, "grad_norm": 0.7973313245158252, "learning_rate": 5.91109528071333e-06, "loss": 0.5948, "step": 2593 }, { "epoch": 2.4896783485357656, "grad_norm": 0.7813854898302107, "learning_rate": 5.9078012014605505e-06, "loss": 0.5987, "step": 2594 }, { "epoch": 2.490638502160346, "grad_norm": 0.939091878596093, "learning_rate": 5.904506714705514e-06, "loss": 0.5913, "step": 2595 }, { "epoch": 2.4915986557849257, "grad_norm": 0.8242209532162841, "learning_rate": 5.901211821927084e-06, "loss": 0.5508, "step": 2596 }, { "epoch": 2.4925588094095055, "grad_norm": 0.9757817372368698, "learning_rate": 5.897916524604299e-06, "loss": 0.5614, "step": 2597 }, { "epoch": 2.4935189630340853, "grad_norm": 0.995191401851482, "learning_rate": 5.894620824216385e-06, "loss": 0.62, "step": 2598 }, { "epoch": 2.4944791166586655, "grad_norm": 0.808094354640379, "learning_rate": 5.8913247222427464e-06, "loss": 0.552, "step": 2599 }, { "epoch": 2.4954392702832453, "grad_norm": 0.7117049324390826, "learning_rate": 5.888028220162967e-06, "loss": 0.4879, "step": 2600 }, { "epoch": 2.496399423907825, "grad_norm": 1.0789118806394233, "learning_rate": 5.884731319456812e-06, "loss": 0.5194, "step": 2601 }, { "epoch": 2.4973595775324053, "grad_norm": 0.8976115974858392, "learning_rate": 5.881434021604224e-06, "loss": 0.47, "step": 2602 }, { "epoch": 2.498319731156985, "grad_norm": 0.7951604044383246, "learning_rate": 5.878136328085328e-06, "loss": 0.6132, "step": 2603 }, { "epoch": 2.499279884781565, "grad_norm": 0.8176679388810381, "learning_rate": 5.874838240380419e-06, "loss": 0.4949, "step": 2604 }, { "epoch": 2.5002400384061447, "grad_norm": 1.0206553439644166, "learning_rate": 5.871539759969978e-06, "loss": 0.6595, "step": 2605 }, { "epoch": 2.501200192030725, "grad_norm": 0.9135933590528469, "learning_rate": 5.8682408883346535e-06, "loss": 0.6286, "step": 2606 }, { "epoch": 2.5021603456553048, "grad_norm": 1.0196737035087236, "learning_rate": 5.864941626955274e-06, "loss": 0.6169, "step": 2607 }, { "epoch": 2.503120499279885, "grad_norm": 0.8539674214757117, "learning_rate": 5.861641977312844e-06, "loss": 0.5314, "step": 2608 }, { "epoch": 2.504080652904465, "grad_norm": 0.8142730437979234, "learning_rate": 5.858341940888543e-06, "loss": 0.5524, "step": 2609 }, { "epoch": 2.5050408065290446, "grad_norm": 0.7009855412887921, "learning_rate": 5.855041519163718e-06, "loss": 0.5173, "step": 2610 }, { "epoch": 2.5060009601536244, "grad_norm": 0.8589489267033282, "learning_rate": 5.851740713619896e-06, "loss": 0.5515, "step": 2611 }, { "epoch": 2.5069611137782046, "grad_norm": 0.9151163237143922, "learning_rate": 5.848439525738772e-06, "loss": 0.5006, "step": 2612 }, { "epoch": 2.5079212674027844, "grad_norm": 1.0074930032169513, "learning_rate": 5.845137957002214e-06, "loss": 0.6238, "step": 2613 }, { "epoch": 2.5088814210273642, "grad_norm": 0.74254594800641, "learning_rate": 5.841836008892262e-06, "loss": 0.528, "step": 2614 }, { "epoch": 2.5098415746519445, "grad_norm": 0.957498573312248, "learning_rate": 5.838533682891123e-06, "loss": 0.5283, "step": 2615 }, { "epoch": 2.5108017282765243, "grad_norm": 1.025646545028794, "learning_rate": 5.835230980481177e-06, "loss": 0.4798, "step": 2616 }, { "epoch": 2.511761881901104, "grad_norm": 1.0758798105827811, "learning_rate": 5.831927903144972e-06, "loss": 0.5858, "step": 2617 }, { "epoch": 2.512722035525684, "grad_norm": 0.8959791703323698, "learning_rate": 5.828624452365226e-06, "loss": 0.5952, "step": 2618 }, { "epoch": 2.513682189150264, "grad_norm": 1.156944894505362, "learning_rate": 5.82532062962482e-06, "loss": 0.5721, "step": 2619 }, { "epoch": 2.514642342774844, "grad_norm": 0.7332856130434292, "learning_rate": 5.822016436406805e-06, "loss": 0.4757, "step": 2620 }, { "epoch": 2.515602496399424, "grad_norm": 0.8117831209037574, "learning_rate": 5.818711874194398e-06, "loss": 0.5984, "step": 2621 }, { "epoch": 2.516562650024004, "grad_norm": 0.9650163653865688, "learning_rate": 5.8154069444709825e-06, "loss": 0.6349, "step": 2622 }, { "epoch": 2.5175228036485837, "grad_norm": 1.0103479711483954, "learning_rate": 5.8121016487201074e-06, "loss": 0.5786, "step": 2623 }, { "epoch": 2.5184829572731635, "grad_norm": 0.8016294156669418, "learning_rate": 5.80879598842548e-06, "loss": 0.5816, "step": 2624 }, { "epoch": 2.5194431108977438, "grad_norm": 0.9679529446312245, "learning_rate": 5.805489965070983e-06, "loss": 0.5869, "step": 2625 }, { "epoch": 2.5204032645223235, "grad_norm": 1.3168855218081659, "learning_rate": 5.8021835801406475e-06, "loss": 0.5603, "step": 2626 }, { "epoch": 2.5213634181469033, "grad_norm": 0.7150028777644813, "learning_rate": 5.798876835118681e-06, "loss": 0.4945, "step": 2627 }, { "epoch": 2.5223235717714836, "grad_norm": 0.6980508226878578, "learning_rate": 5.79556973148944e-06, "loss": 0.5362, "step": 2628 }, { "epoch": 2.5232837253960634, "grad_norm": 0.7784072742764976, "learning_rate": 5.792262270737451e-06, "loss": 0.5365, "step": 2629 }, { "epoch": 2.524243879020643, "grad_norm": 0.8986762473204757, "learning_rate": 5.788954454347397e-06, "loss": 0.6289, "step": 2630 }, { "epoch": 2.525204032645223, "grad_norm": 0.7419724909810675, "learning_rate": 5.7856462838041235e-06, "loss": 0.5645, "step": 2631 }, { "epoch": 2.526164186269803, "grad_norm": 0.9588780475552959, "learning_rate": 5.782337760592631e-06, "loss": 0.5349, "step": 2632 }, { "epoch": 2.527124339894383, "grad_norm": 0.8617805131439098, "learning_rate": 5.7790288861980795e-06, "loss": 0.5904, "step": 2633 }, { "epoch": 2.5280844935189632, "grad_norm": 0.9187555183221969, "learning_rate": 5.77571966210579e-06, "loss": 0.5282, "step": 2634 }, { "epoch": 2.529044647143543, "grad_norm": 0.8157015748524258, "learning_rate": 5.772410089801234e-06, "loss": 0.5137, "step": 2635 }, { "epoch": 2.530004800768123, "grad_norm": 1.1401835335195956, "learning_rate": 5.769100170770048e-06, "loss": 0.6383, "step": 2636 }, { "epoch": 2.5309649543927026, "grad_norm": 0.8916299039532491, "learning_rate": 5.765789906498015e-06, "loss": 0.5458, "step": 2637 }, { "epoch": 2.531925108017283, "grad_norm": 0.7400785817111695, "learning_rate": 5.762479298471078e-06, "loss": 0.4933, "step": 2638 }, { "epoch": 2.5328852616418627, "grad_norm": 0.6826164421018314, "learning_rate": 5.759168348175335e-06, "loss": 0.5146, "step": 2639 }, { "epoch": 2.5338454152664425, "grad_norm": 0.7028000807872307, "learning_rate": 5.755857057097036e-06, "loss": 0.5109, "step": 2640 }, { "epoch": 2.5348055688910227, "grad_norm": 0.8271127193252461, "learning_rate": 5.752545426722583e-06, "loss": 0.6186, "step": 2641 }, { "epoch": 2.5357657225156025, "grad_norm": 0.9090280356733762, "learning_rate": 5.7492334585385325e-06, "loss": 0.5971, "step": 2642 }, { "epoch": 2.5367258761401823, "grad_norm": 0.7688457693964327, "learning_rate": 5.745921154031593e-06, "loss": 0.477, "step": 2643 }, { "epoch": 2.537686029764762, "grad_norm": 1.3510511190987449, "learning_rate": 5.742608514688619e-06, "loss": 0.4879, "step": 2644 }, { "epoch": 2.5386461833893423, "grad_norm": 0.9172984442459522, "learning_rate": 5.739295541996622e-06, "loss": 0.5386, "step": 2645 }, { "epoch": 2.539606337013922, "grad_norm": 0.7757519159147409, "learning_rate": 5.7359822374427565e-06, "loss": 0.5726, "step": 2646 }, { "epoch": 2.5405664906385024, "grad_norm": 0.7642870473006498, "learning_rate": 5.732668602514334e-06, "loss": 0.5369, "step": 2647 }, { "epoch": 2.541526644263082, "grad_norm": 0.7056980866522469, "learning_rate": 5.729354638698806e-06, "loss": 0.5093, "step": 2648 }, { "epoch": 2.542486797887662, "grad_norm": 1.0040681880764843, "learning_rate": 5.726040347483779e-06, "loss": 0.6165, "step": 2649 }, { "epoch": 2.5434469515122418, "grad_norm": 0.7160434328936509, "learning_rate": 5.722725730357001e-06, "loss": 0.4583, "step": 2650 }, { "epoch": 2.544407105136822, "grad_norm": 0.7663627592528942, "learning_rate": 5.719410788806366e-06, "loss": 0.5446, "step": 2651 }, { "epoch": 2.545367258761402, "grad_norm": 1.2333848897919644, "learning_rate": 5.716095524319919e-06, "loss": 0.4747, "step": 2652 }, { "epoch": 2.5463274123859816, "grad_norm": 0.8441339214750823, "learning_rate": 5.712779938385846e-06, "loss": 0.5599, "step": 2653 }, { "epoch": 2.547287566010562, "grad_norm": 1.3150815028154936, "learning_rate": 5.709464032492478e-06, "loss": 0.5284, "step": 2654 }, { "epoch": 2.5482477196351416, "grad_norm": 0.9705168803812315, "learning_rate": 5.706147808128288e-06, "loss": 0.6257, "step": 2655 }, { "epoch": 2.5492078732597214, "grad_norm": 1.1119672371181084, "learning_rate": 5.702831266781898e-06, "loss": 0.5536, "step": 2656 }, { "epoch": 2.550168026884301, "grad_norm": 1.0441494164840175, "learning_rate": 5.699514409942064e-06, "loss": 0.4808, "step": 2657 }, { "epoch": 2.5511281805088815, "grad_norm": 0.999088897102054, "learning_rate": 5.69619723909769e-06, "loss": 0.6354, "step": 2658 }, { "epoch": 2.5520883341334613, "grad_norm": 1.1336520390832063, "learning_rate": 5.6928797557378145e-06, "loss": 0.584, "step": 2659 }, { "epoch": 2.5530484877580415, "grad_norm": 0.8313811831392457, "learning_rate": 5.689561961351624e-06, "loss": 0.5328, "step": 2660 }, { "epoch": 2.5540086413826213, "grad_norm": 0.7882267254109605, "learning_rate": 5.68624385742844e-06, "loss": 0.4417, "step": 2661 }, { "epoch": 2.554968795007201, "grad_norm": 0.8101725749138518, "learning_rate": 5.682925445457726e-06, "loss": 0.5293, "step": 2662 }, { "epoch": 2.555928948631781, "grad_norm": 1.4659801958815954, "learning_rate": 5.679606726929076e-06, "loss": 0.5396, "step": 2663 }, { "epoch": 2.556889102256361, "grad_norm": 0.6337107603966345, "learning_rate": 5.676287703332234e-06, "loss": 0.4517, "step": 2664 }, { "epoch": 2.557849255880941, "grad_norm": 0.9590447793475039, "learning_rate": 5.672968376157071e-06, "loss": 0.5785, "step": 2665 }, { "epoch": 2.558809409505521, "grad_norm": 0.8587741326057172, "learning_rate": 5.6696487468935955e-06, "loss": 0.5862, "step": 2666 }, { "epoch": 2.559769563130101, "grad_norm": 0.8252794481032426, "learning_rate": 5.666328817031958e-06, "loss": 0.4888, "step": 2667 }, { "epoch": 2.5607297167546808, "grad_norm": 0.8792041944115155, "learning_rate": 5.663008588062435e-06, "loss": 0.4866, "step": 2668 }, { "epoch": 2.5616898703792605, "grad_norm": 0.8471839319470963, "learning_rate": 5.659688061475445e-06, "loss": 0.5663, "step": 2669 }, { "epoch": 2.5626500240038403, "grad_norm": 0.8411938499262136, "learning_rate": 5.656367238761536e-06, "loss": 0.4692, "step": 2670 }, { "epoch": 2.5636101776284206, "grad_norm": 0.9384839711947874, "learning_rate": 5.653046121411389e-06, "loss": 0.5642, "step": 2671 }, { "epoch": 2.5645703312530004, "grad_norm": 0.8747630966252606, "learning_rate": 5.649724710915819e-06, "loss": 0.3891, "step": 2672 }, { "epoch": 2.5655304848775806, "grad_norm": 0.8844913659566841, "learning_rate": 5.646403008765772e-06, "loss": 0.5589, "step": 2673 }, { "epoch": 2.5664906385021604, "grad_norm": 1.0288516402390138, "learning_rate": 5.643081016452323e-06, "loss": 0.5854, "step": 2674 }, { "epoch": 2.56745079212674, "grad_norm": 0.7543286976313875, "learning_rate": 5.639758735466682e-06, "loss": 0.5286, "step": 2675 }, { "epoch": 2.56841094575132, "grad_norm": 1.2760102974997547, "learning_rate": 5.636436167300181e-06, "loss": 0.6887, "step": 2676 }, { "epoch": 2.5693710993759002, "grad_norm": 1.0705389315763687, "learning_rate": 5.633113313444288e-06, "loss": 0.6126, "step": 2677 }, { "epoch": 2.57033125300048, "grad_norm": 0.6048591005940751, "learning_rate": 5.629790175390598e-06, "loss": 0.4894, "step": 2678 }, { "epoch": 2.5712914066250603, "grad_norm": 0.9146898493626733, "learning_rate": 5.626466754630829e-06, "loss": 0.5752, "step": 2679 }, { "epoch": 2.57225156024964, "grad_norm": 0.7276420954527769, "learning_rate": 5.623143052656831e-06, "loss": 0.5076, "step": 2680 }, { "epoch": 2.57321171387422, "grad_norm": 0.8573026160550506, "learning_rate": 5.619819070960577e-06, "loss": 0.5449, "step": 2681 }, { "epoch": 2.5741718674987997, "grad_norm": 0.9247452518873407, "learning_rate": 5.616494811034168e-06, "loss": 0.5626, "step": 2682 }, { "epoch": 2.5751320211233795, "grad_norm": 0.9310619202537215, "learning_rate": 5.613170274369827e-06, "loss": 0.5535, "step": 2683 }, { "epoch": 2.5760921747479597, "grad_norm": 1.0016886443328092, "learning_rate": 5.609845462459907e-06, "loss": 0.5882, "step": 2684 }, { "epoch": 2.5770523283725395, "grad_norm": 0.880893723275252, "learning_rate": 5.606520376796875e-06, "loss": 0.5612, "step": 2685 }, { "epoch": 2.5780124819971197, "grad_norm": 0.9422909350394907, "learning_rate": 5.603195018873329e-06, "loss": 0.5805, "step": 2686 }, { "epoch": 2.5789726356216995, "grad_norm": 0.6969268253062391, "learning_rate": 5.59986939018199e-06, "loss": 0.562, "step": 2687 }, { "epoch": 2.5799327892462793, "grad_norm": 1.004850493796453, "learning_rate": 5.596543492215691e-06, "loss": 0.59, "step": 2688 }, { "epoch": 2.580892942870859, "grad_norm": 1.0271757754218298, "learning_rate": 5.5932173264673975e-06, "loss": 0.6588, "step": 2689 }, { "epoch": 2.5818530964954394, "grad_norm": 0.8475803095457829, "learning_rate": 5.589890894430186e-06, "loss": 0.6127, "step": 2690 }, { "epoch": 2.582813250120019, "grad_norm": 0.9914674688062514, "learning_rate": 5.586564197597256e-06, "loss": 0.6322, "step": 2691 }, { "epoch": 2.5837734037445994, "grad_norm": 0.7146581122917471, "learning_rate": 5.583237237461929e-06, "loss": 0.4758, "step": 2692 }, { "epoch": 2.584733557369179, "grad_norm": 0.7108639155808897, "learning_rate": 5.579910015517642e-06, "loss": 0.518, "step": 2693 }, { "epoch": 2.585693710993759, "grad_norm": 0.8065803272947183, "learning_rate": 5.576582533257946e-06, "loss": 0.6289, "step": 2694 }, { "epoch": 2.586653864618339, "grad_norm": 1.257391284449206, "learning_rate": 5.573254792176515e-06, "loss": 0.5209, "step": 2695 }, { "epoch": 2.5876140182429186, "grad_norm": 1.0507890610694985, "learning_rate": 5.569926793767136e-06, "loss": 0.5112, "step": 2696 }, { "epoch": 2.588574171867499, "grad_norm": 0.7154952749546757, "learning_rate": 5.566598539523711e-06, "loss": 0.5414, "step": 2697 }, { "epoch": 2.5895343254920786, "grad_norm": 0.9816709190938574, "learning_rate": 5.563270030940258e-06, "loss": 0.6005, "step": 2698 }, { "epoch": 2.590494479116659, "grad_norm": 0.8737610648639607, "learning_rate": 5.559941269510908e-06, "loss": 0.5102, "step": 2699 }, { "epoch": 2.5914546327412387, "grad_norm": 1.2717539494516743, "learning_rate": 5.556612256729909e-06, "loss": 0.5453, "step": 2700 }, { "epoch": 2.5924147863658185, "grad_norm": 0.9380570284034553, "learning_rate": 5.553282994091617e-06, "loss": 0.5742, "step": 2701 }, { "epoch": 2.5933749399903983, "grad_norm": 0.8811855983397117, "learning_rate": 5.549953483090506e-06, "loss": 0.6113, "step": 2702 }, { "epoch": 2.5943350936149785, "grad_norm": 0.8305522406609445, "learning_rate": 5.546623725221152e-06, "loss": 0.5322, "step": 2703 }, { "epoch": 2.5952952472395583, "grad_norm": 0.6775484163844926, "learning_rate": 5.543293721978251e-06, "loss": 0.5348, "step": 2704 }, { "epoch": 2.5962554008641385, "grad_norm": 1.1063542605939185, "learning_rate": 5.539963474856607e-06, "loss": 0.6181, "step": 2705 }, { "epoch": 2.5972155544887183, "grad_norm": 0.7901809097712382, "learning_rate": 5.536632985351132e-06, "loss": 0.5625, "step": 2706 }, { "epoch": 2.598175708113298, "grad_norm": 1.2684366322016223, "learning_rate": 5.533302254956843e-06, "loss": 0.6313, "step": 2707 }, { "epoch": 2.599135861737878, "grad_norm": 0.7573406720515273, "learning_rate": 5.529971285168873e-06, "loss": 0.5109, "step": 2708 }, { "epoch": 2.600096015362458, "grad_norm": 0.8621648992976847, "learning_rate": 5.526640077482462e-06, "loss": 0.5749, "step": 2709 }, { "epoch": 2.601056168987038, "grad_norm": 1.0427000582019204, "learning_rate": 5.523308633392945e-06, "loss": 0.5928, "step": 2710 }, { "epoch": 2.6020163226116177, "grad_norm": 0.7120470494057578, "learning_rate": 5.519976954395779e-06, "loss": 0.4738, "step": 2711 }, { "epoch": 2.602976476236198, "grad_norm": 0.9549879043221737, "learning_rate": 5.516645041986516e-06, "loss": 0.5531, "step": 2712 }, { "epoch": 2.603936629860778, "grad_norm": 0.7125785142263011, "learning_rate": 5.513312897660815e-06, "loss": 0.5376, "step": 2713 }, { "epoch": 2.6048967834853576, "grad_norm": 0.8156493928289598, "learning_rate": 5.509980522914442e-06, "loss": 0.4535, "step": 2714 }, { "epoch": 2.6058569371099374, "grad_norm": 1.2267262872082871, "learning_rate": 5.506647919243265e-06, "loss": 0.5489, "step": 2715 }, { "epoch": 2.6068170907345176, "grad_norm": 1.1220899948401415, "learning_rate": 5.503315088143251e-06, "loss": 0.6708, "step": 2716 }, { "epoch": 2.6077772443590974, "grad_norm": 1.0267944839277785, "learning_rate": 5.499982031110475e-06, "loss": 0.6267, "step": 2717 }, { "epoch": 2.6087373979836777, "grad_norm": 1.0397177787464587, "learning_rate": 5.496648749641109e-06, "loss": 0.6421, "step": 2718 }, { "epoch": 2.6096975516082574, "grad_norm": 0.9538397423096311, "learning_rate": 5.493315245231429e-06, "loss": 0.6526, "step": 2719 }, { "epoch": 2.6106577052328372, "grad_norm": 0.696548871446055, "learning_rate": 5.489981519377808e-06, "loss": 0.4828, "step": 2720 }, { "epoch": 2.611617858857417, "grad_norm": 0.9569524199815257, "learning_rate": 5.486647573576721e-06, "loss": 0.6291, "step": 2721 }, { "epoch": 2.6125780124819973, "grad_norm": 0.877754376844991, "learning_rate": 5.483313409324741e-06, "loss": 0.4632, "step": 2722 }, { "epoch": 2.613538166106577, "grad_norm": 1.167379628850693, "learning_rate": 5.479979028118536e-06, "loss": 0.6301, "step": 2723 }, { "epoch": 2.614498319731157, "grad_norm": 0.8687488683666513, "learning_rate": 5.476644431454878e-06, "loss": 0.5586, "step": 2724 }, { "epoch": 2.615458473355737, "grad_norm": 0.852266297050393, "learning_rate": 5.473309620830628e-06, "loss": 0.5095, "step": 2725 }, { "epoch": 2.616418626980317, "grad_norm": 1.3953572875847209, "learning_rate": 5.469974597742749e-06, "loss": 0.6421, "step": 2726 }, { "epoch": 2.6173787806048967, "grad_norm": 1.0570854492070743, "learning_rate": 5.4666393636882965e-06, "loss": 0.5744, "step": 2727 }, { "epoch": 2.6183389342294765, "grad_norm": 0.7715356971688643, "learning_rate": 5.463303920164423e-06, "loss": 0.5997, "step": 2728 }, { "epoch": 2.6192990878540567, "grad_norm": 0.9542776861808182, "learning_rate": 5.459968268668372e-06, "loss": 0.673, "step": 2729 }, { "epoch": 2.6202592414786365, "grad_norm": 1.22447262330308, "learning_rate": 5.456632410697481e-06, "loss": 0.4847, "step": 2730 }, { "epoch": 2.6212193951032168, "grad_norm": 1.0092323973394532, "learning_rate": 5.453296347749183e-06, "loss": 0.5165, "step": 2731 }, { "epoch": 2.6221795487277966, "grad_norm": 0.891327606516193, "learning_rate": 5.449960081321e-06, "loss": 0.5879, "step": 2732 }, { "epoch": 2.6231397023523764, "grad_norm": 0.8034507014085214, "learning_rate": 5.446623612910545e-06, "loss": 0.5267, "step": 2733 }, { "epoch": 2.624099855976956, "grad_norm": 1.155725016281046, "learning_rate": 5.443286944015525e-06, "loss": 0.6381, "step": 2734 }, { "epoch": 2.6250600096015364, "grad_norm": 0.7746555055248321, "learning_rate": 5.439950076133734e-06, "loss": 0.4566, "step": 2735 }, { "epoch": 2.626020163226116, "grad_norm": 1.1105464246084977, "learning_rate": 5.4366130107630566e-06, "loss": 0.5964, "step": 2736 }, { "epoch": 2.626980316850696, "grad_norm": 1.346702851556844, "learning_rate": 5.4332757494014675e-06, "loss": 0.6131, "step": 2737 }, { "epoch": 2.6279404704752762, "grad_norm": 0.8239642729893849, "learning_rate": 5.429938293547024e-06, "loss": 0.622, "step": 2738 }, { "epoch": 2.628900624099856, "grad_norm": 0.7634265342859452, "learning_rate": 5.426600644697877e-06, "loss": 0.634, "step": 2739 }, { "epoch": 2.629860777724436, "grad_norm": 0.7761906635749541, "learning_rate": 5.423262804352262e-06, "loss": 0.5009, "step": 2740 }, { "epoch": 2.6308209313490156, "grad_norm": 0.9668939303491332, "learning_rate": 5.419924774008496e-06, "loss": 0.5772, "step": 2741 }, { "epoch": 2.631781084973596, "grad_norm": 0.8888669328634402, "learning_rate": 5.416586555164989e-06, "loss": 0.6036, "step": 2742 }, { "epoch": 2.6327412385981757, "grad_norm": 0.7531127070349425, "learning_rate": 5.4132481493202295e-06, "loss": 0.509, "step": 2743 }, { "epoch": 2.633701392222756, "grad_norm": 1.3345553635935594, "learning_rate": 5.409909557972796e-06, "loss": 0.6929, "step": 2744 }, { "epoch": 2.6346615458473357, "grad_norm": 0.8792540526382455, "learning_rate": 5.406570782621341e-06, "loss": 0.5244, "step": 2745 }, { "epoch": 2.6356216994719155, "grad_norm": 0.9378433751029297, "learning_rate": 5.403231824764609e-06, "loss": 0.6096, "step": 2746 }, { "epoch": 2.6365818530964953, "grad_norm": 0.8570794896699859, "learning_rate": 5.39989268590142e-06, "loss": 0.6032, "step": 2747 }, { "epoch": 2.6375420067210755, "grad_norm": 1.2305376814387052, "learning_rate": 5.396553367530679e-06, "loss": 0.528, "step": 2748 }, { "epoch": 2.6385021603456553, "grad_norm": 1.1737346758075697, "learning_rate": 5.39321387115137e-06, "loss": 0.5435, "step": 2749 }, { "epoch": 2.639462313970235, "grad_norm": 0.7991841020745009, "learning_rate": 5.389874198262559e-06, "loss": 0.5529, "step": 2750 }, { "epoch": 2.6404224675948154, "grad_norm": 0.7162390715111451, "learning_rate": 5.386534350363387e-06, "loss": 0.4799, "step": 2751 }, { "epoch": 2.641382621219395, "grad_norm": 0.8028441497198588, "learning_rate": 5.3831943289530765e-06, "loss": 0.5623, "step": 2752 }, { "epoch": 2.642342774843975, "grad_norm": 0.8272264191594435, "learning_rate": 5.379854135530929e-06, "loss": 0.4712, "step": 2753 }, { "epoch": 2.6433029284685547, "grad_norm": 0.9808884015373625, "learning_rate": 5.376513771596319e-06, "loss": 0.5192, "step": 2754 }, { "epoch": 2.644263082093135, "grad_norm": 0.8629510818598704, "learning_rate": 5.373173238648704e-06, "loss": 0.4896, "step": 2755 }, { "epoch": 2.645223235717715, "grad_norm": 1.0082456714635617, "learning_rate": 5.369832538187609e-06, "loss": 0.5535, "step": 2756 }, { "epoch": 2.646183389342295, "grad_norm": 0.8931387200981459, "learning_rate": 5.366491671712641e-06, "loss": 0.6235, "step": 2757 }, { "epoch": 2.647143542966875, "grad_norm": 1.0157012626889244, "learning_rate": 5.363150640723479e-06, "loss": 0.5965, "step": 2758 }, { "epoch": 2.6481036965914546, "grad_norm": 0.82474890375531, "learning_rate": 5.359809446719878e-06, "loss": 0.5716, "step": 2759 }, { "epoch": 2.6490638502160344, "grad_norm": 0.5815503629908271, "learning_rate": 5.356468091201661e-06, "loss": 0.4926, "step": 2760 }, { "epoch": 2.6500240038406147, "grad_norm": 0.9494571797179733, "learning_rate": 5.353126575668728e-06, "loss": 0.5487, "step": 2761 }, { "epoch": 2.6509841574651944, "grad_norm": 1.1379191314601567, "learning_rate": 5.3497849016210515e-06, "loss": 0.6881, "step": 2762 }, { "epoch": 2.6519443110897742, "grad_norm": 0.9197773667943027, "learning_rate": 5.3464430705586715e-06, "loss": 0.6353, "step": 2763 }, { "epoch": 2.6529044647143545, "grad_norm": 1.43321308880503, "learning_rate": 5.3431010839817e-06, "loss": 0.6188, "step": 2764 }, { "epoch": 2.6538646183389343, "grad_norm": 0.9963285813780347, "learning_rate": 5.339758943390319e-06, "loss": 0.6602, "step": 2765 }, { "epoch": 2.654824771963514, "grad_norm": 1.3299018683041408, "learning_rate": 5.3364166502847796e-06, "loss": 0.6611, "step": 2766 }, { "epoch": 2.655784925588094, "grad_norm": 0.9430970922736763, "learning_rate": 5.3330742061654014e-06, "loss": 0.6448, "step": 2767 }, { "epoch": 2.656745079212674, "grad_norm": 1.2429533695993225, "learning_rate": 5.329731612532575e-06, "loss": 0.4068, "step": 2768 }, { "epoch": 2.657705232837254, "grad_norm": 0.8857770511609304, "learning_rate": 5.32638887088675e-06, "loss": 0.6106, "step": 2769 }, { "epoch": 2.658665386461834, "grad_norm": 0.7099538524559781, "learning_rate": 5.323045982728451e-06, "loss": 0.5865, "step": 2770 }, { "epoch": 2.659625540086414, "grad_norm": 0.8685762372986285, "learning_rate": 5.319702949558263e-06, "loss": 0.5038, "step": 2771 }, { "epoch": 2.6605856937109937, "grad_norm": 1.6950934274610854, "learning_rate": 5.316359772876839e-06, "loss": 0.5865, "step": 2772 }, { "epoch": 2.6615458473355735, "grad_norm": 1.0173679926030197, "learning_rate": 5.313016454184894e-06, "loss": 0.5239, "step": 2773 }, { "epoch": 2.6625060009601538, "grad_norm": 1.1504621404448494, "learning_rate": 5.309672994983209e-06, "loss": 0.5157, "step": 2774 }, { "epoch": 2.6634661545847336, "grad_norm": 0.8027821787544791, "learning_rate": 5.306329396772627e-06, "loss": 0.5587, "step": 2775 }, { "epoch": 2.6644263082093134, "grad_norm": 1.117046500155871, "learning_rate": 5.302985661054053e-06, "loss": 0.5967, "step": 2776 }, { "epoch": 2.6653864618338936, "grad_norm": 0.8043008476324895, "learning_rate": 5.2996417893284556e-06, "loss": 0.6182, "step": 2777 }, { "epoch": 2.6663466154584734, "grad_norm": 1.1953520981292436, "learning_rate": 5.296297783096861e-06, "loss": 0.537, "step": 2778 }, { "epoch": 2.667306769083053, "grad_norm": 1.0510202370973325, "learning_rate": 5.292953643860358e-06, "loss": 0.5821, "step": 2779 }, { "epoch": 2.668266922707633, "grad_norm": 0.9596696605662415, "learning_rate": 5.289609373120096e-06, "loss": 0.4941, "step": 2780 }, { "epoch": 2.6692270763322132, "grad_norm": 1.01909531322049, "learning_rate": 5.286264972377283e-06, "loss": 0.4692, "step": 2781 }, { "epoch": 2.670187229956793, "grad_norm": 1.3035075191164636, "learning_rate": 5.282920443133184e-06, "loss": 0.6848, "step": 2782 }, { "epoch": 2.6711473835813733, "grad_norm": 0.8407625865710462, "learning_rate": 5.279575786889121e-06, "loss": 0.5761, "step": 2783 }, { "epoch": 2.672107537205953, "grad_norm": 0.8099125751828627, "learning_rate": 5.2762310051464775e-06, "loss": 0.5893, "step": 2784 }, { "epoch": 2.673067690830533, "grad_norm": 0.9217477676196206, "learning_rate": 5.272886099406688e-06, "loss": 0.5236, "step": 2785 }, { "epoch": 2.6740278444551127, "grad_norm": 0.7107057784577531, "learning_rate": 5.269541071171246e-06, "loss": 0.4957, "step": 2786 }, { "epoch": 2.674987998079693, "grad_norm": 0.7750401107871964, "learning_rate": 5.266195921941697e-06, "loss": 0.5315, "step": 2787 }, { "epoch": 2.6759481517042727, "grad_norm": 0.81974390476827, "learning_rate": 5.262850653219643e-06, "loss": 0.5554, "step": 2788 }, { "epoch": 2.6769083053288525, "grad_norm": 0.8516527447009047, "learning_rate": 5.259505266506742e-06, "loss": 0.6004, "step": 2789 }, { "epoch": 2.6778684589534327, "grad_norm": 0.8075913820359919, "learning_rate": 5.256159763304703e-06, "loss": 0.5698, "step": 2790 }, { "epoch": 2.6788286125780125, "grad_norm": 0.8300074040078267, "learning_rate": 5.252814145115281e-06, "loss": 0.5496, "step": 2791 }, { "epoch": 2.6797887662025923, "grad_norm": 0.8344602135915483, "learning_rate": 5.249468413440293e-06, "loss": 0.55, "step": 2792 }, { "epoch": 2.680748919827172, "grad_norm": 1.147481442434825, "learning_rate": 5.246122569781601e-06, "loss": 0.6003, "step": 2793 }, { "epoch": 2.6817090734517524, "grad_norm": 0.8018301687449206, "learning_rate": 5.242776615641119e-06, "loss": 0.4884, "step": 2794 }, { "epoch": 2.682669227076332, "grad_norm": 1.1528599334793839, "learning_rate": 5.2394305525208076e-06, "loss": 0.5477, "step": 2795 }, { "epoch": 2.6836293807009124, "grad_norm": 1.424087221751726, "learning_rate": 5.236084381922681e-06, "loss": 0.5055, "step": 2796 }, { "epoch": 2.684589534325492, "grad_norm": 1.6433524199813832, "learning_rate": 5.232738105348801e-06, "loss": 0.581, "step": 2797 }, { "epoch": 2.685549687950072, "grad_norm": 1.0480150981929754, "learning_rate": 5.229391724301271e-06, "loss": 0.5789, "step": 2798 }, { "epoch": 2.686509841574652, "grad_norm": 0.8417437635684826, "learning_rate": 5.22604524028225e-06, "loss": 0.5363, "step": 2799 }, { "epoch": 2.687469995199232, "grad_norm": 1.227301419010782, "learning_rate": 5.222698654793936e-06, "loss": 0.5724, "step": 2800 }, { "epoch": 2.688430148823812, "grad_norm": 0.969113348688189, "learning_rate": 5.219351969338576e-06, "loss": 0.5983, "step": 2801 }, { "epoch": 2.6893903024483916, "grad_norm": 0.7338356808346161, "learning_rate": 5.2160051854184616e-06, "loss": 0.5973, "step": 2802 }, { "epoch": 2.690350456072972, "grad_norm": 1.1642056213545264, "learning_rate": 5.212658304535929e-06, "loss": 0.5578, "step": 2803 }, { "epoch": 2.6913106096975516, "grad_norm": 0.7456935688386179, "learning_rate": 5.209311328193356e-06, "loss": 0.4917, "step": 2804 }, { "epoch": 2.6922707633221314, "grad_norm": 0.9260760596297731, "learning_rate": 5.205964257893166e-06, "loss": 0.5908, "step": 2805 }, { "epoch": 2.6932309169467112, "grad_norm": 1.05446148075854, "learning_rate": 5.202617095137823e-06, "loss": 0.4417, "step": 2806 }, { "epoch": 2.6941910705712915, "grad_norm": 1.1739664940040064, "learning_rate": 5.199269841429831e-06, "loss": 0.6633, "step": 2807 }, { "epoch": 2.6951512241958713, "grad_norm": 1.013728085487111, "learning_rate": 5.195922498271737e-06, "loss": 0.664, "step": 2808 }, { "epoch": 2.6961113778204515, "grad_norm": 1.3244950161921376, "learning_rate": 5.192575067166126e-06, "loss": 0.5272, "step": 2809 }, { "epoch": 2.6970715314450313, "grad_norm": 0.7828956398344186, "learning_rate": 5.189227549615629e-06, "loss": 0.5969, "step": 2810 }, { "epoch": 2.698031685069611, "grad_norm": 0.9929250327626931, "learning_rate": 5.185879947122906e-06, "loss": 0.5879, "step": 2811 }, { "epoch": 2.698991838694191, "grad_norm": 0.7285106200465528, "learning_rate": 5.182532261190663e-06, "loss": 0.4833, "step": 2812 }, { "epoch": 2.699951992318771, "grad_norm": 0.8879568555582249, "learning_rate": 5.179184493321637e-06, "loss": 0.6064, "step": 2813 }, { "epoch": 2.700912145943351, "grad_norm": 1.0254434755180422, "learning_rate": 5.175836645018608e-06, "loss": 0.5584, "step": 2814 }, { "epoch": 2.7018722995679307, "grad_norm": 0.8678759094680042, "learning_rate": 5.1724887177843894e-06, "loss": 0.5772, "step": 2815 }, { "epoch": 2.702832453192511, "grad_norm": 0.7569497101180229, "learning_rate": 5.1691407131218296e-06, "loss": 0.5231, "step": 2816 }, { "epoch": 2.7037926068170908, "grad_norm": 0.7372472593724549, "learning_rate": 5.165792632533811e-06, "loss": 0.5813, "step": 2817 }, { "epoch": 2.7047527604416706, "grad_norm": 0.8291085064256809, "learning_rate": 5.162444477523251e-06, "loss": 0.6509, "step": 2818 }, { "epoch": 2.7057129140662504, "grad_norm": 1.0543670937244367, "learning_rate": 5.159096249593104e-06, "loss": 0.693, "step": 2819 }, { "epoch": 2.7066730676908306, "grad_norm": 0.814380262473815, "learning_rate": 5.1557479502463495e-06, "loss": 0.5024, "step": 2820 }, { "epoch": 2.7076332213154104, "grad_norm": 1.514299892554918, "learning_rate": 5.152399580986005e-06, "loss": 0.6452, "step": 2821 }, { "epoch": 2.7085933749399906, "grad_norm": 1.233518749171904, "learning_rate": 5.149051143315118e-06, "loss": 0.5918, "step": 2822 }, { "epoch": 2.7095535285645704, "grad_norm": 0.9108121896927276, "learning_rate": 5.1457026387367655e-06, "loss": 0.6186, "step": 2823 }, { "epoch": 2.7105136821891502, "grad_norm": 0.7660323924613025, "learning_rate": 5.142354068754057e-06, "loss": 0.5257, "step": 2824 }, { "epoch": 2.71147383581373, "grad_norm": 0.8533368869237873, "learning_rate": 5.139005434870128e-06, "loss": 0.6374, "step": 2825 }, { "epoch": 2.7124339894383103, "grad_norm": 1.029186033165337, "learning_rate": 5.135656738588144e-06, "loss": 0.6697, "step": 2826 }, { "epoch": 2.71339414306289, "grad_norm": 0.8626379360577886, "learning_rate": 5.132307981411301e-06, "loss": 0.5459, "step": 2827 }, { "epoch": 2.71435429668747, "grad_norm": 0.6931533542707502, "learning_rate": 5.128959164842821e-06, "loss": 0.5161, "step": 2828 }, { "epoch": 2.71531445031205, "grad_norm": 0.6616146004167408, "learning_rate": 5.125610290385948e-06, "loss": 0.5439, "step": 2829 }, { "epoch": 2.71627460393663, "grad_norm": 0.8005261665756354, "learning_rate": 5.1222613595439605e-06, "loss": 0.5467, "step": 2830 }, { "epoch": 2.7172347575612097, "grad_norm": 0.9578996071306888, "learning_rate": 5.118912373820154e-06, "loss": 0.516, "step": 2831 }, { "epoch": 2.7181949111857895, "grad_norm": 1.2074716240415104, "learning_rate": 5.1155633347178545e-06, "loss": 0.5961, "step": 2832 }, { "epoch": 2.7191550648103697, "grad_norm": 1.212575483933078, "learning_rate": 5.1122142437404085e-06, "loss": 0.5951, "step": 2833 }, { "epoch": 2.7201152184349495, "grad_norm": 1.0413859689164209, "learning_rate": 5.1088651023911905e-06, "loss": 0.6224, "step": 2834 }, { "epoch": 2.7210753720595298, "grad_norm": 0.9345217099426696, "learning_rate": 5.10551591217359e-06, "loss": 0.5343, "step": 2835 }, { "epoch": 2.7220355256841096, "grad_norm": 0.7110280059583833, "learning_rate": 5.102166674591024e-06, "loss": 0.5106, "step": 2836 }, { "epoch": 2.7229956793086894, "grad_norm": 0.8286452693054055, "learning_rate": 5.09881739114693e-06, "loss": 0.6068, "step": 2837 }, { "epoch": 2.723955832933269, "grad_norm": 0.8487877523784975, "learning_rate": 5.095468063344767e-06, "loss": 0.5825, "step": 2838 }, { "epoch": 2.7249159865578494, "grad_norm": 1.0333355448747579, "learning_rate": 5.09211869268801e-06, "loss": 0.5331, "step": 2839 }, { "epoch": 2.725876140182429, "grad_norm": 0.9580243719565993, "learning_rate": 5.088769280680156e-06, "loss": 0.546, "step": 2840 }, { "epoch": 2.726836293807009, "grad_norm": 0.8517011607392625, "learning_rate": 5.08541982882472e-06, "loss": 0.6324, "step": 2841 }, { "epoch": 2.727796447431589, "grad_norm": 0.7445907032374407, "learning_rate": 5.082070338625238e-06, "loss": 0.5782, "step": 2842 }, { "epoch": 2.728756601056169, "grad_norm": 0.8761243651467072, "learning_rate": 5.078720811585257e-06, "loss": 0.5693, "step": 2843 }, { "epoch": 2.729716754680749, "grad_norm": 0.755296193520602, "learning_rate": 5.0753712492083455e-06, "loss": 0.5016, "step": 2844 }, { "epoch": 2.7306769083053286, "grad_norm": 0.969866450944725, "learning_rate": 5.072021652998086e-06, "loss": 0.53, "step": 2845 }, { "epoch": 2.731637061929909, "grad_norm": 0.9475889133073618, "learning_rate": 5.068672024458076e-06, "loss": 0.5796, "step": 2846 }, { "epoch": 2.7325972155544886, "grad_norm": 1.0766754055284986, "learning_rate": 5.065322365091928e-06, "loss": 0.5744, "step": 2847 }, { "epoch": 2.733557369179069, "grad_norm": 1.0695727351910735, "learning_rate": 5.061972676403269e-06, "loss": 0.5769, "step": 2848 }, { "epoch": 2.7345175228036487, "grad_norm": 0.9550455495161231, "learning_rate": 5.058622959895738e-06, "loss": 0.5826, "step": 2849 }, { "epoch": 2.7354776764282285, "grad_norm": 1.0576624347446582, "learning_rate": 5.055273217072988e-06, "loss": 0.507, "step": 2850 }, { "epoch": 2.7364378300528083, "grad_norm": 0.9831722170722194, "learning_rate": 5.051923449438679e-06, "loss": 0.5763, "step": 2851 }, { "epoch": 2.7373979836773885, "grad_norm": 0.9222923967191309, "learning_rate": 5.048573658496492e-06, "loss": 0.65, "step": 2852 }, { "epoch": 2.7383581373019683, "grad_norm": 1.4532580101608645, "learning_rate": 5.0452238457501065e-06, "loss": 0.6787, "step": 2853 }, { "epoch": 2.739318290926548, "grad_norm": 0.8984842010871488, "learning_rate": 5.0418740127032204e-06, "loss": 0.6217, "step": 2854 }, { "epoch": 2.7402784445511283, "grad_norm": 0.7907235629609742, "learning_rate": 5.038524160859538e-06, "loss": 0.6774, "step": 2855 }, { "epoch": 2.741238598175708, "grad_norm": 0.8375298609341927, "learning_rate": 5.035174291722772e-06, "loss": 0.4945, "step": 2856 }, { "epoch": 2.742198751800288, "grad_norm": 0.9693154996390144, "learning_rate": 5.031824406796642e-06, "loss": 0.6046, "step": 2857 }, { "epoch": 2.7431589054248677, "grad_norm": 0.7577797917617582, "learning_rate": 5.028474507584875e-06, "loss": 0.5425, "step": 2858 }, { "epoch": 2.744119059049448, "grad_norm": 0.8748809701058167, "learning_rate": 5.025124595591206e-06, "loss": 0.5611, "step": 2859 }, { "epoch": 2.7450792126740278, "grad_norm": 0.8946125739314568, "learning_rate": 5.0217746723193775e-06, "loss": 0.5327, "step": 2860 }, { "epoch": 2.746039366298608, "grad_norm": 0.7559350437862631, "learning_rate": 5.018424739273129e-06, "loss": 0.5992, "step": 2861 }, { "epoch": 2.746999519923188, "grad_norm": 0.8655293639789648, "learning_rate": 5.015074797956212e-06, "loss": 0.4905, "step": 2862 }, { "epoch": 2.7479596735477676, "grad_norm": 0.7143900022469163, "learning_rate": 5.01172484987238e-06, "loss": 0.5334, "step": 2863 }, { "epoch": 2.7489198271723474, "grad_norm": 0.6692397530337836, "learning_rate": 5.008374896525387e-06, "loss": 0.5599, "step": 2864 }, { "epoch": 2.7498799807969276, "grad_norm": 1.1242431666956936, "learning_rate": 5.0050249394189945e-06, "loss": 0.6545, "step": 2865 }, { "epoch": 2.7508401344215074, "grad_norm": 0.8436335676052561, "learning_rate": 5.001674980056959e-06, "loss": 0.6221, "step": 2866 }, { "epoch": 2.7518002880460872, "grad_norm": 1.2594540584258287, "learning_rate": 4.998325019943044e-06, "loss": 0.5195, "step": 2867 }, { "epoch": 2.7527604416706675, "grad_norm": 1.3176525355612132, "learning_rate": 4.994975060581007e-06, "loss": 0.6188, "step": 2868 }, { "epoch": 2.7537205952952473, "grad_norm": 1.0517291363862549, "learning_rate": 4.991625103474614e-06, "loss": 0.6679, "step": 2869 }, { "epoch": 2.754680748919827, "grad_norm": 0.9483793969481509, "learning_rate": 4.988275150127622e-06, "loss": 0.5689, "step": 2870 }, { "epoch": 2.755640902544407, "grad_norm": 0.7488083620511894, "learning_rate": 4.984925202043789e-06, "loss": 0.5896, "step": 2871 }, { "epoch": 2.756601056168987, "grad_norm": 0.7822241122699791, "learning_rate": 4.981575260726873e-06, "loss": 0.5796, "step": 2872 }, { "epoch": 2.757561209793567, "grad_norm": 0.852105215582262, "learning_rate": 4.978225327680625e-06, "loss": 0.4771, "step": 2873 }, { "epoch": 2.758521363418147, "grad_norm": 0.7622547716477299, "learning_rate": 4.974875404408794e-06, "loss": 0.5298, "step": 2874 }, { "epoch": 2.759481517042727, "grad_norm": 1.1797915807925194, "learning_rate": 4.971525492415126e-06, "loss": 0.6415, "step": 2875 }, { "epoch": 2.7604416706673067, "grad_norm": 0.9908649378870896, "learning_rate": 4.96817559320336e-06, "loss": 0.5989, "step": 2876 }, { "epoch": 2.7614018242918865, "grad_norm": 0.727773971441581, "learning_rate": 4.9648257082772305e-06, "loss": 0.462, "step": 2877 }, { "epoch": 2.7623619779164668, "grad_norm": 0.7665813611994887, "learning_rate": 4.9614758391404634e-06, "loss": 0.4936, "step": 2878 }, { "epoch": 2.7633221315410466, "grad_norm": 2.0313565893963075, "learning_rate": 4.9581259872967795e-06, "loss": 0.5796, "step": 2879 }, { "epoch": 2.7642822851656264, "grad_norm": 1.6361585628980468, "learning_rate": 4.954776154249896e-06, "loss": 0.6739, "step": 2880 }, { "epoch": 2.7652424387902066, "grad_norm": 0.8737506716117023, "learning_rate": 4.95142634150351e-06, "loss": 0.5366, "step": 2881 }, { "epoch": 2.7662025924147864, "grad_norm": 0.8619233029491494, "learning_rate": 4.94807655056132e-06, "loss": 0.5376, "step": 2882 }, { "epoch": 2.767162746039366, "grad_norm": 1.222420071314913, "learning_rate": 4.944726782927014e-06, "loss": 0.6101, "step": 2883 }, { "epoch": 2.768122899663946, "grad_norm": 0.8119123586281217, "learning_rate": 4.9413770401042625e-06, "loss": 0.5231, "step": 2884 }, { "epoch": 2.769083053288526, "grad_norm": 0.6751884734010303, "learning_rate": 4.938027323596732e-06, "loss": 0.5415, "step": 2885 }, { "epoch": 2.770043206913106, "grad_norm": 0.9800559826668412, "learning_rate": 4.934677634908073e-06, "loss": 0.529, "step": 2886 }, { "epoch": 2.7710033605376863, "grad_norm": 1.1304238990482627, "learning_rate": 4.931327975541925e-06, "loss": 0.4976, "step": 2887 }, { "epoch": 2.771963514162266, "grad_norm": 0.8582155049203667, "learning_rate": 4.927978347001916e-06, "loss": 0.5696, "step": 2888 }, { "epoch": 2.772923667786846, "grad_norm": 0.8266371567869433, "learning_rate": 4.924628750791656e-06, "loss": 0.513, "step": 2889 }, { "epoch": 2.7738838214114256, "grad_norm": 1.0052094998077734, "learning_rate": 4.9212791884147456e-06, "loss": 0.6075, "step": 2890 }, { "epoch": 2.774843975036006, "grad_norm": 0.9114829201450042, "learning_rate": 4.917929661374764e-06, "loss": 0.5253, "step": 2891 }, { "epoch": 2.7758041286605857, "grad_norm": 0.8720671693231004, "learning_rate": 4.9145801711752804e-06, "loss": 0.6322, "step": 2892 }, { "epoch": 2.7767642822851655, "grad_norm": 0.949355809774783, "learning_rate": 4.911230719319846e-06, "loss": 0.6174, "step": 2893 }, { "epoch": 2.7777244359097457, "grad_norm": 0.8706965509936089, "learning_rate": 4.907881307311992e-06, "loss": 0.5672, "step": 2894 }, { "epoch": 2.7786845895343255, "grad_norm": 0.7632234044082503, "learning_rate": 4.904531936655236e-06, "loss": 0.6304, "step": 2895 }, { "epoch": 2.7796447431589053, "grad_norm": 0.830042620304823, "learning_rate": 4.901182608853071e-06, "loss": 0.6767, "step": 2896 }, { "epoch": 2.780604896783485, "grad_norm": 0.6273485293685256, "learning_rate": 4.897833325408977e-06, "loss": 0.5155, "step": 2897 }, { "epoch": 2.7815650504080653, "grad_norm": 1.0782567554890286, "learning_rate": 4.894484087826413e-06, "loss": 0.619, "step": 2898 }, { "epoch": 2.782525204032645, "grad_norm": 0.6745707407350029, "learning_rate": 4.891134897608811e-06, "loss": 0.5203, "step": 2899 }, { "epoch": 2.7834853576572254, "grad_norm": 1.1052704672289617, "learning_rate": 4.887785756259591e-06, "loss": 0.6038, "step": 2900 }, { "epoch": 2.784445511281805, "grad_norm": 0.8899673399176622, "learning_rate": 4.884436665282146e-06, "loss": 0.5844, "step": 2901 }, { "epoch": 2.785405664906385, "grad_norm": 0.8514740099153787, "learning_rate": 4.881087626179847e-06, "loss": 0.5008, "step": 2902 }, { "epoch": 2.7863658185309648, "grad_norm": 0.9908697079145432, "learning_rate": 4.877738640456042e-06, "loss": 0.6125, "step": 2903 }, { "epoch": 2.787325972155545, "grad_norm": 1.1157708298124993, "learning_rate": 4.874389709614053e-06, "loss": 0.6162, "step": 2904 }, { "epoch": 2.788286125780125, "grad_norm": 0.9558929847860761, "learning_rate": 4.871040835157181e-06, "loss": 0.5967, "step": 2905 }, { "epoch": 2.7892462794047046, "grad_norm": 0.7823205240196205, "learning_rate": 4.8676920185887e-06, "loss": 0.5413, "step": 2906 }, { "epoch": 2.790206433029285, "grad_norm": 0.9112072505953971, "learning_rate": 4.8643432614118565e-06, "loss": 0.5376, "step": 2907 }, { "epoch": 2.7911665866538646, "grad_norm": 0.9221591949175044, "learning_rate": 4.860994565129875e-06, "loss": 0.5493, "step": 2908 }, { "epoch": 2.7921267402784444, "grad_norm": 0.8342178127552398, "learning_rate": 4.857645931245946e-06, "loss": 0.6038, "step": 2909 }, { "epoch": 2.7930868939030242, "grad_norm": 1.1359811826595312, "learning_rate": 4.854297361263235e-06, "loss": 0.6395, "step": 2910 }, { "epoch": 2.7940470475276045, "grad_norm": 1.0332376660946696, "learning_rate": 4.850948856684884e-06, "loss": 0.5444, "step": 2911 }, { "epoch": 2.7950072011521843, "grad_norm": 0.8705677465334742, "learning_rate": 4.8476004190139954e-06, "loss": 0.5946, "step": 2912 }, { "epoch": 2.7959673547767645, "grad_norm": 0.7783446420196491, "learning_rate": 4.844252049753651e-06, "loss": 0.4914, "step": 2913 }, { "epoch": 2.7969275084013443, "grad_norm": 0.7214909794465172, "learning_rate": 4.840903750406898e-06, "loss": 0.4987, "step": 2914 }, { "epoch": 2.797887662025924, "grad_norm": 1.0357070280870575, "learning_rate": 4.8375555224767486e-06, "loss": 0.5936, "step": 2915 }, { "epoch": 2.798847815650504, "grad_norm": 0.7979290843200889, "learning_rate": 4.83420736746619e-06, "loss": 0.4591, "step": 2916 }, { "epoch": 2.799807969275084, "grad_norm": 0.8958607464490439, "learning_rate": 4.830859286878172e-06, "loss": 0.5705, "step": 2917 }, { "epoch": 2.800768122899664, "grad_norm": 0.8029765161456159, "learning_rate": 4.8275112822156105e-06, "loss": 0.5934, "step": 2918 }, { "epoch": 2.8017282765242437, "grad_norm": 1.01466541389195, "learning_rate": 4.824163354981393e-06, "loss": 0.5485, "step": 2919 }, { "epoch": 2.802688430148824, "grad_norm": 0.9346347304097536, "learning_rate": 4.820815506678364e-06, "loss": 0.5042, "step": 2920 }, { "epoch": 2.8036485837734038, "grad_norm": 0.9345335439865217, "learning_rate": 4.81746773880934e-06, "loss": 0.5765, "step": 2921 }, { "epoch": 2.8046087373979836, "grad_norm": 0.8416437517951479, "learning_rate": 4.8141200528770955e-06, "loss": 0.6109, "step": 2922 }, { "epoch": 2.8055688910225633, "grad_norm": 1.0731329384073558, "learning_rate": 4.810772450384372e-06, "loss": 0.587, "step": 2923 }, { "epoch": 2.8065290446471436, "grad_norm": 0.7493800845064654, "learning_rate": 4.807424932833875e-06, "loss": 0.5163, "step": 2924 }, { "epoch": 2.8074891982717234, "grad_norm": 0.8498316715115648, "learning_rate": 4.8040775017282644e-06, "loss": 0.379, "step": 2925 }, { "epoch": 2.8084493518963036, "grad_norm": 0.8015163966695312, "learning_rate": 4.8007301585701726e-06, "loss": 0.6054, "step": 2926 }, { "epoch": 2.8094095055208834, "grad_norm": 1.221906514870264, "learning_rate": 4.7973829048621786e-06, "loss": 0.6206, "step": 2927 }, { "epoch": 2.810369659145463, "grad_norm": 1.1146262483513474, "learning_rate": 4.794035742106834e-06, "loss": 0.654, "step": 2928 }, { "epoch": 2.811329812770043, "grad_norm": 0.8760379402641922, "learning_rate": 4.7906886718066445e-06, "loss": 0.5489, "step": 2929 }, { "epoch": 2.8122899663946233, "grad_norm": 0.9144663867492914, "learning_rate": 4.7873416954640725e-06, "loss": 0.5547, "step": 2930 }, { "epoch": 2.813250120019203, "grad_norm": 1.0672192798429738, "learning_rate": 4.783994814581539e-06, "loss": 0.5596, "step": 2931 }, { "epoch": 2.814210273643783, "grad_norm": 1.0850610121084665, "learning_rate": 4.7806480306614256e-06, "loss": 0.5799, "step": 2932 }, { "epoch": 2.815170427268363, "grad_norm": 0.9486281957715099, "learning_rate": 4.7773013452060654e-06, "loss": 0.6342, "step": 2933 }, { "epoch": 2.816130580892943, "grad_norm": 0.8545571788186971, "learning_rate": 4.773954759717752e-06, "loss": 0.6248, "step": 2934 }, { "epoch": 2.8170907345175227, "grad_norm": 1.116432537168795, "learning_rate": 4.7706082756987295e-06, "loss": 0.5421, "step": 2935 }, { "epoch": 2.8180508881421025, "grad_norm": 0.9056797559562542, "learning_rate": 4.7672618946512e-06, "loss": 0.5161, "step": 2936 }, { "epoch": 2.8190110417666827, "grad_norm": 0.7639094773918543, "learning_rate": 4.76391561807732e-06, "loss": 0.5529, "step": 2937 }, { "epoch": 2.8199711953912625, "grad_norm": 0.9346833874453674, "learning_rate": 4.760569447479193e-06, "loss": 0.5205, "step": 2938 }, { "epoch": 2.8209313490158427, "grad_norm": 0.7843549075761868, "learning_rate": 4.757223384358884e-06, "loss": 0.5875, "step": 2939 }, { "epoch": 2.8218915026404225, "grad_norm": 1.260260804215348, "learning_rate": 4.753877430218401e-06, "loss": 0.6112, "step": 2940 }, { "epoch": 2.8228516562650023, "grad_norm": 0.7626271029769138, "learning_rate": 4.750531586559707e-06, "loss": 0.562, "step": 2941 }, { "epoch": 2.823811809889582, "grad_norm": 0.8820854043367795, "learning_rate": 4.7471858548847214e-06, "loss": 0.5656, "step": 2942 }, { "epoch": 2.8247719635141624, "grad_norm": 0.8351561978330843, "learning_rate": 4.743840236695299e-06, "loss": 0.5229, "step": 2943 }, { "epoch": 2.825732117138742, "grad_norm": 0.8965831828807459, "learning_rate": 4.740494733493258e-06, "loss": 0.5219, "step": 2944 }, { "epoch": 2.826692270763322, "grad_norm": 0.6940488116669402, "learning_rate": 4.737149346780358e-06, "loss": 0.5102, "step": 2945 }, { "epoch": 2.827652424387902, "grad_norm": 0.66058898816578, "learning_rate": 4.733804078058305e-06, "loss": 0.4722, "step": 2946 }, { "epoch": 2.828612578012482, "grad_norm": 0.8983980313109161, "learning_rate": 4.730458928828757e-06, "loss": 0.6249, "step": 2947 }, { "epoch": 2.829572731637062, "grad_norm": 0.9495157618258931, "learning_rate": 4.7271139005933144e-06, "loss": 0.5751, "step": 2948 }, { "epoch": 2.8305328852616416, "grad_norm": 0.7619335291084912, "learning_rate": 4.723768994853523e-06, "loss": 0.5488, "step": 2949 }, { "epoch": 2.831493038886222, "grad_norm": 0.801906740604625, "learning_rate": 4.72042421311088e-06, "loss": 0.4483, "step": 2950 }, { "epoch": 2.8324531925108016, "grad_norm": 1.0137318244819593, "learning_rate": 4.717079556866818e-06, "loss": 0.6824, "step": 2951 }, { "epoch": 2.833413346135382, "grad_norm": 1.1079261470228001, "learning_rate": 4.713735027622719e-06, "loss": 0.5802, "step": 2952 }, { "epoch": 2.8343734997599617, "grad_norm": 1.017037603900067, "learning_rate": 4.710390626879906e-06, "loss": 0.7072, "step": 2953 }, { "epoch": 2.8353336533845415, "grad_norm": 1.0667662238792703, "learning_rate": 4.707046356139643e-06, "loss": 0.5673, "step": 2954 }, { "epoch": 2.8362938070091213, "grad_norm": 0.7291841531735604, "learning_rate": 4.703702216903142e-06, "loss": 0.542, "step": 2955 }, { "epoch": 2.8372539606337015, "grad_norm": 1.1436000457113433, "learning_rate": 4.700358210671545e-06, "loss": 0.5482, "step": 2956 }, { "epoch": 2.8382141142582813, "grad_norm": 0.8658554614081735, "learning_rate": 4.697014338945947e-06, "loss": 0.6139, "step": 2957 }, { "epoch": 2.839174267882861, "grad_norm": 0.6637046552641884, "learning_rate": 4.693670603227374e-06, "loss": 0.5361, "step": 2958 }, { "epoch": 2.8401344215074413, "grad_norm": 1.0021447139900141, "learning_rate": 4.690327005016792e-06, "loss": 0.6001, "step": 2959 }, { "epoch": 2.841094575132021, "grad_norm": 0.8463195364449034, "learning_rate": 4.686983545815107e-06, "loss": 0.618, "step": 2960 }, { "epoch": 2.842054728756601, "grad_norm": 0.9846420589399077, "learning_rate": 4.683640227123162e-06, "loss": 0.5783, "step": 2961 }, { "epoch": 2.8430148823811807, "grad_norm": 0.8114746444319065, "learning_rate": 4.680297050441737e-06, "loss": 0.4672, "step": 2962 }, { "epoch": 2.843975036005761, "grad_norm": 0.8248614463268978, "learning_rate": 4.67695401727155e-06, "loss": 0.5374, "step": 2963 }, { "epoch": 2.8449351896303408, "grad_norm": 0.7275507255666933, "learning_rate": 4.6736111291132505e-06, "loss": 0.555, "step": 2964 }, { "epoch": 2.845895343254921, "grad_norm": 0.8507217475573444, "learning_rate": 4.670268387467427e-06, "loss": 0.6096, "step": 2965 }, { "epoch": 2.846855496879501, "grad_norm": 1.416756634605639, "learning_rate": 4.666925793834599e-06, "loss": 0.5841, "step": 2966 }, { "epoch": 2.8478156505040806, "grad_norm": 1.4981370779708199, "learning_rate": 4.663583349715222e-06, "loss": 0.6439, "step": 2967 }, { "epoch": 2.8487758041286604, "grad_norm": 0.9390667547922735, "learning_rate": 4.660241056609685e-06, "loss": 0.6904, "step": 2968 }, { "epoch": 2.8497359577532406, "grad_norm": 0.7379472903163558, "learning_rate": 4.656898916018302e-06, "loss": 0.5679, "step": 2969 }, { "epoch": 2.8506961113778204, "grad_norm": 0.8786255911759271, "learning_rate": 4.653556929441332e-06, "loss": 0.4972, "step": 2970 }, { "epoch": 2.8516562650024007, "grad_norm": 1.0498771434997423, "learning_rate": 4.650215098378949e-06, "loss": 0.5606, "step": 2971 }, { "epoch": 2.8526164186269805, "grad_norm": 0.7615441712466747, "learning_rate": 4.646873424331272e-06, "loss": 0.608, "step": 2972 }, { "epoch": 2.8535765722515603, "grad_norm": 0.8376127664863438, "learning_rate": 4.64353190879834e-06, "loss": 0.5594, "step": 2973 }, { "epoch": 2.85453672587614, "grad_norm": 0.8482226025157907, "learning_rate": 4.640190553280123e-06, "loss": 0.5198, "step": 2974 }, { "epoch": 2.85549687950072, "grad_norm": 0.9236695300453229, "learning_rate": 4.636849359276521e-06, "loss": 0.5556, "step": 2975 }, { "epoch": 2.8564570331253, "grad_norm": 0.6691807258670308, "learning_rate": 4.6335083282873605e-06, "loss": 0.5597, "step": 2976 }, { "epoch": 2.85741718674988, "grad_norm": 0.8570282894336693, "learning_rate": 4.6301674618123924e-06, "loss": 0.564, "step": 2977 }, { "epoch": 2.85837734037446, "grad_norm": 0.8837752882051586, "learning_rate": 4.626826761351299e-06, "loss": 0.4339, "step": 2978 }, { "epoch": 2.85933749399904, "grad_norm": 1.0641035459648969, "learning_rate": 4.623486228403682e-06, "loss": 0.4574, "step": 2979 }, { "epoch": 2.8602976476236197, "grad_norm": 1.4263418496878348, "learning_rate": 4.620145864469073e-06, "loss": 0.5812, "step": 2980 }, { "epoch": 2.8612578012481995, "grad_norm": 0.7903838666253568, "learning_rate": 4.616805671046925e-06, "loss": 0.5498, "step": 2981 }, { "epoch": 2.8622179548727797, "grad_norm": 0.7891249239627228, "learning_rate": 4.613465649636615e-06, "loss": 0.5966, "step": 2982 }, { "epoch": 2.8631781084973595, "grad_norm": 0.9306190716415175, "learning_rate": 4.610125801737443e-06, "loss": 0.6153, "step": 2983 }, { "epoch": 2.86413826212194, "grad_norm": 1.1272173580140326, "learning_rate": 4.6067861288486315e-06, "loss": 0.5308, "step": 2984 }, { "epoch": 2.8650984157465196, "grad_norm": 0.7593395168285015, "learning_rate": 4.603446632469322e-06, "loss": 0.584, "step": 2985 }, { "epoch": 2.8660585693710994, "grad_norm": 0.9344488962104383, "learning_rate": 4.600107314098583e-06, "loss": 0.4931, "step": 2986 }, { "epoch": 2.867018722995679, "grad_norm": 0.9654179658799489, "learning_rate": 4.596768175235393e-06, "loss": 0.493, "step": 2987 }, { "epoch": 2.867978876620259, "grad_norm": 1.080083239715295, "learning_rate": 4.59342921737866e-06, "loss": 0.5757, "step": 2988 }, { "epoch": 2.868939030244839, "grad_norm": 0.8016007211738998, "learning_rate": 4.590090442027206e-06, "loss": 0.5323, "step": 2989 }, { "epoch": 2.869899183869419, "grad_norm": 1.223707336686027, "learning_rate": 4.5867518506797705e-06, "loss": 0.6246, "step": 2990 }, { "epoch": 2.8708593374939992, "grad_norm": 0.749727025370526, "learning_rate": 4.583413444835013e-06, "loss": 0.6186, "step": 2991 }, { "epoch": 2.871819491118579, "grad_norm": 1.019412638566963, "learning_rate": 4.580075225991505e-06, "loss": 0.5343, "step": 2992 }, { "epoch": 2.872779644743159, "grad_norm": 0.6940835270233356, "learning_rate": 4.576737195647741e-06, "loss": 0.5555, "step": 2993 }, { "epoch": 2.8737397983677386, "grad_norm": 0.8045501203739621, "learning_rate": 4.573399355302124e-06, "loss": 0.579, "step": 2994 }, { "epoch": 2.874699951992319, "grad_norm": 0.8621343607978998, "learning_rate": 4.5700617064529774e-06, "loss": 0.6132, "step": 2995 }, { "epoch": 2.8756601056168987, "grad_norm": 0.9533361582500471, "learning_rate": 4.566724250598535e-06, "loss": 0.6758, "step": 2996 }, { "epoch": 2.876620259241479, "grad_norm": 1.1658352964821912, "learning_rate": 4.563386989236944e-06, "loss": 0.5405, "step": 2997 }, { "epoch": 2.8775804128660587, "grad_norm": 0.828908846167536, "learning_rate": 4.560049923866266e-06, "loss": 0.5566, "step": 2998 }, { "epoch": 2.8785405664906385, "grad_norm": 0.8084447589047143, "learning_rate": 4.556713055984477e-06, "loss": 0.4637, "step": 2999 }, { "epoch": 2.8795007201152183, "grad_norm": 0.8910489317318511, "learning_rate": 4.5533763870894555e-06, "loss": 0.6165, "step": 3000 }, { "epoch": 2.8804608737397985, "grad_norm": 0.9434466273564177, "learning_rate": 4.5500399186790034e-06, "loss": 0.5269, "step": 3001 }, { "epoch": 2.8814210273643783, "grad_norm": 0.8028002569770732, "learning_rate": 4.5467036522508185e-06, "loss": 0.5543, "step": 3002 }, { "epoch": 2.882381180988958, "grad_norm": 0.8005250981208207, "learning_rate": 4.54336758930252e-06, "loss": 0.5397, "step": 3003 }, { "epoch": 2.8833413346135384, "grad_norm": 1.272023856138233, "learning_rate": 4.54003173133163e-06, "loss": 0.538, "step": 3004 }, { "epoch": 2.884301488238118, "grad_norm": 0.7254242413564801, "learning_rate": 4.536696079835578e-06, "loss": 0.4655, "step": 3005 }, { "epoch": 2.885261641862698, "grad_norm": 1.2108522639793606, "learning_rate": 4.5333606363117035e-06, "loss": 0.6221, "step": 3006 }, { "epoch": 2.8862217954872778, "grad_norm": 1.1362947746190455, "learning_rate": 4.530025402257252e-06, "loss": 0.5818, "step": 3007 }, { "epoch": 2.887181949111858, "grad_norm": 0.99563399248279, "learning_rate": 4.526690379169374e-06, "loss": 0.6211, "step": 3008 }, { "epoch": 2.888142102736438, "grad_norm": 0.8992984225801237, "learning_rate": 4.5233555685451245e-06, "loss": 0.5328, "step": 3009 }, { "epoch": 2.889102256361018, "grad_norm": 0.6913514117762256, "learning_rate": 4.5200209718814656e-06, "loss": 0.5148, "step": 3010 }, { "epoch": 2.890062409985598, "grad_norm": 0.9611048932583985, "learning_rate": 4.516686590675261e-06, "loss": 0.5696, "step": 3011 }, { "epoch": 2.8910225636101776, "grad_norm": 0.9362407333199204, "learning_rate": 4.513352426423281e-06, "loss": 0.617, "step": 3012 }, { "epoch": 2.8919827172347574, "grad_norm": 0.9375406026194598, "learning_rate": 4.510018480622193e-06, "loss": 0.5913, "step": 3013 }, { "epoch": 2.8929428708593377, "grad_norm": 1.0381307978267604, "learning_rate": 4.506684754768573e-06, "loss": 0.5709, "step": 3014 }, { "epoch": 2.8939030244839175, "grad_norm": 0.84634737157008, "learning_rate": 4.503351250358893e-06, "loss": 0.5387, "step": 3015 }, { "epoch": 2.8948631781084972, "grad_norm": 0.7535779680324264, "learning_rate": 4.500017968889525e-06, "loss": 0.5497, "step": 3016 }, { "epoch": 2.8958233317330775, "grad_norm": 0.8262716126603851, "learning_rate": 4.4966849118567515e-06, "loss": 0.6121, "step": 3017 }, { "epoch": 2.8967834853576573, "grad_norm": 0.7003478699659915, "learning_rate": 4.4933520807567365e-06, "loss": 0.4279, "step": 3018 }, { "epoch": 2.897743638982237, "grad_norm": 0.7395431171730129, "learning_rate": 4.490019477085558e-06, "loss": 0.5443, "step": 3019 }, { "epoch": 2.898703792606817, "grad_norm": 1.0640971019218037, "learning_rate": 4.4866871023391865e-06, "loss": 0.4435, "step": 3020 }, { "epoch": 2.899663946231397, "grad_norm": 0.9605318622007952, "learning_rate": 4.483354958013485e-06, "loss": 0.6334, "step": 3021 }, { "epoch": 2.900624099855977, "grad_norm": 1.1163375869393177, "learning_rate": 4.480023045604223e-06, "loss": 0.6922, "step": 3022 }, { "epoch": 2.901584253480557, "grad_norm": 0.7092268139460385, "learning_rate": 4.476691366607056e-06, "loss": 0.4951, "step": 3023 }, { "epoch": 2.902544407105137, "grad_norm": 0.8715740350263974, "learning_rate": 4.473359922517541e-06, "loss": 0.6357, "step": 3024 }, { "epoch": 2.9035045607297167, "grad_norm": 0.7591438992910845, "learning_rate": 4.4700287148311275e-06, "loss": 0.53, "step": 3025 }, { "epoch": 2.9044647143542965, "grad_norm": 0.7656664611919346, "learning_rate": 4.466697745043158e-06, "loss": 0.5012, "step": 3026 }, { "epoch": 2.905424867978877, "grad_norm": 0.9002248848065717, "learning_rate": 4.463367014648872e-06, "loss": 0.512, "step": 3027 }, { "epoch": 2.9063850216034566, "grad_norm": 0.9116037063478836, "learning_rate": 4.460036525143395e-06, "loss": 0.5439, "step": 3028 }, { "epoch": 2.9073451752280364, "grad_norm": 0.7690930419058877, "learning_rate": 4.4567062780217486e-06, "loss": 0.6383, "step": 3029 }, { "epoch": 2.9083053288526166, "grad_norm": 0.8171136653005105, "learning_rate": 4.45337627477885e-06, "loss": 0.6746, "step": 3030 }, { "epoch": 2.9092654824771964, "grad_norm": 0.8664449445931007, "learning_rate": 4.450046516909496e-06, "loss": 0.5976, "step": 3031 }, { "epoch": 2.910225636101776, "grad_norm": 0.9081380659936392, "learning_rate": 4.446717005908382e-06, "loss": 0.6025, "step": 3032 }, { "epoch": 2.911185789726356, "grad_norm": 0.9237010049216585, "learning_rate": 4.4433877432700915e-06, "loss": 0.5099, "step": 3033 }, { "epoch": 2.9121459433509362, "grad_norm": 0.8566965829999111, "learning_rate": 4.440058730489092e-06, "loss": 0.6571, "step": 3034 }, { "epoch": 2.913106096975516, "grad_norm": 0.7974482458546323, "learning_rate": 4.436729969059745e-06, "loss": 0.5362, "step": 3035 }, { "epoch": 2.9140662506000963, "grad_norm": 1.446983730577126, "learning_rate": 4.433401460476291e-06, "loss": 0.7234, "step": 3036 }, { "epoch": 2.915026404224676, "grad_norm": 1.0673869751636853, "learning_rate": 4.430073206232865e-06, "loss": 0.6945, "step": 3037 }, { "epoch": 2.915986557849256, "grad_norm": 0.8844428989396558, "learning_rate": 4.426745207823486e-06, "loss": 0.5738, "step": 3038 }, { "epoch": 2.9169467114738357, "grad_norm": 0.9824548902743531, "learning_rate": 4.4234174667420546e-06, "loss": 0.6186, "step": 3039 }, { "epoch": 2.917906865098416, "grad_norm": 1.1404685812283493, "learning_rate": 4.42008998448236e-06, "loss": 0.5392, "step": 3040 }, { "epoch": 2.9188670187229957, "grad_norm": 0.8913417675910726, "learning_rate": 4.416762762538072e-06, "loss": 0.4348, "step": 3041 }, { "epoch": 2.9198271723475755, "grad_norm": 0.9812949602993988, "learning_rate": 4.413435802402744e-06, "loss": 0.4612, "step": 3042 }, { "epoch": 2.9207873259721557, "grad_norm": 1.1005912068567407, "learning_rate": 4.410109105569817e-06, "loss": 0.5834, "step": 3043 }, { "epoch": 2.9217474795967355, "grad_norm": 1.4909411444617133, "learning_rate": 4.406782673532604e-06, "loss": 0.5782, "step": 3044 }, { "epoch": 2.9227076332213153, "grad_norm": 0.9076699921927123, "learning_rate": 4.403456507784311e-06, "loss": 0.5743, "step": 3045 }, { "epoch": 2.923667786845895, "grad_norm": 0.9024694028854532, "learning_rate": 4.400130609818012e-06, "loss": 0.5559, "step": 3046 }, { "epoch": 2.9246279404704754, "grad_norm": 0.9120894787811443, "learning_rate": 4.3968049811266705e-06, "loss": 0.5781, "step": 3047 }, { "epoch": 2.925588094095055, "grad_norm": 0.8940010843505528, "learning_rate": 4.393479623203126e-06, "loss": 0.6149, "step": 3048 }, { "epoch": 2.9265482477196354, "grad_norm": 0.8790160387568612, "learning_rate": 4.390154537540095e-06, "loss": 0.4824, "step": 3049 }, { "epoch": 2.927508401344215, "grad_norm": 0.9775137326131264, "learning_rate": 4.386829725630173e-06, "loss": 0.5969, "step": 3050 }, { "epoch": 2.928468554968795, "grad_norm": 0.7856500964514143, "learning_rate": 4.383505188965835e-06, "loss": 0.5618, "step": 3051 }, { "epoch": 2.929428708593375, "grad_norm": 0.9894150821102421, "learning_rate": 4.380180929039424e-06, "loss": 0.663, "step": 3052 }, { "epoch": 2.930388862217955, "grad_norm": 0.7876120545966883, "learning_rate": 4.376856947343171e-06, "loss": 0.5294, "step": 3053 }, { "epoch": 2.931349015842535, "grad_norm": 0.8737122623793914, "learning_rate": 4.3735332453691725e-06, "loss": 0.5053, "step": 3054 }, { "epoch": 2.9323091694671146, "grad_norm": 0.7851955871039885, "learning_rate": 4.370209824609403e-06, "loss": 0.5026, "step": 3055 }, { "epoch": 2.933269323091695, "grad_norm": 0.837045853274039, "learning_rate": 4.366886686555713e-06, "loss": 0.5417, "step": 3056 }, { "epoch": 2.9342294767162747, "grad_norm": 1.0912637846718156, "learning_rate": 4.36356383269982e-06, "loss": 0.5179, "step": 3057 }, { "epoch": 2.9351896303408544, "grad_norm": 0.812281367372158, "learning_rate": 4.360241264533321e-06, "loss": 0.6258, "step": 3058 }, { "epoch": 2.9361497839654342, "grad_norm": 0.955694653154261, "learning_rate": 4.356918983547678e-06, "loss": 0.5727, "step": 3059 }, { "epoch": 2.9371099375900145, "grad_norm": 1.068064198708222, "learning_rate": 4.353596991234228e-06, "loss": 0.5111, "step": 3060 }, { "epoch": 2.9380700912145943, "grad_norm": 1.0820001766772187, "learning_rate": 4.3502752890841825e-06, "loss": 0.4688, "step": 3061 }, { "epoch": 2.9390302448391745, "grad_norm": 0.8983143669354525, "learning_rate": 4.3469538785886115e-06, "loss": 0.5905, "step": 3062 }, { "epoch": 2.9399903984637543, "grad_norm": 0.7696148647546619, "learning_rate": 4.343632761238465e-06, "loss": 0.5309, "step": 3063 }, { "epoch": 2.940950552088334, "grad_norm": 1.2176905977534755, "learning_rate": 4.340311938524556e-06, "loss": 0.5687, "step": 3064 }, { "epoch": 2.941910705712914, "grad_norm": 0.936368456327625, "learning_rate": 4.336991411937566e-06, "loss": 0.606, "step": 3065 }, { "epoch": 2.942870859337494, "grad_norm": 0.7562231172530031, "learning_rate": 4.333671182968045e-06, "loss": 0.5745, "step": 3066 }, { "epoch": 2.943831012962074, "grad_norm": 1.1057145572044464, "learning_rate": 4.330351253106405e-06, "loss": 0.6315, "step": 3067 }, { "epoch": 2.9447911665866537, "grad_norm": 0.9347728427361188, "learning_rate": 4.327031623842931e-06, "loss": 0.5625, "step": 3068 }, { "epoch": 2.945751320211234, "grad_norm": 1.0719838200353187, "learning_rate": 4.323712296667768e-06, "loss": 0.5779, "step": 3069 }, { "epoch": 2.9467114738358138, "grad_norm": 0.889525287859196, "learning_rate": 4.320393273070924e-06, "loss": 0.5328, "step": 3070 }, { "epoch": 2.9476716274603936, "grad_norm": 0.9403944707064635, "learning_rate": 4.317074554542277e-06, "loss": 0.5351, "step": 3071 }, { "epoch": 2.9486317810849734, "grad_norm": 0.7085845055440657, "learning_rate": 4.313756142571561e-06, "loss": 0.4965, "step": 3072 }, { "epoch": 2.9495919347095536, "grad_norm": 0.9387749162552769, "learning_rate": 4.310438038648376e-06, "loss": 0.5497, "step": 3073 }, { "epoch": 2.9505520883341334, "grad_norm": 1.19731984086214, "learning_rate": 4.307120244262189e-06, "loss": 0.6787, "step": 3074 }, { "epoch": 2.9515122419587136, "grad_norm": 1.128910773698787, "learning_rate": 4.303802760902313e-06, "loss": 0.5817, "step": 3075 }, { "epoch": 2.9524723955832934, "grad_norm": 1.3813563040052745, "learning_rate": 4.300485590057939e-06, "loss": 0.5698, "step": 3076 }, { "epoch": 2.9534325492078732, "grad_norm": 1.2513401339478936, "learning_rate": 4.2971687332181035e-06, "loss": 0.6151, "step": 3077 }, { "epoch": 2.954392702832453, "grad_norm": 1.2513189536570497, "learning_rate": 4.293852191871712e-06, "loss": 0.5639, "step": 3078 }, { "epoch": 2.9553528564570333, "grad_norm": 1.091472376510908, "learning_rate": 4.290535967507524e-06, "loss": 0.5459, "step": 3079 }, { "epoch": 2.956313010081613, "grad_norm": 0.6619025476012991, "learning_rate": 4.287220061614155e-06, "loss": 0.5833, "step": 3080 }, { "epoch": 2.957273163706193, "grad_norm": 1.0017398636729826, "learning_rate": 4.283904475680082e-06, "loss": 0.6342, "step": 3081 }, { "epoch": 2.958233317330773, "grad_norm": 1.0464105186756483, "learning_rate": 4.280589211193635e-06, "loss": 0.547, "step": 3082 }, { "epoch": 2.959193470955353, "grad_norm": 1.0840441321269283, "learning_rate": 4.277274269643002e-06, "loss": 0.5698, "step": 3083 }, { "epoch": 2.9601536245799327, "grad_norm": 1.1791850363518777, "learning_rate": 4.273959652516223e-06, "loss": 0.5667, "step": 3084 }, { "epoch": 2.9611137782045125, "grad_norm": 0.678249156146001, "learning_rate": 4.270645361301195e-06, "loss": 0.473, "step": 3085 }, { "epoch": 2.9620739318290927, "grad_norm": 0.7353250020501182, "learning_rate": 4.267331397485666e-06, "loss": 0.5496, "step": 3086 }, { "epoch": 2.9630340854536725, "grad_norm": 0.8418299983878557, "learning_rate": 4.264017762557245e-06, "loss": 0.6052, "step": 3087 }, { "epoch": 2.9639942390782528, "grad_norm": 0.7562683836444333, "learning_rate": 4.260704458003379e-06, "loss": 0.5522, "step": 3088 }, { "epoch": 2.9649543927028326, "grad_norm": 0.723877744455734, "learning_rate": 4.257391485311384e-06, "loss": 0.601, "step": 3089 }, { "epoch": 2.9659145463274124, "grad_norm": 0.706612756518904, "learning_rate": 4.254078845968409e-06, "loss": 0.5709, "step": 3090 }, { "epoch": 2.966874699951992, "grad_norm": 0.9100528871444342, "learning_rate": 4.2507665414614675e-06, "loss": 0.6357, "step": 3091 }, { "epoch": 2.9678348535765724, "grad_norm": 0.820319753091544, "learning_rate": 4.247454573277418e-06, "loss": 0.5518, "step": 3092 }, { "epoch": 2.968795007201152, "grad_norm": 0.8716750597586485, "learning_rate": 4.244142942902965e-06, "loss": 0.5284, "step": 3093 }, { "epoch": 2.969755160825732, "grad_norm": 0.8972905105510052, "learning_rate": 4.240831651824665e-06, "loss": 0.6018, "step": 3094 }, { "epoch": 2.9707153144503122, "grad_norm": 0.8428320241485139, "learning_rate": 4.237520701528923e-06, "loss": 0.5328, "step": 3095 }, { "epoch": 2.971675468074892, "grad_norm": 0.8279012465470964, "learning_rate": 4.234210093501987e-06, "loss": 0.5307, "step": 3096 }, { "epoch": 2.972635621699472, "grad_norm": 0.7849161908969255, "learning_rate": 4.230899829229955e-06, "loss": 0.5878, "step": 3097 }, { "epoch": 2.9735957753240516, "grad_norm": 0.6795106062372063, "learning_rate": 4.2275899101987665e-06, "loss": 0.5192, "step": 3098 }, { "epoch": 2.974555928948632, "grad_norm": 0.948150784818748, "learning_rate": 4.2242803378942115e-06, "loss": 0.6073, "step": 3099 }, { "epoch": 2.9755160825732117, "grad_norm": 1.254213126703427, "learning_rate": 4.220971113801921e-06, "loss": 0.4849, "step": 3100 }, { "epoch": 2.976476236197792, "grad_norm": 0.8367187521085235, "learning_rate": 4.2176622394073705e-06, "loss": 0.4937, "step": 3101 }, { "epoch": 2.9774363898223717, "grad_norm": 0.7421669901491044, "learning_rate": 4.214353716195879e-06, "loss": 0.488, "step": 3102 }, { "epoch": 2.9783965434469515, "grad_norm": 0.9093484213232655, "learning_rate": 4.211045545652604e-06, "loss": 0.5848, "step": 3103 }, { "epoch": 2.9793566970715313, "grad_norm": 1.225503226582091, "learning_rate": 4.207737729262549e-06, "loss": 0.6243, "step": 3104 }, { "epoch": 2.9803168506961115, "grad_norm": 1.0840561030540536, "learning_rate": 4.2044302685105635e-06, "loss": 0.6129, "step": 3105 }, { "epoch": 2.9812770043206913, "grad_norm": 0.665895564639045, "learning_rate": 4.201123164881321e-06, "loss": 0.4967, "step": 3106 }, { "epoch": 2.982237157945271, "grad_norm": 0.8662199921441227, "learning_rate": 4.1978164198593524e-06, "loss": 0.4767, "step": 3107 }, { "epoch": 2.9831973115698514, "grad_norm": 0.9389551775307353, "learning_rate": 4.194510034929019e-06, "loss": 0.4078, "step": 3108 }, { "epoch": 2.984157465194431, "grad_norm": 0.8345461561330111, "learning_rate": 4.19120401157452e-06, "loss": 0.6179, "step": 3109 }, { "epoch": 2.985117618819011, "grad_norm": 0.6941382163144192, "learning_rate": 4.187898351279895e-06, "loss": 0.4984, "step": 3110 }, { "epoch": 2.9860777724435907, "grad_norm": 0.94565927116189, "learning_rate": 4.184593055529018e-06, "loss": 0.595, "step": 3111 }, { "epoch": 2.987037926068171, "grad_norm": 0.7466975912332401, "learning_rate": 4.181288125805603e-06, "loss": 0.545, "step": 3112 }, { "epoch": 2.9879980796927508, "grad_norm": 0.8742800358955037, "learning_rate": 4.177983563593198e-06, "loss": 0.5152, "step": 3113 }, { "epoch": 2.988958233317331, "grad_norm": 0.9084283000576713, "learning_rate": 4.174679370375182e-06, "loss": 0.5375, "step": 3114 }, { "epoch": 2.989918386941911, "grad_norm": 0.8555944934796202, "learning_rate": 4.1713755476347766e-06, "loss": 0.5183, "step": 3115 }, { "epoch": 2.9908785405664906, "grad_norm": 0.9656023499140833, "learning_rate": 4.168072096855029e-06, "loss": 0.6239, "step": 3116 }, { "epoch": 2.9918386941910704, "grad_norm": 0.9131578120977237, "learning_rate": 4.164769019518823e-06, "loss": 0.5911, "step": 3117 }, { "epoch": 2.9927988478156506, "grad_norm": 0.9144669532906555, "learning_rate": 4.16146631710888e-06, "loss": 0.6188, "step": 3118 }, { "epoch": 2.9937590014402304, "grad_norm": 0.685674738894272, "learning_rate": 4.15816399110774e-06, "loss": 0.5194, "step": 3119 }, { "epoch": 2.9947191550648102, "grad_norm": 1.045818791995369, "learning_rate": 4.154862042997789e-06, "loss": 0.5319, "step": 3120 }, { "epoch": 2.9956793086893905, "grad_norm": 0.836619616966976, "learning_rate": 4.151560474261229e-06, "loss": 0.5512, "step": 3121 }, { "epoch": 2.9966394623139703, "grad_norm": 0.870174041400193, "learning_rate": 4.148259286380105e-06, "loss": 0.6073, "step": 3122 }, { "epoch": 2.99759961593855, "grad_norm": 0.8164268925006102, "learning_rate": 4.144958480836283e-06, "loss": 0.5933, "step": 3123 }, { "epoch": 2.99855976956313, "grad_norm": 0.774284373653836, "learning_rate": 4.141658059111459e-06, "loss": 0.5291, "step": 3124 }, { "epoch": 2.99951992318771, "grad_norm": 0.8042296418598113, "learning_rate": 4.138358022687156e-06, "loss": 0.5198, "step": 3125 }, { "epoch": 3.0, "grad_norm": 0.8042296418598113, "learning_rate": 4.135058373044728e-06, "loss": 0.4282, "step": 3126 }, { "epoch": 3.00096015362458, "grad_norm": 1.3914026073398533, "learning_rate": 4.131759111665349e-06, "loss": 0.5079, "step": 3127 }, { "epoch": 3.00192030724916, "grad_norm": 0.864345737734963, "learning_rate": 4.128460240030025e-06, "loss": 0.4914, "step": 3128 }, { "epoch": 3.00288046087374, "grad_norm": 1.0173391487759513, "learning_rate": 4.125161759619582e-06, "loss": 0.525, "step": 3129 }, { "epoch": 3.0038406144983196, "grad_norm": 0.9269098837347737, "learning_rate": 4.121863671914672e-06, "loss": 0.4683, "step": 3130 }, { "epoch": 3.0048007681229, "grad_norm": 1.2393366645365296, "learning_rate": 4.118565978395777e-06, "loss": 0.559, "step": 3131 }, { "epoch": 3.0057609217474797, "grad_norm": 1.4904121246223232, "learning_rate": 4.115268680543189e-06, "loss": 0.4693, "step": 3132 }, { "epoch": 3.0067210753720595, "grad_norm": 1.0216014012309922, "learning_rate": 4.111971779837036e-06, "loss": 0.4868, "step": 3133 }, { "epoch": 3.0076812289966393, "grad_norm": 0.8880499548142906, "learning_rate": 4.108675277757255e-06, "loss": 0.5602, "step": 3134 }, { "epoch": 3.0086413826212195, "grad_norm": 1.262756965000972, "learning_rate": 4.105379175783616e-06, "loss": 0.4705, "step": 3135 }, { "epoch": 3.0096015362457993, "grad_norm": 0.8300903152367692, "learning_rate": 4.102083475395702e-06, "loss": 0.548, "step": 3136 }, { "epoch": 3.010561689870379, "grad_norm": 1.2456932901969473, "learning_rate": 4.098788178072917e-06, "loss": 0.4203, "step": 3137 }, { "epoch": 3.0115218434949593, "grad_norm": 0.7223071003733428, "learning_rate": 4.095493285294486e-06, "loss": 0.5695, "step": 3138 }, { "epoch": 3.012481997119539, "grad_norm": 0.8966358838992559, "learning_rate": 4.092198798539452e-06, "loss": 0.4807, "step": 3139 }, { "epoch": 3.013442150744119, "grad_norm": 0.9028618379600367, "learning_rate": 4.088904719286671e-06, "loss": 0.516, "step": 3140 }, { "epoch": 3.014402304368699, "grad_norm": 1.1098978838201368, "learning_rate": 4.085611049014825e-06, "loss": 0.4907, "step": 3141 }, { "epoch": 3.015362457993279, "grad_norm": 0.8672209126935914, "learning_rate": 4.082317789202403e-06, "loss": 0.4816, "step": 3142 }, { "epoch": 3.0163226116178588, "grad_norm": 1.2905782324195754, "learning_rate": 4.079024941327715e-06, "loss": 0.4351, "step": 3143 }, { "epoch": 3.017282765242439, "grad_norm": 1.0926908943502787, "learning_rate": 4.075732506868889e-06, "loss": 0.5071, "step": 3144 }, { "epoch": 3.018242918867019, "grad_norm": 1.194613743657086, "learning_rate": 4.072440487303859e-06, "loss": 0.5053, "step": 3145 }, { "epoch": 3.0192030724915986, "grad_norm": 0.8082286407264156, "learning_rate": 4.06914888411038e-06, "loss": 0.4925, "step": 3146 }, { "epoch": 3.0201632261161784, "grad_norm": 1.0477322139784369, "learning_rate": 4.065857698766016e-06, "loss": 0.634, "step": 3147 }, { "epoch": 3.0211233797407586, "grad_norm": 1.3708715187760205, "learning_rate": 4.062566932748143e-06, "loss": 0.4107, "step": 3148 }, { "epoch": 3.0220835333653384, "grad_norm": 0.880903779861004, "learning_rate": 4.059276587533959e-06, "loss": 0.5231, "step": 3149 }, { "epoch": 3.023043686989918, "grad_norm": 0.8333157688055226, "learning_rate": 4.0559866646004546e-06, "loss": 0.487, "step": 3150 }, { "epoch": 3.0240038406144985, "grad_norm": 0.7525026515782269, "learning_rate": 4.052697165424448e-06, "loss": 0.5265, "step": 3151 }, { "epoch": 3.0249639942390782, "grad_norm": 1.1265851530012572, "learning_rate": 4.049408091482556e-06, "loss": 0.5928, "step": 3152 }, { "epoch": 3.025924147863658, "grad_norm": 1.3053303706305273, "learning_rate": 4.046119444251211e-06, "loss": 0.4272, "step": 3153 }, { "epoch": 3.0268843014882383, "grad_norm": 0.7231318591392458, "learning_rate": 4.042831225206651e-06, "loss": 0.4864, "step": 3154 }, { "epoch": 3.027844455112818, "grad_norm": 1.0721500002687478, "learning_rate": 4.039543435824923e-06, "loss": 0.5645, "step": 3155 }, { "epoch": 3.028804608737398, "grad_norm": 0.8732134506927817, "learning_rate": 4.036256077581879e-06, "loss": 0.4033, "step": 3156 }, { "epoch": 3.029764762361978, "grad_norm": 1.19449399067152, "learning_rate": 4.032969151953182e-06, "loss": 0.513, "step": 3157 }, { "epoch": 3.030724915986558, "grad_norm": 1.0434294240413988, "learning_rate": 4.029682660414294e-06, "loss": 0.5519, "step": 3158 }, { "epoch": 3.0316850696111377, "grad_norm": 0.8945283450322828, "learning_rate": 4.0263966044404885e-06, "loss": 0.5121, "step": 3159 }, { "epoch": 3.0326452232357175, "grad_norm": 0.8076839859834142, "learning_rate": 4.023110985506839e-06, "loss": 0.4826, "step": 3160 }, { "epoch": 3.0336053768602977, "grad_norm": 1.1599832792609848, "learning_rate": 4.019825805088225e-06, "loss": 0.5749, "step": 3161 }, { "epoch": 3.0345655304848775, "grad_norm": 0.9223889776029057, "learning_rate": 4.0165410646593324e-06, "loss": 0.4523, "step": 3162 }, { "epoch": 3.0355256841094573, "grad_norm": 1.0440677383250325, "learning_rate": 4.01325676569464e-06, "loss": 0.4535, "step": 3163 }, { "epoch": 3.0364858377340376, "grad_norm": 0.9090040021174142, "learning_rate": 4.009972909668442e-06, "loss": 0.5373, "step": 3164 }, { "epoch": 3.0374459913586174, "grad_norm": 0.743795499750293, "learning_rate": 4.006689498054817e-06, "loss": 0.5213, "step": 3165 }, { "epoch": 3.038406144983197, "grad_norm": 1.018710086010761, "learning_rate": 4.00340653232766e-06, "loss": 0.5148, "step": 3166 }, { "epoch": 3.0393662986077774, "grad_norm": 0.891142125993626, "learning_rate": 4.00012401396066e-06, "loss": 0.5213, "step": 3167 }, { "epoch": 3.040326452232357, "grad_norm": 0.7227832463138173, "learning_rate": 3.9968419444273e-06, "loss": 0.5599, "step": 3168 }, { "epoch": 3.041286605856937, "grad_norm": 0.8362108758578106, "learning_rate": 3.993560325200869e-06, "loss": 0.4478, "step": 3169 }, { "epoch": 3.0422467594815172, "grad_norm": 0.8086433172947891, "learning_rate": 3.99027915775445e-06, "loss": 0.4535, "step": 3170 }, { "epoch": 3.043206913106097, "grad_norm": 0.8614908040933269, "learning_rate": 3.986998443560923e-06, "loss": 0.4788, "step": 3171 }, { "epoch": 3.044167066730677, "grad_norm": 0.772173250821362, "learning_rate": 3.9837181840929665e-06, "loss": 0.5569, "step": 3172 }, { "epoch": 3.0451272203552566, "grad_norm": 0.7870481620571897, "learning_rate": 3.980438380823053e-06, "loss": 0.5016, "step": 3173 }, { "epoch": 3.046087373979837, "grad_norm": 0.8917133630374904, "learning_rate": 3.977159035223451e-06, "loss": 0.514, "step": 3174 }, { "epoch": 3.0470475276044167, "grad_norm": 1.154845268876268, "learning_rate": 3.973880148766228e-06, "loss": 0.4282, "step": 3175 }, { "epoch": 3.0480076812289965, "grad_norm": 0.6898135695135046, "learning_rate": 3.970601722923234e-06, "loss": 0.4382, "step": 3176 }, { "epoch": 3.0489678348535767, "grad_norm": 0.8375983233274111, "learning_rate": 3.9673237591661265e-06, "loss": 0.5465, "step": 3177 }, { "epoch": 3.0499279884781565, "grad_norm": 0.9088856121829855, "learning_rate": 3.964046258966342e-06, "loss": 0.4014, "step": 3178 }, { "epoch": 3.0508881421027363, "grad_norm": 0.9316047972646029, "learning_rate": 3.960769223795118e-06, "loss": 0.4679, "step": 3179 }, { "epoch": 3.0518482957273165, "grad_norm": 0.733969715327885, "learning_rate": 3.957492655123483e-06, "loss": 0.4929, "step": 3180 }, { "epoch": 3.0528084493518963, "grad_norm": 0.6293302662204848, "learning_rate": 3.954216554422249e-06, "loss": 0.5658, "step": 3181 }, { "epoch": 3.053768602976476, "grad_norm": 1.0978071380177679, "learning_rate": 3.9509409231620264e-06, "loss": 0.5708, "step": 3182 }, { "epoch": 3.0547287566010564, "grad_norm": 0.800978391266061, "learning_rate": 3.947665762813209e-06, "loss": 0.4451, "step": 3183 }, { "epoch": 3.055688910225636, "grad_norm": 0.899121048380145, "learning_rate": 3.944391074845982e-06, "loss": 0.5359, "step": 3184 }, { "epoch": 3.056649063850216, "grad_norm": 0.9602610419757382, "learning_rate": 3.941116860730319e-06, "loss": 0.5072, "step": 3185 }, { "epoch": 3.0576092174747957, "grad_norm": 0.9530424499269211, "learning_rate": 3.9378431219359755e-06, "loss": 0.4572, "step": 3186 }, { "epoch": 3.058569371099376, "grad_norm": 0.9525910440604959, "learning_rate": 3.934569859932502e-06, "loss": 0.5201, "step": 3187 }, { "epoch": 3.059529524723956, "grad_norm": 0.7543402474701523, "learning_rate": 3.9312970761892305e-06, "loss": 0.527, "step": 3188 }, { "epoch": 3.0604896783485356, "grad_norm": 1.1558345572881445, "learning_rate": 3.9280247721752765e-06, "loss": 0.3989, "step": 3189 }, { "epoch": 3.061449831973116, "grad_norm": 0.9562258282319074, "learning_rate": 3.924752949359545e-06, "loss": 0.4805, "step": 3190 }, { "epoch": 3.0624099855976956, "grad_norm": 0.9656481748733818, "learning_rate": 3.921481609210719e-06, "loss": 0.4478, "step": 3191 }, { "epoch": 3.0633701392222754, "grad_norm": 0.6631237703576387, "learning_rate": 3.918210753197269e-06, "loss": 0.4775, "step": 3192 }, { "epoch": 3.0643302928468557, "grad_norm": 0.811318287639578, "learning_rate": 3.914940382787452e-06, "loss": 0.5628, "step": 3193 }, { "epoch": 3.0652904464714354, "grad_norm": 0.9755508414077494, "learning_rate": 3.911670499449295e-06, "loss": 0.5745, "step": 3194 }, { "epoch": 3.0662506000960152, "grad_norm": 1.0603568485280532, "learning_rate": 3.908401104650621e-06, "loss": 0.4513, "step": 3195 }, { "epoch": 3.0672107537205955, "grad_norm": 1.1704035710427858, "learning_rate": 3.905132199859018e-06, "loss": 0.4949, "step": 3196 }, { "epoch": 3.0681709073451753, "grad_norm": 1.1676750253453283, "learning_rate": 3.901863786541869e-06, "loss": 0.5049, "step": 3197 }, { "epoch": 3.069131060969755, "grad_norm": 0.9417132579418849, "learning_rate": 3.898595866166329e-06, "loss": 0.4524, "step": 3198 }, { "epoch": 3.070091214594335, "grad_norm": 0.8371927514189818, "learning_rate": 3.895328440199329e-06, "loss": 0.5556, "step": 3199 }, { "epoch": 3.071051368218915, "grad_norm": 0.85461658843214, "learning_rate": 3.892061510107586e-06, "loss": 0.4207, "step": 3200 }, { "epoch": 3.072011521843495, "grad_norm": 0.8895443839363464, "learning_rate": 3.888795077357587e-06, "loss": 0.4774, "step": 3201 }, { "epoch": 3.0729716754680747, "grad_norm": 1.1668535060463365, "learning_rate": 3.885529143415599e-06, "loss": 0.4569, "step": 3202 }, { "epoch": 3.073931829092655, "grad_norm": 1.1579290012163137, "learning_rate": 3.882263709747667e-06, "loss": 0.4969, "step": 3203 }, { "epoch": 3.0748919827172347, "grad_norm": 0.9281255997173742, "learning_rate": 3.878998777819607e-06, "loss": 0.5285, "step": 3204 }, { "epoch": 3.0758521363418145, "grad_norm": 1.095535261676153, "learning_rate": 3.8757343490970115e-06, "loss": 0.5432, "step": 3205 }, { "epoch": 3.0768122899663948, "grad_norm": 0.8321071785370201, "learning_rate": 3.872470425045251e-06, "loss": 0.567, "step": 3206 }, { "epoch": 3.0777724435909746, "grad_norm": 0.967497540577389, "learning_rate": 3.869207007129463e-06, "loss": 0.509, "step": 3207 }, { "epoch": 3.0787325972155544, "grad_norm": 0.9012055914495627, "learning_rate": 3.865944096814564e-06, "loss": 0.4056, "step": 3208 }, { "epoch": 3.0796927508401346, "grad_norm": 0.9170221099024487, "learning_rate": 3.862681695565234e-06, "loss": 0.4241, "step": 3209 }, { "epoch": 3.0806529044647144, "grad_norm": 0.6107509265960442, "learning_rate": 3.859419804845935e-06, "loss": 0.4644, "step": 3210 }, { "epoch": 3.081613058089294, "grad_norm": 1.0977476184229413, "learning_rate": 3.856158426120895e-06, "loss": 0.5517, "step": 3211 }, { "epoch": 3.082573211713874, "grad_norm": 0.8000609906181393, "learning_rate": 3.8528975608541095e-06, "loss": 0.459, "step": 3212 }, { "epoch": 3.0835333653384542, "grad_norm": 1.0107033687071696, "learning_rate": 3.849637210509346e-06, "loss": 0.4634, "step": 3213 }, { "epoch": 3.084493518963034, "grad_norm": 0.9741502650664509, "learning_rate": 3.846377376550144e-06, "loss": 0.5579, "step": 3214 }, { "epoch": 3.085453672587614, "grad_norm": 0.8554699959377902, "learning_rate": 3.843118060439805e-06, "loss": 0.4644, "step": 3215 }, { "epoch": 3.086413826212194, "grad_norm": 0.8218207943779344, "learning_rate": 3.839859263641403e-06, "loss": 0.4278, "step": 3216 }, { "epoch": 3.087373979836774, "grad_norm": 1.0652267913236484, "learning_rate": 3.836600987617776e-06, "loss": 0.4098, "step": 3217 }, { "epoch": 3.0883341334613537, "grad_norm": 0.7447256017775651, "learning_rate": 3.833343233831529e-06, "loss": 0.5164, "step": 3218 }, { "epoch": 3.089294287085934, "grad_norm": 0.9343076041667224, "learning_rate": 3.830086003745034e-06, "loss": 0.5548, "step": 3219 }, { "epoch": 3.0902544407105137, "grad_norm": 0.9911176344914455, "learning_rate": 3.826829298820425e-06, "loss": 0.4751, "step": 3220 }, { "epoch": 3.0912145943350935, "grad_norm": 1.332234418840999, "learning_rate": 3.823573120519606e-06, "loss": 0.5675, "step": 3221 }, { "epoch": 3.0921747479596737, "grad_norm": 0.9289772885352412, "learning_rate": 3.8203174703042345e-06, "loss": 0.5367, "step": 3222 }, { "epoch": 3.0931349015842535, "grad_norm": 0.9021107000671617, "learning_rate": 3.817062349635742e-06, "loss": 0.4441, "step": 3223 }, { "epoch": 3.0940950552088333, "grad_norm": 1.1442534252656118, "learning_rate": 3.813807759975317e-06, "loss": 0.5125, "step": 3224 }, { "epoch": 3.095055208833413, "grad_norm": 1.036406839292829, "learning_rate": 3.8105537027839078e-06, "loss": 0.5498, "step": 3225 }, { "epoch": 3.0960153624579934, "grad_norm": 0.8725452963166547, "learning_rate": 3.8073001795222274e-06, "loss": 0.4953, "step": 3226 }, { "epoch": 3.096975516082573, "grad_norm": 0.7765375237093565, "learning_rate": 3.8040471916507467e-06, "loss": 0.5296, "step": 3227 }, { "epoch": 3.097935669707153, "grad_norm": 0.8353693744713051, "learning_rate": 3.8007947406296975e-06, "loss": 0.417, "step": 3228 }, { "epoch": 3.098895823331733, "grad_norm": 0.8376864393602464, "learning_rate": 3.797542827919071e-06, "loss": 0.4805, "step": 3229 }, { "epoch": 3.099855976956313, "grad_norm": 0.8621626110944723, "learning_rate": 3.7942914549786152e-06, "loss": 0.4916, "step": 3230 }, { "epoch": 3.100816130580893, "grad_norm": 0.9181894923402313, "learning_rate": 3.7910406232678363e-06, "loss": 0.4547, "step": 3231 }, { "epoch": 3.101776284205473, "grad_norm": 1.3471651100182613, "learning_rate": 3.7877903342460002e-06, "loss": 0.4974, "step": 3232 }, { "epoch": 3.102736437830053, "grad_norm": 1.0635141780237085, "learning_rate": 3.7845405893721237e-06, "loss": 0.4444, "step": 3233 }, { "epoch": 3.1036965914546326, "grad_norm": 1.053828093547588, "learning_rate": 3.781291390104986e-06, "loss": 0.4633, "step": 3234 }, { "epoch": 3.104656745079213, "grad_norm": 1.0777120753067402, "learning_rate": 3.7780427379031137e-06, "loss": 0.5534, "step": 3235 }, { "epoch": 3.1056168987037926, "grad_norm": 1.0179693817797615, "learning_rate": 3.774794634224794e-06, "loss": 0.5323, "step": 3236 }, { "epoch": 3.1065770523283724, "grad_norm": 0.8686957818219224, "learning_rate": 3.7715470805280695e-06, "loss": 0.5034, "step": 3237 }, { "epoch": 3.1075372059529527, "grad_norm": 0.6627162091041557, "learning_rate": 3.7683000782707258e-06, "loss": 0.5161, "step": 3238 }, { "epoch": 3.1084973595775325, "grad_norm": 0.9745977774370423, "learning_rate": 3.765053628910314e-06, "loss": 0.5441, "step": 3239 }, { "epoch": 3.1094575132021123, "grad_norm": 0.9243441922856489, "learning_rate": 3.7618077339041244e-06, "loss": 0.4801, "step": 3240 }, { "epoch": 3.110417666826692, "grad_norm": 0.6210359592458765, "learning_rate": 3.758562394709208e-06, "loss": 0.5305, "step": 3241 }, { "epoch": 3.1113778204512723, "grad_norm": 1.1632680130925295, "learning_rate": 3.7553176127823643e-06, "loss": 0.6257, "step": 3242 }, { "epoch": 3.112337974075852, "grad_norm": 0.8487320374393228, "learning_rate": 3.7520733895801375e-06, "loss": 0.4664, "step": 3243 }, { "epoch": 3.113298127700432, "grad_norm": 0.9966529973473434, "learning_rate": 3.748829726558827e-06, "loss": 0.4245, "step": 3244 }, { "epoch": 3.114258281325012, "grad_norm": 1.1148040435112274, "learning_rate": 3.745586625174479e-06, "loss": 0.516, "step": 3245 }, { "epoch": 3.115218434949592, "grad_norm": 0.9376150319465978, "learning_rate": 3.7423440868828842e-06, "loss": 0.478, "step": 3246 }, { "epoch": 3.1161785885741717, "grad_norm": 1.1103062466598324, "learning_rate": 3.7391021131395866e-06, "loss": 0.4663, "step": 3247 }, { "epoch": 3.117138742198752, "grad_norm": 1.1970599319039896, "learning_rate": 3.7358607053998717e-06, "loss": 0.48, "step": 3248 }, { "epoch": 3.1180988958233318, "grad_norm": 1.1599830393958144, "learning_rate": 3.732619865118772e-06, "loss": 0.5244, "step": 3249 }, { "epoch": 3.1190590494479116, "grad_norm": 1.116870438372372, "learning_rate": 3.7293795937510703e-06, "loss": 0.429, "step": 3250 }, { "epoch": 3.120019203072492, "grad_norm": 1.090844058021211, "learning_rate": 3.7261398927512836e-06, "loss": 0.6078, "step": 3251 }, { "epoch": 3.1209793566970716, "grad_norm": 1.0477548121792273, "learning_rate": 3.722900763573685e-06, "loss": 0.5762, "step": 3252 }, { "epoch": 3.1219395103216514, "grad_norm": 0.8872541856457361, "learning_rate": 3.71966220767228e-06, "loss": 0.5386, "step": 3253 }, { "epoch": 3.122899663946231, "grad_norm": 0.9152961213631531, "learning_rate": 3.7164242265008244e-06, "loss": 0.413, "step": 3254 }, { "epoch": 3.1238598175708114, "grad_norm": 0.6696519809647296, "learning_rate": 3.7131868215128135e-06, "loss": 0.4788, "step": 3255 }, { "epoch": 3.1248199711953912, "grad_norm": 1.0228293659162584, "learning_rate": 3.7099499941614823e-06, "loss": 0.4968, "step": 3256 }, { "epoch": 3.125780124819971, "grad_norm": 0.9839411687865492, "learning_rate": 3.7067137458998092e-06, "loss": 0.5828, "step": 3257 }, { "epoch": 3.1267402784445513, "grad_norm": 0.9105196642769025, "learning_rate": 3.7034780781805126e-06, "loss": 0.4742, "step": 3258 }, { "epoch": 3.127700432069131, "grad_norm": 1.0060527073959544, "learning_rate": 3.7002429924560458e-06, "loss": 0.4347, "step": 3259 }, { "epoch": 3.128660585693711, "grad_norm": 0.8761829356904088, "learning_rate": 3.6970084901786075e-06, "loss": 0.4636, "step": 3260 }, { "epoch": 3.129620739318291, "grad_norm": 1.0495892402629192, "learning_rate": 3.6937745728001283e-06, "loss": 0.4931, "step": 3261 }, { "epoch": 3.130580892942871, "grad_norm": 1.1370610482021453, "learning_rate": 3.6905412417722807e-06, "loss": 0.4351, "step": 3262 }, { "epoch": 3.1315410465674507, "grad_norm": 0.7815142198993061, "learning_rate": 3.6873084985464734e-06, "loss": 0.5032, "step": 3263 }, { "epoch": 3.132501200192031, "grad_norm": 0.9817196842842449, "learning_rate": 3.6840763445738485e-06, "loss": 0.5319, "step": 3264 }, { "epoch": 3.1334613538166107, "grad_norm": 1.1449671628348084, "learning_rate": 3.6808447813052873e-06, "loss": 0.4935, "step": 3265 }, { "epoch": 3.1344215074411905, "grad_norm": 1.122560784664853, "learning_rate": 3.677613810191401e-06, "loss": 0.5607, "step": 3266 }, { "epoch": 3.1353816610657703, "grad_norm": 0.8920583419676801, "learning_rate": 3.674383432682539e-06, "loss": 0.4683, "step": 3267 }, { "epoch": 3.1363418146903506, "grad_norm": 0.6895196285643278, "learning_rate": 3.671153650228787e-06, "loss": 0.5222, "step": 3268 }, { "epoch": 3.1373019683149304, "grad_norm": 1.0968728205096376, "learning_rate": 3.6679244642799544e-06, "loss": 0.4575, "step": 3269 }, { "epoch": 3.13826212193951, "grad_norm": 0.8529899194511216, "learning_rate": 3.664695876285592e-06, "loss": 0.5433, "step": 3270 }, { "epoch": 3.1392222755640904, "grad_norm": 1.508602516952346, "learning_rate": 3.6614678876949767e-06, "loss": 0.5067, "step": 3271 }, { "epoch": 3.14018242918867, "grad_norm": 1.0203575347718647, "learning_rate": 3.658240499957118e-06, "loss": 0.4919, "step": 3272 }, { "epoch": 3.14114258281325, "grad_norm": 0.8282690336366383, "learning_rate": 3.6550137145207577e-06, "loss": 0.5028, "step": 3273 }, { "epoch": 3.1421027364378302, "grad_norm": 1.0488309016109558, "learning_rate": 3.651787532834362e-06, "loss": 0.4682, "step": 3274 }, { "epoch": 3.14306289006241, "grad_norm": 0.8820974018297362, "learning_rate": 3.648561956346132e-06, "loss": 0.4902, "step": 3275 }, { "epoch": 3.14402304368699, "grad_norm": 1.0044170948955715, "learning_rate": 3.645336986503995e-06, "loss": 0.4822, "step": 3276 }, { "epoch": 3.14498319731157, "grad_norm": 0.9463898105033111, "learning_rate": 3.642112624755603e-06, "loss": 0.5128, "step": 3277 }, { "epoch": 3.14594335093615, "grad_norm": 0.876320474716569, "learning_rate": 3.6388888725483394e-06, "loss": 0.475, "step": 3278 }, { "epoch": 3.1469035045607296, "grad_norm": 1.104805996147048, "learning_rate": 3.6356657313293104e-06, "loss": 0.4025, "step": 3279 }, { "epoch": 3.1478636581853094, "grad_norm": 1.2192516172385965, "learning_rate": 3.6324432025453494e-06, "loss": 0.4758, "step": 3280 }, { "epoch": 3.1488238118098897, "grad_norm": 0.8552174141219044, "learning_rate": 3.62922128764302e-06, "loss": 0.4667, "step": 3281 }, { "epoch": 3.1497839654344695, "grad_norm": 0.9568780009674015, "learning_rate": 3.6259999880685987e-06, "loss": 0.5198, "step": 3282 }, { "epoch": 3.1507441190590493, "grad_norm": 1.182887268254218, "learning_rate": 3.6227793052680983e-06, "loss": 0.5006, "step": 3283 }, { "epoch": 3.1517042726836295, "grad_norm": 0.874155996726093, "learning_rate": 3.6195592406872436e-06, "loss": 0.4712, "step": 3284 }, { "epoch": 3.1526644263082093, "grad_norm": 0.7502605471559851, "learning_rate": 3.6163397957714895e-06, "loss": 0.5501, "step": 3285 }, { "epoch": 3.153624579932789, "grad_norm": 1.0432809854192018, "learning_rate": 3.6131209719660133e-06, "loss": 0.5279, "step": 3286 }, { "epoch": 3.1545847335573693, "grad_norm": 0.9609259689843344, "learning_rate": 3.609902770715705e-06, "loss": 0.51, "step": 3287 }, { "epoch": 3.155544887181949, "grad_norm": 0.661784314221155, "learning_rate": 3.6066851934651847e-06, "loss": 0.5295, "step": 3288 }, { "epoch": 3.156505040806529, "grad_norm": 0.8075938794765092, "learning_rate": 3.6034682416587885e-06, "loss": 0.4221, "step": 3289 }, { "epoch": 3.157465194431109, "grad_norm": 0.8061289056733293, "learning_rate": 3.600251916740569e-06, "loss": 0.4721, "step": 3290 }, { "epoch": 3.158425348055689, "grad_norm": 0.9718561749492596, "learning_rate": 3.597036220154303e-06, "loss": 0.5775, "step": 3291 }, { "epoch": 3.1593855016802688, "grad_norm": 0.8185976831848043, "learning_rate": 3.5938211533434797e-06, "loss": 0.637, "step": 3292 }, { "epoch": 3.160345655304849, "grad_norm": 0.8781819235452001, "learning_rate": 3.590606717751308e-06, "loss": 0.5244, "step": 3293 }, { "epoch": 3.161305808929429, "grad_norm": 1.0624849399914915, "learning_rate": 3.5873929148207187e-06, "loss": 0.5472, "step": 3294 }, { "epoch": 3.1622659625540086, "grad_norm": 0.5735663869623923, "learning_rate": 3.5841797459943473e-06, "loss": 0.4534, "step": 3295 }, { "epoch": 3.1632261161785884, "grad_norm": 0.7799703320397082, "learning_rate": 3.5809672127145563e-06, "loss": 0.4519, "step": 3296 }, { "epoch": 3.1641862698031686, "grad_norm": 0.77672842883097, "learning_rate": 3.5777553164234114e-06, "loss": 0.5461, "step": 3297 }, { "epoch": 3.1651464234277484, "grad_norm": 0.6342728225868007, "learning_rate": 3.5745440585627032e-06, "loss": 0.5536, "step": 3298 }, { "epoch": 3.1661065770523282, "grad_norm": 0.9421309569708362, "learning_rate": 3.571333440573932e-06, "loss": 0.4773, "step": 3299 }, { "epoch": 3.1670667306769085, "grad_norm": 0.9867503372992854, "learning_rate": 3.5681234638983054e-06, "loss": 0.4776, "step": 3300 }, { "epoch": 3.1680268843014883, "grad_norm": 0.7204587451760992, "learning_rate": 3.5649141299767504e-06, "loss": 0.5643, "step": 3301 }, { "epoch": 3.168987037926068, "grad_norm": 0.9054503155855498, "learning_rate": 3.561705440249903e-06, "loss": 0.5225, "step": 3302 }, { "epoch": 3.1699471915506483, "grad_norm": 0.7617256536532794, "learning_rate": 3.558497396158108e-06, "loss": 0.4426, "step": 3303 }, { "epoch": 3.170907345175228, "grad_norm": 0.9198864401471253, "learning_rate": 3.5552899991414248e-06, "loss": 0.4963, "step": 3304 }, { "epoch": 3.171867498799808, "grad_norm": 1.0012469440729712, "learning_rate": 3.5520832506396156e-06, "loss": 0.5399, "step": 3305 }, { "epoch": 3.172827652424388, "grad_norm": 1.1455687705705755, "learning_rate": 3.548877152092158e-06, "loss": 0.5208, "step": 3306 }, { "epoch": 3.173787806048968, "grad_norm": 0.792280805526548, "learning_rate": 3.5456717049382373e-06, "loss": 0.574, "step": 3307 }, { "epoch": 3.1747479596735477, "grad_norm": 0.7136475818493971, "learning_rate": 3.5424669106167408e-06, "loss": 0.5772, "step": 3308 }, { "epoch": 3.1757081132981275, "grad_norm": 0.7257256473343688, "learning_rate": 3.5392627705662686e-06, "loss": 0.4932, "step": 3309 }, { "epoch": 3.1766682669227078, "grad_norm": 0.8129532294294097, "learning_rate": 3.5360592862251242e-06, "loss": 0.5944, "step": 3310 }, { "epoch": 3.1776284205472876, "grad_norm": 0.7934212041476287, "learning_rate": 3.5328564590313155e-06, "loss": 0.466, "step": 3311 }, { "epoch": 3.1785885741718674, "grad_norm": 0.8207315807886126, "learning_rate": 3.529654290422563e-06, "loss": 0.5329, "step": 3312 }, { "epoch": 3.1795487277964476, "grad_norm": 0.7673926637843752, "learning_rate": 3.526452781836279e-06, "loss": 0.4878, "step": 3313 }, { "epoch": 3.1805088814210274, "grad_norm": 0.8177945990280431, "learning_rate": 3.5232519347095935e-06, "loss": 0.5477, "step": 3314 }, { "epoch": 3.181469035045607, "grad_norm": 0.7402865437943232, "learning_rate": 3.5200517504793254e-06, "loss": 0.458, "step": 3315 }, { "epoch": 3.1824291886701874, "grad_norm": 1.0476649417196677, "learning_rate": 3.516852230582008e-06, "loss": 0.5227, "step": 3316 }, { "epoch": 3.183389342294767, "grad_norm": 1.2021534209765687, "learning_rate": 3.513653376453872e-06, "loss": 0.473, "step": 3317 }, { "epoch": 3.184349495919347, "grad_norm": 0.683032447156478, "learning_rate": 3.510455189530845e-06, "loss": 0.4704, "step": 3318 }, { "epoch": 3.1853096495439273, "grad_norm": 0.8039328281310977, "learning_rate": 3.507257671248561e-06, "loss": 0.4645, "step": 3319 }, { "epoch": 3.186269803168507, "grad_norm": 1.1597060414112543, "learning_rate": 3.5040608230423533e-06, "loss": 0.498, "step": 3320 }, { "epoch": 3.187229956793087, "grad_norm": 0.8484751084056502, "learning_rate": 3.5008646463472506e-06, "loss": 0.6071, "step": 3321 }, { "epoch": 3.1881901104176666, "grad_norm": 0.7176727802240697, "learning_rate": 3.497669142597984e-06, "loss": 0.5252, "step": 3322 }, { "epoch": 3.189150264042247, "grad_norm": 0.7885389320003356, "learning_rate": 3.4944743132289793e-06, "loss": 0.5479, "step": 3323 }, { "epoch": 3.1901104176668267, "grad_norm": 0.9351437223270705, "learning_rate": 3.4912801596743615e-06, "loss": 0.5521, "step": 3324 }, { "epoch": 3.1910705712914065, "grad_norm": 1.1533277721665702, "learning_rate": 3.4880866833679555e-06, "loss": 0.4421, "step": 3325 }, { "epoch": 3.1920307249159867, "grad_norm": 0.8074644278091012, "learning_rate": 3.4848938857432727e-06, "loss": 0.5508, "step": 3326 }, { "epoch": 3.1929908785405665, "grad_norm": 0.843448179118432, "learning_rate": 3.4817017682335326e-06, "loss": 0.4795, "step": 3327 }, { "epoch": 3.1939510321651463, "grad_norm": 0.9668607602812582, "learning_rate": 3.478510332271635e-06, "loss": 0.4555, "step": 3328 }, { "epoch": 3.1949111857897265, "grad_norm": 1.123016626023655, "learning_rate": 3.475319579290187e-06, "loss": 0.5384, "step": 3329 }, { "epoch": 3.1958713394143063, "grad_norm": 0.6036919254328519, "learning_rate": 3.4721295107214835e-06, "loss": 0.4709, "step": 3330 }, { "epoch": 3.196831493038886, "grad_norm": 0.8883327680201057, "learning_rate": 3.4689401279975087e-06, "loss": 0.464, "step": 3331 }, { "epoch": 3.1977916466634664, "grad_norm": 1.0836208339816957, "learning_rate": 3.4657514325499454e-06, "loss": 0.5766, "step": 3332 }, { "epoch": 3.198751800288046, "grad_norm": 0.8819861022302214, "learning_rate": 3.4625634258101636e-06, "loss": 0.5155, "step": 3333 }, { "epoch": 3.199711953912626, "grad_norm": 0.6461803833811736, "learning_rate": 3.459376109209226e-06, "loss": 0.5272, "step": 3334 }, { "epoch": 3.2006721075372058, "grad_norm": 1.0353567489322915, "learning_rate": 3.456189484177884e-06, "loss": 0.4751, "step": 3335 }, { "epoch": 3.201632261161786, "grad_norm": 0.6897957831526529, "learning_rate": 3.4530035521465803e-06, "loss": 0.5139, "step": 3336 }, { "epoch": 3.202592414786366, "grad_norm": 0.8622010770764439, "learning_rate": 3.4498183145454445e-06, "loss": 0.4647, "step": 3337 }, { "epoch": 3.2035525684109456, "grad_norm": 0.9368149133855839, "learning_rate": 3.446633772804299e-06, "loss": 0.5086, "step": 3338 }, { "epoch": 3.204512722035526, "grad_norm": 0.7450727274691857, "learning_rate": 3.443449928352646e-06, "loss": 0.4731, "step": 3339 }, { "epoch": 3.2054728756601056, "grad_norm": 0.9922006882627202, "learning_rate": 3.4402667826196845e-06, "loss": 0.5446, "step": 3340 }, { "epoch": 3.2064330292846854, "grad_norm": 0.9242462142317583, "learning_rate": 3.4370843370342887e-06, "loss": 0.5783, "step": 3341 }, { "epoch": 3.2073931829092657, "grad_norm": 0.868573613462201, "learning_rate": 3.433902593025028e-06, "loss": 0.4514, "step": 3342 }, { "epoch": 3.2083533365338455, "grad_norm": 0.7372887870020827, "learning_rate": 3.430721552020154e-06, "loss": 0.6377, "step": 3343 }, { "epoch": 3.2093134901584253, "grad_norm": 0.9123847063589252, "learning_rate": 3.4275412154475996e-06, "loss": 0.5213, "step": 3344 }, { "epoch": 3.2102736437830055, "grad_norm": 1.0897131383025043, "learning_rate": 3.4243615847349855e-06, "loss": 0.526, "step": 3345 }, { "epoch": 3.2112337974075853, "grad_norm": 1.4304237903175665, "learning_rate": 3.4211826613096135e-06, "loss": 0.4695, "step": 3346 }, { "epoch": 3.212193951032165, "grad_norm": 1.1061173359849088, "learning_rate": 3.418004446598467e-06, "loss": 0.4218, "step": 3347 }, { "epoch": 3.213154104656745, "grad_norm": 0.7212509302340454, "learning_rate": 3.4148269420282155e-06, "loss": 0.6019, "step": 3348 }, { "epoch": 3.214114258281325, "grad_norm": 1.0476337689419806, "learning_rate": 3.4116501490252037e-06, "loss": 0.5386, "step": 3349 }, { "epoch": 3.215074411905905, "grad_norm": 0.8195738709038421, "learning_rate": 3.4084740690154617e-06, "loss": 0.5528, "step": 3350 }, { "epoch": 3.2160345655304847, "grad_norm": 0.8797765244330742, "learning_rate": 3.405298703424699e-06, "loss": 0.56, "step": 3351 }, { "epoch": 3.216994719155065, "grad_norm": 1.0895789833687701, "learning_rate": 3.4021240536783e-06, "loss": 0.4484, "step": 3352 }, { "epoch": 3.2179548727796448, "grad_norm": 1.0126110804278703, "learning_rate": 3.3989501212013344e-06, "loss": 0.4698, "step": 3353 }, { "epoch": 3.2189150264042246, "grad_norm": 1.2395511776710193, "learning_rate": 3.3957769074185443e-06, "loss": 0.5045, "step": 3354 }, { "epoch": 3.219875180028805, "grad_norm": 1.211456669672208, "learning_rate": 3.392604413754351e-06, "loss": 0.556, "step": 3355 }, { "epoch": 3.2208353336533846, "grad_norm": 1.034519960530992, "learning_rate": 3.3894326416328575e-06, "loss": 0.4987, "step": 3356 }, { "epoch": 3.2217954872779644, "grad_norm": 1.0067472792479046, "learning_rate": 3.3862615924778326e-06, "loss": 0.4481, "step": 3357 }, { "epoch": 3.2227556409025446, "grad_norm": 1.0355011164356367, "learning_rate": 3.3830912677127313e-06, "loss": 0.5169, "step": 3358 }, { "epoch": 3.2237157945271244, "grad_norm": 0.8182212155459202, "learning_rate": 3.3799216687606727e-06, "loss": 0.426, "step": 3359 }, { "epoch": 3.224675948151704, "grad_norm": 1.335823137837632, "learning_rate": 3.3767527970444614e-06, "loss": 0.5062, "step": 3360 }, { "epoch": 3.225636101776284, "grad_norm": 0.9989039261153168, "learning_rate": 3.3735846539865684e-06, "loss": 0.4021, "step": 3361 }, { "epoch": 3.2265962554008643, "grad_norm": 0.752273233115041, "learning_rate": 3.370417241009138e-06, "loss": 0.483, "step": 3362 }, { "epoch": 3.227556409025444, "grad_norm": 1.0151818605851748, "learning_rate": 3.367250559533989e-06, "loss": 0.5228, "step": 3363 }, { "epoch": 3.228516562650024, "grad_norm": 1.0586404820816278, "learning_rate": 3.3640846109826107e-06, "loss": 0.5388, "step": 3364 }, { "epoch": 3.229476716274604, "grad_norm": 1.2065884971775047, "learning_rate": 3.3609193967761637e-06, "loss": 0.5272, "step": 3365 }, { "epoch": 3.230436869899184, "grad_norm": 0.7894568530740448, "learning_rate": 3.3577549183354797e-06, "loss": 0.4627, "step": 3366 }, { "epoch": 3.2313970235237637, "grad_norm": 0.776445123669472, "learning_rate": 3.3545911770810565e-06, "loss": 0.5327, "step": 3367 }, { "epoch": 3.232357177148344, "grad_norm": 0.7779282347735662, "learning_rate": 3.3514281744330644e-06, "loss": 0.4816, "step": 3368 }, { "epoch": 3.2333173307729237, "grad_norm": 0.6765998515683992, "learning_rate": 3.348265911811346e-06, "loss": 0.5418, "step": 3369 }, { "epoch": 3.2342774843975035, "grad_norm": 0.6694001811797424, "learning_rate": 3.345104390635401e-06, "loss": 0.4071, "step": 3370 }, { "epoch": 3.2352376380220838, "grad_norm": 0.752921339126049, "learning_rate": 3.3419436123244076e-06, "loss": 0.5292, "step": 3371 }, { "epoch": 3.2361977916466635, "grad_norm": 1.3723404940405148, "learning_rate": 3.3387835782972e-06, "loss": 0.4671, "step": 3372 }, { "epoch": 3.2371579452712433, "grad_norm": 1.131903358703648, "learning_rate": 3.3356242899722877e-06, "loss": 0.4623, "step": 3373 }, { "epoch": 3.238118098895823, "grad_norm": 1.0284420134196048, "learning_rate": 3.3324657487678423e-06, "loss": 0.47, "step": 3374 }, { "epoch": 3.2390782525204034, "grad_norm": 0.8802463588007342, "learning_rate": 3.3293079561016957e-06, "loss": 0.5225, "step": 3375 }, { "epoch": 3.240038406144983, "grad_norm": 0.6643893359160115, "learning_rate": 3.3261509133913484e-06, "loss": 0.4205, "step": 3376 }, { "epoch": 3.240998559769563, "grad_norm": 0.7829192755287233, "learning_rate": 3.3229946220539657e-06, "loss": 0.5636, "step": 3377 }, { "epoch": 3.241958713394143, "grad_norm": 0.7256790249482753, "learning_rate": 3.3198390835063685e-06, "loss": 0.3787, "step": 3378 }, { "epoch": 3.242918867018723, "grad_norm": 0.9378987159415216, "learning_rate": 3.3166842991650493e-06, "loss": 0.5291, "step": 3379 }, { "epoch": 3.243879020643303, "grad_norm": 0.9173605215254617, "learning_rate": 3.3135302704461526e-06, "loss": 0.4321, "step": 3380 }, { "epoch": 3.244839174267883, "grad_norm": 0.8323496771266007, "learning_rate": 3.31037699876549e-06, "loss": 0.476, "step": 3381 }, { "epoch": 3.245799327892463, "grad_norm": 0.7816268405721645, "learning_rate": 3.307224485538534e-06, "loss": 0.5163, "step": 3382 }, { "epoch": 3.2467594815170426, "grad_norm": 0.934391035449295, "learning_rate": 3.3040727321804083e-06, "loss": 0.4106, "step": 3383 }, { "epoch": 3.247719635141623, "grad_norm": 1.1009346661352553, "learning_rate": 3.300921740105908e-06, "loss": 0.4829, "step": 3384 }, { "epoch": 3.2486797887662027, "grad_norm": 0.9930613071345377, "learning_rate": 3.297771510729474e-06, "loss": 0.5579, "step": 3385 }, { "epoch": 3.2496399423907825, "grad_norm": 0.7905002153491939, "learning_rate": 3.2946220454652134e-06, "loss": 0.5809, "step": 3386 }, { "epoch": 3.2506000960153623, "grad_norm": 0.772530966928306, "learning_rate": 3.2914733457268876e-06, "loss": 0.4531, "step": 3387 }, { "epoch": 3.2515602496399425, "grad_norm": 0.7859773680138604, "learning_rate": 3.288325412927914e-06, "loss": 0.4746, "step": 3388 }, { "epoch": 3.2525204032645223, "grad_norm": 0.9093991809621209, "learning_rate": 3.285178248481366e-06, "loss": 0.575, "step": 3389 }, { "epoch": 3.253480556889102, "grad_norm": 0.9475375398419091, "learning_rate": 3.282031853799971e-06, "loss": 0.4553, "step": 3390 }, { "epoch": 3.2544407105136823, "grad_norm": 0.863771297579437, "learning_rate": 3.278886230296113e-06, "loss": 0.416, "step": 3391 }, { "epoch": 3.255400864138262, "grad_norm": 0.6949504336027247, "learning_rate": 3.2757413793818294e-06, "loss": 0.4278, "step": 3392 }, { "epoch": 3.256361017762842, "grad_norm": 0.7240433243432793, "learning_rate": 3.2725973024688073e-06, "loss": 0.5347, "step": 3393 }, { "epoch": 3.257321171387422, "grad_norm": 1.109649150541374, "learning_rate": 3.2694540009683926e-06, "loss": 0.5604, "step": 3394 }, { "epoch": 3.258281325012002, "grad_norm": 1.320802575208359, "learning_rate": 3.266311476291579e-06, "loss": 0.4846, "step": 3395 }, { "epoch": 3.2592414786365818, "grad_norm": 1.0957346272157438, "learning_rate": 3.2631697298490105e-06, "loss": 0.4197, "step": 3396 }, { "epoch": 3.260201632261162, "grad_norm": 1.189806487794949, "learning_rate": 3.2600287630509854e-06, "loss": 0.4547, "step": 3397 }, { "epoch": 3.261161785885742, "grad_norm": 1.116824380054104, "learning_rate": 3.256888577307448e-06, "loss": 0.536, "step": 3398 }, { "epoch": 3.2621219395103216, "grad_norm": 1.034639236580772, "learning_rate": 3.2537491740279946e-06, "loss": 0.4892, "step": 3399 }, { "epoch": 3.2630820931349014, "grad_norm": 1.1528435804086117, "learning_rate": 3.250610554621875e-06, "loss": 0.5688, "step": 3400 }, { "epoch": 3.2640422467594816, "grad_norm": 0.7819241030145392, "learning_rate": 3.2474727204979728e-06, "loss": 0.4719, "step": 3401 }, { "epoch": 3.2650024003840614, "grad_norm": 0.7785423362859598, "learning_rate": 3.2443356730648363e-06, "loss": 0.4387, "step": 3402 }, { "epoch": 3.265962554008641, "grad_norm": 0.842106783512466, "learning_rate": 3.2411994137306463e-06, "loss": 0.5768, "step": 3403 }, { "epoch": 3.2669227076332215, "grad_norm": 0.9574601635247647, "learning_rate": 3.23806394390324e-06, "loss": 0.5324, "step": 3404 }, { "epoch": 3.2678828612578013, "grad_norm": 0.9228615236001334, "learning_rate": 3.234929264990096e-06, "loss": 0.4532, "step": 3405 }, { "epoch": 3.268843014882381, "grad_norm": 0.8611448971452202, "learning_rate": 3.231795378398337e-06, "loss": 0.521, "step": 3406 }, { "epoch": 3.2698031685069613, "grad_norm": 0.811645435559627, "learning_rate": 3.2286622855347306e-06, "loss": 0.4654, "step": 3407 }, { "epoch": 3.270763322131541, "grad_norm": 1.1053931276061826, "learning_rate": 3.2255299878056913e-06, "loss": 0.6085, "step": 3408 }, { "epoch": 3.271723475756121, "grad_norm": 0.8453771458214657, "learning_rate": 3.2223984866172715e-06, "loss": 0.4031, "step": 3409 }, { "epoch": 3.272683629380701, "grad_norm": 0.7717872992037439, "learning_rate": 3.21926778337517e-06, "loss": 0.4254, "step": 3410 }, { "epoch": 3.273643783005281, "grad_norm": 0.8436072769537821, "learning_rate": 3.2161378794847255e-06, "loss": 0.4439, "step": 3411 }, { "epoch": 3.2746039366298607, "grad_norm": 1.0584379739565548, "learning_rate": 3.2130087763509165e-06, "loss": 0.4724, "step": 3412 }, { "epoch": 3.2755640902544405, "grad_norm": 0.9944348062017636, "learning_rate": 3.2098804753783696e-06, "loss": 0.4294, "step": 3413 }, { "epoch": 3.2765242438790207, "grad_norm": 0.695885032163286, "learning_rate": 3.2067529779713384e-06, "loss": 0.5089, "step": 3414 }, { "epoch": 3.2774843975036005, "grad_norm": 0.8530386596174464, "learning_rate": 3.2036262855337304e-06, "loss": 0.5143, "step": 3415 }, { "epoch": 3.2784445511281803, "grad_norm": 1.099993787952269, "learning_rate": 3.200500399469077e-06, "loss": 0.4883, "step": 3416 }, { "epoch": 3.2794047047527606, "grad_norm": 0.9009048796187077, "learning_rate": 3.1973753211805597e-06, "loss": 0.5353, "step": 3417 }, { "epoch": 3.2803648583773404, "grad_norm": 0.8558784062220904, "learning_rate": 3.1942510520709928e-06, "loss": 0.4927, "step": 3418 }, { "epoch": 3.28132501200192, "grad_norm": 1.0831290750035352, "learning_rate": 3.1911275935428256e-06, "loss": 0.4996, "step": 3419 }, { "epoch": 3.2822851656265004, "grad_norm": 0.7669297761059098, "learning_rate": 3.1880049469981468e-06, "loss": 0.55, "step": 3420 }, { "epoch": 3.28324531925108, "grad_norm": 1.2246947849045529, "learning_rate": 3.1848831138386772e-06, "loss": 0.6171, "step": 3421 }, { "epoch": 3.28420547287566, "grad_norm": 0.9250400887493104, "learning_rate": 3.1817620954657745e-06, "loss": 0.4461, "step": 3422 }, { "epoch": 3.2851656265002402, "grad_norm": 0.6516881792663238, "learning_rate": 3.178641893280433e-06, "loss": 0.5007, "step": 3423 }, { "epoch": 3.28612578012482, "grad_norm": 1.2988238984954288, "learning_rate": 3.1755225086832755e-06, "loss": 0.5071, "step": 3424 }, { "epoch": 3.2870859337494, "grad_norm": 1.1134319911252994, "learning_rate": 3.17240394307456e-06, "loss": 0.5479, "step": 3425 }, { "epoch": 3.2880460873739796, "grad_norm": 0.6676863557077021, "learning_rate": 3.169286197854181e-06, "loss": 0.5254, "step": 3426 }, { "epoch": 3.28900624099856, "grad_norm": 0.8348614162205359, "learning_rate": 3.1661692744216564e-06, "loss": 0.524, "step": 3427 }, { "epoch": 3.2899663946231397, "grad_norm": 0.8220547007946011, "learning_rate": 3.163053174176145e-06, "loss": 0.5231, "step": 3428 }, { "epoch": 3.2909265482477195, "grad_norm": 1.361712828450403, "learning_rate": 3.159937898516424e-06, "loss": 0.4368, "step": 3429 }, { "epoch": 3.2918867018722997, "grad_norm": 0.8410716905126402, "learning_rate": 3.1568234488409126e-06, "loss": 0.5224, "step": 3430 }, { "epoch": 3.2928468554968795, "grad_norm": 0.7285710542649372, "learning_rate": 3.153709826547654e-06, "loss": 0.5181, "step": 3431 }, { "epoch": 3.2938070091214593, "grad_norm": 1.299433722149376, "learning_rate": 3.1505970330343172e-06, "loss": 0.4216, "step": 3432 }, { "epoch": 3.2947671627460395, "grad_norm": 0.959388343429885, "learning_rate": 3.147485069698204e-06, "loss": 0.493, "step": 3433 }, { "epoch": 3.2957273163706193, "grad_norm": 0.9654883866098786, "learning_rate": 3.1443739379362396e-06, "loss": 0.5514, "step": 3434 }, { "epoch": 3.296687469995199, "grad_norm": 0.8556312090132383, "learning_rate": 3.1412636391449792e-06, "loss": 0.5505, "step": 3435 }, { "epoch": 3.2976476236197794, "grad_norm": 1.218594140846149, "learning_rate": 3.1381541747206034e-06, "loss": 0.5483, "step": 3436 }, { "epoch": 3.298607777244359, "grad_norm": 0.7508880680501319, "learning_rate": 3.1350455460589157e-06, "loss": 0.4334, "step": 3437 }, { "epoch": 3.299567930868939, "grad_norm": 0.7882229302589332, "learning_rate": 3.131937754555347e-06, "loss": 0.568, "step": 3438 }, { "epoch": 3.3005280844935188, "grad_norm": 0.8116215908386524, "learning_rate": 3.128830801604953e-06, "loss": 0.5175, "step": 3439 }, { "epoch": 3.301488238118099, "grad_norm": 0.9233611732212935, "learning_rate": 3.125724688602409e-06, "loss": 0.5427, "step": 3440 }, { "epoch": 3.302448391742679, "grad_norm": 0.9518232099343326, "learning_rate": 3.122619416942019e-06, "loss": 0.5081, "step": 3441 }, { "epoch": 3.3034085453672586, "grad_norm": 0.8974283550891985, "learning_rate": 3.1195149880177027e-06, "loss": 0.5331, "step": 3442 }, { "epoch": 3.304368698991839, "grad_norm": 0.6265937734360616, "learning_rate": 3.1164114032230063e-06, "loss": 0.4694, "step": 3443 }, { "epoch": 3.3053288526164186, "grad_norm": 0.7919058324108978, "learning_rate": 3.1133086639511e-06, "loss": 0.5498, "step": 3444 }, { "epoch": 3.3062890062409984, "grad_norm": 0.7897645369276705, "learning_rate": 3.1102067715947637e-06, "loss": 0.5175, "step": 3445 }, { "epoch": 3.3072491598655787, "grad_norm": 0.7520513555421898, "learning_rate": 3.10710572754641e-06, "loss": 0.5744, "step": 3446 }, { "epoch": 3.3082093134901585, "grad_norm": 0.9411272969391404, "learning_rate": 3.104005533198058e-06, "loss": 0.4893, "step": 3447 }, { "epoch": 3.3091694671147383, "grad_norm": 0.8153676499947207, "learning_rate": 3.100906189941356e-06, "loss": 0.5993, "step": 3448 }, { "epoch": 3.3101296207393185, "grad_norm": 0.9863816255241237, "learning_rate": 3.0978076991675667e-06, "loss": 0.5567, "step": 3449 }, { "epoch": 3.3110897743638983, "grad_norm": 1.1768497615205096, "learning_rate": 3.0947100622675663e-06, "loss": 0.4897, "step": 3450 }, { "epoch": 3.312049927988478, "grad_norm": 1.0624645270504616, "learning_rate": 3.091613280631853e-06, "loss": 0.5683, "step": 3451 }, { "epoch": 3.313010081613058, "grad_norm": 0.8517554304213403, "learning_rate": 3.08851735565054e-06, "loss": 0.5505, "step": 3452 }, { "epoch": 3.313970235237638, "grad_norm": 1.0831025933929703, "learning_rate": 3.0854222887133513e-06, "loss": 0.5024, "step": 3453 }, { "epoch": 3.314930388862218, "grad_norm": 0.6981090077528713, "learning_rate": 3.082328081209633e-06, "loss": 0.4713, "step": 3454 }, { "epoch": 3.3158905424867977, "grad_norm": 0.7955751025026161, "learning_rate": 3.079234734528339e-06, "loss": 0.4737, "step": 3455 }, { "epoch": 3.316850696111378, "grad_norm": 1.0406690945337183, "learning_rate": 3.07614225005804e-06, "loss": 0.5628, "step": 3456 }, { "epoch": 3.3178108497359577, "grad_norm": 1.382225111634664, "learning_rate": 3.0730506291869235e-06, "loss": 0.5435, "step": 3457 }, { "epoch": 3.3187710033605375, "grad_norm": 1.0857516588377127, "learning_rate": 3.069959873302778e-06, "loss": 0.4186, "step": 3458 }, { "epoch": 3.319731156985118, "grad_norm": 0.6749819020558921, "learning_rate": 3.0668699837930175e-06, "loss": 0.5183, "step": 3459 }, { "epoch": 3.3206913106096976, "grad_norm": 0.6842972660031886, "learning_rate": 3.0637809620446547e-06, "loss": 0.4848, "step": 3460 }, { "epoch": 3.3216514642342774, "grad_norm": 0.7675362846136607, "learning_rate": 3.0606928094443216e-06, "loss": 0.5327, "step": 3461 }, { "epoch": 3.3226116178588576, "grad_norm": 0.8134109476892541, "learning_rate": 3.0576055273782574e-06, "loss": 0.3378, "step": 3462 }, { "epoch": 3.3235717714834374, "grad_norm": 1.3362572652083549, "learning_rate": 3.0545191172323076e-06, "loss": 0.5671, "step": 3463 }, { "epoch": 3.324531925108017, "grad_norm": 0.8907410274875643, "learning_rate": 3.0514335803919314e-06, "loss": 0.5362, "step": 3464 }, { "epoch": 3.325492078732597, "grad_norm": 1.0525261207170387, "learning_rate": 3.048348918242191e-06, "loss": 0.5072, "step": 3465 }, { "epoch": 3.3264522323571772, "grad_norm": 1.0038135406142583, "learning_rate": 3.045265132167759e-06, "loss": 0.5217, "step": 3466 }, { "epoch": 3.327412385981757, "grad_norm": 0.7288398302584024, "learning_rate": 3.0421822235529154e-06, "loss": 0.5697, "step": 3467 }, { "epoch": 3.328372539606337, "grad_norm": 1.036767436838622, "learning_rate": 3.0391001937815425e-06, "loss": 0.518, "step": 3468 }, { "epoch": 3.329332693230917, "grad_norm": 1.1153879155493893, "learning_rate": 3.036019044237132e-06, "loss": 0.4592, "step": 3469 }, { "epoch": 3.330292846855497, "grad_norm": 0.9531287024683106, "learning_rate": 3.0329387763027794e-06, "loss": 0.5088, "step": 3470 }, { "epoch": 3.3312530004800767, "grad_norm": 1.017939117408441, "learning_rate": 3.029859391361183e-06, "loss": 0.4069, "step": 3471 }, { "epoch": 3.332213154104657, "grad_norm": 0.8087050732933155, "learning_rate": 3.026780890794647e-06, "loss": 0.4451, "step": 3472 }, { "epoch": 3.3331733077292367, "grad_norm": 0.6411390110678028, "learning_rate": 3.0237032759850736e-06, "loss": 0.4559, "step": 3473 }, { "epoch": 3.3341334613538165, "grad_norm": 1.0845410428197275, "learning_rate": 3.020626548313976e-06, "loss": 0.546, "step": 3474 }, { "epoch": 3.3350936149783967, "grad_norm": 1.0080623592654574, "learning_rate": 3.017550709162464e-06, "loss": 0.4985, "step": 3475 }, { "epoch": 3.3360537686029765, "grad_norm": 1.2428227173963817, "learning_rate": 3.014475759911246e-06, "loss": 0.5593, "step": 3476 }, { "epoch": 3.3370139222275563, "grad_norm": 1.0308468436838147, "learning_rate": 3.0114017019406362e-06, "loss": 0.4977, "step": 3477 }, { "epoch": 3.337974075852136, "grad_norm": 0.9884925899751822, "learning_rate": 3.0083285366305447e-06, "loss": 0.4154, "step": 3478 }, { "epoch": 3.3389342294767164, "grad_norm": 0.6966291224317721, "learning_rate": 3.0052562653604827e-06, "loss": 0.5535, "step": 3479 }, { "epoch": 3.339894383101296, "grad_norm": 0.6621998678079485, "learning_rate": 3.002184889509562e-06, "loss": 0.4889, "step": 3480 }, { "epoch": 3.340854536725876, "grad_norm": 1.0195400339897462, "learning_rate": 2.9991144104564883e-06, "loss": 0.5685, "step": 3481 }, { "epoch": 3.341814690350456, "grad_norm": 0.7657377092012025, "learning_rate": 2.9960448295795667e-06, "loss": 0.4672, "step": 3482 }, { "epoch": 3.342774843975036, "grad_norm": 0.7854520073619666, "learning_rate": 2.9929761482567017e-06, "loss": 0.4749, "step": 3483 }, { "epoch": 3.343734997599616, "grad_norm": 0.8618681156516549, "learning_rate": 2.9899083678653885e-06, "loss": 0.5076, "step": 3484 }, { "epoch": 3.344695151224196, "grad_norm": 0.8593128128683736, "learning_rate": 2.9868414897827236e-06, "loss": 0.519, "step": 3485 }, { "epoch": 3.345655304848776, "grad_norm": 0.8755637011105256, "learning_rate": 2.983775515385393e-06, "loss": 0.4475, "step": 3486 }, { "epoch": 3.3466154584733556, "grad_norm": 0.8355276262381611, "learning_rate": 2.9807104460496807e-06, "loss": 0.5804, "step": 3487 }, { "epoch": 3.347575612097936, "grad_norm": 0.8153811728599661, "learning_rate": 2.9776462831514664e-06, "loss": 0.4686, "step": 3488 }, { "epoch": 3.3485357657225157, "grad_norm": 0.942479724347764, "learning_rate": 2.9745830280662157e-06, "loss": 0.5312, "step": 3489 }, { "epoch": 3.3494959193470955, "grad_norm": 0.8850410383540434, "learning_rate": 2.971520682168996e-06, "loss": 0.575, "step": 3490 }, { "epoch": 3.3504560729716752, "grad_norm": 1.0454981152993945, "learning_rate": 2.968459246834455e-06, "loss": 0.5316, "step": 3491 }, { "epoch": 3.3514162265962555, "grad_norm": 0.944349451047761, "learning_rate": 2.9653987234368443e-06, "loss": 0.5321, "step": 3492 }, { "epoch": 3.3523763802208353, "grad_norm": 0.9774384351085383, "learning_rate": 2.9623391133499986e-06, "loss": 0.4965, "step": 3493 }, { "epoch": 3.353336533845415, "grad_norm": 1.1023229014086853, "learning_rate": 2.959280417947342e-06, "loss": 0.493, "step": 3494 }, { "epoch": 3.3542966874699953, "grad_norm": 0.8318263023054902, "learning_rate": 2.9562226386018935e-06, "loss": 0.5464, "step": 3495 }, { "epoch": 3.355256841094575, "grad_norm": 0.8516873290925602, "learning_rate": 2.953165776686254e-06, "loss": 0.5, "step": 3496 }, { "epoch": 3.356216994719155, "grad_norm": 0.9185689972438491, "learning_rate": 2.9501098335726188e-06, "loss": 0.4564, "step": 3497 }, { "epoch": 3.357177148343735, "grad_norm": 1.0978081174766652, "learning_rate": 2.947054810632768e-06, "loss": 0.5513, "step": 3498 }, { "epoch": 3.358137301968315, "grad_norm": 0.9916279443885602, "learning_rate": 2.9440007092380663e-06, "loss": 0.5173, "step": 3499 }, { "epoch": 3.3590974555928947, "grad_norm": 1.3599075995579364, "learning_rate": 2.9409475307594683e-06, "loss": 0.3948, "step": 3500 }, { "epoch": 3.360057609217475, "grad_norm": 1.244184627333581, "learning_rate": 2.937895276567516e-06, "loss": 0.5084, "step": 3501 }, { "epoch": 3.361017762842055, "grad_norm": 0.8797825761858017, "learning_rate": 2.934843948032329e-06, "loss": 0.578, "step": 3502 }, { "epoch": 3.3619779164666346, "grad_norm": 0.9759906342553212, "learning_rate": 2.931793546523622e-06, "loss": 0.4898, "step": 3503 }, { "epoch": 3.3629380700912144, "grad_norm": 0.996936696832017, "learning_rate": 2.9287440734106802e-06, "loss": 0.4458, "step": 3504 }, { "epoch": 3.3638982237157946, "grad_norm": 1.0674750972867315, "learning_rate": 2.9256955300623847e-06, "loss": 0.5034, "step": 3505 }, { "epoch": 3.3648583773403744, "grad_norm": 1.0230850194070693, "learning_rate": 2.9226479178471933e-06, "loss": 0.4903, "step": 3506 }, { "epoch": 3.365818530964954, "grad_norm": 1.0396022028888299, "learning_rate": 2.919601238133145e-06, "loss": 0.4927, "step": 3507 }, { "epoch": 3.3667786845895344, "grad_norm": 1.3956784077139481, "learning_rate": 2.916555492287865e-06, "loss": 0.54, "step": 3508 }, { "epoch": 3.3677388382141142, "grad_norm": 1.156406582305839, "learning_rate": 2.913510681678552e-06, "loss": 0.464, "step": 3509 }, { "epoch": 3.368698991838694, "grad_norm": 0.6932865225163766, "learning_rate": 2.9104668076719876e-06, "loss": 0.4671, "step": 3510 }, { "epoch": 3.3696591454632743, "grad_norm": 0.7932529596848417, "learning_rate": 2.9074238716345426e-06, "loss": 0.5806, "step": 3511 }, { "epoch": 3.370619299087854, "grad_norm": 0.7652880269376217, "learning_rate": 2.9043818749321506e-06, "loss": 0.5478, "step": 3512 }, { "epoch": 3.371579452712434, "grad_norm": 0.8917902135233862, "learning_rate": 2.9013408189303337e-06, "loss": 0.4926, "step": 3513 }, { "epoch": 3.372539606337014, "grad_norm": 0.968741820064991, "learning_rate": 2.8983007049941904e-06, "loss": 0.5487, "step": 3514 }, { "epoch": 3.373499759961594, "grad_norm": 1.1372626052146027, "learning_rate": 2.8952615344883938e-06, "loss": 0.5654, "step": 3515 }, { "epoch": 3.3744599135861737, "grad_norm": 1.2487876542555232, "learning_rate": 2.8922233087771988e-06, "loss": 0.4128, "step": 3516 }, { "epoch": 3.3754200672107535, "grad_norm": 1.0602055136562398, "learning_rate": 2.889186029224428e-06, "loss": 0.5183, "step": 3517 }, { "epoch": 3.3763802208353337, "grad_norm": 1.145508670554867, "learning_rate": 2.8861496971934843e-06, "loss": 0.605, "step": 3518 }, { "epoch": 3.3773403744599135, "grad_norm": 1.3590378677666575, "learning_rate": 2.883114314047346e-06, "loss": 0.5368, "step": 3519 }, { "epoch": 3.3783005280844933, "grad_norm": 1.031000791586425, "learning_rate": 2.880079881148564e-06, "loss": 0.5099, "step": 3520 }, { "epoch": 3.3792606817090736, "grad_norm": 0.8755915022855161, "learning_rate": 2.8770463998592645e-06, "loss": 0.537, "step": 3521 }, { "epoch": 3.3802208353336534, "grad_norm": 0.8835570132406997, "learning_rate": 2.8740138715411413e-06, "loss": 0.5351, "step": 3522 }, { "epoch": 3.381180988958233, "grad_norm": 1.1075865493525636, "learning_rate": 2.8709822975554636e-06, "loss": 0.4355, "step": 3523 }, { "epoch": 3.3821411425828134, "grad_norm": 0.9052672140792265, "learning_rate": 2.8679516792630783e-06, "loss": 0.577, "step": 3524 }, { "epoch": 3.383101296207393, "grad_norm": 0.7619684650012741, "learning_rate": 2.864922018024392e-06, "loss": 0.563, "step": 3525 }, { "epoch": 3.384061449831973, "grad_norm": 1.2108080617839274, "learning_rate": 2.8618933151993895e-06, "loss": 0.5056, "step": 3526 }, { "epoch": 3.3850216034565532, "grad_norm": 1.151626993434883, "learning_rate": 2.858865572147622e-06, "loss": 0.509, "step": 3527 }, { "epoch": 3.385981757081133, "grad_norm": 0.7510736768749485, "learning_rate": 2.8558387902282124e-06, "loss": 0.454, "step": 3528 }, { "epoch": 3.386941910705713, "grad_norm": 0.901496698534567, "learning_rate": 2.852812970799852e-06, "loss": 0.4688, "step": 3529 }, { "epoch": 3.3879020643302926, "grad_norm": 1.0592626742620839, "learning_rate": 2.849788115220794e-06, "loss": 0.4739, "step": 3530 }, { "epoch": 3.388862217954873, "grad_norm": 1.0171377605924796, "learning_rate": 2.846764224848867e-06, "loss": 0.5761, "step": 3531 }, { "epoch": 3.3898223715794527, "grad_norm": 0.7812381568003919, "learning_rate": 2.843741301041463e-06, "loss": 0.4599, "step": 3532 }, { "epoch": 3.390782525204033, "grad_norm": 0.9885125716220089, "learning_rate": 2.8407193451555398e-06, "loss": 0.5684, "step": 3533 }, { "epoch": 3.3917426788286127, "grad_norm": 0.7743876902015497, "learning_rate": 2.8376983585476226e-06, "loss": 0.4976, "step": 3534 }, { "epoch": 3.3927028324531925, "grad_norm": 1.1112793842969886, "learning_rate": 2.834678342573797e-06, "loss": 0.5602, "step": 3535 }, { "epoch": 3.3936629860777723, "grad_norm": 0.7973403654512881, "learning_rate": 2.8316592985897173e-06, "loss": 0.5008, "step": 3536 }, { "epoch": 3.3946231397023525, "grad_norm": 0.7111486092177683, "learning_rate": 2.8286412279506006e-06, "loss": 0.5474, "step": 3537 }, { "epoch": 3.3955832933269323, "grad_norm": 0.7194223288206409, "learning_rate": 2.8256241320112253e-06, "loss": 0.4018, "step": 3538 }, { "epoch": 3.396543446951512, "grad_norm": 1.0894559249605273, "learning_rate": 2.822608012125937e-06, "loss": 0.4692, "step": 3539 }, { "epoch": 3.3975036005760924, "grad_norm": 1.079182084871087, "learning_rate": 2.8195928696486345e-06, "loss": 0.5163, "step": 3540 }, { "epoch": 3.398463754200672, "grad_norm": 0.7509641994424535, "learning_rate": 2.8165787059327844e-06, "loss": 0.5329, "step": 3541 }, { "epoch": 3.399423907825252, "grad_norm": 0.7956354777330309, "learning_rate": 2.8135655223314166e-06, "loss": 0.4521, "step": 3542 }, { "epoch": 3.4003840614498317, "grad_norm": 0.7809256997555183, "learning_rate": 2.810553320197113e-06, "loss": 0.433, "step": 3543 }, { "epoch": 3.401344215074412, "grad_norm": 1.3458226612097064, "learning_rate": 2.8075421008820197e-06, "loss": 0.4118, "step": 3544 }, { "epoch": 3.4023043686989918, "grad_norm": 1.1106370623238166, "learning_rate": 2.804531865737842e-06, "loss": 0.5724, "step": 3545 }, { "epoch": 3.403264522323572, "grad_norm": 0.7764318003965627, "learning_rate": 2.8015226161158403e-06, "loss": 0.518, "step": 3546 }, { "epoch": 3.404224675948152, "grad_norm": 0.8910584477141742, "learning_rate": 2.798514353366838e-06, "loss": 0.528, "step": 3547 }, { "epoch": 3.4051848295727316, "grad_norm": 0.8553476380341921, "learning_rate": 2.7955070788412085e-06, "loss": 0.4986, "step": 3548 }, { "epoch": 3.4061449831973114, "grad_norm": 1.2040371568380739, "learning_rate": 2.7925007938888867e-06, "loss": 0.4993, "step": 3549 }, { "epoch": 3.4071051368218916, "grad_norm": 0.9106043523414967, "learning_rate": 2.789495499859361e-06, "loss": 0.5927, "step": 3550 }, { "epoch": 3.4080652904464714, "grad_norm": 1.3117938808527245, "learning_rate": 2.786491198101677e-06, "loss": 0.415, "step": 3551 }, { "epoch": 3.4090254440710512, "grad_norm": 1.0392440198605104, "learning_rate": 2.783487889964435e-06, "loss": 0.5212, "step": 3552 }, { "epoch": 3.4099855976956315, "grad_norm": 1.0774810545157363, "learning_rate": 2.7804855767957833e-06, "loss": 0.4808, "step": 3553 }, { "epoch": 3.4109457513202113, "grad_norm": 0.9928465527482729, "learning_rate": 2.7774842599434294e-06, "loss": 0.5564, "step": 3554 }, { "epoch": 3.411905904944791, "grad_norm": 1.0120035286883737, "learning_rate": 2.7744839407546374e-06, "loss": 0.4662, "step": 3555 }, { "epoch": 3.412866058569371, "grad_norm": 0.8581524451552237, "learning_rate": 2.7714846205762134e-06, "loss": 0.573, "step": 3556 }, { "epoch": 3.413826212193951, "grad_norm": 1.0875723236262445, "learning_rate": 2.7684863007545227e-06, "loss": 0.4806, "step": 3557 }, { "epoch": 3.414786365818531, "grad_norm": 1.4235965780549735, "learning_rate": 2.765488982635477e-06, "loss": 0.5257, "step": 3558 }, { "epoch": 3.415746519443111, "grad_norm": 0.7178392925556015, "learning_rate": 2.7624926675645415e-06, "loss": 0.4448, "step": 3559 }, { "epoch": 3.416706673067691, "grad_norm": 1.0196208133885427, "learning_rate": 2.759497356886731e-06, "loss": 0.4947, "step": 3560 }, { "epoch": 3.4176668266922707, "grad_norm": 0.900867016944538, "learning_rate": 2.7565030519466063e-06, "loss": 0.4781, "step": 3561 }, { "epoch": 3.4186269803168505, "grad_norm": 0.807933225207921, "learning_rate": 2.75350975408828e-06, "loss": 0.4503, "step": 3562 }, { "epoch": 3.4195871339414308, "grad_norm": 0.8611822175313192, "learning_rate": 2.7505174646554112e-06, "loss": 0.4667, "step": 3563 }, { "epoch": 3.4205472875660106, "grad_norm": 0.8230677016549365, "learning_rate": 2.7475261849912073e-06, "loss": 0.4239, "step": 3564 }, { "epoch": 3.4215074411905904, "grad_norm": 1.039180896150858, "learning_rate": 2.744535916438423e-06, "loss": 0.52, "step": 3565 }, { "epoch": 3.4224675948151706, "grad_norm": 0.6492514695505134, "learning_rate": 2.7415466603393538e-06, "loss": 0.4927, "step": 3566 }, { "epoch": 3.4234277484397504, "grad_norm": 0.8154229150778218, "learning_rate": 2.738558418035846e-06, "loss": 0.5541, "step": 3567 }, { "epoch": 3.42438790206433, "grad_norm": 1.2559131737840719, "learning_rate": 2.7355711908692942e-06, "loss": 0.522, "step": 3568 }, { "epoch": 3.42534805568891, "grad_norm": 1.4595941884504655, "learning_rate": 2.732584980180628e-06, "loss": 0.4585, "step": 3569 }, { "epoch": 3.4263082093134902, "grad_norm": 1.1880346746428818, "learning_rate": 2.7295997873103286e-06, "loss": 0.4181, "step": 3570 }, { "epoch": 3.42726836293807, "grad_norm": 1.0715235739442563, "learning_rate": 2.726615613598411e-06, "loss": 0.5803, "step": 3571 }, { "epoch": 3.4282285165626503, "grad_norm": 1.417639073085241, "learning_rate": 2.7236324603844467e-06, "loss": 0.3797, "step": 3572 }, { "epoch": 3.42918867018723, "grad_norm": 0.9893285559407268, "learning_rate": 2.7206503290075388e-06, "loss": 0.5229, "step": 3573 }, { "epoch": 3.43014882381181, "grad_norm": 0.9026182823245618, "learning_rate": 2.7176692208063317e-06, "loss": 0.4698, "step": 3574 }, { "epoch": 3.4311089774363897, "grad_norm": 0.9578987648964752, "learning_rate": 2.7146891371190144e-06, "loss": 0.5764, "step": 3575 }, { "epoch": 3.43206913106097, "grad_norm": 0.7648480371731745, "learning_rate": 2.7117100792833144e-06, "loss": 0.466, "step": 3576 }, { "epoch": 3.4330292846855497, "grad_norm": 0.8442896620627242, "learning_rate": 2.7087320486364997e-06, "loss": 0.5981, "step": 3577 }, { "epoch": 3.4339894383101295, "grad_norm": 1.3178415461329567, "learning_rate": 2.7057550465153782e-06, "loss": 0.4852, "step": 3578 }, { "epoch": 3.4349495919347097, "grad_norm": 0.778270193219872, "learning_rate": 2.7027790742562905e-06, "loss": 0.5019, "step": 3579 }, { "epoch": 3.4359097455592895, "grad_norm": 1.091035496873672, "learning_rate": 2.6998041331951196e-06, "loss": 0.5403, "step": 3580 }, { "epoch": 3.4368698991838693, "grad_norm": 0.8329196071192416, "learning_rate": 2.6968302246672863e-06, "loss": 0.6159, "step": 3581 }, { "epoch": 3.437830052808449, "grad_norm": 0.9916598927307053, "learning_rate": 2.693857350007746e-06, "loss": 0.5161, "step": 3582 }, { "epoch": 3.4387902064330294, "grad_norm": 0.9868640342076743, "learning_rate": 2.6908855105509912e-06, "loss": 0.4851, "step": 3583 }, { "epoch": 3.439750360057609, "grad_norm": 0.8628128835413127, "learning_rate": 2.6879147076310457e-06, "loss": 0.4751, "step": 3584 }, { "epoch": 3.4407105136821894, "grad_norm": 0.9098646833431265, "learning_rate": 2.6849449425814713e-06, "loss": 0.5026, "step": 3585 }, { "epoch": 3.441670667306769, "grad_norm": 0.7706998342570753, "learning_rate": 2.6819762167353695e-06, "loss": 0.441, "step": 3586 }, { "epoch": 3.442630820931349, "grad_norm": 0.9935070330364282, "learning_rate": 2.6790085314253645e-06, "loss": 0.5139, "step": 3587 }, { "epoch": 3.4435909745559288, "grad_norm": 1.1505520653539714, "learning_rate": 2.6760418879836186e-06, "loss": 0.495, "step": 3588 }, { "epoch": 3.444551128180509, "grad_norm": 0.7625936029016338, "learning_rate": 2.673076287741828e-06, "loss": 0.5162, "step": 3589 }, { "epoch": 3.445511281805089, "grad_norm": 0.9717662583757145, "learning_rate": 2.670111732031218e-06, "loss": 0.4883, "step": 3590 }, { "epoch": 3.4464714354296686, "grad_norm": 1.1343852786090238, "learning_rate": 2.6671482221825484e-06, "loss": 0.5222, "step": 3591 }, { "epoch": 3.447431589054249, "grad_norm": 1.0595071865149555, "learning_rate": 2.6641857595261035e-06, "loss": 0.491, "step": 3592 }, { "epoch": 3.4483917426788286, "grad_norm": 1.1216393350880205, "learning_rate": 2.661224345391703e-06, "loss": 0.5161, "step": 3593 }, { "epoch": 3.4493518963034084, "grad_norm": 0.8639941414400532, "learning_rate": 2.658263981108693e-06, "loss": 0.4212, "step": 3594 }, { "epoch": 3.4503120499279882, "grad_norm": 0.8701928403516643, "learning_rate": 2.65530466800595e-06, "loss": 0.6313, "step": 3595 }, { "epoch": 3.4512722035525685, "grad_norm": 0.971995363874264, "learning_rate": 2.652346407411881e-06, "loss": 0.4352, "step": 3596 }, { "epoch": 3.4522323571771483, "grad_norm": 1.2290235728022298, "learning_rate": 2.649389200654412e-06, "loss": 0.4801, "step": 3597 }, { "epoch": 3.4531925108017285, "grad_norm": 1.508467802717194, "learning_rate": 2.6464330490610026e-06, "loss": 0.5659, "step": 3598 }, { "epoch": 3.4541526644263083, "grad_norm": 0.8750199120176935, "learning_rate": 2.6434779539586432e-06, "loss": 0.5649, "step": 3599 }, { "epoch": 3.455112818050888, "grad_norm": 0.9893273620929268, "learning_rate": 2.640523916673838e-06, "loss": 0.5263, "step": 3600 }, { "epoch": 3.456072971675468, "grad_norm": 0.6892126018588575, "learning_rate": 2.6375709385326255e-06, "loss": 0.535, "step": 3601 }, { "epoch": 3.457033125300048, "grad_norm": 0.6141912197733196, "learning_rate": 2.6346190208605653e-06, "loss": 0.6198, "step": 3602 }, { "epoch": 3.457993278924628, "grad_norm": 1.0496508173291021, "learning_rate": 2.6316681649827423e-06, "loss": 0.4664, "step": 3603 }, { "epoch": 3.4589534325492077, "grad_norm": 0.7384136042253114, "learning_rate": 2.6287183722237663e-06, "loss": 0.5426, "step": 3604 }, { "epoch": 3.459913586173788, "grad_norm": 1.0470361874567828, "learning_rate": 2.625769643907763e-06, "loss": 0.4697, "step": 3605 }, { "epoch": 3.4608737397983678, "grad_norm": 0.8432154294475551, "learning_rate": 2.622821981358388e-06, "loss": 0.4878, "step": 3606 }, { "epoch": 3.4618338934229476, "grad_norm": 0.8693601168006075, "learning_rate": 2.6198753858988147e-06, "loss": 0.4154, "step": 3607 }, { "epoch": 3.4627940470475274, "grad_norm": 0.7455264384873933, "learning_rate": 2.616929858851739e-06, "loss": 0.5334, "step": 3608 }, { "epoch": 3.4637542006721076, "grad_norm": 0.8218758111196783, "learning_rate": 2.613985401539377e-06, "loss": 0.4773, "step": 3609 }, { "epoch": 3.4647143542966874, "grad_norm": 1.0771470120003843, "learning_rate": 2.6110420152834614e-06, "loss": 0.4915, "step": 3610 }, { "epoch": 3.4656745079212676, "grad_norm": 0.9278610958040304, "learning_rate": 2.6080997014052474e-06, "loss": 0.5116, "step": 3611 }, { "epoch": 3.4666346615458474, "grad_norm": 0.7344754365909257, "learning_rate": 2.6051584612255132e-06, "loss": 0.5239, "step": 3612 }, { "epoch": 3.4675948151704272, "grad_norm": 0.9448904935086336, "learning_rate": 2.6022182960645436e-06, "loss": 0.5503, "step": 3613 }, { "epoch": 3.468554968795007, "grad_norm": 1.041937210054607, "learning_rate": 2.5992792072421526e-06, "loss": 0.4292, "step": 3614 }, { "epoch": 3.4695151224195873, "grad_norm": 1.0530206856510955, "learning_rate": 2.596341196077659e-06, "loss": 0.4572, "step": 3615 }, { "epoch": 3.470475276044167, "grad_norm": 1.5146464227474865, "learning_rate": 2.593404263889911e-06, "loss": 0.4056, "step": 3616 }, { "epoch": 3.471435429668747, "grad_norm": 1.1231037225994904, "learning_rate": 2.5904684119972656e-06, "loss": 0.4553, "step": 3617 }, { "epoch": 3.472395583293327, "grad_norm": 1.0593505022416516, "learning_rate": 2.587533641717592e-06, "loss": 0.4492, "step": 3618 }, { "epoch": 3.473355736917907, "grad_norm": 0.8350651690124183, "learning_rate": 2.5845999543682787e-06, "loss": 0.5121, "step": 3619 }, { "epoch": 3.4743158905424867, "grad_norm": 0.7129866176230729, "learning_rate": 2.581667351266226e-06, "loss": 0.5214, "step": 3620 }, { "epoch": 3.475276044167067, "grad_norm": 0.7777167617458534, "learning_rate": 2.5787358337278493e-06, "loss": 0.4144, "step": 3621 }, { "epoch": 3.4762361977916467, "grad_norm": 0.8406045706919425, "learning_rate": 2.575805403069077e-06, "loss": 0.4232, "step": 3622 }, { "epoch": 3.4771963514162265, "grad_norm": 1.0160726536553775, "learning_rate": 2.5728760606053443e-06, "loss": 0.5367, "step": 3623 }, { "epoch": 3.4781565050408068, "grad_norm": 0.8856408990612038, "learning_rate": 2.5699478076516034e-06, "loss": 0.5727, "step": 3624 }, { "epoch": 3.4791166586653866, "grad_norm": 0.9913320577955356, "learning_rate": 2.5670206455223156e-06, "loss": 0.5419, "step": 3625 }, { "epoch": 3.4800768122899663, "grad_norm": 0.9123026183031779, "learning_rate": 2.5640945755314527e-06, "loss": 0.5146, "step": 3626 }, { "epoch": 3.481036965914546, "grad_norm": 0.8849964021207584, "learning_rate": 2.5611695989924988e-06, "loss": 0.4932, "step": 3627 }, { "epoch": 3.4819971195391264, "grad_norm": 0.689747094200367, "learning_rate": 2.5582457172184405e-06, "loss": 0.5203, "step": 3628 }, { "epoch": 3.482957273163706, "grad_norm": 0.6967335435341256, "learning_rate": 2.5553229315217763e-06, "loss": 0.4446, "step": 3629 }, { "epoch": 3.483917426788286, "grad_norm": 0.9159231226350923, "learning_rate": 2.5524012432145203e-06, "loss": 0.4788, "step": 3630 }, { "epoch": 3.484877580412866, "grad_norm": 1.0239819727477197, "learning_rate": 2.5494806536081806e-06, "loss": 0.5514, "step": 3631 }, { "epoch": 3.485837734037446, "grad_norm": 0.902106341127842, "learning_rate": 2.5465611640137813e-06, "loss": 0.521, "step": 3632 }, { "epoch": 3.486797887662026, "grad_norm": 0.7454098545270077, "learning_rate": 2.543642775741849e-06, "loss": 0.4572, "step": 3633 }, { "epoch": 3.487758041286606, "grad_norm": 0.9004628005323445, "learning_rate": 2.540725490102418e-06, "loss": 0.5425, "step": 3634 }, { "epoch": 3.488718194911186, "grad_norm": 1.0908006837090785, "learning_rate": 2.5378093084050287e-06, "loss": 0.5085, "step": 3635 }, { "epoch": 3.4896783485357656, "grad_norm": 0.8328010009194754, "learning_rate": 2.534894231958719e-06, "loss": 0.4096, "step": 3636 }, { "epoch": 3.490638502160346, "grad_norm": 0.9624294824939009, "learning_rate": 2.531980262072038e-06, "loss": 0.4773, "step": 3637 }, { "epoch": 3.4915986557849257, "grad_norm": 0.6436986758334201, "learning_rate": 2.5290674000530365e-06, "loss": 0.4587, "step": 3638 }, { "epoch": 3.4925588094095055, "grad_norm": 0.7733959065362295, "learning_rate": 2.526155647209265e-06, "loss": 0.5535, "step": 3639 }, { "epoch": 3.4935189630340853, "grad_norm": 0.9306355909174091, "learning_rate": 2.523245004847783e-06, "loss": 0.4378, "step": 3640 }, { "epoch": 3.4944791166586655, "grad_norm": 0.7812828168500796, "learning_rate": 2.520335474275141e-06, "loss": 0.5462, "step": 3641 }, { "epoch": 3.4954392702832453, "grad_norm": 1.0872805304984285, "learning_rate": 2.5174270567973957e-06, "loss": 0.4038, "step": 3642 }, { "epoch": 3.496399423907825, "grad_norm": 0.9864805587951005, "learning_rate": 2.5145197537201116e-06, "loss": 0.4622, "step": 3643 }, { "epoch": 3.4973595775324053, "grad_norm": 0.6764914883267406, "learning_rate": 2.51161356634834e-06, "loss": 0.5598, "step": 3644 }, { "epoch": 3.498319731156985, "grad_norm": 0.7944647745093998, "learning_rate": 2.5087084959866403e-06, "loss": 0.5757, "step": 3645 }, { "epoch": 3.499279884781565, "grad_norm": 0.6812743136484498, "learning_rate": 2.5058045439390634e-06, "loss": 0.4988, "step": 3646 }, { "epoch": 3.5002400384061447, "grad_norm": 1.1060098284758169, "learning_rate": 2.502901711509167e-06, "loss": 0.6064, "step": 3647 }, { "epoch": 3.501200192030725, "grad_norm": 0.9074839624038273, "learning_rate": 2.5000000000000015e-06, "loss": 0.4859, "step": 3648 }, { "epoch": 3.5021603456553048, "grad_norm": 0.9652468482667867, "learning_rate": 2.4970994107141106e-06, "loss": 0.5406, "step": 3649 }, { "epoch": 3.503120499279885, "grad_norm": 0.8923347793736934, "learning_rate": 2.4941999449535393e-06, "loss": 0.5718, "step": 3650 }, { "epoch": 3.504080652904465, "grad_norm": 0.9762099077351241, "learning_rate": 2.491301604019828e-06, "loss": 0.4983, "step": 3651 }, { "epoch": 3.5050408065290446, "grad_norm": 0.7002006590708644, "learning_rate": 2.4884043892140096e-06, "loss": 0.4942, "step": 3652 }, { "epoch": 3.5060009601536244, "grad_norm": 0.8617668250580495, "learning_rate": 2.4855083018366154e-06, "loss": 0.5054, "step": 3653 }, { "epoch": 3.5069611137782046, "grad_norm": 0.7799155404348764, "learning_rate": 2.4826133431876644e-06, "loss": 0.5674, "step": 3654 }, { "epoch": 3.5079212674027844, "grad_norm": 0.908704288105357, "learning_rate": 2.479719514566672e-06, "loss": 0.5713, "step": 3655 }, { "epoch": 3.5088814210273642, "grad_norm": 0.7989600729751257, "learning_rate": 2.4768268172726535e-06, "loss": 0.5449, "step": 3656 }, { "epoch": 3.5098415746519445, "grad_norm": 0.7050516871603852, "learning_rate": 2.473935252604103e-06, "loss": 0.4468, "step": 3657 }, { "epoch": 3.5108017282765243, "grad_norm": 0.7126605873316985, "learning_rate": 2.4710448218590176e-06, "loss": 0.5061, "step": 3658 }, { "epoch": 3.511761881901104, "grad_norm": 0.773016706380144, "learning_rate": 2.468155526334875e-06, "loss": 0.5469, "step": 3659 }, { "epoch": 3.512722035525684, "grad_norm": 0.883093739894347, "learning_rate": 2.465267367328655e-06, "loss": 0.5183, "step": 3660 }, { "epoch": 3.513682189150264, "grad_norm": 0.7528279507027883, "learning_rate": 2.46238034613682e-06, "loss": 0.419, "step": 3661 }, { "epoch": 3.514642342774844, "grad_norm": 0.7578895784275321, "learning_rate": 2.4594944640553207e-06, "loss": 0.5727, "step": 3662 }, { "epoch": 3.515602496399424, "grad_norm": 0.8592879096507339, "learning_rate": 2.4566097223795997e-06, "loss": 0.4087, "step": 3663 }, { "epoch": 3.516562650024004, "grad_norm": 0.617879765210984, "learning_rate": 2.453726122404587e-06, "loss": 0.5206, "step": 3664 }, { "epoch": 3.5175228036485837, "grad_norm": 0.798163685865454, "learning_rate": 2.4508436654247004e-06, "loss": 0.5142, "step": 3665 }, { "epoch": 3.5184829572731635, "grad_norm": 1.1139977130598706, "learning_rate": 2.4479623527338447e-06, "loss": 0.491, "step": 3666 }, { "epoch": 3.5194431108977438, "grad_norm": 0.900770251804858, "learning_rate": 2.4450821856254074e-06, "loss": 0.4961, "step": 3667 }, { "epoch": 3.5204032645223235, "grad_norm": 1.409815649767775, "learning_rate": 2.442203165392266e-06, "loss": 0.4676, "step": 3668 }, { "epoch": 3.5213634181469033, "grad_norm": 1.261496633422323, "learning_rate": 2.439325293326783e-06, "loss": 0.4748, "step": 3669 }, { "epoch": 3.5223235717714836, "grad_norm": 1.2482969121177256, "learning_rate": 2.4364485707208037e-06, "loss": 0.4588, "step": 3670 }, { "epoch": 3.5232837253960634, "grad_norm": 1.2951981015687588, "learning_rate": 2.433572998865659e-06, "loss": 0.5721, "step": 3671 }, { "epoch": 3.524243879020643, "grad_norm": 0.8160937742552223, "learning_rate": 2.430698579052161e-06, "loss": 0.5446, "step": 3672 }, { "epoch": 3.525204032645223, "grad_norm": 1.1317421924827453, "learning_rate": 2.4278253125706044e-06, "loss": 0.5336, "step": 3673 }, { "epoch": 3.526164186269803, "grad_norm": 1.0019032697406478, "learning_rate": 2.424953200710773e-06, "loss": 0.4722, "step": 3674 }, { "epoch": 3.527124339894383, "grad_norm": 1.3388066012195, "learning_rate": 2.4220822447619223e-06, "loss": 0.4809, "step": 3675 }, { "epoch": 3.5280844935189632, "grad_norm": 1.2932050435410463, "learning_rate": 2.419212446012796e-06, "loss": 0.5218, "step": 3676 }, { "epoch": 3.529044647143543, "grad_norm": 1.6177063892883323, "learning_rate": 2.416343805751615e-06, "loss": 0.5601, "step": 3677 }, { "epoch": 3.530004800768123, "grad_norm": 0.8945761413912526, "learning_rate": 2.4134763252660825e-06, "loss": 0.4916, "step": 3678 }, { "epoch": 3.5309649543927026, "grad_norm": 1.014337697043889, "learning_rate": 2.4106100058433814e-06, "loss": 0.5035, "step": 3679 }, { "epoch": 3.531925108017283, "grad_norm": 0.9683609451892855, "learning_rate": 2.4077448487701667e-06, "loss": 0.5077, "step": 3680 }, { "epoch": 3.5328852616418627, "grad_norm": 0.7111854618895629, "learning_rate": 2.4048808553325804e-06, "loss": 0.4937, "step": 3681 }, { "epoch": 3.5338454152664425, "grad_norm": 0.9955333119551943, "learning_rate": 2.4020180268162374e-06, "loss": 0.4639, "step": 3682 }, { "epoch": 3.5348055688910227, "grad_norm": 0.8566022955468208, "learning_rate": 2.3991563645062317e-06, "loss": 0.5673, "step": 3683 }, { "epoch": 3.5357657225156025, "grad_norm": 0.7970159278045398, "learning_rate": 2.3962958696871343e-06, "loss": 0.4185, "step": 3684 }, { "epoch": 3.5367258761401823, "grad_norm": 0.9323417380368659, "learning_rate": 2.393436543642987e-06, "loss": 0.4746, "step": 3685 }, { "epoch": 3.537686029764762, "grad_norm": 0.9205863448228657, "learning_rate": 2.3905783876573108e-06, "loss": 0.5107, "step": 3686 }, { "epoch": 3.5386461833893423, "grad_norm": 0.7852506625698427, "learning_rate": 2.387721403013107e-06, "loss": 0.5268, "step": 3687 }, { "epoch": 3.539606337013922, "grad_norm": 1.008756863375548, "learning_rate": 2.3848655909928392e-06, "loss": 0.5201, "step": 3688 }, { "epoch": 3.5405664906385024, "grad_norm": 1.028152013676738, "learning_rate": 2.382010952878456e-06, "loss": 0.5147, "step": 3689 }, { "epoch": 3.541526644263082, "grad_norm": 1.3104151215510214, "learning_rate": 2.379157489951367e-06, "loss": 0.6212, "step": 3690 }, { "epoch": 3.542486797887662, "grad_norm": 0.9148023421992518, "learning_rate": 2.376305203492468e-06, "loss": 0.4979, "step": 3691 }, { "epoch": 3.5434469515122418, "grad_norm": 0.8206206360033533, "learning_rate": 2.373454094782119e-06, "loss": 0.4723, "step": 3692 }, { "epoch": 3.544407105136822, "grad_norm": 0.8765067913127856, "learning_rate": 2.370604165100148e-06, "loss": 0.5224, "step": 3693 }, { "epoch": 3.545367258761402, "grad_norm": 0.9536286075049744, "learning_rate": 2.367755415725862e-06, "loss": 0.5817, "step": 3694 }, { "epoch": 3.5463274123859816, "grad_norm": 0.8224181906518976, "learning_rate": 2.3649078479380316e-06, "loss": 0.4967, "step": 3695 }, { "epoch": 3.547287566010562, "grad_norm": 0.7753939457857306, "learning_rate": 2.3620614630149013e-06, "loss": 0.4175, "step": 3696 }, { "epoch": 3.5482477196351416, "grad_norm": 1.3529614089198578, "learning_rate": 2.359216262234184e-06, "loss": 0.518, "step": 3697 }, { "epoch": 3.5492078732597214, "grad_norm": 1.066165957305857, "learning_rate": 2.3563722468730567e-06, "loss": 0.4864, "step": 3698 }, { "epoch": 3.550168026884301, "grad_norm": 0.8783087012782463, "learning_rate": 2.3535294182081666e-06, "loss": 0.4779, "step": 3699 }, { "epoch": 3.5511281805088815, "grad_norm": 0.6469575325240258, "learning_rate": 2.350687777515635e-06, "loss": 0.5305, "step": 3700 }, { "epoch": 3.5520883341334613, "grad_norm": 0.877069205950521, "learning_rate": 2.3478473260710387e-06, "loss": 0.5801, "step": 3701 }, { "epoch": 3.5530484877580415, "grad_norm": 1.4787867761290385, "learning_rate": 2.34500806514943e-06, "loss": 0.5167, "step": 3702 }, { "epoch": 3.5540086413826213, "grad_norm": 0.8085516348687933, "learning_rate": 2.3421699960253168e-06, "loss": 0.4817, "step": 3703 }, { "epoch": 3.554968795007201, "grad_norm": 1.0214107316359369, "learning_rate": 2.3393331199726837e-06, "loss": 0.4512, "step": 3704 }, { "epoch": 3.555928948631781, "grad_norm": 0.9533582821718921, "learning_rate": 2.3364974382649737e-06, "loss": 0.4822, "step": 3705 }, { "epoch": 3.556889102256361, "grad_norm": 1.3137794566769483, "learning_rate": 2.3336629521750902e-06, "loss": 0.5385, "step": 3706 }, { "epoch": 3.557849255880941, "grad_norm": 0.957937269032781, "learning_rate": 2.3308296629754067e-06, "loss": 0.4315, "step": 3707 }, { "epoch": 3.558809409505521, "grad_norm": 1.0931829554507413, "learning_rate": 2.327997571937755e-06, "loss": 0.4302, "step": 3708 }, { "epoch": 3.559769563130101, "grad_norm": 0.9907254955611745, "learning_rate": 2.3251666803334316e-06, "loss": 0.4865, "step": 3709 }, { "epoch": 3.5607297167546808, "grad_norm": 0.878512742280391, "learning_rate": 2.3223369894331947e-06, "loss": 0.4491, "step": 3710 }, { "epoch": 3.5616898703792605, "grad_norm": 1.1533792464091386, "learning_rate": 2.3195085005072587e-06, "loss": 0.5231, "step": 3711 }, { "epoch": 3.5626500240038403, "grad_norm": 1.1686786863672762, "learning_rate": 2.3166812148253045e-06, "loss": 0.6011, "step": 3712 }, { "epoch": 3.5636101776284206, "grad_norm": 1.0741460831058993, "learning_rate": 2.313855133656469e-06, "loss": 0.5104, "step": 3713 }, { "epoch": 3.5645703312530004, "grad_norm": 1.0095302657995349, "learning_rate": 2.3110302582693512e-06, "loss": 0.4159, "step": 3714 }, { "epoch": 3.5655304848775806, "grad_norm": 1.0711850960895741, "learning_rate": 2.3082065899320084e-06, "loss": 0.5349, "step": 3715 }, { "epoch": 3.5664906385021604, "grad_norm": 1.1190888987201555, "learning_rate": 2.3053841299119524e-06, "loss": 0.4736, "step": 3716 }, { "epoch": 3.56745079212674, "grad_norm": 0.7695387691904497, "learning_rate": 2.3025628794761544e-06, "loss": 0.4733, "step": 3717 }, { "epoch": 3.56841094575132, "grad_norm": 0.8684309687144539, "learning_rate": 2.2997428398910497e-06, "loss": 0.4659, "step": 3718 }, { "epoch": 3.5693710993759002, "grad_norm": 0.6609736281285014, "learning_rate": 2.2969240124225177e-06, "loss": 0.536, "step": 3719 }, { "epoch": 3.57033125300048, "grad_norm": 0.8672199146659839, "learning_rate": 2.294106398335905e-06, "loss": 0.4329, "step": 3720 }, { "epoch": 3.5712914066250603, "grad_norm": 0.9865372155337937, "learning_rate": 2.2912899988960015e-06, "loss": 0.4646, "step": 3721 }, { "epoch": 3.57225156024964, "grad_norm": 0.901721731271915, "learning_rate": 2.288474815367065e-06, "loss": 0.566, "step": 3722 }, { "epoch": 3.57321171387422, "grad_norm": 0.780424035571715, "learning_rate": 2.2856608490128005e-06, "loss": 0.5697, "step": 3723 }, { "epoch": 3.5741718674987997, "grad_norm": 0.8547075737669523, "learning_rate": 2.2828481010963647e-06, "loss": 0.5232, "step": 3724 }, { "epoch": 3.5751320211233795, "grad_norm": 0.8947170318636488, "learning_rate": 2.280036572880372e-06, "loss": 0.5574, "step": 3725 }, { "epoch": 3.5760921747479597, "grad_norm": 0.6107457636580168, "learning_rate": 2.277226265626887e-06, "loss": 0.5087, "step": 3726 }, { "epoch": 3.5770523283725395, "grad_norm": 0.8289299460491248, "learning_rate": 2.2744171805974257e-06, "loss": 0.4813, "step": 3727 }, { "epoch": 3.5780124819971197, "grad_norm": 0.9132043611995266, "learning_rate": 2.27160931905296e-06, "loss": 0.4788, "step": 3728 }, { "epoch": 3.5789726356216995, "grad_norm": 0.6893891267203847, "learning_rate": 2.268802682253903e-06, "loss": 0.5518, "step": 3729 }, { "epoch": 3.5799327892462793, "grad_norm": 1.018150754606302, "learning_rate": 2.2659972714601264e-06, "loss": 0.5355, "step": 3730 }, { "epoch": 3.580892942870859, "grad_norm": 0.830731792957442, "learning_rate": 2.2631930879309525e-06, "loss": 0.4086, "step": 3731 }, { "epoch": 3.5818530964954394, "grad_norm": 0.9155610244797766, "learning_rate": 2.2603901329251452e-06, "loss": 0.5318, "step": 3732 }, { "epoch": 3.582813250120019, "grad_norm": 0.979185381702044, "learning_rate": 2.2575884077009246e-06, "loss": 0.5434, "step": 3733 }, { "epoch": 3.5837734037445994, "grad_norm": 0.8383099186202485, "learning_rate": 2.2547879135159495e-06, "loss": 0.5063, "step": 3734 }, { "epoch": 3.584733557369179, "grad_norm": 0.7106125244574798, "learning_rate": 2.2519886516273365e-06, "loss": 0.5408, "step": 3735 }, { "epoch": 3.585693710993759, "grad_norm": 1.2162147355165096, "learning_rate": 2.2491906232916457e-06, "loss": 0.5197, "step": 3736 }, { "epoch": 3.586653864618339, "grad_norm": 1.3036271488106448, "learning_rate": 2.246393829764877e-06, "loss": 0.5141, "step": 3737 }, { "epoch": 3.5876140182429186, "grad_norm": 0.8695194744188787, "learning_rate": 2.243598272302484e-06, "loss": 0.4879, "step": 3738 }, { "epoch": 3.588574171867499, "grad_norm": 0.7213683341482486, "learning_rate": 2.240803952159362e-06, "loss": 0.423, "step": 3739 }, { "epoch": 3.5895343254920786, "grad_norm": 0.5807931168413029, "learning_rate": 2.238010870589852e-06, "loss": 0.4888, "step": 3740 }, { "epoch": 3.590494479116659, "grad_norm": 0.9560901338897689, "learning_rate": 2.23521902884774e-06, "loss": 0.5141, "step": 3741 }, { "epoch": 3.5914546327412387, "grad_norm": 1.1504776418479836, "learning_rate": 2.23242842818625e-06, "loss": 0.5516, "step": 3742 }, { "epoch": 3.5924147863658185, "grad_norm": 0.9778711920644858, "learning_rate": 2.2296390698580558e-06, "loss": 0.5525, "step": 3743 }, { "epoch": 3.5933749399903983, "grad_norm": 1.464156655818805, "learning_rate": 2.2268509551152684e-06, "loss": 0.5231, "step": 3744 }, { "epoch": 3.5943350936149785, "grad_norm": 0.7525524673319295, "learning_rate": 2.2240640852094446e-06, "loss": 0.437, "step": 3745 }, { "epoch": 3.5952952472395583, "grad_norm": 1.0224500424555252, "learning_rate": 2.2212784613915816e-06, "loss": 0.4485, "step": 3746 }, { "epoch": 3.5962554008641385, "grad_norm": 0.9273918196642318, "learning_rate": 2.2184940849121123e-06, "loss": 0.4683, "step": 3747 }, { "epoch": 3.5972155544887183, "grad_norm": 0.9072454360300336, "learning_rate": 2.215710957020914e-06, "loss": 0.4124, "step": 3748 }, { "epoch": 3.598175708113298, "grad_norm": 0.9470507298281676, "learning_rate": 2.2129290789673085e-06, "loss": 0.4593, "step": 3749 }, { "epoch": 3.599135861737878, "grad_norm": 0.798450220157891, "learning_rate": 2.2101484520000456e-06, "loss": 0.4498, "step": 3750 }, { "epoch": 3.600096015362458, "grad_norm": 1.0016275792379439, "learning_rate": 2.2073690773673203e-06, "loss": 0.5035, "step": 3751 }, { "epoch": 3.601056168987038, "grad_norm": 0.901017543987506, "learning_rate": 2.2045909563167643e-06, "loss": 0.503, "step": 3752 }, { "epoch": 3.6020163226116177, "grad_norm": 1.1823314404052347, "learning_rate": 2.201814090095448e-06, "loss": 0.4665, "step": 3753 }, { "epoch": 3.602976476236198, "grad_norm": 0.9648450612998336, "learning_rate": 2.1990384799498767e-06, "loss": 0.5059, "step": 3754 }, { "epoch": 3.603936629860778, "grad_norm": 1.3306008898284105, "learning_rate": 2.196264127125989e-06, "loss": 0.4515, "step": 3755 }, { "epoch": 3.6048967834853576, "grad_norm": 0.9395112280774136, "learning_rate": 2.193491032869165e-06, "loss": 0.5894, "step": 3756 }, { "epoch": 3.6058569371099374, "grad_norm": 0.8780686117171458, "learning_rate": 2.190719198424216e-06, "loss": 0.5008, "step": 3757 }, { "epoch": 3.6068170907345176, "grad_norm": 0.8711103282192286, "learning_rate": 2.1879486250353894e-06, "loss": 0.458, "step": 3758 }, { "epoch": 3.6077772443590974, "grad_norm": 0.8614401188558233, "learning_rate": 2.185179313946368e-06, "loss": 0.4313, "step": 3759 }, { "epoch": 3.6087373979836777, "grad_norm": 0.9741804819859328, "learning_rate": 2.1824112664002612e-06, "loss": 0.394, "step": 3760 }, { "epoch": 3.6096975516082574, "grad_norm": 0.9423221546404649, "learning_rate": 2.1796444836396174e-06, "loss": 0.6365, "step": 3761 }, { "epoch": 3.6106577052328372, "grad_norm": 0.805973002121136, "learning_rate": 2.1768789669064205e-06, "loss": 0.4509, "step": 3762 }, { "epoch": 3.611617858857417, "grad_norm": 1.2499606164471577, "learning_rate": 2.174114717442076e-06, "loss": 0.5414, "step": 3763 }, { "epoch": 3.6125780124819973, "grad_norm": 1.4562105546714994, "learning_rate": 2.1713517364874297e-06, "loss": 0.5176, "step": 3764 }, { "epoch": 3.613538166106577, "grad_norm": 0.8131633260907718, "learning_rate": 2.1685900252827485e-06, "loss": 0.4824, "step": 3765 }, { "epoch": 3.614498319731157, "grad_norm": 0.8489572479868623, "learning_rate": 2.1658295850677408e-06, "loss": 0.5433, "step": 3766 }, { "epoch": 3.615458473355737, "grad_norm": 0.8721484251098021, "learning_rate": 2.1630704170815386e-06, "loss": 0.4442, "step": 3767 }, { "epoch": 3.616418626980317, "grad_norm": 0.6595444647226734, "learning_rate": 2.160312522562699e-06, "loss": 0.4943, "step": 3768 }, { "epoch": 3.6173787806048967, "grad_norm": 0.9798962359003458, "learning_rate": 2.1575559027492136e-06, "loss": 0.6011, "step": 3769 }, { "epoch": 3.6183389342294765, "grad_norm": 0.7811120340196949, "learning_rate": 2.1548005588785e-06, "loss": 0.4943, "step": 3770 }, { "epoch": 3.6192990878540567, "grad_norm": 1.0550202007293008, "learning_rate": 2.1520464921874014e-06, "loss": 0.5333, "step": 3771 }, { "epoch": 3.6202592414786365, "grad_norm": 0.991551035940957, "learning_rate": 2.1492937039121914e-06, "loss": 0.5387, "step": 3772 }, { "epoch": 3.6212193951032168, "grad_norm": 1.1161852178044573, "learning_rate": 2.1465421952885633e-06, "loss": 0.5564, "step": 3773 }, { "epoch": 3.6221795487277966, "grad_norm": 1.0309754492529266, "learning_rate": 2.1437919675516407e-06, "loss": 0.4409, "step": 3774 }, { "epoch": 3.6231397023523764, "grad_norm": 0.7667455714304143, "learning_rate": 2.1410430219359757e-06, "loss": 0.5596, "step": 3775 }, { "epoch": 3.624099855976956, "grad_norm": 0.8852559839582709, "learning_rate": 2.1382953596755364e-06, "loss": 0.5501, "step": 3776 }, { "epoch": 3.6250600096015364, "grad_norm": 0.9559269273259989, "learning_rate": 2.1355489820037216e-06, "loss": 0.4009, "step": 3777 }, { "epoch": 3.626020163226116, "grad_norm": 0.9038416840940787, "learning_rate": 2.132803890153346e-06, "loss": 0.4465, "step": 3778 }, { "epoch": 3.626980316850696, "grad_norm": 0.8222066808922783, "learning_rate": 2.1300600853566577e-06, "loss": 0.5137, "step": 3779 }, { "epoch": 3.6279404704752762, "grad_norm": 0.8310140223407402, "learning_rate": 2.12731756884532e-06, "loss": 0.5063, "step": 3780 }, { "epoch": 3.628900624099856, "grad_norm": 0.7528367268921751, "learning_rate": 2.1245763418504173e-06, "loss": 0.5513, "step": 3781 }, { "epoch": 3.629860777724436, "grad_norm": 1.062719611154465, "learning_rate": 2.121836405602458e-06, "loss": 0.4485, "step": 3782 }, { "epoch": 3.6308209313490156, "grad_norm": 1.3632728395176892, "learning_rate": 2.1190977613313707e-06, "loss": 0.4565, "step": 3783 }, { "epoch": 3.631781084973596, "grad_norm": 0.6536388050841562, "learning_rate": 2.1163604102665023e-06, "loss": 0.5619, "step": 3784 }, { "epoch": 3.6327412385981757, "grad_norm": 1.1296770829778275, "learning_rate": 2.113624353636624e-06, "loss": 0.4428, "step": 3785 }, { "epoch": 3.633701392222756, "grad_norm": 1.0302442537947734, "learning_rate": 2.110889592669918e-06, "loss": 0.523, "step": 3786 }, { "epoch": 3.6346615458473357, "grad_norm": 0.7655550036916589, "learning_rate": 2.1081561285939904e-06, "loss": 0.4798, "step": 3787 }, { "epoch": 3.6356216994719155, "grad_norm": 0.9101464793161821, "learning_rate": 2.1054239626358648e-06, "loss": 0.5075, "step": 3788 }, { "epoch": 3.6365818530964953, "grad_norm": 0.9954972107964059, "learning_rate": 2.102693096021981e-06, "loss": 0.509, "step": 3789 }, { "epoch": 3.6375420067210755, "grad_norm": 0.7259179852016796, "learning_rate": 2.0999635299781976e-06, "loss": 0.516, "step": 3790 }, { "epoch": 3.6385021603456553, "grad_norm": 1.179427894096622, "learning_rate": 2.0972352657297845e-06, "loss": 0.5218, "step": 3791 }, { "epoch": 3.639462313970235, "grad_norm": 1.248664872902119, "learning_rate": 2.0945083045014296e-06, "loss": 0.5384, "step": 3792 }, { "epoch": 3.6404224675948154, "grad_norm": 1.1929398294203277, "learning_rate": 2.0917826475172427e-06, "loss": 0.4761, "step": 3793 }, { "epoch": 3.641382621219395, "grad_norm": 0.9232940185422632, "learning_rate": 2.089058296000736e-06, "loss": 0.5139, "step": 3794 }, { "epoch": 3.642342774843975, "grad_norm": 0.8667381531825574, "learning_rate": 2.0863352511748466e-06, "loss": 0.473, "step": 3795 }, { "epoch": 3.6433029284685547, "grad_norm": 0.8534675626451674, "learning_rate": 2.083613514261913e-06, "loss": 0.487, "step": 3796 }, { "epoch": 3.644263082093135, "grad_norm": 0.8345733262433228, "learning_rate": 2.0808930864837007e-06, "loss": 0.5973, "step": 3797 }, { "epoch": 3.645223235717715, "grad_norm": 0.805530079912269, "learning_rate": 2.07817396906138e-06, "loss": 0.436, "step": 3798 }, { "epoch": 3.646183389342295, "grad_norm": 0.6107398624319403, "learning_rate": 2.0754561632155307e-06, "loss": 0.502, "step": 3799 }, { "epoch": 3.647143542966875, "grad_norm": 0.8786037674228829, "learning_rate": 2.072739670166148e-06, "loss": 0.4799, "step": 3800 }, { "epoch": 3.6481036965914546, "grad_norm": 0.8123072021884278, "learning_rate": 2.070024491132636e-06, "loss": 0.4484, "step": 3801 }, { "epoch": 3.6490638502160344, "grad_norm": 0.9471214913446105, "learning_rate": 2.067310627333811e-06, "loss": 0.4869, "step": 3802 }, { "epoch": 3.6500240038406147, "grad_norm": 0.8057018735134795, "learning_rate": 2.064598079987898e-06, "loss": 0.5093, "step": 3803 }, { "epoch": 3.6509841574651944, "grad_norm": 0.8939067811830095, "learning_rate": 2.0618868503125273e-06, "loss": 0.5291, "step": 3804 }, { "epoch": 3.6519443110897742, "grad_norm": 0.8458819485900274, "learning_rate": 2.059176939524741e-06, "loss": 0.5033, "step": 3805 }, { "epoch": 3.6529044647143545, "grad_norm": 0.9676388137263293, "learning_rate": 2.056468348840995e-06, "loss": 0.4391, "step": 3806 }, { "epoch": 3.6538646183389343, "grad_norm": 1.1273483955687906, "learning_rate": 2.0537610794771405e-06, "loss": 0.4386, "step": 3807 }, { "epoch": 3.654824771963514, "grad_norm": 0.7122482833711168, "learning_rate": 2.0510551326484457e-06, "loss": 0.4556, "step": 3808 }, { "epoch": 3.655784925588094, "grad_norm": 0.712547569101207, "learning_rate": 2.0483505095695755e-06, "loss": 0.5216, "step": 3809 }, { "epoch": 3.656745079212674, "grad_norm": 1.0046379890523085, "learning_rate": 2.045647211454613e-06, "loss": 0.4316, "step": 3810 }, { "epoch": 3.657705232837254, "grad_norm": 0.950836245180027, "learning_rate": 2.0429452395170372e-06, "loss": 0.4876, "step": 3811 }, { "epoch": 3.658665386461834, "grad_norm": 1.1570824532671464, "learning_rate": 2.0402445949697335e-06, "loss": 0.5038, "step": 3812 }, { "epoch": 3.659625540086414, "grad_norm": 0.6613174703207048, "learning_rate": 2.037545279024993e-06, "loss": 0.4423, "step": 3813 }, { "epoch": 3.6605856937109937, "grad_norm": 0.8766045273380032, "learning_rate": 2.0348472928945096e-06, "loss": 0.4714, "step": 3814 }, { "epoch": 3.6615458473355735, "grad_norm": 1.4925790651703073, "learning_rate": 2.0321506377893807e-06, "loss": 0.5042, "step": 3815 }, { "epoch": 3.6625060009601538, "grad_norm": 0.6824936782566504, "learning_rate": 2.029455314920108e-06, "loss": 0.6062, "step": 3816 }, { "epoch": 3.6634661545847336, "grad_norm": 1.0284473079522356, "learning_rate": 2.026761325496589e-06, "loss": 0.4608, "step": 3817 }, { "epoch": 3.6644263082093134, "grad_norm": 0.8737621751976463, "learning_rate": 2.024068670728127e-06, "loss": 0.5263, "step": 3818 }, { "epoch": 3.6653864618338936, "grad_norm": 1.5239970531564417, "learning_rate": 2.021377351823432e-06, "loss": 0.5406, "step": 3819 }, { "epoch": 3.6663466154584734, "grad_norm": 1.0913985823587649, "learning_rate": 2.0186873699906017e-06, "loss": 0.5371, "step": 3820 }, { "epoch": 3.667306769083053, "grad_norm": 1.0147215246392702, "learning_rate": 2.0159987264371445e-06, "loss": 0.4678, "step": 3821 }, { "epoch": 3.668266922707633, "grad_norm": 1.0151047528757589, "learning_rate": 2.0133114223699577e-06, "loss": 0.6083, "step": 3822 }, { "epoch": 3.6692270763322132, "grad_norm": 0.845800758576364, "learning_rate": 2.01062545899535e-06, "loss": 0.5399, "step": 3823 }, { "epoch": 3.670187229956793, "grad_norm": 0.967899782689877, "learning_rate": 2.007940837519021e-06, "loss": 0.4805, "step": 3824 }, { "epoch": 3.6711473835813733, "grad_norm": 0.9386487105334532, "learning_rate": 2.0052575591460636e-06, "loss": 0.4358, "step": 3825 }, { "epoch": 3.672107537205953, "grad_norm": 1.0202454054141634, "learning_rate": 2.002575625080977e-06, "loss": 0.4952, "step": 3826 }, { "epoch": 3.673067690830533, "grad_norm": 1.1641621376561073, "learning_rate": 1.9998950365276514e-06, "loss": 0.4552, "step": 3827 }, { "epoch": 3.6740278444551127, "grad_norm": 1.0978379865411254, "learning_rate": 1.997215794689375e-06, "loss": 0.5546, "step": 3828 }, { "epoch": 3.674987998079693, "grad_norm": 0.932590756860919, "learning_rate": 1.9945379007688314e-06, "loss": 0.5267, "step": 3829 }, { "epoch": 3.6759481517042727, "grad_norm": 0.8526061366785054, "learning_rate": 1.991861355968096e-06, "loss": 0.6071, "step": 3830 }, { "epoch": 3.6769083053288525, "grad_norm": 1.0929562709618696, "learning_rate": 1.989186161488643e-06, "loss": 0.4805, "step": 3831 }, { "epoch": 3.6778684589534327, "grad_norm": 1.1469668880248638, "learning_rate": 1.986512318531338e-06, "loss": 0.6377, "step": 3832 }, { "epoch": 3.6788286125780125, "grad_norm": 1.1237868431720337, "learning_rate": 1.9838398282964412e-06, "loss": 0.5392, "step": 3833 }, { "epoch": 3.6797887662025923, "grad_norm": 0.9300839237467747, "learning_rate": 1.9811686919836054e-06, "loss": 0.4478, "step": 3834 }, { "epoch": 3.680748919827172, "grad_norm": 0.6087317909633315, "learning_rate": 1.9784989107918736e-06, "loss": 0.394, "step": 3835 }, { "epoch": 3.6817090734517524, "grad_norm": 1.101411198472666, "learning_rate": 1.97583048591968e-06, "loss": 0.541, "step": 3836 }, { "epoch": 3.682669227076332, "grad_norm": 0.9706455459812657, "learning_rate": 1.9731634185648584e-06, "loss": 0.525, "step": 3837 }, { "epoch": 3.6836293807009124, "grad_norm": 0.9414819617980742, "learning_rate": 1.970497709924622e-06, "loss": 0.4135, "step": 3838 }, { "epoch": 3.684589534325492, "grad_norm": 0.7255741839945409, "learning_rate": 1.9678333611955814e-06, "loss": 0.4791, "step": 3839 }, { "epoch": 3.685549687950072, "grad_norm": 0.8860209982012034, "learning_rate": 1.965170373573729e-06, "loss": 0.4218, "step": 3840 }, { "epoch": 3.686509841574652, "grad_norm": 0.8692154623193221, "learning_rate": 1.9625087482544574e-06, "loss": 0.4708, "step": 3841 }, { "epoch": 3.687469995199232, "grad_norm": 1.2376377492676094, "learning_rate": 1.959848486432542e-06, "loss": 0.5476, "step": 3842 }, { "epoch": 3.688430148823812, "grad_norm": 1.3472429612018666, "learning_rate": 1.957189589302141e-06, "loss": 0.4415, "step": 3843 }, { "epoch": 3.6893903024483916, "grad_norm": 0.7692676212641436, "learning_rate": 1.9545320580568077e-06, "loss": 0.4939, "step": 3844 }, { "epoch": 3.690350456072972, "grad_norm": 0.8563444787390867, "learning_rate": 1.95187589388948e-06, "loss": 0.4737, "step": 3845 }, { "epoch": 3.6913106096975516, "grad_norm": 1.0764521864969325, "learning_rate": 1.9492210979924805e-06, "loss": 0.4698, "step": 3846 }, { "epoch": 3.6922707633221314, "grad_norm": 0.6970163596057671, "learning_rate": 1.9465676715575203e-06, "loss": 0.4836, "step": 3847 }, { "epoch": 3.6932309169467112, "grad_norm": 1.2314538868751899, "learning_rate": 1.9439156157756917e-06, "loss": 0.5437, "step": 3848 }, { "epoch": 3.6941910705712915, "grad_norm": 1.264148736715841, "learning_rate": 1.9412649318374737e-06, "loss": 0.546, "step": 3849 }, { "epoch": 3.6951512241958713, "grad_norm": 0.8813221056715124, "learning_rate": 1.938615620932735e-06, "loss": 0.5143, "step": 3850 }, { "epoch": 3.6961113778204515, "grad_norm": 1.1141665013164042, "learning_rate": 1.9359676842507186e-06, "loss": 0.4644, "step": 3851 }, { "epoch": 3.6970715314450313, "grad_norm": 1.1797905016064496, "learning_rate": 1.933321122980057e-06, "loss": 0.449, "step": 3852 }, { "epoch": 3.698031685069611, "grad_norm": 1.3065796207868745, "learning_rate": 1.9306759383087593e-06, "loss": 0.528, "step": 3853 }, { "epoch": 3.698991838694191, "grad_norm": 1.7185454362374641, "learning_rate": 1.9280321314242252e-06, "loss": 0.58, "step": 3854 }, { "epoch": 3.699951992318771, "grad_norm": 1.1134536733412375, "learning_rate": 1.925389703513232e-06, "loss": 0.5191, "step": 3855 }, { "epoch": 3.700912145943351, "grad_norm": 1.1633486935032888, "learning_rate": 1.922748655761932e-06, "loss": 0.5189, "step": 3856 }, { "epoch": 3.7018722995679307, "grad_norm": 0.8582859646626665, "learning_rate": 1.9201089893558677e-06, "loss": 0.4855, "step": 3857 }, { "epoch": 3.702832453192511, "grad_norm": 0.731351529480996, "learning_rate": 1.9174707054799558e-06, "loss": 0.5739, "step": 3858 }, { "epoch": 3.7037926068170908, "grad_norm": 0.8134139106347049, "learning_rate": 1.9148338053184933e-06, "loss": 0.4813, "step": 3859 }, { "epoch": 3.7047527604416706, "grad_norm": 0.637630133961132, "learning_rate": 1.9121982900551592e-06, "loss": 0.4339, "step": 3860 }, { "epoch": 3.7057129140662504, "grad_norm": 1.0357617590857955, "learning_rate": 1.909564160873004e-06, "loss": 0.5206, "step": 3861 }, { "epoch": 3.7066730676908306, "grad_norm": 1.0096643134205505, "learning_rate": 1.9069314189544602e-06, "loss": 0.4045, "step": 3862 }, { "epoch": 3.7076332213154104, "grad_norm": 1.3071061323390936, "learning_rate": 1.9043000654813433e-06, "loss": 0.5748, "step": 3863 }, { "epoch": 3.7085933749399906, "grad_norm": 1.094882988225478, "learning_rate": 1.9016701016348338e-06, "loss": 0.5549, "step": 3864 }, { "epoch": 3.7095535285645704, "grad_norm": 0.6559291473282551, "learning_rate": 1.8990415285954976e-06, "loss": 0.5199, "step": 3865 }, { "epoch": 3.7105136821891502, "grad_norm": 1.0312713812796972, "learning_rate": 1.896414347543269e-06, "loss": 0.4786, "step": 3866 }, { "epoch": 3.71147383581373, "grad_norm": 1.0414066272525149, "learning_rate": 1.8937885596574656e-06, "loss": 0.4341, "step": 3867 }, { "epoch": 3.7124339894383103, "grad_norm": 1.0667203361160325, "learning_rate": 1.8911641661167762e-06, "loss": 0.6122, "step": 3868 }, { "epoch": 3.71339414306289, "grad_norm": 1.1482650893895865, "learning_rate": 1.8885411680992594e-06, "loss": 0.4497, "step": 3869 }, { "epoch": 3.71435429668747, "grad_norm": 0.8724719277244144, "learning_rate": 1.885919566782352e-06, "loss": 0.4657, "step": 3870 }, { "epoch": 3.71531445031205, "grad_norm": 0.809829396675353, "learning_rate": 1.8832993633428643e-06, "loss": 0.5203, "step": 3871 }, { "epoch": 3.71627460393663, "grad_norm": 0.9702728925598892, "learning_rate": 1.880680558956976e-06, "loss": 0.5098, "step": 3872 }, { "epoch": 3.7172347575612097, "grad_norm": 0.8285262441356138, "learning_rate": 1.878063154800243e-06, "loss": 0.4145, "step": 3873 }, { "epoch": 3.7181949111857895, "grad_norm": 0.9467256728734245, "learning_rate": 1.8754471520475863e-06, "loss": 0.4971, "step": 3874 }, { "epoch": 3.7191550648103697, "grad_norm": 0.9973521186383556, "learning_rate": 1.8728325518733038e-06, "loss": 0.4067, "step": 3875 }, { "epoch": 3.7201152184349495, "grad_norm": 1.016429426767654, "learning_rate": 1.870219355451061e-06, "loss": 0.5132, "step": 3876 }, { "epoch": 3.7210753720595298, "grad_norm": 1.020397908460958, "learning_rate": 1.8676075639538943e-06, "loss": 0.6001, "step": 3877 }, { "epoch": 3.7220355256841096, "grad_norm": 0.6491872289841643, "learning_rate": 1.8649971785542097e-06, "loss": 0.5289, "step": 3878 }, { "epoch": 3.7229956793086894, "grad_norm": 1.0699825324453371, "learning_rate": 1.8623882004237792e-06, "loss": 0.5445, "step": 3879 }, { "epoch": 3.723955832933269, "grad_norm": 0.981831627525532, "learning_rate": 1.8597806307337446e-06, "loss": 0.4924, "step": 3880 }, { "epoch": 3.7249159865578494, "grad_norm": 0.9983605623850992, "learning_rate": 1.8571744706546213e-06, "loss": 0.4744, "step": 3881 }, { "epoch": 3.725876140182429, "grad_norm": 1.1539714534861603, "learning_rate": 1.8545697213562813e-06, "loss": 0.5826, "step": 3882 }, { "epoch": 3.726836293807009, "grad_norm": 0.773477716322822, "learning_rate": 1.8519663840079721e-06, "loss": 0.4513, "step": 3883 }, { "epoch": 3.727796447431589, "grad_norm": 0.9835728359140223, "learning_rate": 1.8493644597782996e-06, "loss": 0.5289, "step": 3884 }, { "epoch": 3.728756601056169, "grad_norm": 0.7732844998622604, "learning_rate": 1.8467639498352442e-06, "loss": 0.4628, "step": 3885 }, { "epoch": 3.729716754680749, "grad_norm": 0.7495765721830818, "learning_rate": 1.8441648553461467e-06, "loss": 0.4002, "step": 3886 }, { "epoch": 3.7306769083053286, "grad_norm": 0.7110130214644875, "learning_rate": 1.8415671774777106e-06, "loss": 0.5651, "step": 3887 }, { "epoch": 3.731637061929909, "grad_norm": 1.0195003179982542, "learning_rate": 1.8389709173960063e-06, "loss": 0.5396, "step": 3888 }, { "epoch": 3.7325972155544886, "grad_norm": 0.970203642172964, "learning_rate": 1.8363760762664678e-06, "loss": 0.3781, "step": 3889 }, { "epoch": 3.733557369179069, "grad_norm": 0.8865914883413661, "learning_rate": 1.8337826552538911e-06, "loss": 0.493, "step": 3890 }, { "epoch": 3.7345175228036487, "grad_norm": 0.9568494040724864, "learning_rate": 1.8311906555224379e-06, "loss": 0.5049, "step": 3891 }, { "epoch": 3.7354776764282285, "grad_norm": 1.405480404671445, "learning_rate": 1.8286000782356245e-06, "loss": 0.4933, "step": 3892 }, { "epoch": 3.7364378300528083, "grad_norm": 1.332862118494347, "learning_rate": 1.826010924556334e-06, "loss": 0.4763, "step": 3893 }, { "epoch": 3.7373979836773885, "grad_norm": 0.9435805405768178, "learning_rate": 1.823423195646815e-06, "loss": 0.6508, "step": 3894 }, { "epoch": 3.7383581373019683, "grad_norm": 0.8841096919558995, "learning_rate": 1.820836892668666e-06, "loss": 0.5122, "step": 3895 }, { "epoch": 3.739318290926548, "grad_norm": 0.8303054673903346, "learning_rate": 1.8182520167828537e-06, "loss": 0.5562, "step": 3896 }, { "epoch": 3.7402784445511283, "grad_norm": 0.7661430423343341, "learning_rate": 1.8156685691496972e-06, "loss": 0.4198, "step": 3897 }, { "epoch": 3.741238598175708, "grad_norm": 0.9340304214360069, "learning_rate": 1.813086550928883e-06, "loss": 0.5237, "step": 3898 }, { "epoch": 3.742198751800288, "grad_norm": 0.789098506927442, "learning_rate": 1.8105059632794514e-06, "loss": 0.4712, "step": 3899 }, { "epoch": 3.7431589054248677, "grad_norm": 0.9476968674777795, "learning_rate": 1.8079268073597978e-06, "loss": 0.4833, "step": 3900 }, { "epoch": 3.744119059049448, "grad_norm": 0.7771674522527977, "learning_rate": 1.8053490843276788e-06, "loss": 0.584, "step": 3901 }, { "epoch": 3.7450792126740278, "grad_norm": 1.1922760552741227, "learning_rate": 1.802772795340207e-06, "loss": 0.4981, "step": 3902 }, { "epoch": 3.746039366298608, "grad_norm": 0.9225242611520136, "learning_rate": 1.8001979415538507e-06, "loss": 0.5374, "step": 3903 }, { "epoch": 3.746999519923188, "grad_norm": 1.210895039226324, "learning_rate": 1.797624524124436e-06, "loss": 0.5565, "step": 3904 }, { "epoch": 3.7479596735477676, "grad_norm": 0.8293825449658596, "learning_rate": 1.7950525442071394e-06, "loss": 0.4715, "step": 3905 }, { "epoch": 3.7489198271723474, "grad_norm": 0.9150719166673643, "learning_rate": 1.792482002956495e-06, "loss": 0.6145, "step": 3906 }, { "epoch": 3.7498799807969276, "grad_norm": 0.8435478877454587, "learning_rate": 1.7899129015263966e-06, "loss": 0.5219, "step": 3907 }, { "epoch": 3.7508401344215074, "grad_norm": 1.2239904494244633, "learning_rate": 1.7873452410700809e-06, "loss": 0.4548, "step": 3908 }, { "epoch": 3.7518002880460872, "grad_norm": 1.4776534255346432, "learning_rate": 1.7847790227401469e-06, "loss": 0.4298, "step": 3909 }, { "epoch": 3.7527604416706675, "grad_norm": 1.5371849396012824, "learning_rate": 1.7822142476885374e-06, "loss": 0.4574, "step": 3910 }, { "epoch": 3.7537205952952473, "grad_norm": 1.3696055417818767, "learning_rate": 1.7796509170665571e-06, "loss": 0.4419, "step": 3911 }, { "epoch": 3.754680748919827, "grad_norm": 0.9947768907015963, "learning_rate": 1.777089032024859e-06, "loss": 0.4961, "step": 3912 }, { "epoch": 3.755640902544407, "grad_norm": 1.2208406991900391, "learning_rate": 1.774528593713441e-06, "loss": 0.5168, "step": 3913 }, { "epoch": 3.756601056168987, "grad_norm": 0.9590784994678724, "learning_rate": 1.771969603281659e-06, "loss": 0.46, "step": 3914 }, { "epoch": 3.757561209793567, "grad_norm": 1.0144858966026968, "learning_rate": 1.7694120618782169e-06, "loss": 0.5545, "step": 3915 }, { "epoch": 3.758521363418147, "grad_norm": 1.0714777016327064, "learning_rate": 1.7668559706511667e-06, "loss": 0.5044, "step": 3916 }, { "epoch": 3.759481517042727, "grad_norm": 1.0249707744643457, "learning_rate": 1.7643013307479129e-06, "loss": 0.575, "step": 3917 }, { "epoch": 3.7604416706673067, "grad_norm": 0.8975668341057544, "learning_rate": 1.7617481433152022e-06, "loss": 0.455, "step": 3918 }, { "epoch": 3.7614018242918865, "grad_norm": 0.8442498635538498, "learning_rate": 1.759196409499135e-06, "loss": 0.5656, "step": 3919 }, { "epoch": 3.7623619779164668, "grad_norm": 0.9402170543477671, "learning_rate": 1.756646130445157e-06, "loss": 0.4787, "step": 3920 }, { "epoch": 3.7633221315410466, "grad_norm": 0.792867325595836, "learning_rate": 1.7540973072980621e-06, "loss": 0.4743, "step": 3921 }, { "epoch": 3.7642822851656264, "grad_norm": 0.74014746416629, "learning_rate": 1.7515499412019903e-06, "loss": 0.4715, "step": 3922 }, { "epoch": 3.7652424387902066, "grad_norm": 0.7661059624290066, "learning_rate": 1.7490040333004248e-06, "loss": 0.4644, "step": 3923 }, { "epoch": 3.7662025924147864, "grad_norm": 0.7095609286610052, "learning_rate": 1.7464595847361954e-06, "loss": 0.4752, "step": 3924 }, { "epoch": 3.767162746039366, "grad_norm": 0.8377510859907462, "learning_rate": 1.7439165966514831e-06, "loss": 0.5284, "step": 3925 }, { "epoch": 3.768122899663946, "grad_norm": 1.3596345274413744, "learning_rate": 1.741375070187804e-06, "loss": 0.5248, "step": 3926 }, { "epoch": 3.769083053288526, "grad_norm": 0.9590166115638962, "learning_rate": 1.7388350064860242e-06, "loss": 0.5599, "step": 3927 }, { "epoch": 3.770043206913106, "grad_norm": 1.123474703938837, "learning_rate": 1.7362964066863469e-06, "loss": 0.4809, "step": 3928 }, { "epoch": 3.7710033605376863, "grad_norm": 0.9792094682755929, "learning_rate": 1.7337592719283276e-06, "loss": 0.3935, "step": 3929 }, { "epoch": 3.771963514162266, "grad_norm": 0.8467937923166816, "learning_rate": 1.7312236033508584e-06, "loss": 0.461, "step": 3930 }, { "epoch": 3.772923667786846, "grad_norm": 0.7629751276325175, "learning_rate": 1.7286894020921707e-06, "loss": 0.4526, "step": 3931 }, { "epoch": 3.7738838214114256, "grad_norm": 1.0893133864748819, "learning_rate": 1.726156669289842e-06, "loss": 0.4785, "step": 3932 }, { "epoch": 3.774843975036006, "grad_norm": 0.8169990841230694, "learning_rate": 1.7236254060807889e-06, "loss": 0.4976, "step": 3933 }, { "epoch": 3.7758041286605857, "grad_norm": 1.3315746034540517, "learning_rate": 1.7210956136012684e-06, "loss": 0.5507, "step": 3934 }, { "epoch": 3.7767642822851655, "grad_norm": 1.1531061527213524, "learning_rate": 1.718567292986879e-06, "loss": 0.5359, "step": 3935 }, { "epoch": 3.7777244359097457, "grad_norm": 1.069209587836764, "learning_rate": 1.7160404453725543e-06, "loss": 0.564, "step": 3936 }, { "epoch": 3.7786845895343255, "grad_norm": 0.877787838423237, "learning_rate": 1.7135150718925674e-06, "loss": 0.4798, "step": 3937 }, { "epoch": 3.7796447431589053, "grad_norm": 0.980446226617164, "learning_rate": 1.7109911736805385e-06, "loss": 0.4235, "step": 3938 }, { "epoch": 3.780604896783485, "grad_norm": 0.8332725010112277, "learning_rate": 1.708468751869412e-06, "loss": 0.5777, "step": 3939 }, { "epoch": 3.7815650504080653, "grad_norm": 0.839455576568235, "learning_rate": 1.7059478075914799e-06, "loss": 0.3816, "step": 3940 }, { "epoch": 3.782525204032645, "grad_norm": 0.8308335203093796, "learning_rate": 1.703428341978362e-06, "loss": 0.3922, "step": 3941 }, { "epoch": 3.7834853576572254, "grad_norm": 0.8720602154683017, "learning_rate": 1.7009103561610252e-06, "loss": 0.5259, "step": 3942 }, { "epoch": 3.784445511281805, "grad_norm": 0.8396319264789237, "learning_rate": 1.6983938512697655e-06, "loss": 0.5187, "step": 3943 }, { "epoch": 3.785405664906385, "grad_norm": 1.2153076777193155, "learning_rate": 1.695878828434212e-06, "loss": 0.4491, "step": 3944 }, { "epoch": 3.7863658185309648, "grad_norm": 0.9017508959055832, "learning_rate": 1.6933652887833336e-06, "loss": 0.5266, "step": 3945 }, { "epoch": 3.787325972155545, "grad_norm": 1.1217420762898962, "learning_rate": 1.6908532334454308e-06, "loss": 0.5397, "step": 3946 }, { "epoch": 3.788286125780125, "grad_norm": 1.075595805235117, "learning_rate": 1.6883426635481392e-06, "loss": 0.4826, "step": 3947 }, { "epoch": 3.7892462794047046, "grad_norm": 1.7298604084620675, "learning_rate": 1.6858335802184278e-06, "loss": 0.4751, "step": 3948 }, { "epoch": 3.790206433029285, "grad_norm": 0.8920390130387915, "learning_rate": 1.683325984582594e-06, "loss": 0.6113, "step": 3949 }, { "epoch": 3.7911665866538646, "grad_norm": 1.0016530194467204, "learning_rate": 1.680819877766272e-06, "loss": 0.546, "step": 3950 }, { "epoch": 3.7921267402784444, "grad_norm": 0.8973720059119742, "learning_rate": 1.678315260894427e-06, "loss": 0.4363, "step": 3951 }, { "epoch": 3.7930868939030242, "grad_norm": 1.2173803865656694, "learning_rate": 1.6758121350913548e-06, "loss": 0.5671, "step": 3952 }, { "epoch": 3.7940470475276045, "grad_norm": 1.0248503977017867, "learning_rate": 1.6733105014806834e-06, "loss": 0.5448, "step": 3953 }, { "epoch": 3.7950072011521843, "grad_norm": 0.8371532250587892, "learning_rate": 1.6708103611853638e-06, "loss": 0.5431, "step": 3954 }, { "epoch": 3.7959673547767645, "grad_norm": 0.92653153525797, "learning_rate": 1.6683117153276878e-06, "loss": 0.5096, "step": 3955 }, { "epoch": 3.7969275084013443, "grad_norm": 0.9489627747914922, "learning_rate": 1.6658145650292707e-06, "loss": 0.5014, "step": 3956 }, { "epoch": 3.797887662025924, "grad_norm": 0.7115458462812435, "learning_rate": 1.6633189114110531e-06, "loss": 0.5041, "step": 3957 }, { "epoch": 3.798847815650504, "grad_norm": 0.6883585750815618, "learning_rate": 1.6608247555933093e-06, "loss": 0.4055, "step": 3958 }, { "epoch": 3.799807969275084, "grad_norm": 0.6792706130981496, "learning_rate": 1.6583320986956386e-06, "loss": 0.5148, "step": 3959 }, { "epoch": 3.800768122899664, "grad_norm": 0.8082689878482988, "learning_rate": 1.6558409418369686e-06, "loss": 0.5072, "step": 3960 }, { "epoch": 3.8017282765242437, "grad_norm": 0.8345913934236219, "learning_rate": 1.6533512861355543e-06, "loss": 0.54, "step": 3961 }, { "epoch": 3.802688430148824, "grad_norm": 0.8308422417419711, "learning_rate": 1.6508631327089725e-06, "loss": 0.5066, "step": 3962 }, { "epoch": 3.8036485837734038, "grad_norm": 0.962840113531976, "learning_rate": 1.64837648267413e-06, "loss": 0.5085, "step": 3963 }, { "epoch": 3.8046087373979836, "grad_norm": 1.3599082059850294, "learning_rate": 1.6458913371472573e-06, "loss": 0.5418, "step": 3964 }, { "epoch": 3.8055688910225633, "grad_norm": 1.0137791162202294, "learning_rate": 1.6434076972439106e-06, "loss": 0.404, "step": 3965 }, { "epoch": 3.8065290446471436, "grad_norm": 0.7261659668116449, "learning_rate": 1.6409255640789695e-06, "loss": 0.5355, "step": 3966 }, { "epoch": 3.8074891982717234, "grad_norm": 0.8509204828594858, "learning_rate": 1.6384449387666351e-06, "loss": 0.5768, "step": 3967 }, { "epoch": 3.8084493518963036, "grad_norm": 0.8492677879618364, "learning_rate": 1.6359658224204328e-06, "loss": 0.5317, "step": 3968 }, { "epoch": 3.8094095055208834, "grad_norm": 0.7868860112515507, "learning_rate": 1.6334882161532161e-06, "loss": 0.4166, "step": 3969 }, { "epoch": 3.810369659145463, "grad_norm": 0.6317808329102411, "learning_rate": 1.6310121210771524e-06, "loss": 0.4819, "step": 3970 }, { "epoch": 3.811329812770043, "grad_norm": 0.8887292768292944, "learning_rate": 1.628537538303736e-06, "loss": 0.4521, "step": 3971 }, { "epoch": 3.8122899663946233, "grad_norm": 0.9314952014622512, "learning_rate": 1.6260644689437767e-06, "loss": 0.4272, "step": 3972 }, { "epoch": 3.813250120019203, "grad_norm": 1.0309132340900338, "learning_rate": 1.6235929141074137e-06, "loss": 0.5188, "step": 3973 }, { "epoch": 3.814210273643783, "grad_norm": 1.1245915534611188, "learning_rate": 1.6211228749041007e-06, "loss": 0.4374, "step": 3974 }, { "epoch": 3.815170427268363, "grad_norm": 0.8338821105644686, "learning_rate": 1.6186543524426101e-06, "loss": 0.5085, "step": 3975 }, { "epoch": 3.816130580892943, "grad_norm": 1.060829973576904, "learning_rate": 1.616187347831036e-06, "loss": 0.4844, "step": 3976 }, { "epoch": 3.8170907345175227, "grad_norm": 1.6254162759465194, "learning_rate": 1.6137218621767902e-06, "loss": 0.4298, "step": 3977 }, { "epoch": 3.8180508881421025, "grad_norm": 0.9594296394454627, "learning_rate": 1.6112578965866039e-06, "loss": 0.3622, "step": 3978 }, { "epoch": 3.8190110417666827, "grad_norm": 0.9037026211071861, "learning_rate": 1.608795452166525e-06, "loss": 0.4604, "step": 3979 }, { "epoch": 3.8199711953912625, "grad_norm": 1.0826126897222599, "learning_rate": 1.6063345300219162e-06, "loss": 0.5239, "step": 3980 }, { "epoch": 3.8209313490158427, "grad_norm": 1.062514252631848, "learning_rate": 1.6038751312574592e-06, "loss": 0.5156, "step": 3981 }, { "epoch": 3.8218915026404225, "grad_norm": 1.0637719499829976, "learning_rate": 1.6014172569771564e-06, "loss": 0.4564, "step": 3982 }, { "epoch": 3.8228516562650023, "grad_norm": 0.9524200333710778, "learning_rate": 1.598960908284316e-06, "loss": 0.4701, "step": 3983 }, { "epoch": 3.823811809889582, "grad_norm": 0.7196953339599067, "learning_rate": 1.5965060862815713e-06, "loss": 0.4703, "step": 3984 }, { "epoch": 3.8247719635141624, "grad_norm": 0.8034908878456822, "learning_rate": 1.5940527920708594e-06, "loss": 0.4628, "step": 3985 }, { "epoch": 3.825732117138742, "grad_norm": 0.8228327104772886, "learning_rate": 1.5916010267534433e-06, "loss": 0.5074, "step": 3986 }, { "epoch": 3.826692270763322, "grad_norm": 0.7196991294216143, "learning_rate": 1.5891507914298936e-06, "loss": 0.5034, "step": 3987 }, { "epoch": 3.827652424387902, "grad_norm": 0.8669095658003314, "learning_rate": 1.5867020872000927e-06, "loss": 0.5488, "step": 3988 }, { "epoch": 3.828612578012482, "grad_norm": 0.8431788612460692, "learning_rate": 1.5842549151632386e-06, "loss": 0.497, "step": 3989 }, { "epoch": 3.829572731637062, "grad_norm": 1.2173514566583228, "learning_rate": 1.5818092764178405e-06, "loss": 0.4624, "step": 3990 }, { "epoch": 3.8305328852616416, "grad_norm": 1.1300922073861408, "learning_rate": 1.5793651720617203e-06, "loss": 0.4408, "step": 3991 }, { "epoch": 3.831493038886222, "grad_norm": 0.849119702241092, "learning_rate": 1.576922603192011e-06, "loss": 0.459, "step": 3992 }, { "epoch": 3.8324531925108016, "grad_norm": 0.9532453128742033, "learning_rate": 1.5744815709051532e-06, "loss": 0.4991, "step": 3993 }, { "epoch": 3.833413346135382, "grad_norm": 0.880895326834877, "learning_rate": 1.5720420762969012e-06, "loss": 0.4891, "step": 3994 }, { "epoch": 3.8343734997599617, "grad_norm": 0.7457340539207377, "learning_rate": 1.5696041204623185e-06, "loss": 0.4759, "step": 3995 }, { "epoch": 3.8353336533845415, "grad_norm": 1.1290385986618707, "learning_rate": 1.5671677044957772e-06, "loss": 0.4945, "step": 3996 }, { "epoch": 3.8362938070091213, "grad_norm": 0.7904451184249721, "learning_rate": 1.5647328294909598e-06, "loss": 0.4889, "step": 3997 }, { "epoch": 3.8372539606337015, "grad_norm": 0.7871893328803605, "learning_rate": 1.5622994965408527e-06, "loss": 0.4907, "step": 3998 }, { "epoch": 3.8382141142582813, "grad_norm": 0.7930320329296842, "learning_rate": 1.5598677067377533e-06, "loss": 0.5124, "step": 3999 }, { "epoch": 3.839174267882861, "grad_norm": 1.0584156881285172, "learning_rate": 1.55743746117327e-06, "loss": 0.62, "step": 4000 }, { "epoch": 3.8401344215074413, "grad_norm": 1.531691679976701, "learning_rate": 1.5550087609383102e-06, "loss": 0.4722, "step": 4001 }, { "epoch": 3.841094575132021, "grad_norm": 0.8064444775512112, "learning_rate": 1.5525816071230932e-06, "loss": 0.5301, "step": 4002 }, { "epoch": 3.842054728756601, "grad_norm": 1.3569989328809873, "learning_rate": 1.5501560008171412e-06, "loss": 0.4396, "step": 4003 }, { "epoch": 3.8430148823811807, "grad_norm": 1.190611598047288, "learning_rate": 1.5477319431092835e-06, "loss": 0.6478, "step": 4004 }, { "epoch": 3.843975036005761, "grad_norm": 0.8426592371979329, "learning_rate": 1.5453094350876563e-06, "loss": 0.4969, "step": 4005 }, { "epoch": 3.8449351896303408, "grad_norm": 0.9834120497446485, "learning_rate": 1.5428884778396929e-06, "loss": 0.4846, "step": 4006 }, { "epoch": 3.845895343254921, "grad_norm": 1.4558191430806184, "learning_rate": 1.540469072452137e-06, "loss": 0.5361, "step": 4007 }, { "epoch": 3.846855496879501, "grad_norm": 0.9432916196340416, "learning_rate": 1.5380512200110347e-06, "loss": 0.4738, "step": 4008 }, { "epoch": 3.8478156505040806, "grad_norm": 0.7562002502713544, "learning_rate": 1.5356349216017336e-06, "loss": 0.5353, "step": 4009 }, { "epoch": 3.8487758041286604, "grad_norm": 0.9560068632844432, "learning_rate": 1.5332201783088857e-06, "loss": 0.5642, "step": 4010 }, { "epoch": 3.8497359577532406, "grad_norm": 0.9503726443536183, "learning_rate": 1.5308069912164403e-06, "loss": 0.5654, "step": 4011 }, { "epoch": 3.8506961113778204, "grad_norm": 0.926339616851514, "learning_rate": 1.5283953614076514e-06, "loss": 0.5086, "step": 4012 }, { "epoch": 3.8516562650024007, "grad_norm": 0.8949309574164176, "learning_rate": 1.5259852899650784e-06, "loss": 0.548, "step": 4013 }, { "epoch": 3.8526164186269805, "grad_norm": 1.239308660939615, "learning_rate": 1.5235767779705717e-06, "loss": 0.4782, "step": 4014 }, { "epoch": 3.8535765722515603, "grad_norm": 0.8406925180403649, "learning_rate": 1.5211698265052905e-06, "loss": 0.54, "step": 4015 }, { "epoch": 3.85453672587614, "grad_norm": 0.7441058389511468, "learning_rate": 1.518764436649684e-06, "loss": 0.5072, "step": 4016 }, { "epoch": 3.85549687950072, "grad_norm": 0.838940990206563, "learning_rate": 1.5163606094835115e-06, "loss": 0.5018, "step": 4017 }, { "epoch": 3.8564570331253, "grad_norm": 0.8284475080973874, "learning_rate": 1.5139583460858249e-06, "loss": 0.4968, "step": 4018 }, { "epoch": 3.85741718674988, "grad_norm": 0.9490834900003989, "learning_rate": 1.5115576475349715e-06, "loss": 0.5118, "step": 4019 }, { "epoch": 3.85837734037446, "grad_norm": 0.9284873572079144, "learning_rate": 1.5091585149086008e-06, "loss": 0.4975, "step": 4020 }, { "epoch": 3.85933749399904, "grad_norm": 0.9885169917968071, "learning_rate": 1.5067609492836572e-06, "loss": 0.5653, "step": 4021 }, { "epoch": 3.8602976476236197, "grad_norm": 0.7493394483901892, "learning_rate": 1.5043649517363835e-06, "loss": 0.4509, "step": 4022 }, { "epoch": 3.8612578012481995, "grad_norm": 0.7258304267343755, "learning_rate": 1.5019705233423177e-06, "loss": 0.4588, "step": 4023 }, { "epoch": 3.8622179548727797, "grad_norm": 0.9597960644184329, "learning_rate": 1.4995776651762906e-06, "loss": 0.3987, "step": 4024 }, { "epoch": 3.8631781084973595, "grad_norm": 1.061226017484637, "learning_rate": 1.4971863783124308e-06, "loss": 0.5635, "step": 4025 }, { "epoch": 3.86413826212194, "grad_norm": 0.9100474651906226, "learning_rate": 1.4947966638241662e-06, "loss": 0.5032, "step": 4026 }, { "epoch": 3.8650984157465196, "grad_norm": 0.6572458779185517, "learning_rate": 1.4924085227842084e-06, "loss": 0.5332, "step": 4027 }, { "epoch": 3.8660585693710994, "grad_norm": 0.6883941979229021, "learning_rate": 1.490021956264573e-06, "loss": 0.4728, "step": 4028 }, { "epoch": 3.867018722995679, "grad_norm": 0.7428986981191085, "learning_rate": 1.487636965336558e-06, "loss": 0.5382, "step": 4029 }, { "epoch": 3.867978876620259, "grad_norm": 0.7818197189442377, "learning_rate": 1.4852535510707662e-06, "loss": 0.5526, "step": 4030 }, { "epoch": 3.868939030244839, "grad_norm": 0.6655106349576536, "learning_rate": 1.4828717145370863e-06, "loss": 0.5939, "step": 4031 }, { "epoch": 3.869899183869419, "grad_norm": 0.8778246017653477, "learning_rate": 1.4804914568046957e-06, "loss": 0.4215, "step": 4032 }, { "epoch": 3.8708593374939992, "grad_norm": 0.8008089033683079, "learning_rate": 1.478112778942069e-06, "loss": 0.5221, "step": 4033 }, { "epoch": 3.871819491118579, "grad_norm": 0.8391307627321779, "learning_rate": 1.4757356820169694e-06, "loss": 0.44, "step": 4034 }, { "epoch": 3.872779644743159, "grad_norm": 1.000512620055828, "learning_rate": 1.4733601670964493e-06, "loss": 0.5579, "step": 4035 }, { "epoch": 3.8737397983677386, "grad_norm": 1.1053310541873387, "learning_rate": 1.4709862352468541e-06, "loss": 0.4797, "step": 4036 }, { "epoch": 3.874699951992319, "grad_norm": 0.8787789354638498, "learning_rate": 1.4686138875338136e-06, "loss": 0.5055, "step": 4037 }, { "epoch": 3.8756601056168987, "grad_norm": 1.2295553990559487, "learning_rate": 1.4662431250222504e-06, "loss": 0.3979, "step": 4038 }, { "epoch": 3.876620259241479, "grad_norm": 1.1267618351504367, "learning_rate": 1.4638739487763743e-06, "loss": 0.5119, "step": 4039 }, { "epoch": 3.8775804128660587, "grad_norm": 0.9789956541509062, "learning_rate": 1.4615063598596834e-06, "loss": 0.5757, "step": 4040 }, { "epoch": 3.8785405664906385, "grad_norm": 1.0848468206695112, "learning_rate": 1.4591403593349652e-06, "loss": 0.5139, "step": 4041 }, { "epoch": 3.8795007201152183, "grad_norm": 0.7083853796100976, "learning_rate": 1.4567759482642884e-06, "loss": 0.492, "step": 4042 }, { "epoch": 3.8804608737397985, "grad_norm": 1.0266688914876478, "learning_rate": 1.4544131277090118e-06, "loss": 0.477, "step": 4043 }, { "epoch": 3.8814210273643783, "grad_norm": 1.0793111231165864, "learning_rate": 1.4520518987297854e-06, "loss": 0.5481, "step": 4044 }, { "epoch": 3.882381180988958, "grad_norm": 1.0146801356114863, "learning_rate": 1.449692262386534e-06, "loss": 0.4931, "step": 4045 }, { "epoch": 3.8833413346135384, "grad_norm": 0.8556146852163357, "learning_rate": 1.447334219738477e-06, "loss": 0.4414, "step": 4046 }, { "epoch": 3.884301488238118, "grad_norm": 1.0433990431148177, "learning_rate": 1.4449777718441106e-06, "loss": 0.4425, "step": 4047 }, { "epoch": 3.885261641862698, "grad_norm": 0.7698546890522099, "learning_rate": 1.4426229197612217e-06, "loss": 0.3708, "step": 4048 }, { "epoch": 3.8862217954872778, "grad_norm": 0.8193396999682893, "learning_rate": 1.44026966454688e-06, "loss": 0.4207, "step": 4049 }, { "epoch": 3.887181949111858, "grad_norm": 1.1820553222279078, "learning_rate": 1.4379180072574335e-06, "loss": 0.4476, "step": 4050 }, { "epoch": 3.888142102736438, "grad_norm": 1.1015436868484454, "learning_rate": 1.4355679489485158e-06, "loss": 0.4536, "step": 4051 }, { "epoch": 3.889102256361018, "grad_norm": 0.8592257345120017, "learning_rate": 1.433219490675045e-06, "loss": 0.4735, "step": 4052 }, { "epoch": 3.890062409985598, "grad_norm": 0.7529162782095901, "learning_rate": 1.4308726334912176e-06, "loss": 0.5939, "step": 4053 }, { "epoch": 3.8910225636101776, "grad_norm": 0.9512340887006165, "learning_rate": 1.4285273784505154e-06, "loss": 0.5053, "step": 4054 }, { "epoch": 3.8919827172347574, "grad_norm": 0.9664577228503398, "learning_rate": 1.4261837266056938e-06, "loss": 0.4811, "step": 4055 }, { "epoch": 3.8929428708593377, "grad_norm": 0.9165423112341573, "learning_rate": 1.4238416790087945e-06, "loss": 0.3525, "step": 4056 }, { "epoch": 3.8939030244839175, "grad_norm": 0.6672949306102828, "learning_rate": 1.4215012367111408e-06, "loss": 0.5034, "step": 4057 }, { "epoch": 3.8948631781084972, "grad_norm": 0.7339697233455503, "learning_rate": 1.4191624007633286e-06, "loss": 0.5713, "step": 4058 }, { "epoch": 3.8958233317330775, "grad_norm": 0.6478474759490287, "learning_rate": 1.41682517221524e-06, "loss": 0.4932, "step": 4059 }, { "epoch": 3.8967834853576573, "grad_norm": 1.0445984349903144, "learning_rate": 1.4144895521160257e-06, "loss": 0.4394, "step": 4060 }, { "epoch": 3.897743638982237, "grad_norm": 0.7747001842790043, "learning_rate": 1.412155541514126e-06, "loss": 0.468, "step": 4061 }, { "epoch": 3.898703792606817, "grad_norm": 1.1077792006134408, "learning_rate": 1.4098231414572534e-06, "loss": 0.4751, "step": 4062 }, { "epoch": 3.899663946231397, "grad_norm": 0.8863241269922753, "learning_rate": 1.4074923529923945e-06, "loss": 0.5335, "step": 4063 }, { "epoch": 3.900624099855977, "grad_norm": 0.9755575641784038, "learning_rate": 1.4051631771658159e-06, "loss": 0.546, "step": 4064 }, { "epoch": 3.901584253480557, "grad_norm": 0.9713851611478814, "learning_rate": 1.402835615023061e-06, "loss": 0.5361, "step": 4065 }, { "epoch": 3.902544407105137, "grad_norm": 0.8524790834573509, "learning_rate": 1.400509667608947e-06, "loss": 0.5216, "step": 4066 }, { "epoch": 3.9035045607297167, "grad_norm": 0.9607175424128133, "learning_rate": 1.398185335967568e-06, "loss": 0.5404, "step": 4067 }, { "epoch": 3.9044647143542965, "grad_norm": 1.315686683110279, "learning_rate": 1.3958626211422887e-06, "loss": 0.4337, "step": 4068 }, { "epoch": 3.905424867978877, "grad_norm": 1.0079185127561066, "learning_rate": 1.3935415241757517e-06, "loss": 0.4973, "step": 4069 }, { "epoch": 3.9063850216034566, "grad_norm": 1.2226448395548501, "learning_rate": 1.3912220461098763e-06, "loss": 0.5499, "step": 4070 }, { "epoch": 3.9073451752280364, "grad_norm": 1.0292836148666729, "learning_rate": 1.3889041879858472e-06, "loss": 0.4741, "step": 4071 }, { "epoch": 3.9083053288526166, "grad_norm": 0.6858365544964726, "learning_rate": 1.3865879508441294e-06, "loss": 0.4561, "step": 4072 }, { "epoch": 3.9092654824771964, "grad_norm": 0.6934576720441407, "learning_rate": 1.384273335724452e-06, "loss": 0.4912, "step": 4073 }, { "epoch": 3.910225636101776, "grad_norm": 0.8014688337643019, "learning_rate": 1.3819603436658258e-06, "loss": 0.5354, "step": 4074 }, { "epoch": 3.911185789726356, "grad_norm": 1.3380777949204323, "learning_rate": 1.3796489757065284e-06, "loss": 0.539, "step": 4075 }, { "epoch": 3.9121459433509362, "grad_norm": 1.1560516154236573, "learning_rate": 1.3773392328841046e-06, "loss": 0.5476, "step": 4076 }, { "epoch": 3.913106096975516, "grad_norm": 0.9448042190895921, "learning_rate": 1.3750311162353752e-06, "loss": 0.5713, "step": 4077 }, { "epoch": 3.9140662506000963, "grad_norm": 0.874176814752014, "learning_rate": 1.3727246267964288e-06, "loss": 0.493, "step": 4078 }, { "epoch": 3.915026404224676, "grad_norm": 0.9702593845198296, "learning_rate": 1.3704197656026242e-06, "loss": 0.5237, "step": 4079 }, { "epoch": 3.915986557849256, "grad_norm": 0.7931895198579765, "learning_rate": 1.36811653368859e-06, "loss": 0.5156, "step": 4080 }, { "epoch": 3.9169467114738357, "grad_norm": 1.0132392439034934, "learning_rate": 1.3658149320882196e-06, "loss": 0.5917, "step": 4081 }, { "epoch": 3.917906865098416, "grad_norm": 1.0823767063657017, "learning_rate": 1.3635149618346793e-06, "loss": 0.4862, "step": 4082 }, { "epoch": 3.9188670187229957, "grad_norm": 0.8842536918038679, "learning_rate": 1.3612166239604003e-06, "loss": 0.4722, "step": 4083 }, { "epoch": 3.9198271723475755, "grad_norm": 0.9777547323510339, "learning_rate": 1.3589199194970825e-06, "loss": 0.4745, "step": 4084 }, { "epoch": 3.9207873259721557, "grad_norm": 1.1117169701710892, "learning_rate": 1.3566248494756934e-06, "loss": 0.4392, "step": 4085 }, { "epoch": 3.9217474795967355, "grad_norm": 1.0518037078640194, "learning_rate": 1.3543314149264625e-06, "loss": 0.5288, "step": 4086 }, { "epoch": 3.9227076332213153, "grad_norm": 0.8261835850961528, "learning_rate": 1.352039616878888e-06, "loss": 0.5543, "step": 4087 }, { "epoch": 3.923667786845895, "grad_norm": 0.7840536247769636, "learning_rate": 1.3497494563617375e-06, "loss": 0.4935, "step": 4088 }, { "epoch": 3.9246279404704754, "grad_norm": 0.6728269097782464, "learning_rate": 1.347460934403036e-06, "loss": 0.4857, "step": 4089 }, { "epoch": 3.925588094095055, "grad_norm": 1.1381595374137483, "learning_rate": 1.3451740520300787e-06, "loss": 0.4454, "step": 4090 }, { "epoch": 3.9265482477196354, "grad_norm": 1.1502955128886065, "learning_rate": 1.3428888102694187e-06, "loss": 0.5387, "step": 4091 }, { "epoch": 3.927508401344215, "grad_norm": 1.0192402382396137, "learning_rate": 1.3406052101468808e-06, "loss": 0.5601, "step": 4092 }, { "epoch": 3.928468554968795, "grad_norm": 1.118821648234648, "learning_rate": 1.338323252687549e-06, "loss": 0.6253, "step": 4093 }, { "epoch": 3.929428708593375, "grad_norm": 1.517502295285788, "learning_rate": 1.3360429389157664e-06, "loss": 0.5395, "step": 4094 }, { "epoch": 3.930388862217955, "grad_norm": 1.143197226226651, "learning_rate": 1.3337642698551428e-06, "loss": 0.5007, "step": 4095 }, { "epoch": 3.931349015842535, "grad_norm": 0.8774525599156622, "learning_rate": 1.331487246528549e-06, "loss": 0.4775, "step": 4096 }, { "epoch": 3.9323091694671146, "grad_norm": 0.7829465703530466, "learning_rate": 1.3292118699581153e-06, "loss": 0.5132, "step": 4097 }, { "epoch": 3.933269323091695, "grad_norm": 0.9145353274096759, "learning_rate": 1.3269381411652366e-06, "loss": 0.6108, "step": 4098 }, { "epoch": 3.9342294767162747, "grad_norm": 0.8368762705927574, "learning_rate": 1.324666061170562e-06, "loss": 0.479, "step": 4099 }, { "epoch": 3.9351896303408544, "grad_norm": 0.6345177403318302, "learning_rate": 1.3223956309940038e-06, "loss": 0.4872, "step": 4100 }, { "epoch": 3.9361497839654342, "grad_norm": 0.9968764329522467, "learning_rate": 1.3201268516547382e-06, "loss": 0.4334, "step": 4101 }, { "epoch": 3.9371099375900145, "grad_norm": 0.8643237007278449, "learning_rate": 1.317859724171192e-06, "loss": 0.509, "step": 4102 }, { "epoch": 3.9380700912145943, "grad_norm": 0.7057858010688522, "learning_rate": 1.3155942495610574e-06, "loss": 0.4739, "step": 4103 }, { "epoch": 3.9390302448391745, "grad_norm": 1.076812607787758, "learning_rate": 1.3133304288412768e-06, "loss": 0.4699, "step": 4104 }, { "epoch": 3.9399903984637543, "grad_norm": 0.9127467652918438, "learning_rate": 1.3110682630280603e-06, "loss": 0.4463, "step": 4105 }, { "epoch": 3.940950552088334, "grad_norm": 0.9940347120235042, "learning_rate": 1.3088077531368686e-06, "loss": 0.5, "step": 4106 }, { "epoch": 3.941910705712914, "grad_norm": 0.8252325841763662, "learning_rate": 1.3065489001824194e-06, "loss": 0.5934, "step": 4107 }, { "epoch": 3.942870859337494, "grad_norm": 6.668022645191015, "learning_rate": 1.3042917051786873e-06, "loss": 0.5605, "step": 4108 }, { "epoch": 3.943831012962074, "grad_norm": 0.8369858205159718, "learning_rate": 1.3020361691389043e-06, "loss": 0.536, "step": 4109 }, { "epoch": 3.9447911665866537, "grad_norm": 0.9719080285072355, "learning_rate": 1.299782293075556e-06, "loss": 0.4841, "step": 4110 }, { "epoch": 3.945751320211234, "grad_norm": 1.1548662029039942, "learning_rate": 1.297530078000384e-06, "loss": 0.4529, "step": 4111 }, { "epoch": 3.9467114738358138, "grad_norm": 1.1225664389401018, "learning_rate": 1.295279524924382e-06, "loss": 0.5078, "step": 4112 }, { "epoch": 3.9476716274603936, "grad_norm": 0.6338948909403576, "learning_rate": 1.2930306348577982e-06, "loss": 0.5871, "step": 4113 }, { "epoch": 3.9486317810849734, "grad_norm": 0.7777845943911945, "learning_rate": 1.2907834088101402e-06, "loss": 0.429, "step": 4114 }, { "epoch": 3.9495919347095536, "grad_norm": 0.7048656243278643, "learning_rate": 1.2885378477901595e-06, "loss": 0.6198, "step": 4115 }, { "epoch": 3.9505520883341334, "grad_norm": 0.9511247387143971, "learning_rate": 1.2862939528058665e-06, "loss": 0.5078, "step": 4116 }, { "epoch": 3.9515122419587136, "grad_norm": 1.0941770113790252, "learning_rate": 1.284051724864518e-06, "loss": 0.6075, "step": 4117 }, { "epoch": 3.9524723955832934, "grad_norm": 1.0193678884114714, "learning_rate": 1.2818111649726306e-06, "loss": 0.5144, "step": 4118 }, { "epoch": 3.9534325492078732, "grad_norm": 1.0781491984680291, "learning_rate": 1.2795722741359672e-06, "loss": 0.4499, "step": 4119 }, { "epoch": 3.954392702832453, "grad_norm": 0.978198329802974, "learning_rate": 1.277335053359539e-06, "loss": 0.4499, "step": 4120 }, { "epoch": 3.9553528564570333, "grad_norm": 0.9951234970847984, "learning_rate": 1.2750995036476133e-06, "loss": 0.6097, "step": 4121 }, { "epoch": 3.956313010081613, "grad_norm": 0.7262713186206744, "learning_rate": 1.2728656260037025e-06, "loss": 0.5515, "step": 4122 }, { "epoch": 3.957273163706193, "grad_norm": 0.8868082003556855, "learning_rate": 1.2706334214305716e-06, "loss": 0.5081, "step": 4123 }, { "epoch": 3.958233317330773, "grad_norm": 0.6324265663514611, "learning_rate": 1.2684028909302348e-06, "loss": 0.4445, "step": 4124 }, { "epoch": 3.959193470955353, "grad_norm": 0.8837637283725256, "learning_rate": 1.2661740355039499e-06, "loss": 0.5139, "step": 4125 }, { "epoch": 3.9601536245799327, "grad_norm": 0.8727953739636891, "learning_rate": 1.2639468561522284e-06, "loss": 0.4948, "step": 4126 }, { "epoch": 3.9611137782045125, "grad_norm": 0.9902783586410273, "learning_rate": 1.2617213538748268e-06, "loss": 0.4954, "step": 4127 }, { "epoch": 3.9620739318290927, "grad_norm": 0.8243768851310336, "learning_rate": 1.2594975296707495e-06, "loss": 0.5251, "step": 4128 }, { "epoch": 3.9630340854536725, "grad_norm": 1.0973337213430947, "learning_rate": 1.2572753845382484e-06, "loss": 0.532, "step": 4129 }, { "epoch": 3.9639942390782528, "grad_norm": 1.1692911078372372, "learning_rate": 1.255054919474819e-06, "loss": 0.4613, "step": 4130 }, { "epoch": 3.9649543927028326, "grad_norm": 0.6939501768750728, "learning_rate": 1.252836135477203e-06, "loss": 0.4773, "step": 4131 }, { "epoch": 3.9659145463274124, "grad_norm": 1.0726479897146066, "learning_rate": 1.2506190335413943e-06, "loss": 0.4609, "step": 4132 }, { "epoch": 3.966874699951992, "grad_norm": 0.7107826571347109, "learning_rate": 1.2484036146626205e-06, "loss": 0.3925, "step": 4133 }, { "epoch": 3.9678348535765724, "grad_norm": 0.7611573165361655, "learning_rate": 1.2461898798353639e-06, "loss": 0.5052, "step": 4134 }, { "epoch": 3.968795007201152, "grad_norm": 0.8241813944972161, "learning_rate": 1.2439778300533412e-06, "loss": 0.4283, "step": 4135 }, { "epoch": 3.969755160825732, "grad_norm": 0.7795946845774496, "learning_rate": 1.241767466309522e-06, "loss": 0.4264, "step": 4136 }, { "epoch": 3.9707153144503122, "grad_norm": 0.8645816681151549, "learning_rate": 1.2395587895961158e-06, "loss": 0.4766, "step": 4137 }, { "epoch": 3.971675468074892, "grad_norm": 1.0091395186950214, "learning_rate": 1.2373518009045705e-06, "loss": 0.4608, "step": 4138 }, { "epoch": 3.972635621699472, "grad_norm": 0.7468576690690266, "learning_rate": 1.2351465012255808e-06, "loss": 0.5611, "step": 4139 }, { "epoch": 3.9735957753240516, "grad_norm": 0.688052761846315, "learning_rate": 1.232942891549083e-06, "loss": 0.4234, "step": 4140 }, { "epoch": 3.974555928948632, "grad_norm": 0.8874276873845143, "learning_rate": 1.2307409728642539e-06, "loss": 0.5827, "step": 4141 }, { "epoch": 3.9755160825732117, "grad_norm": 0.9126251319418551, "learning_rate": 1.2285407461595122e-06, "loss": 0.4643, "step": 4142 }, { "epoch": 3.976476236197792, "grad_norm": 1.1013667422263231, "learning_rate": 1.2263422124225132e-06, "loss": 0.4744, "step": 4143 }, { "epoch": 3.9774363898223717, "grad_norm": 0.9601760543878562, "learning_rate": 1.2241453726401559e-06, "loss": 0.5087, "step": 4144 }, { "epoch": 3.9783965434469515, "grad_norm": 0.6377301879955056, "learning_rate": 1.2219502277985817e-06, "loss": 0.4916, "step": 4145 }, { "epoch": 3.9793566970715313, "grad_norm": 1.0228823666868765, "learning_rate": 1.2197567788831638e-06, "loss": 0.5073, "step": 4146 }, { "epoch": 3.9803168506961115, "grad_norm": 0.8788753700110128, "learning_rate": 1.2175650268785205e-06, "loss": 0.4786, "step": 4147 }, { "epoch": 3.9812770043206913, "grad_norm": 0.9823384309895182, "learning_rate": 1.215374972768502e-06, "loss": 0.5393, "step": 4148 }, { "epoch": 3.982237157945271, "grad_norm": 0.7094996955855833, "learning_rate": 1.213186617536204e-06, "loss": 0.5501, "step": 4149 }, { "epoch": 3.9831973115698514, "grad_norm": 0.7421069497091106, "learning_rate": 1.2109999621639561e-06, "loss": 0.5212, "step": 4150 }, { "epoch": 3.984157465194431, "grad_norm": 0.843124073727083, "learning_rate": 1.2088150076333222e-06, "loss": 0.5294, "step": 4151 }, { "epoch": 3.985117618819011, "grad_norm": 0.8606077195221448, "learning_rate": 1.2066317549251055e-06, "loss": 0.5914, "step": 4152 }, { "epoch": 3.9860777724435907, "grad_norm": 0.9096256381569566, "learning_rate": 1.2044502050193451e-06, "loss": 0.431, "step": 4153 }, { "epoch": 3.987037926068171, "grad_norm": 0.6312502620001125, "learning_rate": 1.2022703588953155e-06, "loss": 0.5369, "step": 4154 }, { "epoch": 3.9879980796927508, "grad_norm": 1.0986856387121426, "learning_rate": 1.2000922175315278e-06, "loss": 0.4213, "step": 4155 }, { "epoch": 3.988958233317331, "grad_norm": 0.9163146079518515, "learning_rate": 1.197915781905723e-06, "loss": 0.4495, "step": 4156 }, { "epoch": 3.989918386941911, "grad_norm": 0.8024004504712938, "learning_rate": 1.1957410529948803e-06, "loss": 0.4562, "step": 4157 }, { "epoch": 3.9908785405664906, "grad_norm": 0.708415794627345, "learning_rate": 1.1935680317752173e-06, "loss": 0.5089, "step": 4158 }, { "epoch": 3.9918386941910704, "grad_norm": 0.9608001633175752, "learning_rate": 1.1913967192221742e-06, "loss": 0.4828, "step": 4159 }, { "epoch": 3.9927988478156506, "grad_norm": 1.0250316293770443, "learning_rate": 1.189227116310434e-06, "loss": 0.5149, "step": 4160 }, { "epoch": 3.9937590014402304, "grad_norm": 0.9253603863825774, "learning_rate": 1.1870592240139033e-06, "loss": 0.4553, "step": 4161 }, { "epoch": 3.9947191550648102, "grad_norm": 0.7190734965988832, "learning_rate": 1.18489304330573e-06, "loss": 0.4231, "step": 4162 }, { "epoch": 3.9956793086893905, "grad_norm": 0.7412743409639089, "learning_rate": 1.18272857515829e-06, "loss": 0.4453, "step": 4163 }, { "epoch": 3.9966394623139703, "grad_norm": 0.7484620379584466, "learning_rate": 1.1805658205431868e-06, "loss": 0.5023, "step": 4164 }, { "epoch": 3.99759961593855, "grad_norm": 0.9753464624004037, "learning_rate": 1.1784047804312598e-06, "loss": 0.4607, "step": 4165 }, { "epoch": 3.99855976956313, "grad_norm": 0.7916677760565596, "learning_rate": 1.1762454557925756e-06, "loss": 0.561, "step": 4166 }, { "epoch": 3.99951992318771, "grad_norm": 0.9833772665095369, "learning_rate": 1.174087847596433e-06, "loss": 0.4724, "step": 4167 }, { "epoch": 4.0, "grad_norm": 1.6938529657017545, "learning_rate": 1.1719319568113602e-06, "loss": 0.5145, "step": 4168 }, { "epoch": 4.00096015362458, "grad_norm": 1.1539304671261217, "learning_rate": 1.1697777844051105e-06, "loss": 0.5072, "step": 4169 }, { "epoch": 4.00192030724916, "grad_norm": 0.9143544500057557, "learning_rate": 1.1676253313446706e-06, "loss": 0.4725, "step": 4170 }, { "epoch": 4.002880460873739, "grad_norm": 1.0322584515499804, "learning_rate": 1.165474598596254e-06, "loss": 0.485, "step": 4171 }, { "epoch": 4.00384061449832, "grad_norm": 1.3629202990256055, "learning_rate": 1.1633255871253013e-06, "loss": 0.4665, "step": 4172 }, { "epoch": 4.0048007681229, "grad_norm": 0.7596769664819089, "learning_rate": 1.161178297896482e-06, "loss": 0.5165, "step": 4173 }, { "epoch": 4.00576092174748, "grad_norm": 0.9555076797423135, "learning_rate": 1.1590327318736887e-06, "loss": 0.4299, "step": 4174 }, { "epoch": 4.0067210753720595, "grad_norm": 1.074560141095779, "learning_rate": 1.1568888900200432e-06, "loss": 0.3898, "step": 4175 }, { "epoch": 4.007681228996639, "grad_norm": 0.9201872523872426, "learning_rate": 1.1547467732978967e-06, "loss": 0.5534, "step": 4176 }, { "epoch": 4.008641382621219, "grad_norm": 0.9241005760724432, "learning_rate": 1.152606382668819e-06, "loss": 0.4662, "step": 4177 }, { "epoch": 4.0096015362458, "grad_norm": 0.7896750955770455, "learning_rate": 1.1504677190936115e-06, "loss": 0.4872, "step": 4178 }, { "epoch": 4.0105616898703795, "grad_norm": 1.2723253763526183, "learning_rate": 1.148330783532292e-06, "loss": 0.4211, "step": 4179 }, { "epoch": 4.011521843494959, "grad_norm": 0.9055537232892794, "learning_rate": 1.1461955769441141e-06, "loss": 0.5142, "step": 4180 }, { "epoch": 4.012481997119539, "grad_norm": 0.8557207766605038, "learning_rate": 1.144062100287548e-06, "loss": 0.5453, "step": 4181 }, { "epoch": 4.013442150744119, "grad_norm": 0.9212212543157423, "learning_rate": 1.1419303545202854e-06, "loss": 0.4758, "step": 4182 }, { "epoch": 4.014402304368699, "grad_norm": 0.8168956833798467, "learning_rate": 1.1398003405992468e-06, "loss": 0.4278, "step": 4183 }, { "epoch": 4.0153624579932785, "grad_norm": 0.8082862244256038, "learning_rate": 1.1376720594805719e-06, "loss": 0.597, "step": 4184 }, { "epoch": 4.016322611617859, "grad_norm": 0.9679712627597954, "learning_rate": 1.1355455121196234e-06, "loss": 0.4101, "step": 4185 }, { "epoch": 4.017282765242439, "grad_norm": 0.6058133441784157, "learning_rate": 1.1334206994709874e-06, "loss": 0.4213, "step": 4186 }, { "epoch": 4.018242918867019, "grad_norm": 0.9023568368491182, "learning_rate": 1.1312976224884665e-06, "loss": 0.5639, "step": 4187 }, { "epoch": 4.019203072491599, "grad_norm": 1.0085670282068113, "learning_rate": 1.1291762821250867e-06, "loss": 0.4627, "step": 4188 }, { "epoch": 4.020163226116178, "grad_norm": 0.8458624618021306, "learning_rate": 1.1270566793331006e-06, "loss": 0.4709, "step": 4189 }, { "epoch": 4.021123379740758, "grad_norm": 0.9441036174142976, "learning_rate": 1.1249388150639702e-06, "loss": 0.4592, "step": 4190 }, { "epoch": 4.022083533365339, "grad_norm": 0.9278962166452477, "learning_rate": 1.1228226902683847e-06, "loss": 0.4924, "step": 4191 }, { "epoch": 4.023043686989919, "grad_norm": 0.6778659874057104, "learning_rate": 1.1207083058962465e-06, "loss": 0.4532, "step": 4192 }, { "epoch": 4.0240038406144985, "grad_norm": 0.830408478039141, "learning_rate": 1.1185956628966832e-06, "loss": 0.6059, "step": 4193 }, { "epoch": 4.024963994239078, "grad_norm": 0.8624573611467153, "learning_rate": 1.1164847622180392e-06, "loss": 0.4384, "step": 4194 }, { "epoch": 4.025924147863658, "grad_norm": 1.1640620813685674, "learning_rate": 1.1143756048078707e-06, "loss": 0.463, "step": 4195 }, { "epoch": 4.026884301488238, "grad_norm": 0.937207865989249, "learning_rate": 1.1122681916129586e-06, "loss": 0.5145, "step": 4196 }, { "epoch": 4.027844455112818, "grad_norm": 1.0777598574697018, "learning_rate": 1.1101625235792978e-06, "loss": 0.4205, "step": 4197 }, { "epoch": 4.028804608737398, "grad_norm": 0.7468889856667972, "learning_rate": 1.1080586016521e-06, "loss": 0.4918, "step": 4198 }, { "epoch": 4.029764762361978, "grad_norm": 0.7710631717480465, "learning_rate": 1.1059564267757945e-06, "loss": 0.5365, "step": 4199 }, { "epoch": 4.030724915986558, "grad_norm": 1.1507518855430348, "learning_rate": 1.1038559998940228e-06, "loss": 0.4568, "step": 4200 }, { "epoch": 4.031685069611138, "grad_norm": 0.9799742825591378, "learning_rate": 1.1017573219496453e-06, "loss": 0.5058, "step": 4201 }, { "epoch": 4.0326452232357175, "grad_norm": 0.8113758825887305, "learning_rate": 1.0996603938847361e-06, "loss": 0.5, "step": 4202 }, { "epoch": 4.033605376860297, "grad_norm": 1.1964146174166044, "learning_rate": 1.0975652166405836e-06, "loss": 0.3491, "step": 4203 }, { "epoch": 4.034565530484878, "grad_norm": 0.9723097238498822, "learning_rate": 1.0954717911576923e-06, "loss": 0.4498, "step": 4204 }, { "epoch": 4.035525684109458, "grad_norm": 0.7510846025239971, "learning_rate": 1.0933801183757742e-06, "loss": 0.4715, "step": 4205 }, { "epoch": 4.036485837734038, "grad_norm": 1.1542383714339781, "learning_rate": 1.0912901992337626e-06, "loss": 0.4704, "step": 4206 }, { "epoch": 4.037445991358617, "grad_norm": 0.8329792894128829, "learning_rate": 1.0892020346698e-06, "loss": 0.4527, "step": 4207 }, { "epoch": 4.038406144983197, "grad_norm": 0.8610457700025649, "learning_rate": 1.087115625621239e-06, "loss": 0.4867, "step": 4208 }, { "epoch": 4.039366298607777, "grad_norm": 1.0646426069950743, "learning_rate": 1.0850309730246472e-06, "loss": 0.4425, "step": 4209 }, { "epoch": 4.040326452232357, "grad_norm": 0.9480031500872488, "learning_rate": 1.0829480778158025e-06, "loss": 0.5185, "step": 4210 }, { "epoch": 4.041286605856937, "grad_norm": 0.7102734175501262, "learning_rate": 1.0808669409296951e-06, "loss": 0.4798, "step": 4211 }, { "epoch": 4.042246759481517, "grad_norm": 0.8572814246111411, "learning_rate": 1.0787875633005256e-06, "loss": 0.4716, "step": 4212 }, { "epoch": 4.043206913106097, "grad_norm": 0.95907881231672, "learning_rate": 1.0767099458617025e-06, "loss": 0.4394, "step": 4213 }, { "epoch": 4.044167066730677, "grad_norm": 0.6397664778907053, "learning_rate": 1.0746340895458475e-06, "loss": 0.4113, "step": 4214 }, { "epoch": 4.045127220355257, "grad_norm": 0.8442205341498596, "learning_rate": 1.0725599952847894e-06, "loss": 0.5372, "step": 4215 }, { "epoch": 4.046087373979836, "grad_norm": 0.7066590879110352, "learning_rate": 1.070487664009567e-06, "loss": 0.433, "step": 4216 }, { "epoch": 4.047047527604417, "grad_norm": 1.4116513441653262, "learning_rate": 1.0684170966504298e-06, "loss": 0.4187, "step": 4217 }, { "epoch": 4.048007681228997, "grad_norm": 1.0476119328658156, "learning_rate": 1.0663482941368303e-06, "loss": 0.4821, "step": 4218 }, { "epoch": 4.048967834853577, "grad_norm": 0.8986363632527578, "learning_rate": 1.064281257397432e-06, "loss": 0.5106, "step": 4219 }, { "epoch": 4.0499279884781565, "grad_norm": 0.5848278502147763, "learning_rate": 1.062215987360109e-06, "loss": 0.372, "step": 4220 }, { "epoch": 4.050888142102736, "grad_norm": 0.8067830570206104, "learning_rate": 1.0601524849519357e-06, "loss": 0.4328, "step": 4221 }, { "epoch": 4.051848295727316, "grad_norm": 0.8172466496173983, "learning_rate": 1.0580907510991982e-06, "loss": 0.4436, "step": 4222 }, { "epoch": 4.052808449351896, "grad_norm": 0.8510403126932888, "learning_rate": 1.0560307867273838e-06, "loss": 0.4488, "step": 4223 }, { "epoch": 4.053768602976477, "grad_norm": 0.8939959511310954, "learning_rate": 1.0539725927611915e-06, "loss": 0.4707, "step": 4224 }, { "epoch": 4.054728756601056, "grad_norm": 1.097732915391469, "learning_rate": 1.0519161701245234e-06, "loss": 0.4419, "step": 4225 }, { "epoch": 4.055688910225636, "grad_norm": 0.9414107199067336, "learning_rate": 1.049861519740482e-06, "loss": 0.4053, "step": 4226 }, { "epoch": 4.056649063850216, "grad_norm": 0.7692216143188121, "learning_rate": 1.04780864253138e-06, "loss": 0.519, "step": 4227 }, { "epoch": 4.057609217474796, "grad_norm": 0.9046576567901469, "learning_rate": 1.0457575394187314e-06, "loss": 0.5017, "step": 4228 }, { "epoch": 4.0585693710993755, "grad_norm": 0.6800740592495405, "learning_rate": 1.043708211323255e-06, "loss": 0.5698, "step": 4229 }, { "epoch": 4.059529524723956, "grad_norm": 0.8518586085062039, "learning_rate": 1.0416606591648737e-06, "loss": 0.4237, "step": 4230 }, { "epoch": 4.060489678348536, "grad_norm": 0.8049777669058564, "learning_rate": 1.039614883862709e-06, "loss": 0.5237, "step": 4231 }, { "epoch": 4.061449831973116, "grad_norm": 0.7380969206473762, "learning_rate": 1.037570886335087e-06, "loss": 0.4513, "step": 4232 }, { "epoch": 4.062409985597696, "grad_norm": 0.8104125300962116, "learning_rate": 1.0355286674995413e-06, "loss": 0.4317, "step": 4233 }, { "epoch": 4.063370139222275, "grad_norm": 1.7782902506722207, "learning_rate": 1.0334882282727971e-06, "loss": 0.4466, "step": 4234 }, { "epoch": 4.064330292846855, "grad_norm": 0.6227892399636336, "learning_rate": 1.0314495695707887e-06, "loss": 0.4635, "step": 4235 }, { "epoch": 4.065290446471435, "grad_norm": 1.206657340412036, "learning_rate": 1.029412692308645e-06, "loss": 0.5248, "step": 4236 }, { "epoch": 4.066250600096016, "grad_norm": 0.9520714482122036, "learning_rate": 1.0273775974007017e-06, "loss": 0.4674, "step": 4237 }, { "epoch": 4.0672107537205955, "grad_norm": 1.0451390691584277, "learning_rate": 1.0253442857604911e-06, "loss": 0.4735, "step": 4238 }, { "epoch": 4.068170907345175, "grad_norm": 0.8584802060696435, "learning_rate": 1.0233127583007424e-06, "loss": 0.4711, "step": 4239 }, { "epoch": 4.069131060969755, "grad_norm": 0.8191885531969274, "learning_rate": 1.0212830159333886e-06, "loss": 0.5384, "step": 4240 }, { "epoch": 4.070091214594335, "grad_norm": 0.704955843803098, "learning_rate": 1.0192550595695584e-06, "loss": 0.4861, "step": 4241 }, { "epoch": 4.071051368218915, "grad_norm": 0.9285317454672152, "learning_rate": 1.0172288901195805e-06, "loss": 0.4692, "step": 4242 }, { "epoch": 4.072011521843495, "grad_norm": 0.9366770450684629, "learning_rate": 1.0152045084929813e-06, "loss": 0.5207, "step": 4243 }, { "epoch": 4.072971675468075, "grad_norm": 1.0353293198175209, "learning_rate": 1.0131819155984818e-06, "loss": 0.5531, "step": 4244 }, { "epoch": 4.073931829092655, "grad_norm": 0.7861120113264909, "learning_rate": 1.0111611123440041e-06, "loss": 0.6202, "step": 4245 }, { "epoch": 4.074891982717235, "grad_norm": 0.895773491424911, "learning_rate": 1.0091420996366646e-06, "loss": 0.4866, "step": 4246 }, { "epoch": 4.0758521363418145, "grad_norm": 0.8205116343024726, "learning_rate": 1.0071248783827764e-06, "loss": 0.4826, "step": 4247 }, { "epoch": 4.076812289966394, "grad_norm": 0.7833944739122715, "learning_rate": 1.0051094494878505e-06, "loss": 0.3644, "step": 4248 }, { "epoch": 4.077772443590974, "grad_norm": 0.9628065314565997, "learning_rate": 1.0030958138565883e-06, "loss": 0.3343, "step": 4249 }, { "epoch": 4.078732597215555, "grad_norm": 0.984369185351619, "learning_rate": 1.001083972392889e-06, "loss": 0.4362, "step": 4250 }, { "epoch": 4.079692750840135, "grad_norm": 1.0160691774395112, "learning_rate": 9.990739259998517e-07, "loss": 0.4633, "step": 4251 }, { "epoch": 4.080652904464714, "grad_norm": 0.7616597411507633, "learning_rate": 9.970656755797598e-07, "loss": 0.4539, "step": 4252 }, { "epoch": 4.081613058089294, "grad_norm": 1.0308720375810965, "learning_rate": 9.950592220340972e-07, "loss": 0.4174, "step": 4253 }, { "epoch": 4.082573211713874, "grad_norm": 0.9366606922695732, "learning_rate": 9.930545662635393e-07, "loss": 0.3122, "step": 4254 }, { "epoch": 4.083533365338454, "grad_norm": 0.7853739679679238, "learning_rate": 9.91051709167955e-07, "loss": 0.406, "step": 4255 }, { "epoch": 4.0844935189630345, "grad_norm": 0.8599436147272986, "learning_rate": 9.890506516464066e-07, "loss": 0.5416, "step": 4256 }, { "epoch": 4.085453672587614, "grad_norm": 0.8859114214953786, "learning_rate": 9.870513945971444e-07, "loss": 0.4915, "step": 4257 }, { "epoch": 4.086413826212194, "grad_norm": 1.1696280752395467, "learning_rate": 9.85053938917615e-07, "loss": 0.5, "step": 4258 }, { "epoch": 4.087373979836774, "grad_norm": 0.7001670062951464, "learning_rate": 9.83058285504455e-07, "loss": 0.4348, "step": 4259 }, { "epoch": 4.088334133461354, "grad_norm": 0.7192788413320232, "learning_rate": 9.81064435253492e-07, "loss": 0.4893, "step": 4260 }, { "epoch": 4.0892942870859335, "grad_norm": 1.146164988467551, "learning_rate": 9.790723890597452e-07, "loss": 0.5112, "step": 4261 }, { "epoch": 4.090254440710513, "grad_norm": 0.7480297952766546, "learning_rate": 9.770821478174207e-07, "loss": 0.4192, "step": 4262 }, { "epoch": 4.091214594335094, "grad_norm": 0.8711655788307289, "learning_rate": 9.750937124199162e-07, "loss": 0.4879, "step": 4263 }, { "epoch": 4.092174747959674, "grad_norm": 1.0012640834399809, "learning_rate": 9.731070837598227e-07, "loss": 0.5164, "step": 4264 }, { "epoch": 4.0931349015842535, "grad_norm": 0.9186048383946273, "learning_rate": 9.71122262728913e-07, "loss": 0.4958, "step": 4265 }, { "epoch": 4.094095055208833, "grad_norm": 1.1469334593114975, "learning_rate": 9.69139250218154e-07, "loss": 0.4295, "step": 4266 }, { "epoch": 4.095055208833413, "grad_norm": 0.7952206320872112, "learning_rate": 9.67158047117696e-07, "loss": 0.4581, "step": 4267 }, { "epoch": 4.096015362457993, "grad_norm": 0.8984375111747237, "learning_rate": 9.651786543168834e-07, "loss": 0.5113, "step": 4268 }, { "epoch": 4.096975516082574, "grad_norm": 0.9714299591132362, "learning_rate": 9.63201072704244e-07, "loss": 0.4817, "step": 4269 }, { "epoch": 4.097935669707153, "grad_norm": 0.7950624342110573, "learning_rate": 9.612253031674906e-07, "loss": 0.4939, "step": 4270 }, { "epoch": 4.098895823331733, "grad_norm": 0.8614617507188326, "learning_rate": 9.592513465935272e-07, "loss": 0.434, "step": 4271 }, { "epoch": 4.099855976956313, "grad_norm": 0.8426443989971301, "learning_rate": 9.572792038684404e-07, "loss": 0.4709, "step": 4272 }, { "epoch": 4.100816130580893, "grad_norm": 0.6624183774748574, "learning_rate": 9.553088758775048e-07, "loss": 0.4759, "step": 4273 }, { "epoch": 4.101776284205473, "grad_norm": 0.7661087546093349, "learning_rate": 9.533403635051813e-07, "loss": 0.5181, "step": 4274 }, { "epoch": 4.102736437830052, "grad_norm": 0.9245750734522984, "learning_rate": 9.513736676351104e-07, "loss": 0.4944, "step": 4275 }, { "epoch": 4.103696591454633, "grad_norm": 1.3121510477887328, "learning_rate": 9.494087891501213e-07, "loss": 0.4627, "step": 4276 }, { "epoch": 4.104656745079213, "grad_norm": 0.8022418272899837, "learning_rate": 9.474457289322314e-07, "loss": 0.4855, "step": 4277 }, { "epoch": 4.105616898703793, "grad_norm": 0.8662195173198269, "learning_rate": 9.454844878626329e-07, "loss": 0.4359, "step": 4278 }, { "epoch": 4.1065770523283724, "grad_norm": 0.7493396642862538, "learning_rate": 9.435250668217088e-07, "loss": 0.5218, "step": 4279 }, { "epoch": 4.107537205952952, "grad_norm": 0.8327281146822996, "learning_rate": 9.415674666890178e-07, "loss": 0.4459, "step": 4280 }, { "epoch": 4.108497359577532, "grad_norm": 1.2242627554228445, "learning_rate": 9.396116883433104e-07, "loss": 0.5209, "step": 4281 }, { "epoch": 4.109457513202113, "grad_norm": 1.4372822271900842, "learning_rate": 9.376577326625147e-07, "loss": 0.4729, "step": 4282 }, { "epoch": 4.1104176668266925, "grad_norm": 1.0274139110204858, "learning_rate": 9.35705600523737e-07, "loss": 0.4691, "step": 4283 }, { "epoch": 4.111377820451272, "grad_norm": 0.7671183915220752, "learning_rate": 9.337552928032706e-07, "loss": 0.3918, "step": 4284 }, { "epoch": 4.112337974075852, "grad_norm": 0.7156218932684426, "learning_rate": 9.318068103765871e-07, "loss": 0.5243, "step": 4285 }, { "epoch": 4.113298127700432, "grad_norm": 0.7539853904800545, "learning_rate": 9.2986015411834e-07, "loss": 0.5466, "step": 4286 }, { "epoch": 4.114258281325012, "grad_norm": 0.7650315660174182, "learning_rate": 9.279153249023637e-07, "loss": 0.5296, "step": 4287 }, { "epoch": 4.1152184349495915, "grad_norm": 0.6141921977917617, "learning_rate": 9.259723236016688e-07, "loss": 0.4998, "step": 4288 }, { "epoch": 4.116178588574172, "grad_norm": 0.7625743283929614, "learning_rate": 9.240311510884487e-07, "loss": 0.4782, "step": 4289 }, { "epoch": 4.117138742198752, "grad_norm": 0.9151386460801915, "learning_rate": 9.220918082340752e-07, "loss": 0.4313, "step": 4290 }, { "epoch": 4.118098895823332, "grad_norm": 0.840569351065396, "learning_rate": 9.201542959090987e-07, "loss": 0.506, "step": 4291 }, { "epoch": 4.119059049447912, "grad_norm": 0.7355049900821172, "learning_rate": 9.182186149832484e-07, "loss": 0.4359, "step": 4292 }, { "epoch": 4.120019203072491, "grad_norm": 0.8457178710081827, "learning_rate": 9.162847663254292e-07, "loss": 0.3839, "step": 4293 }, { "epoch": 4.120979356697071, "grad_norm": 0.7711435667260678, "learning_rate": 9.143527508037243e-07, "loss": 0.5454, "step": 4294 }, { "epoch": 4.121939510321652, "grad_norm": 0.7766304923887546, "learning_rate": 9.124225692853995e-07, "loss": 0.3722, "step": 4295 }, { "epoch": 4.122899663946232, "grad_norm": 0.825856847296665, "learning_rate": 9.104942226368879e-07, "loss": 0.4735, "step": 4296 }, { "epoch": 4.123859817570811, "grad_norm": 0.6325046297797443, "learning_rate": 9.085677117238068e-07, "loss": 0.4722, "step": 4297 }, { "epoch": 4.124819971195391, "grad_norm": 0.8937885246422357, "learning_rate": 9.066430374109419e-07, "loss": 0.4358, "step": 4298 }, { "epoch": 4.125780124819971, "grad_norm": 0.821947775699068, "learning_rate": 9.047202005622641e-07, "loss": 0.4436, "step": 4299 }, { "epoch": 4.126740278444551, "grad_norm": 0.8617806608188856, "learning_rate": 9.027992020409127e-07, "loss": 0.4664, "step": 4300 }, { "epoch": 4.1277004320691315, "grad_norm": 0.8830033901566686, "learning_rate": 9.008800427092018e-07, "loss": 0.4328, "step": 4301 }, { "epoch": 4.128660585693711, "grad_norm": 0.768360225407334, "learning_rate": 8.989627234286225e-07, "loss": 0.4498, "step": 4302 }, { "epoch": 4.129620739318291, "grad_norm": 0.7585497820708124, "learning_rate": 8.970472450598394e-07, "loss": 0.4258, "step": 4303 }, { "epoch": 4.130580892942871, "grad_norm": 0.7125367319827364, "learning_rate": 8.951336084626894e-07, "loss": 0.4871, "step": 4304 }, { "epoch": 4.131541046567451, "grad_norm": 0.8138114556520635, "learning_rate": 8.932218144961858e-07, "loss": 0.408, "step": 4305 }, { "epoch": 4.1325012001920305, "grad_norm": 1.6344124990396554, "learning_rate": 8.913118640185086e-07, "loss": 0.4964, "step": 4306 }, { "epoch": 4.13346135381661, "grad_norm": 0.8870832539280857, "learning_rate": 8.894037578870152e-07, "loss": 0.4615, "step": 4307 }, { "epoch": 4.134421507441191, "grad_norm": 0.8323421083831397, "learning_rate": 8.874974969582379e-07, "loss": 0.4399, "step": 4308 }, { "epoch": 4.135381661065771, "grad_norm": 0.6770666907230487, "learning_rate": 8.855930820878727e-07, "loss": 0.5118, "step": 4309 }, { "epoch": 4.136341814690351, "grad_norm": 1.1217520877098008, "learning_rate": 8.836905141307933e-07, "loss": 0.5279, "step": 4310 }, { "epoch": 4.13730196831493, "grad_norm": 1.1683437841138862, "learning_rate": 8.817897939410391e-07, "loss": 0.3961, "step": 4311 }, { "epoch": 4.13826212193951, "grad_norm": 0.9121737466871386, "learning_rate": 8.79890922371826e-07, "loss": 0.4268, "step": 4312 }, { "epoch": 4.13922227556409, "grad_norm": 1.0825734114766223, "learning_rate": 8.779939002755378e-07, "loss": 0.5086, "step": 4313 }, { "epoch": 4.14018242918867, "grad_norm": 0.7058690209354211, "learning_rate": 8.760987285037248e-07, "loss": 0.5396, "step": 4314 }, { "epoch": 4.14114258281325, "grad_norm": 1.0887993670110685, "learning_rate": 8.742054079071105e-07, "loss": 0.5357, "step": 4315 }, { "epoch": 4.14210273643783, "grad_norm": 0.7908687290380587, "learning_rate": 8.723139393355861e-07, "loss": 0.5221, "step": 4316 }, { "epoch": 4.14306289006241, "grad_norm": 0.7787199820644941, "learning_rate": 8.704243236382115e-07, "loss": 0.5133, "step": 4317 }, { "epoch": 4.14402304368699, "grad_norm": 0.8755567146894928, "learning_rate": 8.685365616632175e-07, "loss": 0.5147, "step": 4318 }, { "epoch": 4.14498319731157, "grad_norm": 0.9744866097019197, "learning_rate": 8.666506542579961e-07, "loss": 0.4148, "step": 4319 }, { "epoch": 4.145943350936149, "grad_norm": 1.2128733968602146, "learning_rate": 8.64766602269112e-07, "loss": 0.4477, "step": 4320 }, { "epoch": 4.14690350456073, "grad_norm": 0.8923813920518286, "learning_rate": 8.628844065422992e-07, "loss": 0.5004, "step": 4321 }, { "epoch": 4.14786365818531, "grad_norm": 0.8029226879284792, "learning_rate": 8.610040679224518e-07, "loss": 0.4542, "step": 4322 }, { "epoch": 4.14882381180989, "grad_norm": 0.9389217170208888, "learning_rate": 8.591255872536364e-07, "loss": 0.4687, "step": 4323 }, { "epoch": 4.1497839654344695, "grad_norm": 0.8253540734816273, "learning_rate": 8.572489653790789e-07, "loss": 0.4946, "step": 4324 }, { "epoch": 4.150744119059049, "grad_norm": 0.7968428193142545, "learning_rate": 8.553742031411783e-07, "loss": 0.5109, "step": 4325 }, { "epoch": 4.151704272683629, "grad_norm": 1.1219768932478438, "learning_rate": 8.535013013814952e-07, "loss": 0.4747, "step": 4326 }, { "epoch": 4.15266442630821, "grad_norm": 0.767338725546203, "learning_rate": 8.51630260940754e-07, "loss": 0.4998, "step": 4327 }, { "epoch": 4.1536245799327896, "grad_norm": 0.6505929662959145, "learning_rate": 8.497610826588443e-07, "loss": 0.4058, "step": 4328 }, { "epoch": 4.154584733557369, "grad_norm": 0.9407647229586665, "learning_rate": 8.478937673748211e-07, "loss": 0.4992, "step": 4329 }, { "epoch": 4.155544887181949, "grad_norm": 0.7971087794781138, "learning_rate": 8.460283159269023e-07, "loss": 0.5093, "step": 4330 }, { "epoch": 4.156505040806529, "grad_norm": 0.8475213918648988, "learning_rate": 8.441647291524701e-07, "loss": 0.5298, "step": 4331 }, { "epoch": 4.157465194431109, "grad_norm": 0.847727761994274, "learning_rate": 8.423030078880667e-07, "loss": 0.5553, "step": 4332 }, { "epoch": 4.1584253480556885, "grad_norm": 1.2961891524317506, "learning_rate": 8.404431529693996e-07, "loss": 0.3215, "step": 4333 }, { "epoch": 4.159385501680269, "grad_norm": 1.152377675111764, "learning_rate": 8.385851652313376e-07, "loss": 0.4268, "step": 4334 }, { "epoch": 4.160345655304849, "grad_norm": 1.0738454260959522, "learning_rate": 8.367290455079125e-07, "loss": 0.4092, "step": 4335 }, { "epoch": 4.161305808929429, "grad_norm": 1.0682889207352273, "learning_rate": 8.348747946323165e-07, "loss": 0.4629, "step": 4336 }, { "epoch": 4.162265962554009, "grad_norm": 0.7219985699925275, "learning_rate": 8.330224134369014e-07, "loss": 0.5092, "step": 4337 }, { "epoch": 4.163226116178588, "grad_norm": 0.7096120508659397, "learning_rate": 8.311719027531817e-07, "loss": 0.4698, "step": 4338 }, { "epoch": 4.164186269803168, "grad_norm": 1.060298995154078, "learning_rate": 8.293232634118342e-07, "loss": 0.5496, "step": 4339 }, { "epoch": 4.165146423427748, "grad_norm": 0.8292507823145545, "learning_rate": 8.274764962426912e-07, "loss": 0.518, "step": 4340 }, { "epoch": 4.166106577052329, "grad_norm": 1.1022186888661414, "learning_rate": 8.256316020747479e-07, "loss": 0.4605, "step": 4341 }, { "epoch": 4.1670667306769085, "grad_norm": 0.7517558286438395, "learning_rate": 8.237885817361552e-07, "loss": 0.4304, "step": 4342 }, { "epoch": 4.168026884301488, "grad_norm": 1.03738810409489, "learning_rate": 8.219474360542273e-07, "loss": 0.4224, "step": 4343 }, { "epoch": 4.168987037926068, "grad_norm": 1.0943932061314, "learning_rate": 8.201081658554361e-07, "loss": 0.492, "step": 4344 }, { "epoch": 4.169947191550648, "grad_norm": 0.9398439037911546, "learning_rate": 8.182707719654076e-07, "loss": 0.4646, "step": 4345 }, { "epoch": 4.170907345175228, "grad_norm": 1.0557831973962883, "learning_rate": 8.164352552089288e-07, "loss": 0.542, "step": 4346 }, { "epoch": 4.171867498799808, "grad_norm": 1.4960873820519247, "learning_rate": 8.146016164099446e-07, "loss": 0.461, "step": 4347 }, { "epoch": 4.172827652424388, "grad_norm": 0.8567828032635374, "learning_rate": 8.127698563915554e-07, "loss": 0.4481, "step": 4348 }, { "epoch": 4.173787806048968, "grad_norm": 0.7622107744532078, "learning_rate": 8.109399759760194e-07, "loss": 0.4997, "step": 4349 }, { "epoch": 4.174747959673548, "grad_norm": 0.8747903054407352, "learning_rate": 8.091119759847493e-07, "loss": 0.5031, "step": 4350 }, { "epoch": 4.1757081132981275, "grad_norm": 1.0021226223131108, "learning_rate": 8.07285857238313e-07, "loss": 0.4993, "step": 4351 }, { "epoch": 4.176668266922707, "grad_norm": 0.9766512456871406, "learning_rate": 8.054616205564403e-07, "loss": 0.4825, "step": 4352 }, { "epoch": 4.177628420547288, "grad_norm": 1.0802923243607432, "learning_rate": 8.036392667580084e-07, "loss": 0.4546, "step": 4353 }, { "epoch": 4.178588574171868, "grad_norm": 1.010929786017079, "learning_rate": 8.018187966610535e-07, "loss": 0.4515, "step": 4354 }, { "epoch": 4.179548727796448, "grad_norm": 0.9918745605264955, "learning_rate": 8.000002110827626e-07, "loss": 0.4636, "step": 4355 }, { "epoch": 4.180508881421027, "grad_norm": 0.8661703820219883, "learning_rate": 7.981835108394825e-07, "loss": 0.3948, "step": 4356 }, { "epoch": 4.181469035045607, "grad_norm": 1.2545037805770396, "learning_rate": 7.963686967467111e-07, "loss": 0.4836, "step": 4357 }, { "epoch": 4.182429188670187, "grad_norm": 0.9265969199895173, "learning_rate": 7.945557696190959e-07, "loss": 0.4095, "step": 4358 }, { "epoch": 4.183389342294767, "grad_norm": 3.115606527446755, "learning_rate": 7.927447302704422e-07, "loss": 0.5301, "step": 4359 }, { "epoch": 4.1843494959193475, "grad_norm": 1.0485314410278528, "learning_rate": 7.909355795137058e-07, "loss": 0.4459, "step": 4360 }, { "epoch": 4.185309649543927, "grad_norm": 1.2953861869293886, "learning_rate": 7.891283181609955e-07, "loss": 0.5164, "step": 4361 }, { "epoch": 4.186269803168507, "grad_norm": 0.7719535161294295, "learning_rate": 7.873229470235722e-07, "loss": 0.417, "step": 4362 }, { "epoch": 4.187229956793087, "grad_norm": 0.824735022509581, "learning_rate": 7.855194669118466e-07, "loss": 0.4054, "step": 4363 }, { "epoch": 4.188190110417667, "grad_norm": 1.0284034668454236, "learning_rate": 7.837178786353811e-07, "loss": 0.4261, "step": 4364 }, { "epoch": 4.189150264042246, "grad_norm": 0.8399700517628584, "learning_rate": 7.81918183002891e-07, "loss": 0.4403, "step": 4365 }, { "epoch": 4.190110417666826, "grad_norm": 0.7126618321222217, "learning_rate": 7.801203808222396e-07, "loss": 0.611, "step": 4366 }, { "epoch": 4.191070571291407, "grad_norm": 1.001273513839597, "learning_rate": 7.783244729004425e-07, "loss": 0.5058, "step": 4367 }, { "epoch": 4.192030724915987, "grad_norm": 0.6602405090667326, "learning_rate": 7.765304600436596e-07, "loss": 0.4714, "step": 4368 }, { "epoch": 4.1929908785405665, "grad_norm": 1.0064861188020917, "learning_rate": 7.747383430572081e-07, "loss": 0.4423, "step": 4369 }, { "epoch": 4.193951032165146, "grad_norm": 0.860066880866288, "learning_rate": 7.729481227455498e-07, "loss": 0.5771, "step": 4370 }, { "epoch": 4.194911185789726, "grad_norm": 0.9938875209699272, "learning_rate": 7.711597999122939e-07, "loss": 0.4275, "step": 4371 }, { "epoch": 4.195871339414306, "grad_norm": 0.9475380261905647, "learning_rate": 7.693733753601995e-07, "loss": 0.4705, "step": 4372 }, { "epoch": 4.196831493038887, "grad_norm": 0.6777795504758184, "learning_rate": 7.675888498911743e-07, "loss": 0.4663, "step": 4373 }, { "epoch": 4.197791646663466, "grad_norm": 0.820381803242833, "learning_rate": 7.658062243062725e-07, "loss": 0.4398, "step": 4374 }, { "epoch": 4.198751800288046, "grad_norm": 0.8205873457761504, "learning_rate": 7.640254994056967e-07, "loss": 0.4839, "step": 4375 }, { "epoch": 4.199711953912626, "grad_norm": 0.9249288979695429, "learning_rate": 7.622466759887931e-07, "loss": 0.4871, "step": 4376 }, { "epoch": 4.200672107537206, "grad_norm": 0.7963004776443171, "learning_rate": 7.604697548540574e-07, "loss": 0.4488, "step": 4377 }, { "epoch": 4.201632261161786, "grad_norm": 1.6560719975654141, "learning_rate": 7.58694736799131e-07, "loss": 0.513, "step": 4378 }, { "epoch": 4.202592414786366, "grad_norm": 0.7712617552366521, "learning_rate": 7.569216226207999e-07, "loss": 0.5747, "step": 4379 }, { "epoch": 4.203552568410946, "grad_norm": 1.2749893507046564, "learning_rate": 7.551504131149972e-07, "loss": 0.4065, "step": 4380 }, { "epoch": 4.204512722035526, "grad_norm": 1.138356409499607, "learning_rate": 7.533811090767984e-07, "loss": 0.4511, "step": 4381 }, { "epoch": 4.205472875660106, "grad_norm": 0.8625175238355286, "learning_rate": 7.516137113004246e-07, "loss": 0.526, "step": 4382 }, { "epoch": 4.206433029284685, "grad_norm": 0.845388470480478, "learning_rate": 7.498482205792457e-07, "loss": 0.431, "step": 4383 }, { "epoch": 4.207393182909265, "grad_norm": 0.7775346592468414, "learning_rate": 7.480846377057677e-07, "loss": 0.4577, "step": 4384 }, { "epoch": 4.208353336533845, "grad_norm": 0.7676577597817846, "learning_rate": 7.463229634716462e-07, "loss": 0.5536, "step": 4385 }, { "epoch": 4.209313490158426, "grad_norm": 0.9391733616825814, "learning_rate": 7.445631986676755e-07, "loss": 0.4566, "step": 4386 }, { "epoch": 4.2102736437830055, "grad_norm": 0.8207675678167223, "learning_rate": 7.428053440837973e-07, "loss": 0.4589, "step": 4387 }, { "epoch": 4.211233797407585, "grad_norm": 0.685119671730811, "learning_rate": 7.410494005090946e-07, "loss": 0.4513, "step": 4388 }, { "epoch": 4.212193951032165, "grad_norm": 0.6290608161807941, "learning_rate": 7.392953687317894e-07, "loss": 0.4975, "step": 4389 }, { "epoch": 4.213154104656745, "grad_norm": 0.8141291742065926, "learning_rate": 7.375432495392487e-07, "loss": 0.5275, "step": 4390 }, { "epoch": 4.214114258281325, "grad_norm": 1.0202345411802702, "learning_rate": 7.3579304371798e-07, "loss": 0.3515, "step": 4391 }, { "epoch": 4.215074411905905, "grad_norm": 0.7988872422748813, "learning_rate": 7.34044752053632e-07, "loss": 0.4601, "step": 4392 }, { "epoch": 4.216034565530485, "grad_norm": 1.0528319863858824, "learning_rate": 7.322983753309959e-07, "loss": 0.5201, "step": 4393 }, { "epoch": 4.216994719155065, "grad_norm": 0.8963725342897527, "learning_rate": 7.305539143339985e-07, "loss": 0.4996, "step": 4394 }, { "epoch": 4.217954872779645, "grad_norm": 0.9520218032590222, "learning_rate": 7.2881136984571e-07, "loss": 0.5024, "step": 4395 }, { "epoch": 4.2189150264042246, "grad_norm": 1.2751249391834814, "learning_rate": 7.270707426483431e-07, "loss": 0.4423, "step": 4396 }, { "epoch": 4.219875180028804, "grad_norm": 1.0622299586682014, "learning_rate": 7.253320335232439e-07, "loss": 0.4336, "step": 4397 }, { "epoch": 4.220835333653384, "grad_norm": 1.4411009074751626, "learning_rate": 7.235952432509024e-07, "loss": 0.4069, "step": 4398 }, { "epoch": 4.221795487277965, "grad_norm": 0.7220123461518433, "learning_rate": 7.218603726109413e-07, "loss": 0.431, "step": 4399 }, { "epoch": 4.222755640902545, "grad_norm": 0.8489695730275131, "learning_rate": 7.201274223821298e-07, "loss": 0.4992, "step": 4400 }, { "epoch": 4.223715794527124, "grad_norm": 1.084387040260046, "learning_rate": 7.183963933423699e-07, "loss": 0.4029, "step": 4401 }, { "epoch": 4.224675948151704, "grad_norm": 0.8836643441127857, "learning_rate": 7.166672862687002e-07, "loss": 0.4644, "step": 4402 }, { "epoch": 4.225636101776284, "grad_norm": 0.9251979928275806, "learning_rate": 7.149401019372993e-07, "loss": 0.3911, "step": 4403 }, { "epoch": 4.226596255400864, "grad_norm": 0.6623154418889334, "learning_rate": 7.132148411234818e-07, "loss": 0.4293, "step": 4404 }, { "epoch": 4.2275564090254445, "grad_norm": 0.9607753439217901, "learning_rate": 7.11491504601699e-07, "loss": 0.418, "step": 4405 }, { "epoch": 4.228516562650024, "grad_norm": 0.8400285074996446, "learning_rate": 7.097700931455397e-07, "loss": 0.4838, "step": 4406 }, { "epoch": 4.229476716274604, "grad_norm": 0.869906713492405, "learning_rate": 7.080506075277243e-07, "loss": 0.4196, "step": 4407 }, { "epoch": 4.230436869899184, "grad_norm": 0.78217598029145, "learning_rate": 7.063330485201126e-07, "loss": 0.4558, "step": 4408 }, { "epoch": 4.231397023523764, "grad_norm": 1.026585125193294, "learning_rate": 7.046174168936987e-07, "loss": 0.5347, "step": 4409 }, { "epoch": 4.2323571771483435, "grad_norm": 0.935930442380596, "learning_rate": 7.029037134186112e-07, "loss": 0.4831, "step": 4410 }, { "epoch": 4.233317330772923, "grad_norm": 1.089511121063807, "learning_rate": 7.011919388641148e-07, "loss": 0.3792, "step": 4411 }, { "epoch": 4.234277484397504, "grad_norm": 0.7453956411864053, "learning_rate": 6.994820939986025e-07, "loss": 0.4921, "step": 4412 }, { "epoch": 4.235237638022084, "grad_norm": 0.8874270381582819, "learning_rate": 6.977741795896098e-07, "loss": 0.4153, "step": 4413 }, { "epoch": 4.2361977916466635, "grad_norm": 1.087065078202454, "learning_rate": 6.960681964038008e-07, "loss": 0.4397, "step": 4414 }, { "epoch": 4.237157945271243, "grad_norm": 0.9169909507681471, "learning_rate": 6.943641452069705e-07, "loss": 0.5316, "step": 4415 }, { "epoch": 4.238118098895823, "grad_norm": 1.186971621258769, "learning_rate": 6.92662026764051e-07, "loss": 0.4295, "step": 4416 }, { "epoch": 4.239078252520403, "grad_norm": 1.32476972327495, "learning_rate": 6.909618418391045e-07, "loss": 0.3649, "step": 4417 }, { "epoch": 4.240038406144984, "grad_norm": 1.054021487971042, "learning_rate": 6.892635911953261e-07, "loss": 0.3892, "step": 4418 }, { "epoch": 4.240998559769563, "grad_norm": 1.1690928889788352, "learning_rate": 6.875672755950441e-07, "loss": 0.3629, "step": 4419 }, { "epoch": 4.241958713394143, "grad_norm": 0.9097208202795484, "learning_rate": 6.858728957997129e-07, "loss": 0.3652, "step": 4420 }, { "epoch": 4.242918867018723, "grad_norm": 0.7638190743544068, "learning_rate": 6.841804525699236e-07, "loss": 0.4123, "step": 4421 }, { "epoch": 4.243879020643303, "grad_norm": 0.9966072046985093, "learning_rate": 6.824899466653961e-07, "loss": 0.432, "step": 4422 }, { "epoch": 4.244839174267883, "grad_norm": 1.0805086823850416, "learning_rate": 6.808013788449791e-07, "loss": 0.5114, "step": 4423 }, { "epoch": 4.245799327892462, "grad_norm": 0.7798114386977149, "learning_rate": 6.79114749866655e-07, "loss": 0.4961, "step": 4424 }, { "epoch": 4.246759481517043, "grad_norm": 0.8219751028343714, "learning_rate": 6.774300604875312e-07, "loss": 0.4724, "step": 4425 }, { "epoch": 4.247719635141623, "grad_norm": 1.4883983781026549, "learning_rate": 6.757473114638458e-07, "loss": 0.4702, "step": 4426 }, { "epoch": 4.248679788766203, "grad_norm": 1.134005859662446, "learning_rate": 6.740665035509714e-07, "loss": 0.5138, "step": 4427 }, { "epoch": 4.2496399423907825, "grad_norm": 1.0998448370573124, "learning_rate": 6.723876375034e-07, "loss": 0.3811, "step": 4428 }, { "epoch": 4.250600096015362, "grad_norm": 0.9389674256200014, "learning_rate": 6.707107140747604e-07, "loss": 0.4426, "step": 4429 }, { "epoch": 4.251560249639942, "grad_norm": 0.985661098251187, "learning_rate": 6.690357340178011e-07, "loss": 0.4644, "step": 4430 }, { "epoch": 4.252520403264523, "grad_norm": 1.220460080902016, "learning_rate": 6.673626980844067e-07, "loss": 0.42, "step": 4431 }, { "epoch": 4.2534805568891025, "grad_norm": 1.0071907319693232, "learning_rate": 6.656916070255842e-07, "loss": 0.4016, "step": 4432 }, { "epoch": 4.254440710513682, "grad_norm": 0.8029164973606743, "learning_rate": 6.640224615914681e-07, "loss": 0.5782, "step": 4433 }, { "epoch": 4.255400864138262, "grad_norm": 0.7690745994205141, "learning_rate": 6.62355262531319e-07, "loss": 0.5373, "step": 4434 }, { "epoch": 4.256361017762842, "grad_norm": 0.9772486253433565, "learning_rate": 6.606900105935265e-07, "loss": 0.4978, "step": 4435 }, { "epoch": 4.257321171387422, "grad_norm": 0.7508888553487655, "learning_rate": 6.590267065256029e-07, "loss": 0.5495, "step": 4436 }, { "epoch": 4.2582813250120015, "grad_norm": 0.7610758909385129, "learning_rate": 6.573653510741901e-07, "loss": 0.4297, "step": 4437 }, { "epoch": 4.259241478636582, "grad_norm": 0.7982185442746573, "learning_rate": 6.557059449850494e-07, "loss": 0.3922, "step": 4438 }, { "epoch": 4.260201632261162, "grad_norm": 1.130222566677482, "learning_rate": 6.540484890030724e-07, "loss": 0.4802, "step": 4439 }, { "epoch": 4.261161785885742, "grad_norm": 0.9679865657760506, "learning_rate": 6.523929838722726e-07, "loss": 0.4735, "step": 4440 }, { "epoch": 4.262121939510322, "grad_norm": 1.0377674551638958, "learning_rate": 6.507394303357894e-07, "loss": 0.4616, "step": 4441 }, { "epoch": 4.263082093134901, "grad_norm": 0.8631384041859957, "learning_rate": 6.49087829135886e-07, "loss": 0.4308, "step": 4442 }, { "epoch": 4.264042246759481, "grad_norm": 0.8072355570768749, "learning_rate": 6.474381810139446e-07, "loss": 0.5053, "step": 4443 }, { "epoch": 4.265002400384062, "grad_norm": 0.8788681483869911, "learning_rate": 6.457904867104792e-07, "loss": 0.5138, "step": 4444 }, { "epoch": 4.265962554008642, "grad_norm": 1.0743621413658222, "learning_rate": 6.441447469651213e-07, "loss": 0.5347, "step": 4445 }, { "epoch": 4.2669227076332215, "grad_norm": 0.9925344611080406, "learning_rate": 6.42500962516624e-07, "loss": 0.4976, "step": 4446 }, { "epoch": 4.267882861257801, "grad_norm": 1.2729966140310827, "learning_rate": 6.408591341028652e-07, "loss": 0.5172, "step": 4447 }, { "epoch": 4.268843014882381, "grad_norm": 0.8741285369750141, "learning_rate": 6.392192624608451e-07, "loss": 0.4115, "step": 4448 }, { "epoch": 4.269803168506961, "grad_norm": 0.812933852705911, "learning_rate": 6.375813483266835e-07, "loss": 0.4908, "step": 4449 }, { "epoch": 4.270763322131541, "grad_norm": 1.2483819198121047, "learning_rate": 6.35945392435624e-07, "loss": 0.4544, "step": 4450 }, { "epoch": 4.271723475756121, "grad_norm": 0.8966423733019381, "learning_rate": 6.343113955220276e-07, "loss": 0.5126, "step": 4451 }, { "epoch": 4.272683629380701, "grad_norm": 1.0674688619290444, "learning_rate": 6.326793583193797e-07, "loss": 0.4928, "step": 4452 }, { "epoch": 4.273643783005281, "grad_norm": 0.9244167092262191, "learning_rate": 6.310492815602831e-07, "loss": 0.5357, "step": 4453 }, { "epoch": 4.274603936629861, "grad_norm": 1.467401024130711, "learning_rate": 6.294211659764626e-07, "loss": 0.3694, "step": 4454 }, { "epoch": 4.2755640902544405, "grad_norm": 0.8196226011914708, "learning_rate": 6.277950122987631e-07, "loss": 0.4935, "step": 4455 }, { "epoch": 4.27652424387902, "grad_norm": 0.7129239257653729, "learning_rate": 6.261708212571438e-07, "loss": 0.4355, "step": 4456 }, { "epoch": 4.277484397503601, "grad_norm": 0.7921120497699682, "learning_rate": 6.245485935806905e-07, "loss": 0.4585, "step": 4457 }, { "epoch": 4.278444551128181, "grad_norm": 0.9343360189529025, "learning_rate": 6.229283299976041e-07, "loss": 0.4531, "step": 4458 }, { "epoch": 4.279404704752761, "grad_norm": 0.6915552402041864, "learning_rate": 6.213100312352005e-07, "loss": 0.5298, "step": 4459 }, { "epoch": 4.28036485837734, "grad_norm": 0.9054697171146824, "learning_rate": 6.196936980199181e-07, "loss": 0.5182, "step": 4460 }, { "epoch": 4.28132501200192, "grad_norm": 0.7914502526194045, "learning_rate": 6.18079331077312e-07, "loss": 0.4252, "step": 4461 }, { "epoch": 4.2822851656265, "grad_norm": 1.0489721548620352, "learning_rate": 6.164669311320543e-07, "loss": 0.4643, "step": 4462 }, { "epoch": 4.28324531925108, "grad_norm": 0.8095526680438995, "learning_rate": 6.148564989079347e-07, "loss": 0.5541, "step": 4463 }, { "epoch": 4.2842054728756604, "grad_norm": 1.218818422865881, "learning_rate": 6.13248035127857e-07, "loss": 0.5408, "step": 4464 }, { "epoch": 4.28516562650024, "grad_norm": 0.6931912008846165, "learning_rate": 6.116415405138443e-07, "loss": 0.4213, "step": 4465 }, { "epoch": 4.28612578012482, "grad_norm": 0.8299836549871974, "learning_rate": 6.100370157870356e-07, "loss": 0.5716, "step": 4466 }, { "epoch": 4.2870859337494, "grad_norm": 1.1782594552826686, "learning_rate": 6.08434461667684e-07, "loss": 0.4432, "step": 4467 }, { "epoch": 4.28804608737398, "grad_norm": 0.7800983051206877, "learning_rate": 6.06833878875161e-07, "loss": 0.4372, "step": 4468 }, { "epoch": 4.289006240998559, "grad_norm": 1.1497685967656894, "learning_rate": 6.052352681279478e-07, "loss": 0.4038, "step": 4469 }, { "epoch": 4.28996639462314, "grad_norm": 1.1204947925935986, "learning_rate": 6.036386301436448e-07, "loss": 0.3768, "step": 4470 }, { "epoch": 4.29092654824772, "grad_norm": 0.6621458651316617, "learning_rate": 6.020439656389676e-07, "loss": 0.4478, "step": 4471 }, { "epoch": 4.2918867018723, "grad_norm": 1.0452608456425494, "learning_rate": 6.004512753297421e-07, "loss": 0.4595, "step": 4472 }, { "epoch": 4.2928468554968795, "grad_norm": 0.8421692926451698, "learning_rate": 5.988605599309116e-07, "loss": 0.4687, "step": 4473 }, { "epoch": 4.293807009121459, "grad_norm": 1.0464934306835219, "learning_rate": 5.97271820156528e-07, "loss": 0.4558, "step": 4474 }, { "epoch": 4.294767162746039, "grad_norm": 0.7944724393593119, "learning_rate": 5.95685056719763e-07, "loss": 0.4375, "step": 4475 }, { "epoch": 4.295727316370619, "grad_norm": 0.8012877913611959, "learning_rate": 5.941002703328974e-07, "loss": 0.4891, "step": 4476 }, { "epoch": 4.2966874699952, "grad_norm": 0.8757810295284323, "learning_rate": 5.925174617073226e-07, "loss": 0.5408, "step": 4477 }, { "epoch": 4.297647623619779, "grad_norm": 0.7990762925228717, "learning_rate": 5.909366315535464e-07, "loss": 0.4196, "step": 4478 }, { "epoch": 4.298607777244359, "grad_norm": 0.8835567156433072, "learning_rate": 5.893577805811856e-07, "loss": 0.457, "step": 4479 }, { "epoch": 4.299567930868939, "grad_norm": 0.7282047944844012, "learning_rate": 5.877809094989706e-07, "loss": 0.377, "step": 4480 }, { "epoch": 4.300528084493519, "grad_norm": 0.8699456018372639, "learning_rate": 5.862060190147417e-07, "loss": 0.4783, "step": 4481 }, { "epoch": 4.3014882381180986, "grad_norm": 0.7712568878792925, "learning_rate": 5.84633109835449e-07, "loss": 0.4555, "step": 4482 }, { "epoch": 4.302448391742679, "grad_norm": 0.8843266215878033, "learning_rate": 5.830621826671562e-07, "loss": 0.4135, "step": 4483 }, { "epoch": 4.303408545367259, "grad_norm": 0.7399080432717265, "learning_rate": 5.814932382150346e-07, "loss": 0.4789, "step": 4484 }, { "epoch": 4.304368698991839, "grad_norm": 0.9292161938453902, "learning_rate": 5.799262771833675e-07, "loss": 0.4545, "step": 4485 }, { "epoch": 4.305328852616419, "grad_norm": 1.1018156639356123, "learning_rate": 5.783613002755478e-07, "loss": 0.4362, "step": 4486 }, { "epoch": 4.306289006240998, "grad_norm": 1.0080983476057794, "learning_rate": 5.767983081940731e-07, "loss": 0.4636, "step": 4487 }, { "epoch": 4.307249159865578, "grad_norm": 1.3863123158174753, "learning_rate": 5.752373016405577e-07, "loss": 0.425, "step": 4488 }, { "epoch": 4.308209313490158, "grad_norm": 0.7567497571502831, "learning_rate": 5.736782813157204e-07, "loss": 0.4693, "step": 4489 }, { "epoch": 4.309169467114739, "grad_norm": 1.5559827631453185, "learning_rate": 5.721212479193871e-07, "loss": 0.4857, "step": 4490 }, { "epoch": 4.3101296207393185, "grad_norm": 0.8998571790440143, "learning_rate": 5.705662021504943e-07, "loss": 0.4346, "step": 4491 }, { "epoch": 4.311089774363898, "grad_norm": 0.5980696393031621, "learning_rate": 5.690131447070851e-07, "loss": 0.4699, "step": 4492 }, { "epoch": 4.312049927988478, "grad_norm": 1.0882365392272517, "learning_rate": 5.674620762863109e-07, "loss": 0.4303, "step": 4493 }, { "epoch": 4.313010081613058, "grad_norm": 0.821618249232528, "learning_rate": 5.659129975844302e-07, "loss": 0.4926, "step": 4494 }, { "epoch": 4.313970235237638, "grad_norm": 1.1219676264514598, "learning_rate": 5.643659092968062e-07, "loss": 0.4735, "step": 4495 }, { "epoch": 4.314930388862218, "grad_norm": 0.9621581591145948, "learning_rate": 5.628208121179107e-07, "loss": 0.5299, "step": 4496 }, { "epoch": 4.315890542486798, "grad_norm": 0.8281082360261939, "learning_rate": 5.612777067413228e-07, "loss": 0.4892, "step": 4497 }, { "epoch": 4.316850696111378, "grad_norm": 1.3136339030860675, "learning_rate": 5.597365938597243e-07, "loss": 0.3928, "step": 4498 }, { "epoch": 4.317810849735958, "grad_norm": 1.0080204663616312, "learning_rate": 5.581974741649066e-07, "loss": 0.5206, "step": 4499 }, { "epoch": 4.3187710033605375, "grad_norm": 0.8284239838708933, "learning_rate": 5.566603483477607e-07, "loss": 0.4997, "step": 4500 }, { "epoch": 4.319731156985117, "grad_norm": 0.804925506952228, "learning_rate": 5.551252170982874e-07, "loss": 0.4661, "step": 4501 }, { "epoch": 4.320691310609698, "grad_norm": 0.7923155287077945, "learning_rate": 5.535920811055928e-07, "loss": 0.4448, "step": 4502 }, { "epoch": 4.321651464234278, "grad_norm": 0.6971368638152375, "learning_rate": 5.520609410578826e-07, "loss": 0.4488, "step": 4503 }, { "epoch": 4.322611617858858, "grad_norm": 0.684304261140314, "learning_rate": 5.505317976424712e-07, "loss": 0.4371, "step": 4504 }, { "epoch": 4.323571771483437, "grad_norm": 1.247621335569597, "learning_rate": 5.49004651545772e-07, "loss": 0.4187, "step": 4505 }, { "epoch": 4.324531925108017, "grad_norm": 0.8936914887723734, "learning_rate": 5.474795034533064e-07, "loss": 0.465, "step": 4506 }, { "epoch": 4.325492078732597, "grad_norm": 0.9405662533695208, "learning_rate": 5.459563540496987e-07, "loss": 0.3957, "step": 4507 }, { "epoch": 4.326452232357177, "grad_norm": 0.817863308989369, "learning_rate": 5.444352040186712e-07, "loss": 0.5023, "step": 4508 }, { "epoch": 4.3274123859817575, "grad_norm": 1.0480037735225862, "learning_rate": 5.429160540430533e-07, "loss": 0.4936, "step": 4509 }, { "epoch": 4.328372539606337, "grad_norm": 0.6202808215377612, "learning_rate": 5.413989048047752e-07, "loss": 0.4273, "step": 4510 }, { "epoch": 4.329332693230917, "grad_norm": 1.3024313152690254, "learning_rate": 5.398837569848686e-07, "loss": 0.4234, "step": 4511 }, { "epoch": 4.330292846855497, "grad_norm": 1.1572435702348396, "learning_rate": 5.38370611263469e-07, "loss": 0.4054, "step": 4512 }, { "epoch": 4.331253000480077, "grad_norm": 0.6591713119372713, "learning_rate": 5.368594683198087e-07, "loss": 0.4267, "step": 4513 }, { "epoch": 4.3322131541046565, "grad_norm": 1.0996615860256063, "learning_rate": 5.353503288322243e-07, "loss": 0.5346, "step": 4514 }, { "epoch": 4.333173307729236, "grad_norm": 0.7637184204045931, "learning_rate": 5.338431934781535e-07, "loss": 0.4951, "step": 4515 }, { "epoch": 4.334133461353817, "grad_norm": 0.7196208593657246, "learning_rate": 5.323380629341324e-07, "loss": 0.5921, "step": 4516 }, { "epoch": 4.335093614978397, "grad_norm": 1.3163930598044908, "learning_rate": 5.308349378757988e-07, "loss": 0.4675, "step": 4517 }, { "epoch": 4.3360537686029765, "grad_norm": 0.8562905385120201, "learning_rate": 5.293338189778874e-07, "loss": 0.4977, "step": 4518 }, { "epoch": 4.337013922227556, "grad_norm": 0.7388603843738253, "learning_rate": 5.278347069142364e-07, "loss": 0.5266, "step": 4519 }, { "epoch": 4.337974075852136, "grad_norm": 1.1049342929228856, "learning_rate": 5.263376023577821e-07, "loss": 0.3999, "step": 4520 }, { "epoch": 4.338934229476716, "grad_norm": 1.2271292924347845, "learning_rate": 5.248425059805557e-07, "loss": 0.5543, "step": 4521 }, { "epoch": 4.339894383101297, "grad_norm": 0.6470632673877259, "learning_rate": 5.23349418453692e-07, "loss": 0.4306, "step": 4522 }, { "epoch": 4.340854536725876, "grad_norm": 1.001627720960343, "learning_rate": 5.218583404474214e-07, "loss": 0.451, "step": 4523 }, { "epoch": 4.341814690350456, "grad_norm": 1.0526281089745693, "learning_rate": 5.203692726310738e-07, "loss": 0.3466, "step": 4524 }, { "epoch": 4.342774843975036, "grad_norm": 0.9769541577862471, "learning_rate": 5.188822156730756e-07, "loss": 0.524, "step": 4525 }, { "epoch": 4.343734997599616, "grad_norm": 0.9050474962394479, "learning_rate": 5.173971702409491e-07, "loss": 0.4561, "step": 4526 }, { "epoch": 4.344695151224196, "grad_norm": 1.0298561002940563, "learning_rate": 5.159141370013165e-07, "loss": 0.4831, "step": 4527 }, { "epoch": 4.345655304848776, "grad_norm": 0.7536993355792181, "learning_rate": 5.144331166198963e-07, "loss": 0.4678, "step": 4528 }, { "epoch": 4.346615458473356, "grad_norm": 0.9586263221215651, "learning_rate": 5.129541097615009e-07, "loss": 0.423, "step": 4529 }, { "epoch": 4.347575612097936, "grad_norm": 0.9980567874170491, "learning_rate": 5.114771170900434e-07, "loss": 0.5299, "step": 4530 }, { "epoch": 4.348535765722516, "grad_norm": 1.0851419919907197, "learning_rate": 5.100021392685267e-07, "loss": 0.5491, "step": 4531 }, { "epoch": 4.3494959193470955, "grad_norm": 1.0971351260734432, "learning_rate": 5.08529176959055e-07, "loss": 0.5203, "step": 4532 }, { "epoch": 4.350456072971675, "grad_norm": 0.6963690682946552, "learning_rate": 5.070582308228256e-07, "loss": 0.4938, "step": 4533 }, { "epoch": 4.351416226596255, "grad_norm": 1.3846834233880791, "learning_rate": 5.05589301520128e-07, "loss": 0.4458, "step": 4534 }, { "epoch": 4.352376380220836, "grad_norm": 0.7901158393364703, "learning_rate": 5.041223897103509e-07, "loss": 0.555, "step": 4535 }, { "epoch": 4.3533365338454155, "grad_norm": 0.8853943444413988, "learning_rate": 5.026574960519747e-07, "loss": 0.4415, "step": 4536 }, { "epoch": 4.354296687469995, "grad_norm": 0.602231456393096, "learning_rate": 5.011946212025747e-07, "loss": 0.4516, "step": 4537 }, { "epoch": 4.355256841094575, "grad_norm": 1.034319500036709, "learning_rate": 4.997337658188207e-07, "loss": 0.4796, "step": 4538 }, { "epoch": 4.356216994719155, "grad_norm": 0.8471901058554425, "learning_rate": 4.982749305564727e-07, "loss": 0.5111, "step": 4539 }, { "epoch": 4.357177148343735, "grad_norm": 1.1897784259517195, "learning_rate": 4.968181160703877e-07, "loss": 0.4267, "step": 4540 }, { "epoch": 4.3581373019683145, "grad_norm": 0.8225370602498585, "learning_rate": 4.953633230145144e-07, "loss": 0.4843, "step": 4541 }, { "epoch": 4.359097455592895, "grad_norm": 0.9771849003424808, "learning_rate": 4.93910552041893e-07, "loss": 0.434, "step": 4542 }, { "epoch": 4.360057609217475, "grad_norm": 1.8970678010727016, "learning_rate": 4.924598038046591e-07, "loss": 0.403, "step": 4543 }, { "epoch": 4.361017762842055, "grad_norm": 0.9197146860608542, "learning_rate": 4.910110789540346e-07, "loss": 0.498, "step": 4544 }, { "epoch": 4.361977916466635, "grad_norm": 0.7968305932461855, "learning_rate": 4.895643781403375e-07, "loss": 0.4719, "step": 4545 }, { "epoch": 4.362938070091214, "grad_norm": 0.6999752667824466, "learning_rate": 4.881197020129797e-07, "loss": 0.4791, "step": 4546 }, { "epoch": 4.363898223715794, "grad_norm": 0.9555862836316861, "learning_rate": 4.866770512204566e-07, "loss": 0.4408, "step": 4547 }, { "epoch": 4.364858377340375, "grad_norm": 0.9653493122305095, "learning_rate": 4.852364264103621e-07, "loss": 0.5599, "step": 4548 }, { "epoch": 4.365818530964955, "grad_norm": 0.7845407699383455, "learning_rate": 4.837978282293731e-07, "loss": 0.4619, "step": 4549 }, { "epoch": 4.366778684589534, "grad_norm": 0.8718294623941049, "learning_rate": 4.823612573232644e-07, "loss": 0.433, "step": 4550 }, { "epoch": 4.367738838214114, "grad_norm": 0.9194479203769943, "learning_rate": 4.809267143368979e-07, "loss": 0.5156, "step": 4551 }, { "epoch": 4.368698991838694, "grad_norm": 0.8345914137501059, "learning_rate": 4.794941999142222e-07, "loss": 0.4435, "step": 4552 }, { "epoch": 4.369659145463274, "grad_norm": 0.9238645929509001, "learning_rate": 4.780637146982781e-07, "loss": 0.3996, "step": 4553 }, { "epoch": 4.3706192990878545, "grad_norm": 1.0074710991685167, "learning_rate": 4.7663525933119625e-07, "loss": 0.4779, "step": 4554 }, { "epoch": 4.371579452712434, "grad_norm": 0.9157918117212657, "learning_rate": 4.7520883445419387e-07, "loss": 0.5406, "step": 4555 }, { "epoch": 4.372539606337014, "grad_norm": 1.0741625200909546, "learning_rate": 4.737844407075798e-07, "loss": 0.5226, "step": 4556 }, { "epoch": 4.373499759961594, "grad_norm": 0.8694928604792933, "learning_rate": 4.723620787307465e-07, "loss": 0.4507, "step": 4557 }, { "epoch": 4.374459913586174, "grad_norm": 0.9572401396513633, "learning_rate": 4.7094174916217817e-07, "loss": 0.496, "step": 4558 }, { "epoch": 4.3754200672107535, "grad_norm": 1.1804178171890836, "learning_rate": 4.69523452639446e-07, "loss": 0.4166, "step": 4559 }, { "epoch": 4.376380220835333, "grad_norm": 0.9552319711141223, "learning_rate": 4.6810718979920734e-07, "loss": 0.545, "step": 4560 }, { "epoch": 4.377340374459914, "grad_norm": 1.5943527428062239, "learning_rate": 4.666929612772092e-07, "loss": 0.3888, "step": 4561 }, { "epoch": 4.378300528084494, "grad_norm": 0.9569683006132433, "learning_rate": 4.6528076770828045e-07, "loss": 0.506, "step": 4562 }, { "epoch": 4.379260681709074, "grad_norm": 0.7046810563343769, "learning_rate": 4.638706097263429e-07, "loss": 0.4455, "step": 4563 }, { "epoch": 4.380220835333653, "grad_norm": 0.881022983806149, "learning_rate": 4.6246248796440074e-07, "loss": 0.4612, "step": 4564 }, { "epoch": 4.381180988958233, "grad_norm": 0.8102500045547492, "learning_rate": 4.6105640305454337e-07, "loss": 0.4543, "step": 4565 }, { "epoch": 4.382141142582813, "grad_norm": 0.8440162390991032, "learning_rate": 4.5965235562794754e-07, "loss": 0.5022, "step": 4566 }, { "epoch": 4.383101296207393, "grad_norm": 1.0947030774986928, "learning_rate": 4.582503463148763e-07, "loss": 0.3652, "step": 4567 }, { "epoch": 4.384061449831973, "grad_norm": 0.9337672182894257, "learning_rate": 4.5685037574467574e-07, "loss": 0.5946, "step": 4568 }, { "epoch": 4.385021603456553, "grad_norm": 0.7064859438554193, "learning_rate": 4.5545244454577863e-07, "loss": 0.4323, "step": 4569 }, { "epoch": 4.385981757081133, "grad_norm": 1.000058748152272, "learning_rate": 4.540565533457003e-07, "loss": 0.5023, "step": 4570 }, { "epoch": 4.386941910705713, "grad_norm": 1.0433341050500553, "learning_rate": 4.5266270277104116e-07, "loss": 0.3952, "step": 4571 }, { "epoch": 4.387902064330293, "grad_norm": 0.7206654477429731, "learning_rate": 4.512708934474863e-07, "loss": 0.4906, "step": 4572 }, { "epoch": 4.388862217954872, "grad_norm": 1.1667308775532657, "learning_rate": 4.498811259998048e-07, "loss": 0.4071, "step": 4573 }, { "epoch": 4.389822371579453, "grad_norm": 1.219979134468435, "learning_rate": 4.4849340105184823e-07, "loss": 0.5091, "step": 4574 }, { "epoch": 4.390782525204033, "grad_norm": 0.7307838722136581, "learning_rate": 4.4710771922654936e-07, "loss": 0.5876, "step": 4575 }, { "epoch": 4.391742678828613, "grad_norm": 1.1638844953959897, "learning_rate": 4.4572408114592845e-07, "loss": 0.4845, "step": 4576 }, { "epoch": 4.3927028324531925, "grad_norm": 1.1670921867820807, "learning_rate": 4.4434248743108587e-07, "loss": 0.4701, "step": 4577 }, { "epoch": 4.393662986077772, "grad_norm": 1.0497465760887867, "learning_rate": 4.4296293870220276e-07, "loss": 0.4417, "step": 4578 }, { "epoch": 4.394623139702352, "grad_norm": 1.2671245642199545, "learning_rate": 4.415854355785443e-07, "loss": 0.492, "step": 4579 }, { "epoch": 4.395583293326933, "grad_norm": 0.9739920129023947, "learning_rate": 4.4020997867845695e-07, "loss": 0.5318, "step": 4580 }, { "epoch": 4.396543446951513, "grad_norm": 1.0100999504654438, "learning_rate": 4.388365686193685e-07, "loss": 0.6324, "step": 4581 }, { "epoch": 4.397503600576092, "grad_norm": 0.6921004776603408, "learning_rate": 4.3746520601778974e-07, "loss": 0.5197, "step": 4582 }, { "epoch": 4.398463754200672, "grad_norm": 0.7095110685196567, "learning_rate": 4.3609589148930766e-07, "loss": 0.5405, "step": 4583 }, { "epoch": 4.399423907825252, "grad_norm": 0.6861492353586163, "learning_rate": 4.3472862564859496e-07, "loss": 0.424, "step": 4584 }, { "epoch": 4.400384061449832, "grad_norm": 0.9001261664034099, "learning_rate": 4.3336340910940186e-07, "loss": 0.5082, "step": 4585 }, { "epoch": 4.4013442150744115, "grad_norm": 0.9244674335376083, "learning_rate": 4.320002424845604e-07, "loss": 0.3346, "step": 4586 }, { "epoch": 4.402304368698992, "grad_norm": 0.9018396794387906, "learning_rate": 4.3063912638598106e-07, "loss": 0.524, "step": 4587 }, { "epoch": 4.403264522323572, "grad_norm": 0.9750173671254913, "learning_rate": 4.292800614246534e-07, "loss": 0.4108, "step": 4588 }, { "epoch": 4.404224675948152, "grad_norm": 1.1370322844230303, "learning_rate": 4.2792304821064833e-07, "loss": 0.4698, "step": 4589 }, { "epoch": 4.405184829572732, "grad_norm": 0.9769893079805538, "learning_rate": 4.265680873531136e-07, "loss": 0.4174, "step": 4590 }, { "epoch": 4.406144983197311, "grad_norm": 0.8320208507652684, "learning_rate": 4.2521517946027715e-07, "loss": 0.5072, "step": 4591 }, { "epoch": 4.407105136821891, "grad_norm": 1.0570553352945782, "learning_rate": 4.2386432513944587e-07, "loss": 0.4892, "step": 4592 }, { "epoch": 4.408065290446471, "grad_norm": 0.7613706769138198, "learning_rate": 4.225155249970014e-07, "loss": 0.492, "step": 4593 }, { "epoch": 4.409025444071052, "grad_norm": 1.1258563357947768, "learning_rate": 4.2116877963840786e-07, "loss": 0.4209, "step": 4594 }, { "epoch": 4.4099855976956315, "grad_norm": 0.9304462398903209, "learning_rate": 4.198240896682049e-07, "loss": 0.4894, "step": 4595 }, { "epoch": 4.410945751320211, "grad_norm": 0.9174315436451229, "learning_rate": 4.18481455690008e-07, "loss": 0.3816, "step": 4596 }, { "epoch": 4.411905904944791, "grad_norm": 1.0960306716177322, "learning_rate": 4.1714087830651226e-07, "loss": 0.5308, "step": 4597 }, { "epoch": 4.412866058569371, "grad_norm": 0.7414363795650125, "learning_rate": 4.1580235811948753e-07, "loss": 0.4396, "step": 4598 }, { "epoch": 4.413826212193951, "grad_norm": 0.9442338959600525, "learning_rate": 4.14465895729782e-07, "loss": 0.4193, "step": 4599 }, { "epoch": 4.414786365818531, "grad_norm": 0.6792246417737602, "learning_rate": 4.131314917373208e-07, "loss": 0.3959, "step": 4600 }, { "epoch": 4.415746519443111, "grad_norm": 1.1085040693383075, "learning_rate": 4.1179914674110035e-07, "loss": 0.4461, "step": 4601 }, { "epoch": 4.416706673067691, "grad_norm": 0.6979880567121513, "learning_rate": 4.1046886133919825e-07, "loss": 0.4779, "step": 4602 }, { "epoch": 4.417666826692271, "grad_norm": 0.6955177206767592, "learning_rate": 4.0914063612876473e-07, "loss": 0.4794, "step": 4603 }, { "epoch": 4.4186269803168505, "grad_norm": 0.6966952054797234, "learning_rate": 4.0781447170602617e-07, "loss": 0.5354, "step": 4604 }, { "epoch": 4.41958713394143, "grad_norm": 0.6860239507508614, "learning_rate": 4.064903686662841e-07, "loss": 0.4851, "step": 4605 }, { "epoch": 4.420547287566011, "grad_norm": 0.8027421159840069, "learning_rate": 4.0516832760391256e-07, "loss": 0.5406, "step": 4606 }, { "epoch": 4.421507441190591, "grad_norm": 0.9960355129250067, "learning_rate": 4.0384834911236294e-07, "loss": 0.4064, "step": 4607 }, { "epoch": 4.422467594815171, "grad_norm": 1.00955609928315, "learning_rate": 4.025304337841607e-07, "loss": 0.4818, "step": 4608 }, { "epoch": 4.42342774843975, "grad_norm": 1.288264867095413, "learning_rate": 4.012145822109015e-07, "loss": 0.4916, "step": 4609 }, { "epoch": 4.42438790206433, "grad_norm": 0.8627261732561112, "learning_rate": 3.9990079498325786e-07, "loss": 0.3826, "step": 4610 }, { "epoch": 4.42534805568891, "grad_norm": 0.7365287820158971, "learning_rate": 3.985890726909747e-07, "loss": 0.4866, "step": 4611 }, { "epoch": 4.42630820931349, "grad_norm": 0.841081032382601, "learning_rate": 3.97279415922871e-07, "loss": 0.5195, "step": 4612 }, { "epoch": 4.4272683629380705, "grad_norm": 0.8236988899765743, "learning_rate": 3.9597182526683763e-07, "loss": 0.4433, "step": 4613 }, { "epoch": 4.42822851656265, "grad_norm": 0.9128688542339323, "learning_rate": 3.946663013098373e-07, "loss": 0.5101, "step": 4614 }, { "epoch": 4.42918867018723, "grad_norm": 0.9178548549379767, "learning_rate": 3.9336284463790566e-07, "loss": 0.495, "step": 4615 }, { "epoch": 4.43014882381181, "grad_norm": 0.8589950112269935, "learning_rate": 3.9206145583615076e-07, "loss": 0.3856, "step": 4616 }, { "epoch": 4.43110897743639, "grad_norm": 0.7430108356925259, "learning_rate": 3.90762135488752e-07, "loss": 0.5004, "step": 4617 }, { "epoch": 4.4320691310609694, "grad_norm": 1.0085382949051884, "learning_rate": 3.894648841789617e-07, "loss": 0.5009, "step": 4618 }, { "epoch": 4.433029284685549, "grad_norm": 0.859294648538662, "learning_rate": 3.881697024890996e-07, "loss": 0.5052, "step": 4619 }, { "epoch": 4.43398943831013, "grad_norm": 0.9012964131044574, "learning_rate": 3.8687659100056063e-07, "loss": 0.4207, "step": 4620 }, { "epoch": 4.43494959193471, "grad_norm": 0.8795221396358546, "learning_rate": 3.8558555029380937e-07, "loss": 0.6013, "step": 4621 }, { "epoch": 4.4359097455592895, "grad_norm": 0.7558484153378103, "learning_rate": 3.8429658094837896e-07, "loss": 0.4036, "step": 4622 }, { "epoch": 4.436869899183869, "grad_norm": 0.8012093300071795, "learning_rate": 3.830096835428737e-07, "loss": 0.4862, "step": 4623 }, { "epoch": 4.437830052808449, "grad_norm": 0.9818170043074644, "learning_rate": 3.8172485865496876e-07, "loss": 0.406, "step": 4624 }, { "epoch": 4.438790206433029, "grad_norm": 0.7639358753906593, "learning_rate": 3.8044210686140784e-07, "loss": 0.474, "step": 4625 }, { "epoch": 4.43975036005761, "grad_norm": 1.5501917579921887, "learning_rate": 3.791614287380063e-07, "loss": 0.504, "step": 4626 }, { "epoch": 4.440710513682189, "grad_norm": 0.8727154498756585, "learning_rate": 3.778828248596439e-07, "loss": 0.5413, "step": 4627 }, { "epoch": 4.441670667306769, "grad_norm": 0.6540098005510645, "learning_rate": 3.766062958002742e-07, "loss": 0.4489, "step": 4628 }, { "epoch": 4.442630820931349, "grad_norm": 1.7968404549353034, "learning_rate": 3.7533184213291665e-07, "loss": 0.521, "step": 4629 }, { "epoch": 4.443590974555929, "grad_norm": 1.078098124757566, "learning_rate": 3.7405946442966023e-07, "loss": 0.3743, "step": 4630 }, { "epoch": 4.444551128180509, "grad_norm": 0.9878419796315229, "learning_rate": 3.727891632616626e-07, "loss": 0.4775, "step": 4631 }, { "epoch": 4.445511281805089, "grad_norm": 0.9586246600855642, "learning_rate": 3.7152093919914644e-07, "loss": 0.514, "step": 4632 }, { "epoch": 4.446471435429669, "grad_norm": 0.7990135577799713, "learning_rate": 3.702547928114053e-07, "loss": 0.4187, "step": 4633 }, { "epoch": 4.447431589054249, "grad_norm": 0.8466734052211173, "learning_rate": 3.689907246667984e-07, "loss": 0.4618, "step": 4634 }, { "epoch": 4.448391742678829, "grad_norm": 0.7815511380687299, "learning_rate": 3.677287353327519e-07, "loss": 0.5996, "step": 4635 }, { "epoch": 4.449351896303408, "grad_norm": 0.8561517403146738, "learning_rate": 3.6646882537576144e-07, "loss": 0.4412, "step": 4636 }, { "epoch": 4.450312049927988, "grad_norm": 1.100369671945681, "learning_rate": 3.652109953613836e-07, "loss": 0.5255, "step": 4637 }, { "epoch": 4.451272203552568, "grad_norm": 1.0898813742995235, "learning_rate": 3.639552458542478e-07, "loss": 0.4507, "step": 4638 }, { "epoch": 4.452232357177149, "grad_norm": 0.8423057123424196, "learning_rate": 3.627015774180459e-07, "loss": 0.5209, "step": 4639 }, { "epoch": 4.4531925108017285, "grad_norm": 0.8884012103405522, "learning_rate": 3.614499906155355e-07, "loss": 0.4154, "step": 4640 }, { "epoch": 4.454152664426308, "grad_norm": 0.5849490187784098, "learning_rate": 3.602004860085406e-07, "loss": 0.3428, "step": 4641 }, { "epoch": 4.455112818050888, "grad_norm": 0.6899709894862649, "learning_rate": 3.589530641579514e-07, "loss": 0.4491, "step": 4642 }, { "epoch": 4.456072971675468, "grad_norm": 0.7497577166145533, "learning_rate": 3.577077256237216e-07, "loss": 0.5787, "step": 4643 }, { "epoch": 4.457033125300048, "grad_norm": 0.7805055586647875, "learning_rate": 3.564644709648707e-07, "loss": 0.4892, "step": 4644 }, { "epoch": 4.4579932789246275, "grad_norm": 0.8294065662880638, "learning_rate": 3.5522330073948177e-07, "loss": 0.5218, "step": 4645 }, { "epoch": 4.458953432549208, "grad_norm": 0.8416989703173368, "learning_rate": 3.53984215504703e-07, "loss": 0.4541, "step": 4646 }, { "epoch": 4.459913586173788, "grad_norm": 1.3813349607390601, "learning_rate": 3.5274721581674664e-07, "loss": 0.3864, "step": 4647 }, { "epoch": 4.460873739798368, "grad_norm": 1.1274931996623323, "learning_rate": 3.5151230223088915e-07, "loss": 0.4255, "step": 4648 }, { "epoch": 4.461833893422948, "grad_norm": 0.7803692504497751, "learning_rate": 3.5027947530147045e-07, "loss": 0.4549, "step": 4649 }, { "epoch": 4.462794047047527, "grad_norm": 1.0158645651031917, "learning_rate": 3.4904873558189055e-07, "loss": 0.5275, "step": 4650 }, { "epoch": 4.463754200672107, "grad_norm": 0.7948541240835973, "learning_rate": 3.4782008362461817e-07, "loss": 0.5183, "step": 4651 }, { "epoch": 4.464714354296688, "grad_norm": 0.6927706512951384, "learning_rate": 3.4659351998118204e-07, "loss": 0.5501, "step": 4652 }, { "epoch": 4.465674507921268, "grad_norm": 0.777592563157979, "learning_rate": 3.4536904520217184e-07, "loss": 0.3367, "step": 4653 }, { "epoch": 4.466634661545847, "grad_norm": 1.153242911318788, "learning_rate": 3.441466598372423e-07, "loss": 0.4849, "step": 4654 }, { "epoch": 4.467594815170427, "grad_norm": 0.8391757539405112, "learning_rate": 3.429263644351083e-07, "loss": 0.5179, "step": 4655 }, { "epoch": 4.468554968795007, "grad_norm": 0.9675327561836184, "learning_rate": 3.417081595435484e-07, "loss": 0.4937, "step": 4656 }, { "epoch": 4.469515122419587, "grad_norm": 0.9962024920945941, "learning_rate": 3.4049204570940185e-07, "loss": 0.4631, "step": 4657 }, { "epoch": 4.4704752760441675, "grad_norm": 1.0010798787563218, "learning_rate": 3.392780234785681e-07, "loss": 0.4, "step": 4658 }, { "epoch": 4.471435429668747, "grad_norm": 1.1081626224729024, "learning_rate": 3.3806609339600894e-07, "loss": 0.4579, "step": 4659 }, { "epoch": 4.472395583293327, "grad_norm": 0.8551963664465597, "learning_rate": 3.36856256005747e-07, "loss": 0.3826, "step": 4660 }, { "epoch": 4.473355736917907, "grad_norm": 0.7853608476590042, "learning_rate": 3.356485118508657e-07, "loss": 0.4655, "step": 4661 }, { "epoch": 4.474315890542487, "grad_norm": 0.9169614818928382, "learning_rate": 3.3444286147350955e-07, "loss": 0.4251, "step": 4662 }, { "epoch": 4.4752760441670665, "grad_norm": 1.4729234571023984, "learning_rate": 3.332393054148786e-07, "loss": 0.3657, "step": 4663 }, { "epoch": 4.476236197791646, "grad_norm": 0.8262459139422418, "learning_rate": 3.3203784421524056e-07, "loss": 0.6147, "step": 4664 }, { "epoch": 4.477196351416227, "grad_norm": 1.1515809489003457, "learning_rate": 3.3083847841391514e-07, "loss": 0.3568, "step": 4665 }, { "epoch": 4.478156505040807, "grad_norm": 1.137528199914808, "learning_rate": 3.296412085492867e-07, "loss": 0.4909, "step": 4666 }, { "epoch": 4.4791166586653866, "grad_norm": 0.8420517951195922, "learning_rate": 3.2844603515879647e-07, "loss": 0.5013, "step": 4667 }, { "epoch": 4.480076812289966, "grad_norm": 0.792544017473837, "learning_rate": 3.272529587789447e-07, "loss": 0.4057, "step": 4668 }, { "epoch": 4.481036965914546, "grad_norm": 0.9807372537032139, "learning_rate": 3.2606197994529076e-07, "loss": 0.4225, "step": 4669 }, { "epoch": 4.481997119539126, "grad_norm": 0.7743790858360675, "learning_rate": 3.2487309919245313e-07, "loss": 0.5227, "step": 4670 }, { "epoch": 4.482957273163706, "grad_norm": 0.6685528032885935, "learning_rate": 3.236863170541066e-07, "loss": 0.469, "step": 4671 }, { "epoch": 4.483917426788286, "grad_norm": 1.1198481486403835, "learning_rate": 3.22501634062985e-07, "loss": 0.3881, "step": 4672 }, { "epoch": 4.484877580412866, "grad_norm": 1.1130303130155539, "learning_rate": 3.213190507508801e-07, "loss": 0.4082, "step": 4673 }, { "epoch": 4.485837734037446, "grad_norm": 0.7486465635055354, "learning_rate": 3.2013856764864126e-07, "loss": 0.5588, "step": 4674 }, { "epoch": 4.486797887662026, "grad_norm": 0.6757741513511389, "learning_rate": 3.1896018528617456e-07, "loss": 0.4911, "step": 4675 }, { "epoch": 4.487758041286606, "grad_norm": 1.195101585782354, "learning_rate": 3.17783904192443e-07, "loss": 0.4878, "step": 4676 }, { "epoch": 4.488718194911185, "grad_norm": 1.3178635543310662, "learning_rate": 3.166097248954669e-07, "loss": 0.4155, "step": 4677 }, { "epoch": 4.489678348535766, "grad_norm": 1.0809763891034094, "learning_rate": 3.15437647922322e-07, "loss": 0.4846, "step": 4678 }, { "epoch": 4.490638502160346, "grad_norm": 0.8835463822276124, "learning_rate": 3.1426767379914234e-07, "loss": 0.4543, "step": 4679 }, { "epoch": 4.491598655784926, "grad_norm": 1.2018887930842341, "learning_rate": 3.1309980305111674e-07, "loss": 0.3971, "step": 4680 }, { "epoch": 4.4925588094095055, "grad_norm": 0.7662212468284699, "learning_rate": 3.119340362024875e-07, "loss": 0.5208, "step": 4681 }, { "epoch": 4.493518963034085, "grad_norm": 1.2157695002492968, "learning_rate": 3.1077037377655827e-07, "loss": 0.374, "step": 4682 }, { "epoch": 4.494479116658665, "grad_norm": 0.7498832957178717, "learning_rate": 3.096088162956834e-07, "loss": 0.4827, "step": 4683 }, { "epoch": 4.495439270283246, "grad_norm": 1.0511882500538048, "learning_rate": 3.084493642812725e-07, "loss": 0.5083, "step": 4684 }, { "epoch": 4.4963994239078255, "grad_norm": 0.6360742922544534, "learning_rate": 3.072920182537925e-07, "loss": 0.5096, "step": 4685 }, { "epoch": 4.497359577532405, "grad_norm": 0.7539362692116929, "learning_rate": 3.0613677873276293e-07, "loss": 0.4374, "step": 4686 }, { "epoch": 4.498319731156985, "grad_norm": 1.0196258934449975, "learning_rate": 3.049836462367589e-07, "loss": 0.4875, "step": 4687 }, { "epoch": 4.499279884781565, "grad_norm": 0.7671046097249988, "learning_rate": 3.038326212834103e-07, "loss": 0.3707, "step": 4688 }, { "epoch": 4.500240038406145, "grad_norm": 1.3051760779676957, "learning_rate": 3.0268370438939767e-07, "loss": 0.5097, "step": 4689 }, { "epoch": 4.5012001920307245, "grad_norm": 1.0068035258640196, "learning_rate": 3.015368960704584e-07, "loss": 0.3587, "step": 4690 }, { "epoch": 4.502160345655305, "grad_norm": 0.9182913667498468, "learning_rate": 3.003921968413831e-07, "loss": 0.391, "step": 4691 }, { "epoch": 4.503120499279885, "grad_norm": 0.6839597491410151, "learning_rate": 2.9924960721601393e-07, "loss": 0.4111, "step": 4692 }, { "epoch": 4.504080652904465, "grad_norm": 0.6062812543525237, "learning_rate": 2.98109127707249e-07, "loss": 0.4979, "step": 4693 }, { "epoch": 4.505040806529045, "grad_norm": 1.6712296005201992, "learning_rate": 2.9697075882703486e-07, "loss": 0.3678, "step": 4694 }, { "epoch": 4.506000960153624, "grad_norm": 1.2656523423447643, "learning_rate": 2.95834501086375e-07, "loss": 0.3403, "step": 4695 }, { "epoch": 4.506961113778204, "grad_norm": 0.7182542113899034, "learning_rate": 2.9470035499532425e-07, "loss": 0.4803, "step": 4696 }, { "epoch": 4.507921267402784, "grad_norm": 0.8083309731853684, "learning_rate": 2.9356832106298584e-07, "loss": 0.5434, "step": 4697 }, { "epoch": 4.508881421027365, "grad_norm": 0.8070624614512283, "learning_rate": 2.9243839979751943e-07, "loss": 0.4848, "step": 4698 }, { "epoch": 4.5098415746519445, "grad_norm": 0.6901320451610856, "learning_rate": 2.913105917061343e-07, "loss": 0.5119, "step": 4699 }, { "epoch": 4.510801728276524, "grad_norm": 0.938817686044144, "learning_rate": 2.9018489729509103e-07, "loss": 0.4944, "step": 4700 }, { "epoch": 4.511761881901104, "grad_norm": 0.9621225759041194, "learning_rate": 2.8906131706970375e-07, "loss": 0.4456, "step": 4701 }, { "epoch": 4.512722035525684, "grad_norm": 0.6706475140518059, "learning_rate": 2.879398515343329e-07, "loss": 0.3987, "step": 4702 }, { "epoch": 4.5136821891502645, "grad_norm": 0.7172578880417797, "learning_rate": 2.8682050119239357e-07, "loss": 0.5275, "step": 4703 }, { "epoch": 4.514642342774844, "grad_norm": 0.9188994424947045, "learning_rate": 2.857032665463499e-07, "loss": 0.5248, "step": 4704 }, { "epoch": 4.515602496399424, "grad_norm": 1.5332656773967392, "learning_rate": 2.8458814809771694e-07, "loss": 0.3083, "step": 4705 }, { "epoch": 4.516562650024004, "grad_norm": 1.0640230364705583, "learning_rate": 2.834751463470603e-07, "loss": 0.4018, "step": 4706 }, { "epoch": 4.517522803648584, "grad_norm": 1.0159681497064863, "learning_rate": 2.823642617939926e-07, "loss": 0.4722, "step": 4707 }, { "epoch": 4.5184829572731635, "grad_norm": 0.971650897517889, "learning_rate": 2.8125549493717984e-07, "loss": 0.5032, "step": 4708 }, { "epoch": 4.519443110897743, "grad_norm": 0.8656553211977246, "learning_rate": 2.8014884627433404e-07, "loss": 0.4263, "step": 4709 }, { "epoch": 4.520403264522324, "grad_norm": 0.8346499955777299, "learning_rate": 2.7904431630221937e-07, "loss": 0.495, "step": 4710 }, { "epoch": 4.521363418146904, "grad_norm": 0.9087800840046435, "learning_rate": 2.7794190551664604e-07, "loss": 0.5161, "step": 4711 }, { "epoch": 4.522323571771484, "grad_norm": 1.173634663890911, "learning_rate": 2.7684161441247584e-07, "loss": 0.4699, "step": 4712 }, { "epoch": 4.523283725396063, "grad_norm": 0.7419171110406887, "learning_rate": 2.75743443483617e-07, "loss": 0.5054, "step": 4713 }, { "epoch": 4.524243879020643, "grad_norm": 1.2898707211294664, "learning_rate": 2.746473932230281e-07, "loss": 0.3692, "step": 4714 }, { "epoch": 4.525204032645223, "grad_norm": 0.7192520699752167, "learning_rate": 2.7355346412271223e-07, "loss": 0.5467, "step": 4715 }, { "epoch": 4.526164186269803, "grad_norm": 0.5891164599564385, "learning_rate": 2.7246165667372314e-07, "loss": 0.4376, "step": 4716 }, { "epoch": 4.5271243398943835, "grad_norm": 0.6464015190574306, "learning_rate": 2.713719713661617e-07, "loss": 0.4209, "step": 4717 }, { "epoch": 4.528084493518963, "grad_norm": 0.9479346133824019, "learning_rate": 2.702844086891765e-07, "loss": 0.4696, "step": 4718 }, { "epoch": 4.529044647143543, "grad_norm": 0.7692969366860155, "learning_rate": 2.6919896913096264e-07, "loss": 0.4611, "step": 4719 }, { "epoch": 4.530004800768123, "grad_norm": 1.0359383875589283, "learning_rate": 2.6811565317876154e-07, "loss": 0.444, "step": 4720 }, { "epoch": 4.530964954392703, "grad_norm": 1.0248052735375186, "learning_rate": 2.67034461318863e-07, "loss": 0.4366, "step": 4721 }, { "epoch": 4.531925108017282, "grad_norm": 1.1272436048737384, "learning_rate": 2.659553940366016e-07, "loss": 0.4304, "step": 4722 }, { "epoch": 4.532885261641862, "grad_norm": 0.9019522064521596, "learning_rate": 2.6487845181635994e-07, "loss": 0.525, "step": 4723 }, { "epoch": 4.533845415266443, "grad_norm": 0.9234870254059822, "learning_rate": 2.638036351415668e-07, "loss": 0.4601, "step": 4724 }, { "epoch": 4.534805568891023, "grad_norm": 0.8687407370663205, "learning_rate": 2.627309444946929e-07, "loss": 0.4933, "step": 4725 }, { "epoch": 4.5357657225156025, "grad_norm": 0.8459281040179891, "learning_rate": 2.6166038035725984e-07, "loss": 0.5413, "step": 4726 }, { "epoch": 4.536725876140182, "grad_norm": 0.8841305990803782, "learning_rate": 2.605919432098331e-07, "loss": 0.3532, "step": 4727 }, { "epoch": 4.537686029764762, "grad_norm": 1.0102372656047471, "learning_rate": 2.5952563353202064e-07, "loss": 0.4388, "step": 4728 }, { "epoch": 4.538646183389343, "grad_norm": 0.9032394542987505, "learning_rate": 2.584614518024786e-07, "loss": 0.4495, "step": 4729 }, { "epoch": 4.539606337013923, "grad_norm": 0.9432588549658888, "learning_rate": 2.573993984989065e-07, "loss": 0.4541, "step": 4730 }, { "epoch": 4.540566490638502, "grad_norm": 0.8724593004942726, "learning_rate": 2.5633947409804907e-07, "loss": 0.409, "step": 4731 }, { "epoch": 4.541526644263082, "grad_norm": 1.1052258369387433, "learning_rate": 2.552816790756951e-07, "loss": 0.4399, "step": 4732 }, { "epoch": 4.542486797887662, "grad_norm": 0.7675083430181567, "learning_rate": 2.542260139066771e-07, "loss": 0.4575, "step": 4733 }, { "epoch": 4.543446951512242, "grad_norm": 0.7713623491162881, "learning_rate": 2.5317247906487153e-07, "loss": 0.5163, "step": 4734 }, { "epoch": 4.544407105136822, "grad_norm": 1.0789676620499011, "learning_rate": 2.521210750231995e-07, "loss": 0.3863, "step": 4735 }, { "epoch": 4.545367258761402, "grad_norm": 1.0669315532429429, "learning_rate": 2.510718022536246e-07, "loss": 0.4941, "step": 4736 }, { "epoch": 4.546327412385982, "grad_norm": 0.844428458674825, "learning_rate": 2.5002466122715505e-07, "loss": 0.4533, "step": 4737 }, { "epoch": 4.547287566010562, "grad_norm": 0.6939243911661667, "learning_rate": 2.489796524138394e-07, "loss": 0.5189, "step": 4738 }, { "epoch": 4.548247719635142, "grad_norm": 0.9969894967281693, "learning_rate": 2.479367762827739e-07, "loss": 0.4618, "step": 4739 }, { "epoch": 4.549207873259721, "grad_norm": 0.7637929507797864, "learning_rate": 2.46896033302092e-07, "loss": 0.4996, "step": 4740 }, { "epoch": 4.550168026884301, "grad_norm": 0.7851750631607476, "learning_rate": 2.4585742393897274e-07, "loss": 0.4976, "step": 4741 }, { "epoch": 4.551128180508881, "grad_norm": 1.0312808976811394, "learning_rate": 2.44820948659637e-07, "loss": 0.4924, "step": 4742 }, { "epoch": 4.552088334133462, "grad_norm": 1.0400191857524224, "learning_rate": 2.4378660792934715e-07, "loss": 0.5179, "step": 4743 }, { "epoch": 4.5530484877580415, "grad_norm": 1.2076558245920348, "learning_rate": 2.4275440221240786e-07, "loss": 0.3755, "step": 4744 }, { "epoch": 4.554008641382621, "grad_norm": 0.9128186493411803, "learning_rate": 2.4172433197216625e-07, "loss": 0.5148, "step": 4745 }, { "epoch": 4.554968795007201, "grad_norm": 0.9067777170258913, "learning_rate": 2.4069639767100805e-07, "loss": 0.4911, "step": 4746 }, { "epoch": 4.555928948631781, "grad_norm": 1.0544591804805923, "learning_rate": 2.39670599770363e-07, "loss": 0.4504, "step": 4747 }, { "epoch": 4.556889102256361, "grad_norm": 0.8134398528367948, "learning_rate": 2.3864693873070054e-07, "loss": 0.5119, "step": 4748 }, { "epoch": 4.5578492558809405, "grad_norm": 1.1035659512279834, "learning_rate": 2.376254150115309e-07, "loss": 0.431, "step": 4749 }, { "epoch": 4.558809409505521, "grad_norm": 0.9355214217900575, "learning_rate": 2.3660602907140672e-07, "loss": 0.4852, "step": 4750 }, { "epoch": 4.559769563130101, "grad_norm": 0.7778085591343207, "learning_rate": 2.3558878136791697e-07, "loss": 0.4386, "step": 4751 }, { "epoch": 4.560729716754681, "grad_norm": 0.8091957709150189, "learning_rate": 2.3457367235769468e-07, "loss": 0.4961, "step": 4752 }, { "epoch": 4.5616898703792605, "grad_norm": 0.8401014394680011, "learning_rate": 2.335607024964115e-07, "loss": 0.4629, "step": 4753 }, { "epoch": 4.56265002400384, "grad_norm": 1.1427515221415718, "learning_rate": 2.3254987223877868e-07, "loss": 0.4028, "step": 4754 }, { "epoch": 4.563610177628421, "grad_norm": 0.8130044821575086, "learning_rate": 2.315411820385477e-07, "loss": 0.4676, "step": 4755 }, { "epoch": 4.564570331253001, "grad_norm": 1.1102126965423615, "learning_rate": 2.3053463234850637e-07, "loss": 0.4153, "step": 4756 }, { "epoch": 4.565530484877581, "grad_norm": 0.9045753663444001, "learning_rate": 2.295302236204866e-07, "loss": 0.4839, "step": 4757 }, { "epoch": 4.56649063850216, "grad_norm": 1.151621913188701, "learning_rate": 2.2852795630535661e-07, "loss": 0.3932, "step": 4758 }, { "epoch": 4.56745079212674, "grad_norm": 1.1218351342496518, "learning_rate": 2.2752783085302265e-07, "loss": 0.4528, "step": 4759 }, { "epoch": 4.56841094575132, "grad_norm": 1.1965647553079988, "learning_rate": 2.2652984771243002e-07, "loss": 0.5009, "step": 4760 }, { "epoch": 4.5693710993759, "grad_norm": 0.7020404279474035, "learning_rate": 2.2553400733156317e-07, "loss": 0.3711, "step": 4761 }, { "epoch": 4.5703312530004805, "grad_norm": 0.6823830644467418, "learning_rate": 2.2454031015744448e-07, "loss": 0.4168, "step": 4762 }, { "epoch": 4.57129140662506, "grad_norm": 1.1343757919829704, "learning_rate": 2.2354875663613384e-07, "loss": 0.4302, "step": 4763 }, { "epoch": 4.57225156024964, "grad_norm": 0.9150536772255851, "learning_rate": 2.2255934721272908e-07, "loss": 0.5316, "step": 4764 }, { "epoch": 4.57321171387422, "grad_norm": 0.8233389878629878, "learning_rate": 2.2157208233136495e-07, "loss": 0.5666, "step": 4765 }, { "epoch": 4.5741718674988, "grad_norm": 0.8247008911171559, "learning_rate": 2.2058696243521417e-07, "loss": 0.4692, "step": 4766 }, { "epoch": 4.5751320211233795, "grad_norm": 0.683449050367792, "learning_rate": 2.196039879664874e-07, "loss": 0.5222, "step": 4767 }, { "epoch": 4.576092174747959, "grad_norm": 0.8511141953016819, "learning_rate": 2.186231593664312e-07, "loss": 0.4331, "step": 4768 }, { "epoch": 4.57705232837254, "grad_norm": 0.6789161559672479, "learning_rate": 2.1764447707532722e-07, "loss": 0.4822, "step": 4769 }, { "epoch": 4.57801248199712, "grad_norm": 0.8818329694649779, "learning_rate": 2.1666794153249792e-07, "loss": 0.4361, "step": 4770 }, { "epoch": 4.5789726356216995, "grad_norm": 1.0087502951853904, "learning_rate": 2.1569355317629937e-07, "loss": 0.5464, "step": 4771 }, { "epoch": 4.579932789246279, "grad_norm": 0.7777760382954403, "learning_rate": 2.1472131244412275e-07, "loss": 0.5847, "step": 4772 }, { "epoch": 4.580892942870859, "grad_norm": 0.966316147543303, "learning_rate": 2.1375121977239676e-07, "loss": 0.4613, "step": 4773 }, { "epoch": 4.581853096495439, "grad_norm": 0.6382434791293033, "learning_rate": 2.127832755965864e-07, "loss": 0.45, "step": 4774 }, { "epoch": 4.582813250120019, "grad_norm": 0.9543762278583675, "learning_rate": 2.118174803511913e-07, "loss": 0.4378, "step": 4775 }, { "epoch": 4.583773403744599, "grad_norm": 1.297359116534785, "learning_rate": 2.108538344697464e-07, "loss": 0.5273, "step": 4776 }, { "epoch": 4.584733557369179, "grad_norm": 0.8149093033855317, "learning_rate": 2.0989233838482225e-07, "loss": 0.5626, "step": 4777 }, { "epoch": 4.585693710993759, "grad_norm": 0.6734588212250174, "learning_rate": 2.0893299252802367e-07, "loss": 0.4695, "step": 4778 }, { "epoch": 4.586653864618339, "grad_norm": 1.0707368908782673, "learning_rate": 2.0797579732999063e-07, "loss": 0.5054, "step": 4779 }, { "epoch": 4.587614018242919, "grad_norm": 0.771159363037436, "learning_rate": 2.0702075322039884e-07, "loss": 0.464, "step": 4780 }, { "epoch": 4.588574171867499, "grad_norm": 1.0842871874625841, "learning_rate": 2.0606786062795702e-07, "loss": 0.361, "step": 4781 }, { "epoch": 4.589534325492079, "grad_norm": 0.8242750397847514, "learning_rate": 2.051171199804075e-07, "loss": 0.4474, "step": 4782 }, { "epoch": 4.590494479116659, "grad_norm": 0.8661603858539377, "learning_rate": 2.0416853170452945e-07, "loss": 0.4202, "step": 4783 }, { "epoch": 4.591454632741239, "grad_norm": 1.1009181018412044, "learning_rate": 2.032220962261322e-07, "loss": 0.4642, "step": 4784 }, { "epoch": 4.5924147863658185, "grad_norm": 0.8521428216513732, "learning_rate": 2.022778139700615e-07, "loss": 0.539, "step": 4785 }, { "epoch": 4.593374939990398, "grad_norm": 0.8782385996772073, "learning_rate": 2.0133568536019498e-07, "loss": 0.4526, "step": 4786 }, { "epoch": 4.594335093614978, "grad_norm": 0.728819189910761, "learning_rate": 2.0039571081944431e-07, "loss": 0.5046, "step": 4787 }, { "epoch": 4.595295247239559, "grad_norm": 0.6908645986446862, "learning_rate": 1.994578907697542e-07, "loss": 0.4635, "step": 4788 }, { "epoch": 4.5962554008641385, "grad_norm": 0.7239590556516556, "learning_rate": 1.985222256321029e-07, "loss": 0.5011, "step": 4789 }, { "epoch": 4.597215554488718, "grad_norm": 0.6992615611626155, "learning_rate": 1.9758871582649887e-07, "loss": 0.4225, "step": 4790 }, { "epoch": 4.598175708113298, "grad_norm": 0.7477909272915941, "learning_rate": 1.9665736177198524e-07, "loss": 0.5016, "step": 4791 }, { "epoch": 4.599135861737878, "grad_norm": 0.5852397069300285, "learning_rate": 1.9572816388663706e-07, "loss": 0.4782, "step": 4792 }, { "epoch": 4.600096015362458, "grad_norm": 0.6795732585149922, "learning_rate": 1.9480112258756123e-07, "loss": 0.4264, "step": 4793 }, { "epoch": 4.6010561689870375, "grad_norm": 1.0424908489632772, "learning_rate": 1.9387623829089764e-07, "loss": 0.4098, "step": 4794 }, { "epoch": 4.602016322611618, "grad_norm": 1.1258873471238093, "learning_rate": 1.929535114118153e-07, "loss": 0.4663, "step": 4795 }, { "epoch": 4.602976476236198, "grad_norm": 0.7737434687564749, "learning_rate": 1.920329423645173e-07, "loss": 0.4821, "step": 4796 }, { "epoch": 4.603936629860778, "grad_norm": 0.77680780946051, "learning_rate": 1.9111453156223704e-07, "loss": 0.4634, "step": 4797 }, { "epoch": 4.604896783485358, "grad_norm": 0.752619610286257, "learning_rate": 1.901982794172391e-07, "loss": 0.4401, "step": 4798 }, { "epoch": 4.605856937109937, "grad_norm": 0.623663575815771, "learning_rate": 1.8928418634082068e-07, "loss": 0.4224, "step": 4799 }, { "epoch": 4.606817090734517, "grad_norm": 1.1663145412228972, "learning_rate": 1.8837225274330572e-07, "loss": 0.3879, "step": 4800 }, { "epoch": 4.607777244359097, "grad_norm": 0.7821619769224282, "learning_rate": 1.87462479034054e-07, "loss": 0.4415, "step": 4801 }, { "epoch": 4.608737397983678, "grad_norm": 0.7597480012068021, "learning_rate": 1.8655486562145275e-07, "loss": 0.4316, "step": 4802 }, { "epoch": 4.6096975516082574, "grad_norm": 0.7908639854694701, "learning_rate": 1.8564941291291883e-07, "loss": 0.4867, "step": 4803 }, { "epoch": 4.610657705232837, "grad_norm": 1.2731579014946899, "learning_rate": 1.8474612131490098e-07, "loss": 0.4279, "step": 4804 }, { "epoch": 4.611617858857417, "grad_norm": 1.1338829709548, "learning_rate": 1.8384499123287713e-07, "loss": 0.4988, "step": 4805 }, { "epoch": 4.612578012481997, "grad_norm": 1.026486931798522, "learning_rate": 1.8294602307135478e-07, "loss": 0.4501, "step": 4806 }, { "epoch": 4.6135381661065775, "grad_norm": 0.879253608256226, "learning_rate": 1.820492172338717e-07, "loss": 0.4459, "step": 4807 }, { "epoch": 4.614498319731157, "grad_norm": 0.7208890204361037, "learning_rate": 1.8115457412299363e-07, "loss": 0.5169, "step": 4808 }, { "epoch": 4.615458473355737, "grad_norm": 0.7176158639511218, "learning_rate": 1.802620941403166e-07, "loss": 0.445, "step": 4809 }, { "epoch": 4.616418626980317, "grad_norm": 0.7986408640668442, "learning_rate": 1.7937177768646508e-07, "loss": 0.4092, "step": 4810 }, { "epoch": 4.617378780604897, "grad_norm": 1.2649292981915783, "learning_rate": 1.7848362516109274e-07, "loss": 0.5669, "step": 4811 }, { "epoch": 4.6183389342294765, "grad_norm": 0.8264850094608479, "learning_rate": 1.775976369628829e-07, "loss": 0.5261, "step": 4812 }, { "epoch": 4.619299087854056, "grad_norm": 1.780894721312399, "learning_rate": 1.7671381348954408e-07, "loss": 0.4884, "step": 4813 }, { "epoch": 4.620259241478637, "grad_norm": 1.2125357178474412, "learning_rate": 1.7583215513781725e-07, "loss": 0.4991, "step": 4814 }, { "epoch": 4.621219395103217, "grad_norm": 0.7796660099952121, "learning_rate": 1.749526623034681e-07, "loss": 0.441, "step": 4815 }, { "epoch": 4.622179548727797, "grad_norm": 0.9412558043841397, "learning_rate": 1.7407533538129185e-07, "loss": 0.473, "step": 4816 }, { "epoch": 4.623139702352376, "grad_norm": 0.8834313089525544, "learning_rate": 1.732001747651113e-07, "loss": 0.4847, "step": 4817 }, { "epoch": 4.624099855976956, "grad_norm": 0.9342326436801752, "learning_rate": 1.7232718084777666e-07, "loss": 0.4966, "step": 4818 }, { "epoch": 4.625060009601536, "grad_norm": 0.8869872255097427, "learning_rate": 1.7145635402116556e-07, "loss": 0.4997, "step": 4819 }, { "epoch": 4.626020163226116, "grad_norm": 0.7343155788857423, "learning_rate": 1.7058769467618364e-07, "loss": 0.4718, "step": 4820 }, { "epoch": 4.626980316850696, "grad_norm": 1.0343334998313038, "learning_rate": 1.697212032027612e-07, "loss": 0.4688, "step": 4821 }, { "epoch": 4.627940470475276, "grad_norm": 0.8774014984648517, "learning_rate": 1.688568799898582e-07, "loss": 0.5466, "step": 4822 }, { "epoch": 4.628900624099856, "grad_norm": 1.2092542456435136, "learning_rate": 1.6799472542545935e-07, "loss": 0.4811, "step": 4823 }, { "epoch": 4.629860777724436, "grad_norm": 1.0947438390709796, "learning_rate": 1.671347398965767e-07, "loss": 0.4487, "step": 4824 }, { "epoch": 4.630820931349016, "grad_norm": 0.8213806164104841, "learning_rate": 1.6627692378924975e-07, "loss": 0.558, "step": 4825 }, { "epoch": 4.631781084973595, "grad_norm": 0.7484481672710845, "learning_rate": 1.6542127748854108e-07, "loss": 0.3927, "step": 4826 }, { "epoch": 4.632741238598175, "grad_norm": 0.796767688414836, "learning_rate": 1.6456780137854346e-07, "loss": 0.4015, "step": 4827 }, { "epoch": 4.633701392222756, "grad_norm": 1.3928191923661586, "learning_rate": 1.6371649584237093e-07, "loss": 0.479, "step": 4828 }, { "epoch": 4.634661545847336, "grad_norm": 0.6454665026150745, "learning_rate": 1.6286736126216674e-07, "loss": 0.4682, "step": 4829 }, { "epoch": 4.6356216994719155, "grad_norm": 0.6455355125501806, "learning_rate": 1.6202039801909763e-07, "loss": 0.3548, "step": 4830 }, { "epoch": 4.636581853096495, "grad_norm": 0.9085148533262426, "learning_rate": 1.6117560649335674e-07, "loss": 0.4407, "step": 4831 }, { "epoch": 4.637542006721075, "grad_norm": 0.7721552643927505, "learning_rate": 1.6033298706416234e-07, "loss": 0.4934, "step": 4832 }, { "epoch": 4.638502160345656, "grad_norm": 0.9456271705400285, "learning_rate": 1.594925401097569e-07, "loss": 0.4245, "step": 4833 }, { "epoch": 4.639462313970236, "grad_norm": 1.094174709793503, "learning_rate": 1.5865426600740807e-07, "loss": 0.4958, "step": 4834 }, { "epoch": 4.640422467594815, "grad_norm": 0.8985390532527195, "learning_rate": 1.5781816513340764e-07, "loss": 0.5027, "step": 4835 }, { "epoch": 4.641382621219395, "grad_norm": 0.9136205643411824, "learning_rate": 1.5698423786307314e-07, "loss": 0.5454, "step": 4836 }, { "epoch": 4.642342774843975, "grad_norm": 1.1498712056252538, "learning_rate": 1.5615248457074516e-07, "loss": 0.5188, "step": 4837 }, { "epoch": 4.643302928468555, "grad_norm": 0.921017605728109, "learning_rate": 1.5532290562978948e-07, "loss": 0.4847, "step": 4838 }, { "epoch": 4.6442630820931345, "grad_norm": 0.7836231829508422, "learning_rate": 1.5449550141259427e-07, "loss": 0.4647, "step": 4839 }, { "epoch": 4.645223235717715, "grad_norm": 0.7229241544695276, "learning_rate": 1.5367027229057307e-07, "loss": 0.4434, "step": 4840 }, { "epoch": 4.646183389342295, "grad_norm": 0.8614280266249061, "learning_rate": 1.5284721863416175e-07, "loss": 0.5153, "step": 4841 }, { "epoch": 4.647143542966875, "grad_norm": 0.966668221990195, "learning_rate": 1.5202634081282142e-07, "loss": 0.4778, "step": 4842 }, { "epoch": 4.648103696591455, "grad_norm": 0.8083574737641841, "learning_rate": 1.512076391950351e-07, "loss": 0.5092, "step": 4843 }, { "epoch": 4.649063850216034, "grad_norm": 0.7907727037034538, "learning_rate": 1.5039111414830721e-07, "loss": 0.4234, "step": 4844 }, { "epoch": 4.650024003840614, "grad_norm": 1.1132714655087823, "learning_rate": 1.4957676603917005e-07, "loss": 0.4184, "step": 4845 }, { "epoch": 4.650984157465194, "grad_norm": 0.761468142150567, "learning_rate": 1.4876459523317455e-07, "loss": 0.4825, "step": 4846 }, { "epoch": 4.651944311089775, "grad_norm": 0.9018311591315953, "learning_rate": 1.479546020948952e-07, "loss": 0.5989, "step": 4847 }, { "epoch": 4.6529044647143545, "grad_norm": 0.8246386173486321, "learning_rate": 1.4714678698792894e-07, "loss": 0.467, "step": 4848 }, { "epoch": 4.653864618338934, "grad_norm": 0.877331407781171, "learning_rate": 1.4634115027489627e-07, "loss": 0.5553, "step": 4849 }, { "epoch": 4.654824771963514, "grad_norm": 0.8759248347129419, "learning_rate": 1.4553769231743842e-07, "loss": 0.4319, "step": 4850 }, { "epoch": 4.655784925588094, "grad_norm": 0.9947497674330859, "learning_rate": 1.447364134762197e-07, "loss": 0.4894, "step": 4851 }, { "epoch": 4.656745079212674, "grad_norm": 1.095430463710392, "learning_rate": 1.439373141109246e-07, "loss": 0.4956, "step": 4852 }, { "epoch": 4.6577052328372535, "grad_norm": 0.880449805903669, "learning_rate": 1.4314039458026063e-07, "loss": 0.5042, "step": 4853 }, { "epoch": 4.658665386461834, "grad_norm": 0.8347036662896851, "learning_rate": 1.4234565524195664e-07, "loss": 0.4955, "step": 4854 }, { "epoch": 4.659625540086414, "grad_norm": 0.9523828512609825, "learning_rate": 1.4155309645276283e-07, "loss": 0.5845, "step": 4855 }, { "epoch": 4.660585693710994, "grad_norm": 0.9838894785586744, "learning_rate": 1.407627185684507e-07, "loss": 0.4898, "step": 4856 }, { "epoch": 4.6615458473355735, "grad_norm": 0.7056530180050824, "learning_rate": 1.3997452194381034e-07, "loss": 0.4624, "step": 4857 }, { "epoch": 4.662506000960153, "grad_norm": 1.0784382902145464, "learning_rate": 1.3918850693265763e-07, "loss": 0.4025, "step": 4858 }, { "epoch": 4.663466154584734, "grad_norm": 0.8221240969863406, "learning_rate": 1.3840467388782418e-07, "loss": 0.527, "step": 4859 }, { "epoch": 4.664426308209314, "grad_norm": 1.3059110361781876, "learning_rate": 1.3762302316116527e-07, "loss": 0.4683, "step": 4860 }, { "epoch": 4.665386461833894, "grad_norm": 0.7837723082023593, "learning_rate": 1.3684355510355518e-07, "loss": 0.4922, "step": 4861 }, { "epoch": 4.666346615458473, "grad_norm": 0.7812647222641647, "learning_rate": 1.3606627006488905e-07, "loss": 0.4607, "step": 4862 }, { "epoch": 4.667306769083053, "grad_norm": 0.8548248908311814, "learning_rate": 1.3529116839408162e-07, "loss": 0.5134, "step": 4863 }, { "epoch": 4.668266922707633, "grad_norm": 0.8493774460821467, "learning_rate": 1.3451825043906852e-07, "loss": 0.4879, "step": 4864 }, { "epoch": 4.669227076332213, "grad_norm": 0.7093178179020236, "learning_rate": 1.337475165468033e-07, "loss": 0.46, "step": 4865 }, { "epoch": 4.6701872299567935, "grad_norm": 0.7493640643407073, "learning_rate": 1.3297896706326031e-07, "loss": 0.5582, "step": 4866 }, { "epoch": 4.671147383581373, "grad_norm": 0.7148209646673517, "learning_rate": 1.3221260233343357e-07, "loss": 0.479, "step": 4867 }, { "epoch": 4.672107537205953, "grad_norm": 0.6975236701396929, "learning_rate": 1.3144842270133573e-07, "loss": 0.4208, "step": 4868 }, { "epoch": 4.673067690830533, "grad_norm": 0.9562081640837607, "learning_rate": 1.3068642850999957e-07, "loss": 0.4804, "step": 4869 }, { "epoch": 4.674027844455113, "grad_norm": 0.7814994527415347, "learning_rate": 1.2992662010147428e-07, "loss": 0.4878, "step": 4870 }, { "epoch": 4.6749879980796925, "grad_norm": 0.8589730121289582, "learning_rate": 1.2916899781683257e-07, "loss": 0.5536, "step": 4871 }, { "epoch": 4.675948151704272, "grad_norm": 0.9654822738462338, "learning_rate": 1.2841356199616072e-07, "loss": 0.3784, "step": 4872 }, { "epoch": 4.676908305328853, "grad_norm": 1.0814511867336374, "learning_rate": 1.276603129785664e-07, "loss": 0.4587, "step": 4873 }, { "epoch": 4.677868458953433, "grad_norm": 1.1047210142629533, "learning_rate": 1.2690925110217523e-07, "loss": 0.462, "step": 4874 }, { "epoch": 4.6788286125780125, "grad_norm": 0.8984278246256303, "learning_rate": 1.2616037670413084e-07, "loss": 0.5308, "step": 4875 }, { "epoch": 4.679788766202592, "grad_norm": 0.7064280857197117, "learning_rate": 1.254136901205949e-07, "loss": 0.4533, "step": 4876 }, { "epoch": 4.680748919827172, "grad_norm": 0.9284766262429089, "learning_rate": 1.2466919168674763e-07, "loss": 0.5291, "step": 4877 }, { "epoch": 4.681709073451752, "grad_norm": 0.9522401386363464, "learning_rate": 1.2392688173678503e-07, "loss": 0.4365, "step": 4878 }, { "epoch": 4.682669227076333, "grad_norm": 0.6865477157022447, "learning_rate": 1.2318676060392333e-07, "loss": 0.5396, "step": 4879 }, { "epoch": 4.683629380700912, "grad_norm": 0.8340581018169749, "learning_rate": 1.2244882862039398e-07, "loss": 0.4261, "step": 4880 }, { "epoch": 4.684589534325492, "grad_norm": 1.2276388737538286, "learning_rate": 1.217130861174476e-07, "loss": 0.4184, "step": 4881 }, { "epoch": 4.685549687950072, "grad_norm": 1.4069756422539599, "learning_rate": 1.209795334253516e-07, "loss": 0.4771, "step": 4882 }, { "epoch": 4.686509841574652, "grad_norm": 0.9611062377587603, "learning_rate": 1.2024817087338826e-07, "loss": 0.5076, "step": 4883 }, { "epoch": 4.687469995199232, "grad_norm": 0.7338524028885621, "learning_rate": 1.1951899878985983e-07, "loss": 0.5094, "step": 4884 }, { "epoch": 4.688430148823812, "grad_norm": 0.82700826182432, "learning_rate": 1.1879201750208291e-07, "loss": 0.3701, "step": 4885 }, { "epoch": 4.689390302448392, "grad_norm": 0.8516933098146412, "learning_rate": 1.1806722733639253e-07, "loss": 0.5654, "step": 4886 }, { "epoch": 4.690350456072972, "grad_norm": 0.7822210087466362, "learning_rate": 1.17344628618139e-07, "loss": 0.5388, "step": 4887 }, { "epoch": 4.691310609697552, "grad_norm": 1.4244648386054402, "learning_rate": 1.1662422167168841e-07, "loss": 0.4063, "step": 4888 }, { "epoch": 4.692270763322131, "grad_norm": 0.8503773771340302, "learning_rate": 1.159060068204243e-07, "loss": 0.4371, "step": 4889 }, { "epoch": 4.693230916946711, "grad_norm": 0.6497386660923653, "learning_rate": 1.1518998438674656e-07, "loss": 0.4253, "step": 4890 }, { "epoch": 4.694191070571291, "grad_norm": 1.473129218487771, "learning_rate": 1.1447615469206863e-07, "loss": 0.4482, "step": 4891 }, { "epoch": 4.695151224195872, "grad_norm": 1.1053581231553824, "learning_rate": 1.1376451805682143e-07, "loss": 0.475, "step": 4892 }, { "epoch": 4.6961113778204515, "grad_norm": 0.8266879406497135, "learning_rate": 1.1305507480045108e-07, "loss": 0.4343, "step": 4893 }, { "epoch": 4.697071531445031, "grad_norm": 1.1339389529742367, "learning_rate": 1.1234782524141896e-07, "loss": 0.4457, "step": 4894 }, { "epoch": 4.698031685069611, "grad_norm": 0.8842516185066394, "learning_rate": 1.1164276969720278e-07, "loss": 0.4986, "step": 4895 }, { "epoch": 4.698991838694191, "grad_norm": 0.9333957202616086, "learning_rate": 1.109399084842927e-07, "loss": 0.4628, "step": 4896 }, { "epoch": 4.699951992318771, "grad_norm": 0.7114761489083549, "learning_rate": 1.1023924191819691e-07, "loss": 0.5052, "step": 4897 }, { "epoch": 4.7009121459433505, "grad_norm": 1.030538067458471, "learning_rate": 1.0954077031343658e-07, "loss": 0.4821, "step": 4898 }, { "epoch": 4.701872299567931, "grad_norm": 0.8244351718619999, "learning_rate": 1.0884449398354868e-07, "loss": 0.448, "step": 4899 }, { "epoch": 4.702832453192511, "grad_norm": 0.7343335831696222, "learning_rate": 1.0815041324108378e-07, "loss": 0.478, "step": 4900 }, { "epoch": 4.703792606817091, "grad_norm": 0.8091342767516712, "learning_rate": 1.0745852839760706e-07, "loss": 0.5496, "step": 4901 }, { "epoch": 4.704752760441671, "grad_norm": 1.015854027523989, "learning_rate": 1.0676883976369956e-07, "loss": 0.4654, "step": 4902 }, { "epoch": 4.70571291406625, "grad_norm": 1.112947492118363, "learning_rate": 1.0608134764895306e-07, "loss": 0.4679, "step": 4903 }, { "epoch": 4.70667306769083, "grad_norm": 0.8425322499620592, "learning_rate": 1.053960523619768e-07, "loss": 0.4901, "step": 4904 }, { "epoch": 4.707633221315411, "grad_norm": 1.021677202356927, "learning_rate": 1.0471295421039251e-07, "loss": 0.4785, "step": 4905 }, { "epoch": 4.708593374939991, "grad_norm": 0.6612719474425689, "learning_rate": 1.040320535008349e-07, "loss": 0.4777, "step": 4906 }, { "epoch": 4.70955352856457, "grad_norm": 0.8934134881919478, "learning_rate": 1.0335335053895391e-07, "loss": 0.4914, "step": 4907 }, { "epoch": 4.71051368218915, "grad_norm": 0.7732442055893796, "learning_rate": 1.0267684562941194e-07, "loss": 0.4743, "step": 4908 }, { "epoch": 4.71147383581373, "grad_norm": 0.9184455296382198, "learning_rate": 1.0200253907588387e-07, "loss": 0.3837, "step": 4909 }, { "epoch": 4.71243398943831, "grad_norm": 1.0690904592588744, "learning_rate": 1.0133043118105923e-07, "loss": 0.4922, "step": 4910 }, { "epoch": 4.7133941430628905, "grad_norm": 0.938251008746643, "learning_rate": 1.0066052224664058e-07, "loss": 0.5332, "step": 4911 }, { "epoch": 4.71435429668747, "grad_norm": 1.1566428106921338, "learning_rate": 9.999281257334181e-08, "loss": 0.4455, "step": 4912 }, { "epoch": 4.71531445031205, "grad_norm": 0.9407582367377498, "learning_rate": 9.932730246089262e-08, "loss": 0.4596, "step": 4913 }, { "epoch": 4.71627460393663, "grad_norm": 0.8537229926795888, "learning_rate": 9.86639922080307e-08, "loss": 0.5092, "step": 4914 }, { "epoch": 4.71723475756121, "grad_norm": 1.3073999046883817, "learning_rate": 9.800288211251119e-08, "loss": 0.4739, "step": 4915 }, { "epoch": 4.7181949111857895, "grad_norm": 0.7893406336989749, "learning_rate": 9.734397247109784e-08, "loss": 0.5704, "step": 4916 }, { "epoch": 4.719155064810369, "grad_norm": 0.7418109026797135, "learning_rate": 9.668726357956903e-08, "loss": 0.4531, "step": 4917 }, { "epoch": 4.72011521843495, "grad_norm": 0.769438312324943, "learning_rate": 9.603275573271397e-08, "loss": 0.4831, "step": 4918 }, { "epoch": 4.72107537205953, "grad_norm": 0.7158282199736148, "learning_rate": 9.538044922433432e-08, "loss": 0.4411, "step": 4919 }, { "epoch": 4.72203552568411, "grad_norm": 0.7489700099376845, "learning_rate": 9.47303443472436e-08, "loss": 0.4644, "step": 4920 }, { "epoch": 4.722995679308689, "grad_norm": 1.0246217040288321, "learning_rate": 9.408244139326728e-08, "loss": 0.519, "step": 4921 }, { "epoch": 4.723955832933269, "grad_norm": 0.9362821408557138, "learning_rate": 9.34367406532416e-08, "loss": 0.5182, "step": 4922 }, { "epoch": 4.724915986557849, "grad_norm": 0.8628528264294609, "learning_rate": 9.279324241701415e-08, "loss": 0.4838, "step": 4923 }, { "epoch": 4.725876140182429, "grad_norm": 0.9593629758080178, "learning_rate": 9.215194697344554e-08, "loss": 0.4303, "step": 4924 }, { "epoch": 4.726836293807009, "grad_norm": 0.9472698292436696, "learning_rate": 9.151285461040604e-08, "loss": 0.4776, "step": 4925 }, { "epoch": 4.727796447431589, "grad_norm": 1.180414790474059, "learning_rate": 9.08759656147773e-08, "loss": 0.4142, "step": 4926 }, { "epoch": 4.728756601056169, "grad_norm": 1.1543958933151506, "learning_rate": 9.024128027245171e-08, "loss": 0.4227, "step": 4927 }, { "epoch": 4.729716754680749, "grad_norm": 0.8418240290460491, "learning_rate": 8.960879886833307e-08, "loss": 0.4941, "step": 4928 }, { "epoch": 4.730676908305329, "grad_norm": 1.121641982570022, "learning_rate": 8.897852168633537e-08, "loss": 0.5032, "step": 4929 }, { "epoch": 4.731637061929908, "grad_norm": 0.689173547084386, "learning_rate": 8.835044900938339e-08, "loss": 0.4172, "step": 4930 }, { "epoch": 4.732597215554489, "grad_norm": 0.8224930144132759, "learning_rate": 8.772458111941273e-08, "loss": 0.5359, "step": 4931 }, { "epoch": 4.733557369179069, "grad_norm": 0.8974096719608029, "learning_rate": 8.7100918297367e-08, "loss": 0.4473, "step": 4932 }, { "epoch": 4.734517522803649, "grad_norm": 1.261369201646506, "learning_rate": 8.647946082320446e-08, "loss": 0.4295, "step": 4933 }, { "epoch": 4.7354776764282285, "grad_norm": 0.9461690687099619, "learning_rate": 8.586020897588864e-08, "loss": 0.4879, "step": 4934 }, { "epoch": 4.736437830052808, "grad_norm": 0.7437199408755303, "learning_rate": 8.524316303339609e-08, "loss": 0.4366, "step": 4935 }, { "epoch": 4.737397983677388, "grad_norm": 0.87560220977995, "learning_rate": 8.462832327271131e-08, "loss": 0.4783, "step": 4936 }, { "epoch": 4.738358137301969, "grad_norm": 0.664039598931437, "learning_rate": 8.401568996983022e-08, "loss": 0.4536, "step": 4937 }, { "epoch": 4.7393182909265485, "grad_norm": 1.512044609401102, "learning_rate": 8.340526339975674e-08, "loss": 0.4779, "step": 4938 }, { "epoch": 4.740278444551128, "grad_norm": 0.8280531878650521, "learning_rate": 8.279704383650555e-08, "loss": 0.4966, "step": 4939 }, { "epoch": 4.741238598175708, "grad_norm": 0.8034959824129843, "learning_rate": 8.219103155309938e-08, "loss": 0.4366, "step": 4940 }, { "epoch": 4.742198751800288, "grad_norm": 0.7778338190749002, "learning_rate": 8.158722682157005e-08, "loss": 0.4716, "step": 4941 }, { "epoch": 4.743158905424868, "grad_norm": 0.7382862304510764, "learning_rate": 8.098562991296022e-08, "loss": 0.4733, "step": 4942 }, { "epoch": 4.7441190590494475, "grad_norm": 1.0256814012234086, "learning_rate": 8.03862410973194e-08, "loss": 0.4019, "step": 4943 }, { "epoch": 4.745079212674028, "grad_norm": 1.1961856477847672, "learning_rate": 7.978906064370739e-08, "loss": 0.3558, "step": 4944 }, { "epoch": 4.746039366298608, "grad_norm": 1.0068538553501971, "learning_rate": 7.919408882019142e-08, "loss": 0.4606, "step": 4945 }, { "epoch": 4.746999519923188, "grad_norm": 0.9701572382659532, "learning_rate": 7.86013258938495e-08, "loss": 0.5174, "step": 4946 }, { "epoch": 4.747959673547768, "grad_norm": 0.8199967552058781, "learning_rate": 7.801077213076491e-08, "loss": 0.5098, "step": 4947 }, { "epoch": 4.748919827172347, "grad_norm": 0.8801552591656422, "learning_rate": 7.742242779603115e-08, "loss": 0.4225, "step": 4948 }, { "epoch": 4.749879980796927, "grad_norm": 1.2189548565917663, "learning_rate": 7.683629315375086e-08, "loss": 0.3989, "step": 4949 }, { "epoch": 4.750840134421507, "grad_norm": 0.8659751086212341, "learning_rate": 7.625236846703243e-08, "loss": 0.496, "step": 4950 }, { "epoch": 4.751800288046088, "grad_norm": 0.8861190774499543, "learning_rate": 7.5670653997994e-08, "loss": 0.5854, "step": 4951 }, { "epoch": 4.7527604416706675, "grad_norm": 0.9237955274992801, "learning_rate": 7.509115000776112e-08, "loss": 0.5676, "step": 4952 }, { "epoch": 4.753720595295247, "grad_norm": 0.6718156397663602, "learning_rate": 7.451385675646627e-08, "loss": 0.4396, "step": 4953 }, { "epoch": 4.754680748919827, "grad_norm": 0.8008359030887666, "learning_rate": 7.393877450325105e-08, "loss": 0.5962, "step": 4954 }, { "epoch": 4.755640902544407, "grad_norm": 0.7170357285890014, "learning_rate": 7.336590350626282e-08, "loss": 0.5425, "step": 4955 }, { "epoch": 4.756601056168987, "grad_norm": 1.1865480772117256, "learning_rate": 7.279524402265758e-08, "loss": 0.5007, "step": 4956 }, { "epoch": 4.757561209793567, "grad_norm": 0.8437494665421912, "learning_rate": 7.222679630859875e-08, "loss": 0.3586, "step": 4957 }, { "epoch": 4.758521363418147, "grad_norm": 1.0829142975967316, "learning_rate": 7.166056061925609e-08, "loss": 0.4684, "step": 4958 }, { "epoch": 4.759481517042727, "grad_norm": 1.0202087626630225, "learning_rate": 7.109653720880682e-08, "loss": 0.4667, "step": 4959 }, { "epoch": 4.760441670667307, "grad_norm": 0.7550456336651408, "learning_rate": 7.053472633043457e-08, "loss": 0.4739, "step": 4960 }, { "epoch": 4.7614018242918865, "grad_norm": 1.2140812328519452, "learning_rate": 6.99751282363309e-08, "loss": 0.4687, "step": 4961 }, { "epoch": 4.762361977916466, "grad_norm": 0.8080546332046133, "learning_rate": 6.941774317769268e-08, "loss": 0.571, "step": 4962 }, { "epoch": 4.763322131541047, "grad_norm": 0.940346680530264, "learning_rate": 6.886257140472475e-08, "loss": 0.4791, "step": 4963 }, { "epoch": 4.764282285165627, "grad_norm": 1.14643468068821, "learning_rate": 6.830961316663775e-08, "loss": 0.4041, "step": 4964 }, { "epoch": 4.765242438790207, "grad_norm": 0.7823073347385977, "learning_rate": 6.775886871164872e-08, "loss": 0.5073, "step": 4965 }, { "epoch": 4.766202592414786, "grad_norm": 1.1095226589243397, "learning_rate": 6.721033828698098e-08, "loss": 0.3889, "step": 4966 }, { "epoch": 4.767162746039366, "grad_norm": 1.0056179452037173, "learning_rate": 6.666402213886313e-08, "loss": 0.5448, "step": 4967 }, { "epoch": 4.768122899663946, "grad_norm": 0.7091729033826617, "learning_rate": 6.61199205125318e-08, "loss": 0.4126, "step": 4968 }, { "epoch": 4.769083053288526, "grad_norm": 0.9284212795456186, "learning_rate": 6.557803365222714e-08, "loss": 0.3979, "step": 4969 }, { "epoch": 4.7700432069131065, "grad_norm": 0.9563652909516197, "learning_rate": 6.503836180119849e-08, "loss": 0.479, "step": 4970 }, { "epoch": 4.771003360537686, "grad_norm": 0.7922387820662058, "learning_rate": 6.45009052016965e-08, "loss": 0.5066, "step": 4971 }, { "epoch": 4.771963514162266, "grad_norm": 1.2360986639211033, "learning_rate": 6.396566409498039e-08, "loss": 0.3931, "step": 4972 }, { "epoch": 4.772923667786846, "grad_norm": 0.8275497104319335, "learning_rate": 6.343263872131467e-08, "loss": 0.5341, "step": 4973 }, { "epoch": 4.773883821411426, "grad_norm": 0.8750418759943561, "learning_rate": 6.290182931996846e-08, "loss": 0.445, "step": 4974 }, { "epoch": 4.774843975036005, "grad_norm": 1.0412825850309353, "learning_rate": 6.237323612921675e-08, "loss": 0.5649, "step": 4975 }, { "epoch": 4.775804128660585, "grad_norm": 1.1621605557043333, "learning_rate": 6.184685938633861e-08, "loss": 0.4751, "step": 4976 }, { "epoch": 4.776764282285166, "grad_norm": 0.8011871116806153, "learning_rate": 6.132269932761947e-08, "loss": 0.4392, "step": 4977 }, { "epoch": 4.777724435909746, "grad_norm": 0.8324538365577582, "learning_rate": 6.080075618834891e-08, "loss": 0.5265, "step": 4978 }, { "epoch": 4.7786845895343255, "grad_norm": 1.0999603144630465, "learning_rate": 6.028103020282228e-08, "loss": 0.4844, "step": 4979 }, { "epoch": 4.779644743158905, "grad_norm": 0.9124356617632493, "learning_rate": 5.976352160433796e-08, "loss": 0.3613, "step": 4980 }, { "epoch": 4.780604896783485, "grad_norm": 1.0146591251843575, "learning_rate": 5.9248230625200686e-08, "loss": 0.4134, "step": 4981 }, { "epoch": 4.781565050408066, "grad_norm": 0.9205544006000713, "learning_rate": 5.873515749671932e-08, "loss": 0.5402, "step": 4982 }, { "epoch": 4.782525204032646, "grad_norm": 0.6657244581809003, "learning_rate": 5.822430244920685e-08, "loss": 0.4384, "step": 4983 }, { "epoch": 4.783485357657225, "grad_norm": 1.5906939189474125, "learning_rate": 5.771566571197984e-08, "loss": 0.457, "step": 4984 }, { "epoch": 4.784445511281805, "grad_norm": 1.0701723936235072, "learning_rate": 5.72092475133601e-08, "loss": 0.4553, "step": 4985 }, { "epoch": 4.785405664906385, "grad_norm": 0.7044982190577206, "learning_rate": 5.670504808067412e-08, "loss": 0.433, "step": 4986 }, { "epoch": 4.786365818530965, "grad_norm": 1.2603276075559597, "learning_rate": 5.620306764025085e-08, "loss": 0.4111, "step": 4987 }, { "epoch": 4.787325972155545, "grad_norm": 0.919062816832004, "learning_rate": 5.57033064174245e-08, "loss": 0.5614, "step": 4988 }, { "epoch": 4.788286125780125, "grad_norm": 1.2645312362742496, "learning_rate": 5.5205764636532266e-08, "loss": 0.3781, "step": 4989 }, { "epoch": 4.789246279404705, "grad_norm": 0.991322050725091, "learning_rate": 5.471044252091551e-08, "loss": 0.5066, "step": 4990 }, { "epoch": 4.790206433029285, "grad_norm": 0.9084597214259297, "learning_rate": 5.421734029291858e-08, "loss": 0.4545, "step": 4991 }, { "epoch": 4.791166586653865, "grad_norm": 0.6821374597160126, "learning_rate": 5.372645817389055e-08, "loss": 0.451, "step": 4992 }, { "epoch": 4.792126740278444, "grad_norm": 1.053695359929555, "learning_rate": 5.323779638418236e-08, "loss": 0.5404, "step": 4993 }, { "epoch": 4.793086893903024, "grad_norm": 0.7886863910632661, "learning_rate": 5.275135514314911e-08, "loss": 0.528, "step": 4994 }, { "epoch": 4.794047047527604, "grad_norm": 1.2859563219941301, "learning_rate": 5.226713466915001e-08, "loss": 0.5139, "step": 4995 }, { "epoch": 4.795007201152185, "grad_norm": 0.9073178280370419, "learning_rate": 5.1785135179545643e-08, "loss": 0.4358, "step": 4996 }, { "epoch": 4.7959673547767645, "grad_norm": 0.8762045642437355, "learning_rate": 5.130535689070071e-08, "loss": 0.4237, "step": 4997 }, { "epoch": 4.796927508401344, "grad_norm": 0.7149744643920591, "learning_rate": 5.0827800017982376e-08, "loss": 0.4455, "step": 4998 }, { "epoch": 4.797887662025924, "grad_norm": 0.7984685611981643, "learning_rate": 5.035246477576083e-08, "loss": 0.4654, "step": 4999 }, { "epoch": 4.798847815650504, "grad_norm": 0.9263700635820409, "learning_rate": 4.9879351377408734e-08, "loss": 0.4954, "step": 5000 }, { "epoch": 4.799807969275084, "grad_norm": 0.8184313588919915, "learning_rate": 4.940846003530231e-08, "loss": 0.5416, "step": 5001 }, { "epoch": 4.8007681228996635, "grad_norm": 1.0958726385223756, "learning_rate": 4.893979096081969e-08, "loss": 0.4289, "step": 5002 }, { "epoch": 4.801728276524244, "grad_norm": 0.7288032870470014, "learning_rate": 4.847334436434037e-08, "loss": 0.4295, "step": 5003 }, { "epoch": 4.802688430148824, "grad_norm": 1.0620953278309875, "learning_rate": 4.800912045524797e-08, "loss": 0.5554, "step": 5004 }, { "epoch": 4.803648583773404, "grad_norm": 1.0960683630537096, "learning_rate": 4.7547119441927466e-08, "loss": 0.4026, "step": 5005 }, { "epoch": 4.8046087373979836, "grad_norm": 0.694752760895144, "learning_rate": 4.7087341531766864e-08, "loss": 0.5784, "step": 5006 }, { "epoch": 4.805568891022563, "grad_norm": 0.6360772039373124, "learning_rate": 4.662978693115439e-08, "loss": 0.3846, "step": 5007 }, { "epoch": 4.806529044647144, "grad_norm": 1.2563847284543348, "learning_rate": 4.617445584548186e-08, "loss": 0.3679, "step": 5008 }, { "epoch": 4.807489198271724, "grad_norm": 0.8784235069381005, "learning_rate": 4.5721348479143e-08, "loss": 0.4477, "step": 5009 }, { "epoch": 4.808449351896304, "grad_norm": 0.9448701898155665, "learning_rate": 4.527046503553234e-08, "loss": 0.5089, "step": 5010 }, { "epoch": 4.809409505520883, "grad_norm": 0.8734286527203673, "learning_rate": 4.4821805717046864e-08, "loss": 0.4671, "step": 5011 }, { "epoch": 4.810369659145463, "grad_norm": 1.3815433354915312, "learning_rate": 4.437537072508491e-08, "loss": 0.4623, "step": 5012 }, { "epoch": 4.811329812770043, "grad_norm": 0.8570931623985958, "learning_rate": 4.393116026004618e-08, "loss": 0.3566, "step": 5013 }, { "epoch": 4.812289966394623, "grad_norm": 0.8827034072794893, "learning_rate": 4.3489174521332833e-08, "loss": 0.4527, "step": 5014 }, { "epoch": 4.8132501200192035, "grad_norm": 0.8668709257476908, "learning_rate": 4.304941370734616e-08, "loss": 0.4103, "step": 5015 }, { "epoch": 4.814210273643783, "grad_norm": 1.0239660273019386, "learning_rate": 4.261187801549105e-08, "loss": 0.4958, "step": 5016 }, { "epoch": 4.815170427268363, "grad_norm": 0.805172783151744, "learning_rate": 4.217656764217204e-08, "loss": 0.4527, "step": 5017 }, { "epoch": 4.816130580892943, "grad_norm": 0.8666808085584814, "learning_rate": 4.1743482782796166e-08, "loss": 0.4198, "step": 5018 }, { "epoch": 4.817090734517523, "grad_norm": 0.8823896951774227, "learning_rate": 4.1312623631770134e-08, "loss": 0.5335, "step": 5019 }, { "epoch": 4.8180508881421025, "grad_norm": 0.811124679433711, "learning_rate": 4.08839903825009e-08, "loss": 0.4881, "step": 5020 }, { "epoch": 4.819011041766682, "grad_norm": 0.9637931726672481, "learning_rate": 4.045758322739957e-08, "loss": 0.465, "step": 5021 }, { "epoch": 4.819971195391263, "grad_norm": 1.0514526507239297, "learning_rate": 4.003340235787412e-08, "loss": 0.4122, "step": 5022 }, { "epoch": 4.820931349015843, "grad_norm": 1.4772682497596772, "learning_rate": 3.961144796433502e-08, "loss": 0.5117, "step": 5023 }, { "epoch": 4.8218915026404225, "grad_norm": 0.8489723886169143, "learning_rate": 3.919172023619355e-08, "loss": 0.5661, "step": 5024 }, { "epoch": 4.822851656265002, "grad_norm": 0.6379632069220469, "learning_rate": 3.8774219361860655e-08, "loss": 0.4027, "step": 5025 }, { "epoch": 4.823811809889582, "grad_norm": 1.088323922788695, "learning_rate": 3.835894552874808e-08, "loss": 0.3751, "step": 5026 }, { "epoch": 4.824771963514162, "grad_norm": 1.5130286415402192, "learning_rate": 3.794589892326839e-08, "loss": 0.4304, "step": 5027 }, { "epoch": 4.825732117138742, "grad_norm": 0.8133224200222654, "learning_rate": 3.75350797308327e-08, "loss": 0.5123, "step": 5028 }, { "epoch": 4.826692270763322, "grad_norm": 0.9685785174816965, "learning_rate": 3.712648813585407e-08, "loss": 0.4862, "step": 5029 }, { "epoch": 4.827652424387902, "grad_norm": 1.1088180158783543, "learning_rate": 3.672012432174465e-08, "loss": 0.4504, "step": 5030 }, { "epoch": 4.828612578012482, "grad_norm": 0.905477622468321, "learning_rate": 3.631598847091688e-08, "loss": 0.4809, "step": 5031 }, { "epoch": 4.829572731637062, "grad_norm": 1.0686369926429558, "learning_rate": 3.591408076478342e-08, "loss": 0.5604, "step": 5032 }, { "epoch": 4.830532885261642, "grad_norm": 0.8783633770904773, "learning_rate": 3.5514401383755505e-08, "loss": 0.5164, "step": 5033 }, { "epoch": 4.831493038886222, "grad_norm": 1.1908322515763323, "learning_rate": 3.511695050724573e-08, "loss": 0.4847, "step": 5034 }, { "epoch": 4.832453192510802, "grad_norm": 0.6639111049133329, "learning_rate": 3.4721728313664715e-08, "loss": 0.4204, "step": 5035 }, { "epoch": 4.833413346135382, "grad_norm": 1.0152311159961058, "learning_rate": 3.432873498042444e-08, "loss": 0.3685, "step": 5036 }, { "epoch": 4.834373499759962, "grad_norm": 1.5755089344448918, "learning_rate": 3.3937970683934893e-08, "loss": 0.4696, "step": 5037 }, { "epoch": 4.8353336533845415, "grad_norm": 0.785061334154648, "learning_rate": 3.354943559960577e-08, "loss": 0.4218, "step": 5038 }, { "epoch": 4.836293807009121, "grad_norm": 0.7993859814086511, "learning_rate": 3.3163129901847e-08, "loss": 0.5004, "step": 5039 }, { "epoch": 4.837253960633701, "grad_norm": 1.1050876526150255, "learning_rate": 3.277905376406654e-08, "loss": 0.559, "step": 5040 }, { "epoch": 4.838214114258282, "grad_norm": 0.9049107043871759, "learning_rate": 3.2397207358672024e-08, "loss": 0.6156, "step": 5041 }, { "epoch": 4.8391742678828615, "grad_norm": 0.686011684005548, "learning_rate": 3.201759085707079e-08, "loss": 0.477, "step": 5042 }, { "epoch": 4.840134421507441, "grad_norm": 0.7869729502575403, "learning_rate": 3.164020442966764e-08, "loss": 0.5772, "step": 5043 }, { "epoch": 4.841094575132021, "grad_norm": 1.034613748293376, "learning_rate": 3.126504824586818e-08, "loss": 0.4344, "step": 5044 }, { "epoch": 4.842054728756601, "grad_norm": 0.8153226470707756, "learning_rate": 3.0892122474076025e-08, "loss": 0.4826, "step": 5045 }, { "epoch": 4.843014882381181, "grad_norm": 0.8145225151977139, "learning_rate": 3.0521427281693386e-08, "loss": 0.4219, "step": 5046 }, { "epoch": 4.8439750360057605, "grad_norm": 0.6010595816398245, "learning_rate": 3.015296283512048e-08, "loss": 0.3939, "step": 5047 }, { "epoch": 4.844935189630341, "grad_norm": 0.8195882115283996, "learning_rate": 2.9786729299758343e-08, "loss": 0.3932, "step": 5048 }, { "epoch": 4.845895343254921, "grad_norm": 0.7539184060421315, "learning_rate": 2.942272684000491e-08, "loss": 0.4097, "step": 5049 }, { "epoch": 4.846855496879501, "grad_norm": 1.0966662412983454, "learning_rate": 2.9060955619256702e-08, "loss": 0.4862, "step": 5050 }, { "epoch": 4.847815650504081, "grad_norm": 0.8749361830347784, "learning_rate": 2.8701415799908815e-08, "loss": 0.5242, "step": 5051 }, { "epoch": 4.84877580412866, "grad_norm": 1.2860526392388858, "learning_rate": 2.8344107543356035e-08, "loss": 0.4043, "step": 5052 }, { "epoch": 4.84973595775324, "grad_norm": 0.8110950594219928, "learning_rate": 2.7989031009988398e-08, "loss": 0.5271, "step": 5053 }, { "epoch": 4.85069611137782, "grad_norm": 0.8848391052648131, "learning_rate": 2.763618635919729e-08, "loss": 0.3485, "step": 5054 }, { "epoch": 4.851656265002401, "grad_norm": 0.7747552297239565, "learning_rate": 2.7285573749370465e-08, "loss": 0.452, "step": 5055 }, { "epoch": 4.8526164186269805, "grad_norm": 1.3102051183254013, "learning_rate": 2.693719333789424e-08, "loss": 0.4201, "step": 5056 }, { "epoch": 4.85357657225156, "grad_norm": 0.9552772202494914, "learning_rate": 2.6591045281152418e-08, "loss": 0.4698, "step": 5057 }, { "epoch": 4.85453672587614, "grad_norm": 1.1995285303177305, "learning_rate": 2.6247129734528474e-08, "loss": 0.4346, "step": 5058 }, { "epoch": 4.85549687950072, "grad_norm": 0.8788920733157192, "learning_rate": 2.5905446852400594e-08, "loss": 0.3942, "step": 5059 }, { "epoch": 4.8564570331253005, "grad_norm": 0.8367351840135419, "learning_rate": 2.5565996788148306e-08, "loss": 0.5217, "step": 5060 }, { "epoch": 4.85741718674988, "grad_norm": 0.82476437121515, "learning_rate": 2.5228779694145832e-08, "loss": 0.5134, "step": 5061 }, { "epoch": 4.85837734037446, "grad_norm": 0.7459025248479949, "learning_rate": 2.4893795721766534e-08, "loss": 0.5024, "step": 5062 }, { "epoch": 4.85933749399904, "grad_norm": 1.120863883567557, "learning_rate": 2.4561045021382346e-08, "loss": 0.4406, "step": 5063 }, { "epoch": 4.86029764762362, "grad_norm": 0.8923836227154965, "learning_rate": 2.4230527742359344e-08, "loss": 0.4199, "step": 5064 }, { "epoch": 4.8612578012481995, "grad_norm": 1.2461189100322032, "learning_rate": 2.3902244033065512e-08, "loss": 0.5249, "step": 5065 }, { "epoch": 4.862217954872779, "grad_norm": 0.8039568062576846, "learning_rate": 2.3576194040861866e-08, "loss": 0.4537, "step": 5066 }, { "epoch": 4.86317810849736, "grad_norm": 0.7772701266569403, "learning_rate": 2.3252377912110214e-08, "loss": 0.6216, "step": 5067 }, { "epoch": 4.86413826212194, "grad_norm": 0.9922178420004221, "learning_rate": 2.2930795792167615e-08, "loss": 0.5247, "step": 5068 }, { "epoch": 4.86509841574652, "grad_norm": 0.8920607171356024, "learning_rate": 2.2611447825389154e-08, "loss": 0.5441, "step": 5069 }, { "epoch": 4.866058569371099, "grad_norm": 0.8370676475861539, "learning_rate": 2.2294334155125718e-08, "loss": 0.3805, "step": 5070 }, { "epoch": 4.867018722995679, "grad_norm": 0.8751656053386265, "learning_rate": 2.1979454923727327e-08, "loss": 0.4978, "step": 5071 }, { "epoch": 4.867978876620259, "grad_norm": 0.7329426147654167, "learning_rate": 2.1666810272539806e-08, "loss": 0.4437, "step": 5072 }, { "epoch": 4.868939030244839, "grad_norm": 0.7310267642686481, "learning_rate": 2.1356400341905337e-08, "loss": 0.458, "step": 5073 }, { "epoch": 4.8698991838694194, "grad_norm": 0.685749175952488, "learning_rate": 2.1048225271164125e-08, "loss": 0.4738, "step": 5074 }, { "epoch": 4.870859337493999, "grad_norm": 0.6780626829279383, "learning_rate": 2.0742285198652734e-08, "loss": 0.4444, "step": 5075 }, { "epoch": 4.871819491118579, "grad_norm": 0.9208037157719249, "learning_rate": 2.0438580261704087e-08, "loss": 0.4213, "step": 5076 }, { "epoch": 4.872779644743159, "grad_norm": 0.8621891672765598, "learning_rate": 2.0137110596648025e-08, "loss": 0.4847, "step": 5077 }, { "epoch": 4.873739798367739, "grad_norm": 0.8679446534570034, "learning_rate": 1.9837876338811292e-08, "loss": 0.4614, "step": 5078 }, { "epoch": 4.874699951992318, "grad_norm": 1.1081451153390953, "learning_rate": 1.9540877622516995e-08, "loss": 0.4884, "step": 5079 }, { "epoch": 4.875660105616898, "grad_norm": 0.6868183585800003, "learning_rate": 1.924611458108461e-08, "loss": 0.4319, "step": 5080 }, { "epoch": 4.876620259241479, "grad_norm": 0.8488697516076139, "learning_rate": 1.8953587346829948e-08, "loss": 0.4856, "step": 5081 }, { "epoch": 4.877580412866059, "grad_norm": 1.1114265992748857, "learning_rate": 1.8663296051065762e-08, "loss": 0.4148, "step": 5082 }, { "epoch": 4.8785405664906385, "grad_norm": 0.7089885396766048, "learning_rate": 1.8375240824100028e-08, "loss": 0.5714, "step": 5083 }, { "epoch": 4.879500720115218, "grad_norm": 0.6388661592086452, "learning_rate": 1.8089421795238205e-08, "loss": 0.4226, "step": 5084 }, { "epoch": 4.880460873739798, "grad_norm": 0.8585380345833423, "learning_rate": 1.7805839092781553e-08, "loss": 0.4763, "step": 5085 }, { "epoch": 4.881421027364379, "grad_norm": 0.8060933028224324, "learning_rate": 1.7524492844026576e-08, "loss": 0.3919, "step": 5086 }, { "epoch": 4.882381180988959, "grad_norm": 0.819272184495507, "learning_rate": 1.724538317526725e-08, "loss": 0.4848, "step": 5087 }, { "epoch": 4.883341334613538, "grad_norm": 0.8535309344010886, "learning_rate": 1.69685102117928e-08, "loss": 0.5648, "step": 5088 }, { "epoch": 4.884301488238118, "grad_norm": 0.5856782275084846, "learning_rate": 1.6693874077888805e-08, "loss": 0.3805, "step": 5089 }, { "epoch": 4.885261641862698, "grad_norm": 1.0606217235743496, "learning_rate": 1.6421474896835544e-08, "loss": 0.3973, "step": 5090 }, { "epoch": 4.886221795487278, "grad_norm": 0.9576571605679899, "learning_rate": 1.615131279091131e-08, "loss": 0.5756, "step": 5091 }, { "epoch": 4.8871819491118575, "grad_norm": 0.9920998253355734, "learning_rate": 1.588338788138799e-08, "loss": 0.4903, "step": 5092 }, { "epoch": 4.888142102736438, "grad_norm": 1.149776312794814, "learning_rate": 1.5617700288534377e-08, "loss": 0.4998, "step": 5093 }, { "epoch": 4.889102256361018, "grad_norm": 0.8613121303596031, "learning_rate": 1.5354250131615622e-08, "loss": 0.4883, "step": 5094 }, { "epoch": 4.890062409985598, "grad_norm": 0.7896390829081027, "learning_rate": 1.5093037528890463e-08, "loss": 0.4528, "step": 5095 }, { "epoch": 4.891022563610178, "grad_norm": 0.8329425606945656, "learning_rate": 1.4834062597615107e-08, "loss": 0.4239, "step": 5096 }, { "epoch": 4.891982717234757, "grad_norm": 0.8435450032879405, "learning_rate": 1.4577325454041003e-08, "loss": 0.4787, "step": 5097 }, { "epoch": 4.892942870859337, "grad_norm": 0.9431452378758758, "learning_rate": 1.4322826213414298e-08, "loss": 0.4287, "step": 5098 }, { "epoch": 4.893903024483917, "grad_norm": 0.8482404864543186, "learning_rate": 1.4070564989976387e-08, "loss": 0.5874, "step": 5099 }, { "epoch": 4.894863178108498, "grad_norm": 0.8751048756117361, "learning_rate": 1.3820541896965579e-08, "loss": 0.5286, "step": 5100 }, { "epoch": 4.8958233317330775, "grad_norm": 0.7512896247181582, "learning_rate": 1.3572757046614315e-08, "loss": 0.5061, "step": 5101 }, { "epoch": 4.896783485357657, "grad_norm": 0.8802932429511584, "learning_rate": 1.3327210550150293e-08, "loss": 0.4083, "step": 5102 }, { "epoch": 4.897743638982237, "grad_norm": 1.044254494543596, "learning_rate": 1.308390251779701e-08, "loss": 0.3932, "step": 5103 }, { "epoch": 4.898703792606817, "grad_norm": 0.9537696102243438, "learning_rate": 1.2842833058772653e-08, "loss": 0.4714, "step": 5104 }, { "epoch": 4.899663946231397, "grad_norm": 1.1274917495148677, "learning_rate": 1.260400228129066e-08, "loss": 0.4375, "step": 5105 }, { "epoch": 4.9006240998559765, "grad_norm": 0.7727089472035121, "learning_rate": 1.2367410292560277e-08, "loss": 0.4887, "step": 5106 }, { "epoch": 4.901584253480557, "grad_norm": 0.8138366837602501, "learning_rate": 1.2133057198784326e-08, "loss": 0.4113, "step": 5107 }, { "epoch": 4.902544407105137, "grad_norm": 0.8173689739226606, "learning_rate": 1.1900943105160878e-08, "loss": 0.5035, "step": 5108 }, { "epoch": 4.903504560729717, "grad_norm": 0.7640020133476769, "learning_rate": 1.1671068115884921e-08, "loss": 0.4552, "step": 5109 }, { "epoch": 4.9044647143542965, "grad_norm": 1.1631209192265621, "learning_rate": 1.1443432334144466e-08, "loss": 0.4643, "step": 5110 }, { "epoch": 4.905424867978876, "grad_norm": 1.0361614115675732, "learning_rate": 1.1218035862122223e-08, "loss": 0.489, "step": 5111 }, { "epoch": 4.906385021603457, "grad_norm": 0.7100888504553781, "learning_rate": 1.09948788009967e-08, "loss": 0.4329, "step": 5112 }, { "epoch": 4.907345175228037, "grad_norm": 1.174802904259139, "learning_rate": 1.0773961250939991e-08, "loss": 0.4546, "step": 5113 }, { "epoch": 4.908305328852617, "grad_norm": 0.9384848367244667, "learning_rate": 1.0555283311121101e-08, "loss": 0.3966, "step": 5114 }, { "epoch": 4.909265482477196, "grad_norm": 0.8545888719025044, "learning_rate": 1.0338845079700954e-08, "loss": 0.4478, "step": 5115 }, { "epoch": 4.910225636101776, "grad_norm": 1.085202315372236, "learning_rate": 1.0124646653836279e-08, "loss": 0.5051, "step": 5116 }, { "epoch": 4.911185789726356, "grad_norm": 1.1700438220948715, "learning_rate": 9.912688129679603e-09, "loss": 0.4989, "step": 5117 }, { "epoch": 4.912145943350936, "grad_norm": 1.2635212414162413, "learning_rate": 9.702969602375378e-09, "loss": 0.3333, "step": 5118 }, { "epoch": 4.9131060969755165, "grad_norm": 0.9090582123965726, "learning_rate": 9.495491166065519e-09, "loss": 0.5242, "step": 5119 }, { "epoch": 4.914066250600096, "grad_norm": 0.678867477711088, "learning_rate": 9.290252913883856e-09, "loss": 0.4338, "step": 5120 }, { "epoch": 4.915026404224676, "grad_norm": 0.7433315101175434, "learning_rate": 9.087254937960032e-09, "loss": 0.4472, "step": 5121 }, { "epoch": 4.915986557849256, "grad_norm": 1.0052295753102116, "learning_rate": 8.886497329417265e-09, "loss": 0.4453, "step": 5122 }, { "epoch": 4.916946711473836, "grad_norm": 1.5543759729540403, "learning_rate": 8.687980178374023e-09, "loss": 0.4052, "step": 5123 }, { "epoch": 4.9179068650984155, "grad_norm": 0.8822778230536781, "learning_rate": 8.491703573942356e-09, "loss": 0.4596, "step": 5124 }, { "epoch": 4.918867018722995, "grad_norm": 0.7639090306243572, "learning_rate": 8.29766760422901e-09, "loss": 0.4896, "step": 5125 }, { "epoch": 4.919827172347576, "grad_norm": 0.7266883933724783, "learning_rate": 8.105872356333755e-09, "loss": 0.5727, "step": 5126 }, { "epoch": 4.920787325972156, "grad_norm": 1.4097059179736806, "learning_rate": 7.916317916352168e-09, "loss": 0.3866, "step": 5127 }, { "epoch": 4.9217474795967355, "grad_norm": 0.6798079942580292, "learning_rate": 7.729004369373405e-09, "loss": 0.3778, "step": 5128 }, { "epoch": 4.922707633221315, "grad_norm": 0.9391685409993102, "learning_rate": 7.543931799479653e-09, "loss": 0.4589, "step": 5129 }, { "epoch": 4.923667786845895, "grad_norm": 1.072829547964442, "learning_rate": 7.3611002897489015e-09, "loss": 0.3958, "step": 5130 }, { "epoch": 4.924627940470475, "grad_norm": 1.1254629074791682, "learning_rate": 7.180509922251611e-09, "loss": 0.4314, "step": 5131 }, { "epoch": 4.925588094095055, "grad_norm": 1.315501393563341, "learning_rate": 7.002160778053491e-09, "loss": 0.5644, "step": 5132 }, { "epoch": 4.926548247719635, "grad_norm": 1.0749259043442854, "learning_rate": 6.826052937212724e-09, "loss": 0.4612, "step": 5133 }, { "epoch": 4.927508401344215, "grad_norm": 1.0390371548317028, "learning_rate": 6.652186478783296e-09, "loss": 0.4058, "step": 5134 }, { "epoch": 4.928468554968795, "grad_norm": 0.8710531338405143, "learning_rate": 6.480561480811109e-09, "loss": 0.5546, "step": 5135 }, { "epoch": 4.929428708593375, "grad_norm": 0.8009250943425297, "learning_rate": 6.311178020336761e-09, "loss": 0.4988, "step": 5136 }, { "epoch": 4.930388862217955, "grad_norm": 0.7755860777109317, "learning_rate": 6.14403617339554e-09, "loss": 0.5272, "step": 5137 }, { "epoch": 4.931349015842535, "grad_norm": 0.9997612045820652, "learning_rate": 5.979136015015208e-09, "loss": 0.5651, "step": 5138 }, { "epoch": 4.932309169467115, "grad_norm": 0.8164776785972812, "learning_rate": 5.8164776192171094e-09, "loss": 0.4453, "step": 5139 }, { "epoch": 4.933269323091695, "grad_norm": 0.9948561291861142, "learning_rate": 5.656061059017837e-09, "loss": 0.4136, "step": 5140 }, { "epoch": 4.934229476716275, "grad_norm": 0.9881020194934127, "learning_rate": 5.4978864064259005e-09, "loss": 0.4539, "step": 5141 }, { "epoch": 4.9351896303408544, "grad_norm": 0.8851233956524345, "learning_rate": 5.3419537324445044e-09, "loss": 0.5711, "step": 5142 }, { "epoch": 4.936149783965434, "grad_norm": 0.8070868680362038, "learning_rate": 5.18826310706988e-09, "loss": 0.4534, "step": 5143 }, { "epoch": 4.937109937590014, "grad_norm": 1.11313069125113, "learning_rate": 5.0368145992929495e-09, "loss": 0.504, "step": 5144 }, { "epoch": 4.938070091214595, "grad_norm": 1.3763513113759136, "learning_rate": 4.887608277096001e-09, "loss": 0.4471, "step": 5145 }, { "epoch": 4.9390302448391745, "grad_norm": 0.7891805588992028, "learning_rate": 4.740644207457678e-09, "loss": 0.4588, "step": 5146 }, { "epoch": 4.939990398463754, "grad_norm": 0.9107277578234102, "learning_rate": 4.5959224563474345e-09, "loss": 0.435, "step": 5147 }, { "epoch": 4.940950552088334, "grad_norm": 0.9942782400179924, "learning_rate": 4.4534430887299694e-09, "loss": 0.3507, "step": 5148 }, { "epoch": 4.941910705712914, "grad_norm": 1.2363295084137553, "learning_rate": 4.3132061685619005e-09, "loss": 0.4415, "step": 5149 }, { "epoch": 4.942870859337494, "grad_norm": 0.8561909512422351, "learning_rate": 4.1752117587950945e-09, "loss": 0.4284, "step": 5150 }, { "epoch": 4.9438310129620735, "grad_norm": 1.068297368094186, "learning_rate": 4.0394599213733346e-09, "loss": 0.4155, "step": 5151 }, { "epoch": 4.944791166586654, "grad_norm": 0.7346866571499964, "learning_rate": 3.905950717233986e-09, "loss": 0.4429, "step": 5152 }, { "epoch": 4.945751320211234, "grad_norm": 0.8411895497929577, "learning_rate": 3.774684206308554e-09, "loss": 0.4912, "step": 5153 }, { "epoch": 4.946711473835814, "grad_norm": 0.8010052261164681, "learning_rate": 3.6456604475199055e-09, "loss": 0.4619, "step": 5154 }, { "epoch": 4.947671627460394, "grad_norm": 0.8400258989946399, "learning_rate": 3.5188794987867093e-09, "loss": 0.504, "step": 5155 }, { "epoch": 4.948631781084973, "grad_norm": 0.7596043537068934, "learning_rate": 3.394341417018998e-09, "loss": 0.4589, "step": 5156 }, { "epoch": 4.949591934709553, "grad_norm": 0.7935425958986553, "learning_rate": 3.272046258120942e-09, "loss": 0.4646, "step": 5157 }, { "epoch": 4.950552088334134, "grad_norm": 0.8096511936269737, "learning_rate": 3.151994076988629e-09, "loss": 0.4886, "step": 5158 }, { "epoch": 4.951512241958714, "grad_norm": 0.6654480953556278, "learning_rate": 3.034184927513395e-09, "loss": 0.4458, "step": 5159 }, { "epoch": 4.952472395583293, "grad_norm": 1.0451122295393336, "learning_rate": 2.9186188625779376e-09, "loss": 0.4961, "step": 5160 }, { "epoch": 4.953432549207873, "grad_norm": 1.3452085990181215, "learning_rate": 2.805295934058538e-09, "loss": 0.5307, "step": 5161 }, { "epoch": 4.954392702832453, "grad_norm": 0.8497550813263275, "learning_rate": 2.694216192824506e-09, "loss": 0.4462, "step": 5162 }, { "epoch": 4.955352856457033, "grad_norm": 0.7271518882120606, "learning_rate": 2.5853796887387316e-09, "loss": 0.5373, "step": 5163 }, { "epoch": 4.9563130100816135, "grad_norm": 0.8543370435296792, "learning_rate": 2.47878647065658e-09, "loss": 0.4518, "step": 5164 }, { "epoch": 4.957273163706193, "grad_norm": 0.7008640021221415, "learning_rate": 2.374436586426443e-09, "loss": 0.4286, "step": 5165 }, { "epoch": 4.958233317330773, "grad_norm": 0.8999383727575183, "learning_rate": 2.27233008289085e-09, "loss": 0.5294, "step": 5166 }, { "epoch": 4.959193470955353, "grad_norm": 1.0503727172712398, "learning_rate": 2.1724670058825838e-09, "loss": 0.4519, "step": 5167 }, { "epoch": 4.960153624579933, "grad_norm": 0.8283415818201451, "learning_rate": 2.074847400230784e-09, "loss": 0.4257, "step": 5168 }, { "epoch": 4.9611137782045125, "grad_norm": 0.7504663896588, "learning_rate": 1.9794713097548433e-09, "loss": 0.4744, "step": 5169 }, { "epoch": 4.962073931829092, "grad_norm": 0.8240841632639144, "learning_rate": 1.886338777268293e-09, "loss": 0.469, "step": 5170 }, { "epoch": 4.963034085453673, "grad_norm": 0.8011922742424361, "learning_rate": 1.7954498445771352e-09, "loss": 0.4655, "step": 5171 }, { "epoch": 4.963994239078253, "grad_norm": 0.9210823610071658, "learning_rate": 1.7068045524815114e-09, "loss": 0.414, "step": 5172 }, { "epoch": 4.964954392702833, "grad_norm": 0.865016279582713, "learning_rate": 1.6204029407718147e-09, "loss": 0.4163, "step": 5173 }, { "epoch": 4.965914546327412, "grad_norm": 0.8646282322120049, "learning_rate": 1.5362450482336865e-09, "loss": 0.4694, "step": 5174 }, { "epoch": 4.966874699951992, "grad_norm": 0.8169822765589153, "learning_rate": 1.4543309126446858e-09, "loss": 0.4918, "step": 5175 }, { "epoch": 4.967834853576572, "grad_norm": 1.7232991590262372, "learning_rate": 1.374660570774844e-09, "loss": 0.3388, "step": 5176 }, { "epoch": 4.968795007201152, "grad_norm": 0.8541615452779091, "learning_rate": 1.2972340583877751e-09, "loss": 0.4335, "step": 5177 }, { "epoch": 4.969755160825732, "grad_norm": 0.6797051178798209, "learning_rate": 1.2220514102390114e-09, "loss": 0.4165, "step": 5178 }, { "epoch": 4.970715314450312, "grad_norm": 1.0262637389377658, "learning_rate": 1.149112660077667e-09, "loss": 0.4676, "step": 5179 }, { "epoch": 4.971675468074892, "grad_norm": 0.7421054755148897, "learning_rate": 1.0784178406442192e-09, "loss": 0.46, "step": 5180 }, { "epoch": 4.972635621699472, "grad_norm": 1.1365365071368903, "learning_rate": 1.0099669836743931e-09, "loss": 0.5276, "step": 5181 }, { "epoch": 4.973595775324052, "grad_norm": 1.3472759278876016, "learning_rate": 9.43760119893611e-10, "loss": 0.4095, "step": 5182 }, { "epoch": 4.974555928948631, "grad_norm": 0.8711268366446056, "learning_rate": 8.797972790219877e-10, "loss": 0.4924, "step": 5183 }, { "epoch": 4.975516082573212, "grad_norm": 0.6838886711693785, "learning_rate": 8.180784897715565e-10, "loss": 0.4519, "step": 5184 }, { "epoch": 4.976476236197792, "grad_norm": 1.0480564205182308, "learning_rate": 7.586037798479329e-10, "loss": 0.4435, "step": 5185 }, { "epoch": 4.977436389822372, "grad_norm": 0.7549657624308767, "learning_rate": 7.013731759475395e-10, "loss": 0.4603, "step": 5186 }, { "epoch": 4.9783965434469515, "grad_norm": 1.018818779200841, "learning_rate": 6.463867037614923e-10, "loss": 0.5548, "step": 5187 }, { "epoch": 4.979356697071531, "grad_norm": 0.8234643660805117, "learning_rate": 5.936443879728249e-10, "loss": 0.4909, "step": 5188 }, { "epoch": 4.980316850696111, "grad_norm": 0.9697875738641131, "learning_rate": 5.431462522559327e-10, "loss": 0.4031, "step": 5189 }, { "epoch": 4.981277004320692, "grad_norm": 0.7128262057139834, "learning_rate": 4.948923192793497e-10, "loss": 0.5013, "step": 5190 }, { "epoch": 4.9822371579452716, "grad_norm": 0.9845202916586991, "learning_rate": 4.4888261070408223e-10, "loss": 0.4801, "step": 5191 }, { "epoch": 4.983197311569851, "grad_norm": 1.1075035329556777, "learning_rate": 4.0511714718305396e-10, "loss": 0.481, "step": 5192 }, { "epoch": 4.984157465194431, "grad_norm": 0.8312873372852295, "learning_rate": 3.6359594836277156e-10, "loss": 0.4659, "step": 5193 }, { "epoch": 4.985117618819011, "grad_norm": 0.7914542650428149, "learning_rate": 3.24319032880549e-10, "loss": 0.4832, "step": 5194 }, { "epoch": 4.986077772443591, "grad_norm": 1.1879593629474872, "learning_rate": 2.872864183678381e-10, "loss": 0.4562, "step": 5195 }, { "epoch": 4.9870379260681705, "grad_norm": 0.9399864110395678, "learning_rate": 2.5249812144856335e-10, "loss": 0.4489, "step": 5196 }, { "epoch": 4.987998079692751, "grad_norm": 1.1999031183265763, "learning_rate": 2.1995415773856666e-10, "loss": 0.4985, "step": 5197 }, { "epoch": 4.988958233317331, "grad_norm": 0.8866133890399338, "learning_rate": 1.8965454184671773e-10, "loss": 0.4801, "step": 5198 }, { "epoch": 4.989918386941911, "grad_norm": 1.0498947021334344, "learning_rate": 1.615992873732486e-10, "loss": 0.5237, "step": 5199 }, { "epoch": 4.990878540566491, "grad_norm": 0.8661200216855813, "learning_rate": 1.3578840691308438e-10, "loss": 0.3666, "step": 5200 }, { "epoch": 4.99183869419107, "grad_norm": 1.1827533203092808, "learning_rate": 1.1222191205195743e-10, "loss": 0.407, "step": 5201 }, { "epoch": 4.99279884781565, "grad_norm": 0.7920351871220386, "learning_rate": 9.089981336807275e-11, "loss": 0.5039, "step": 5202 }, { "epoch": 4.99375900144023, "grad_norm": 0.6780231131044385, "learning_rate": 7.182212043377323e-11, "loss": 0.4799, "step": 5203 }, { "epoch": 4.994719155064811, "grad_norm": 0.876842056595737, "learning_rate": 5.4988841812209095e-11, "loss": 0.4384, "step": 5204 }, { "epoch": 4.9956793086893905, "grad_norm": 0.87548861306638, "learning_rate": 4.039998505900311e-11, "loss": 0.5204, "step": 5205 }, { "epoch": 4.99663946231397, "grad_norm": 0.9076014685320334, "learning_rate": 2.8055556724471132e-11, "loss": 0.4943, "step": 5206 }, { "epoch": 4.99759961593855, "grad_norm": 0.9884805995321474, "learning_rate": 1.7955562348626054e-11, "loss": 0.3779, "step": 5207 }, { "epoch": 4.99855976956313, "grad_norm": 1.0691057041032068, "learning_rate": 1.0100006465618706e-11, "loss": 0.4783, "step": 5208 }, { "epoch": 4.99951992318771, "grad_norm": 0.6219173867139143, "learning_rate": 4.4888926020725254e-12, "loss": 0.3704, "step": 5209 }, { "epoch": 5.0, "grad_norm": 0.6219173867139143, "learning_rate": 1.1222232765284447e-12, "loss": 0.4816, "step": 5210 }, { "epoch": 5.0, "step": 5210, "total_flos": 620043550064640.0, "train_loss": 0.5824799254565229, "train_runtime": 137583.5891, "train_samples_per_second": 0.303, "train_steps_per_second": 0.038 } ], "logging_steps": 1, "max_steps": 5210, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 620043550064640.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }