{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.95195530726257, "eval_steps": 112, "global_step": 669, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004464285714285714, "grad_norm": 5.096107617240597, "learning_rate": 5.000000000000001e-07, "loss": 0.81, "step": 1 }, { "epoch": 0.004464285714285714, "eval_loss": 0.7919066548347473, "eval_runtime": 32.2179, "eval_samples_per_second": 82.873, "eval_steps_per_second": 5.183, "step": 1 }, { "epoch": 0.008928571428571428, "grad_norm": 4.9041704402978805, "learning_rate": 1.0000000000000002e-06, "loss": 0.7839, "step": 2 }, { "epoch": 0.013392857142857142, "grad_norm": 4.977917588021941, "learning_rate": 1.5e-06, "loss": 0.7745, "step": 3 }, { "epoch": 0.017857142857142856, "grad_norm": 4.7373014121550705, "learning_rate": 2.0000000000000003e-06, "loss": 0.7681, "step": 4 }, { "epoch": 0.022321428571428572, "grad_norm": 4.563297129857777, "learning_rate": 2.5e-06, "loss": 0.7589, "step": 5 }, { "epoch": 0.026785714285714284, "grad_norm": 3.5264108754930787, "learning_rate": 3e-06, "loss": 0.7629, "step": 6 }, { "epoch": 0.03125, "grad_norm": 3.4126703272457166, "learning_rate": 3.5e-06, "loss": 0.7181, "step": 7 }, { "epoch": 0.03571428571428571, "grad_norm": 3.6971429493758636, "learning_rate": 4.000000000000001e-06, "loss": 0.667, "step": 8 }, { "epoch": 0.04017857142857143, "grad_norm": 1.940701767077778, "learning_rate": 4.5e-06, "loss": 0.6629, "step": 9 }, { "epoch": 0.044642857142857144, "grad_norm": 12.40862535155878, "learning_rate": 5e-06, "loss": 0.6832, "step": 10 }, { "epoch": 0.049107142857142856, "grad_norm": 2.8235547786373867, "learning_rate": 4.992447129909366e-06, "loss": 0.6213, "step": 11 }, { "epoch": 0.05357142857142857, "grad_norm": 1.734602557718062, "learning_rate": 4.984894259818732e-06, "loss": 0.6279, "step": 12 }, { "epoch": 0.05803571428571429, "grad_norm": 1.3840828717613172, "learning_rate": 4.977341389728097e-06, "loss": 0.6134, "step": 13 }, { "epoch": 0.0625, "grad_norm": 1.1262172615918475, "learning_rate": 4.969788519637463e-06, "loss": 0.6029, "step": 14 }, { "epoch": 0.06696428571428571, "grad_norm": 0.9263739512436016, "learning_rate": 4.962235649546828e-06, "loss": 0.6067, "step": 15 }, { "epoch": 0.07142857142857142, "grad_norm": 0.8374584146721384, "learning_rate": 4.954682779456194e-06, "loss": 0.6014, "step": 16 }, { "epoch": 0.07589285714285714, "grad_norm": 0.7757487058764371, "learning_rate": 4.9471299093655595e-06, "loss": 0.5906, "step": 17 }, { "epoch": 0.08035714285714286, "grad_norm": 0.6531787929864147, "learning_rate": 4.939577039274925e-06, "loss": 0.5589, "step": 18 }, { "epoch": 0.08482142857142858, "grad_norm": 0.7223853317191319, "learning_rate": 4.93202416918429e-06, "loss": 0.5877, "step": 19 }, { "epoch": 0.08928571428571429, "grad_norm": 0.6793848952870568, "learning_rate": 4.924471299093656e-06, "loss": 0.6075, "step": 20 }, { "epoch": 0.09375, "grad_norm": 0.65689704768199, "learning_rate": 4.9169184290030215e-06, "loss": 0.5659, "step": 21 }, { "epoch": 0.09821428571428571, "grad_norm": 0.6553424447930819, "learning_rate": 4.909365558912387e-06, "loss": 0.5728, "step": 22 }, { "epoch": 0.10267857142857142, "grad_norm": 0.6095768945247507, "learning_rate": 4.901812688821753e-06, "loss": 0.5716, "step": 23 }, { "epoch": 0.10714285714285714, "grad_norm": 0.5874011280839848, "learning_rate": 4.894259818731118e-06, "loss": 0.5736, "step": 24 }, { "epoch": 0.11160714285714286, "grad_norm": 0.642646780124128, "learning_rate": 4.8867069486404835e-06, "loss": 0.5889, "step": 25 }, { "epoch": 0.11607142857142858, "grad_norm": 0.5585168252549827, "learning_rate": 4.879154078549849e-06, "loss": 0.5761, "step": 26 }, { "epoch": 0.12053571428571429, "grad_norm": 0.4948199084353006, "learning_rate": 4.871601208459215e-06, "loss": 0.545, "step": 27 }, { "epoch": 0.125, "grad_norm": 0.580699538693282, "learning_rate": 4.864048338368581e-06, "loss": 0.5874, "step": 28 }, { "epoch": 0.12946428571428573, "grad_norm": 0.5483036703790811, "learning_rate": 4.8564954682779455e-06, "loss": 0.553, "step": 29 }, { "epoch": 0.13392857142857142, "grad_norm": 0.5207516065535861, "learning_rate": 4.848942598187312e-06, "loss": 0.54, "step": 30 }, { "epoch": 0.13839285714285715, "grad_norm": 0.5453894595978196, "learning_rate": 4.841389728096677e-06, "loss": 0.5779, "step": 31 }, { "epoch": 0.14285714285714285, "grad_norm": 0.5223897707340578, "learning_rate": 4.833836858006043e-06, "loss": 0.5518, "step": 32 }, { "epoch": 0.14732142857142858, "grad_norm": 0.5473812522488812, "learning_rate": 4.826283987915408e-06, "loss": 0.5831, "step": 33 }, { "epoch": 0.15178571428571427, "grad_norm": 0.5437546013329395, "learning_rate": 4.818731117824774e-06, "loss": 0.5884, "step": 34 }, { "epoch": 0.15625, "grad_norm": 0.5305579139277316, "learning_rate": 4.81117824773414e-06, "loss": 0.5398, "step": 35 }, { "epoch": 0.16071428571428573, "grad_norm": 0.5258419287632591, "learning_rate": 4.803625377643505e-06, "loss": 0.5593, "step": 36 }, { "epoch": 0.16517857142857142, "grad_norm": 0.5158851354579528, "learning_rate": 4.79607250755287e-06, "loss": 0.5564, "step": 37 }, { "epoch": 0.16964285714285715, "grad_norm": 0.4824844062486759, "learning_rate": 4.788519637462236e-06, "loss": 0.5616, "step": 38 }, { "epoch": 0.17410714285714285, "grad_norm": 0.5173226727977146, "learning_rate": 4.780966767371602e-06, "loss": 0.5674, "step": 39 }, { "epoch": 0.17857142857142858, "grad_norm": 0.4966629400190002, "learning_rate": 4.773413897280967e-06, "loss": 0.5833, "step": 40 }, { "epoch": 0.18303571428571427, "grad_norm": 0.48186435126437244, "learning_rate": 4.765861027190333e-06, "loss": 0.5571, "step": 41 }, { "epoch": 0.1875, "grad_norm": 0.5029558558326376, "learning_rate": 4.758308157099698e-06, "loss": 0.5358, "step": 42 }, { "epoch": 0.19196428571428573, "grad_norm": 0.5249190794688006, "learning_rate": 4.750755287009064e-06, "loss": 0.5542, "step": 43 }, { "epoch": 0.19642857142857142, "grad_norm": 0.4622817114582695, "learning_rate": 4.743202416918429e-06, "loss": 0.54, "step": 44 }, { "epoch": 0.20089285714285715, "grad_norm": 0.49789824735563454, "learning_rate": 4.735649546827795e-06, "loss": 0.5308, "step": 45 }, { "epoch": 0.20535714285714285, "grad_norm": 0.51199116739647, "learning_rate": 4.728096676737161e-06, "loss": 0.5784, "step": 46 }, { "epoch": 0.20982142857142858, "grad_norm": 0.47152096107780506, "learning_rate": 4.720543806646526e-06, "loss": 0.5654, "step": 47 }, { "epoch": 0.21428571428571427, "grad_norm": 0.5233581267403502, "learning_rate": 4.712990936555891e-06, "loss": 0.5727, "step": 48 }, { "epoch": 0.21875, "grad_norm": 0.517908769266997, "learning_rate": 4.705438066465257e-06, "loss": 0.5258, "step": 49 }, { "epoch": 0.22321428571428573, "grad_norm": 0.5103782222366013, "learning_rate": 4.697885196374623e-06, "loss": 0.564, "step": 50 }, { "epoch": 0.22767857142857142, "grad_norm": 0.5894196264675328, "learning_rate": 4.6903323262839885e-06, "loss": 0.5355, "step": 51 }, { "epoch": 0.23214285714285715, "grad_norm": 0.48136408513935, "learning_rate": 4.682779456193353e-06, "loss": 0.5714, "step": 52 }, { "epoch": 0.23660714285714285, "grad_norm": 0.5996790928794941, "learning_rate": 4.67522658610272e-06, "loss": 0.5385, "step": 53 }, { "epoch": 0.24107142857142858, "grad_norm": 0.5231061994660855, "learning_rate": 4.667673716012085e-06, "loss": 0.5429, "step": 54 }, { "epoch": 0.24553571428571427, "grad_norm": 0.5416056944426403, "learning_rate": 4.6601208459214505e-06, "loss": 0.5423, "step": 55 }, { "epoch": 0.25, "grad_norm": 0.5734105721889117, "learning_rate": 4.652567975830816e-06, "loss": 0.5756, "step": 56 }, { "epoch": 0.2544642857142857, "grad_norm": 0.5318541491738474, "learning_rate": 4.645015105740182e-06, "loss": 0.5486, "step": 57 }, { "epoch": 0.25892857142857145, "grad_norm": 0.4839987879630258, "learning_rate": 4.637462235649548e-06, "loss": 0.5517, "step": 58 }, { "epoch": 0.26339285714285715, "grad_norm": 0.5843387907103592, "learning_rate": 4.6299093655589125e-06, "loss": 0.5288, "step": 59 }, { "epoch": 0.26785714285714285, "grad_norm": 0.5496996598474941, "learning_rate": 4.622356495468278e-06, "loss": 0.542, "step": 60 }, { "epoch": 0.27232142857142855, "grad_norm": 0.5192508705264864, "learning_rate": 4.614803625377644e-06, "loss": 0.5291, "step": 61 }, { "epoch": 0.2767857142857143, "grad_norm": 0.5918952697319948, "learning_rate": 4.60725075528701e-06, "loss": 0.5229, "step": 62 }, { "epoch": 0.28125, "grad_norm": 0.5914173604947244, "learning_rate": 4.5996978851963745e-06, "loss": 0.5754, "step": 63 }, { "epoch": 0.2857142857142857, "grad_norm": 0.5298896547584293, "learning_rate": 4.592145015105741e-06, "loss": 0.5687, "step": 64 }, { "epoch": 0.29017857142857145, "grad_norm": 0.5671535875314645, "learning_rate": 4.584592145015106e-06, "loss": 0.5513, "step": 65 }, { "epoch": 0.29464285714285715, "grad_norm": 0.5234306194196137, "learning_rate": 4.577039274924472e-06, "loss": 0.5226, "step": 66 }, { "epoch": 0.29910714285714285, "grad_norm": 0.4959853502285333, "learning_rate": 4.569486404833837e-06, "loss": 0.5335, "step": 67 }, { "epoch": 0.30357142857142855, "grad_norm": 0.47729026424742405, "learning_rate": 4.561933534743202e-06, "loss": 0.5248, "step": 68 }, { "epoch": 0.3080357142857143, "grad_norm": 0.47294297292566195, "learning_rate": 4.554380664652569e-06, "loss": 0.5588, "step": 69 }, { "epoch": 0.3125, "grad_norm": 0.5275786368006364, "learning_rate": 4.5468277945619336e-06, "loss": 0.5483, "step": 70 }, { "epoch": 0.3169642857142857, "grad_norm": 0.5057705930103896, "learning_rate": 4.539274924471299e-06, "loss": 0.5236, "step": 71 }, { "epoch": 0.32142857142857145, "grad_norm": 0.49625696780589473, "learning_rate": 4.531722054380665e-06, "loss": 0.5221, "step": 72 }, { "epoch": 0.32589285714285715, "grad_norm": 0.49548627058675154, "learning_rate": 4.524169184290031e-06, "loss": 0.5176, "step": 73 }, { "epoch": 0.33035714285714285, "grad_norm": 0.5011897814029462, "learning_rate": 4.516616314199396e-06, "loss": 0.5431, "step": 74 }, { "epoch": 0.33482142857142855, "grad_norm": 0.5168007116628185, "learning_rate": 4.509063444108761e-06, "loss": 0.5289, "step": 75 }, { "epoch": 0.3392857142857143, "grad_norm": 0.5199488476439844, "learning_rate": 4.501510574018128e-06, "loss": 0.5437, "step": 76 }, { "epoch": 0.34375, "grad_norm": 0.5776348374651958, "learning_rate": 4.493957703927493e-06, "loss": 0.5424, "step": 77 }, { "epoch": 0.3482142857142857, "grad_norm": 0.5785274375251012, "learning_rate": 4.486404833836858e-06, "loss": 0.5252, "step": 78 }, { "epoch": 0.35267857142857145, "grad_norm": 0.5624252036929408, "learning_rate": 4.478851963746224e-06, "loss": 0.5502, "step": 79 }, { "epoch": 0.35714285714285715, "grad_norm": 0.5630273292563954, "learning_rate": 4.47129909365559e-06, "loss": 0.5249, "step": 80 }, { "epoch": 0.36160714285714285, "grad_norm": 0.5229643469397499, "learning_rate": 4.463746223564955e-06, "loss": 0.572, "step": 81 }, { "epoch": 0.36607142857142855, "grad_norm": 0.5401109169653766, "learning_rate": 4.45619335347432e-06, "loss": 0.5174, "step": 82 }, { "epoch": 0.3705357142857143, "grad_norm": 0.572853251261138, "learning_rate": 4.448640483383686e-06, "loss": 0.5336, "step": 83 }, { "epoch": 0.375, "grad_norm": 0.49833694539628726, "learning_rate": 4.441087613293052e-06, "loss": 0.5205, "step": 84 }, { "epoch": 0.3794642857142857, "grad_norm": 0.5167418841184229, "learning_rate": 4.4335347432024175e-06, "loss": 0.5405, "step": 85 }, { "epoch": 0.38392857142857145, "grad_norm": 0.5243612918158316, "learning_rate": 4.425981873111782e-06, "loss": 0.5204, "step": 86 }, { "epoch": 0.38839285714285715, "grad_norm": 0.5159624173670619, "learning_rate": 4.418429003021149e-06, "loss": 0.5315, "step": 87 }, { "epoch": 0.39285714285714285, "grad_norm": 0.5227918139330733, "learning_rate": 4.410876132930514e-06, "loss": 0.5418, "step": 88 }, { "epoch": 0.39732142857142855, "grad_norm": 0.48045718180033986, "learning_rate": 4.4033232628398795e-06, "loss": 0.506, "step": 89 }, { "epoch": 0.4017857142857143, "grad_norm": 0.5791314708394477, "learning_rate": 4.395770392749245e-06, "loss": 0.596, "step": 90 }, { "epoch": 0.40625, "grad_norm": 0.6076262289549809, "learning_rate": 4.38821752265861e-06, "loss": 0.5312, "step": 91 }, { "epoch": 0.4107142857142857, "grad_norm": 0.4780377894473343, "learning_rate": 4.380664652567977e-06, "loss": 0.549, "step": 92 }, { "epoch": 0.41517857142857145, "grad_norm": 0.5549693296410254, "learning_rate": 4.3731117824773415e-06, "loss": 0.5159, "step": 93 }, { "epoch": 0.41964285714285715, "grad_norm": 0.5097636365128929, "learning_rate": 4.365558912386707e-06, "loss": 0.5612, "step": 94 }, { "epoch": 0.42410714285714285, "grad_norm": 0.5021825140627123, "learning_rate": 4.358006042296073e-06, "loss": 0.5653, "step": 95 }, { "epoch": 0.42857142857142855, "grad_norm": 0.537945728158161, "learning_rate": 4.350453172205439e-06, "loss": 0.528, "step": 96 }, { "epoch": 0.4330357142857143, "grad_norm": 0.4846132130879552, "learning_rate": 4.342900302114804e-06, "loss": 0.5395, "step": 97 }, { "epoch": 0.4375, "grad_norm": 0.5206116049646974, "learning_rate": 4.335347432024169e-06, "loss": 0.5445, "step": 98 }, { "epoch": 0.4419642857142857, "grad_norm": 0.551262765339688, "learning_rate": 4.327794561933535e-06, "loss": 0.5346, "step": 99 }, { "epoch": 0.44642857142857145, "grad_norm": 0.4984265695742978, "learning_rate": 4.3202416918429006e-06, "loss": 0.5161, "step": 100 }, { "epoch": 0.45089285714285715, "grad_norm": 0.49943465819192145, "learning_rate": 4.312688821752266e-06, "loss": 0.5164, "step": 101 }, { "epoch": 0.45535714285714285, "grad_norm": 0.48826983665318036, "learning_rate": 4.305135951661632e-06, "loss": 0.552, "step": 102 }, { "epoch": 0.45982142857142855, "grad_norm": 0.5320671379410685, "learning_rate": 4.297583081570998e-06, "loss": 0.5512, "step": 103 }, { "epoch": 0.4642857142857143, "grad_norm": 0.5027439491810191, "learning_rate": 4.2900302114803626e-06, "loss": 0.5728, "step": 104 }, { "epoch": 0.46875, "grad_norm": 0.4875858103308158, "learning_rate": 4.282477341389728e-06, "loss": 0.5306, "step": 105 }, { "epoch": 0.4732142857142857, "grad_norm": 0.5205285237402679, "learning_rate": 4.274924471299094e-06, "loss": 0.5274, "step": 106 }, { "epoch": 0.47767857142857145, "grad_norm": 0.47059444782816534, "learning_rate": 4.26737160120846e-06, "loss": 0.5092, "step": 107 }, { "epoch": 0.48214285714285715, "grad_norm": 0.4998056822147382, "learning_rate": 4.259818731117825e-06, "loss": 0.5481, "step": 108 }, { "epoch": 0.48660714285714285, "grad_norm": 0.5403989848602658, "learning_rate": 4.25226586102719e-06, "loss": 0.518, "step": 109 }, { "epoch": 0.49107142857142855, "grad_norm": 0.48519680312136604, "learning_rate": 4.244712990936557e-06, "loss": 0.5522, "step": 110 }, { "epoch": 0.4955357142857143, "grad_norm": 0.5655590239975465, "learning_rate": 4.237160120845922e-06, "loss": 0.526, "step": 111 }, { "epoch": 0.5, "grad_norm": 0.49558151987556875, "learning_rate": 4.229607250755287e-06, "loss": 0.5275, "step": 112 }, { "epoch": 0.5, "eval_loss": 0.5346882343292236, "eval_runtime": 32.2399, "eval_samples_per_second": 82.817, "eval_steps_per_second": 5.18, "step": 112 }, { "epoch": 0.5050279329608939, "grad_norm": 0.5662939933016762, "learning_rate": 4.218512898330804e-06, "loss": 0.515, "step": 113 }, { "epoch": 0.5094972067039106, "grad_norm": 0.5225180760232837, "learning_rate": 4.2109256449165405e-06, "loss": 0.5038, "step": 114 }, { "epoch": 0.5139664804469274, "grad_norm": 0.5377410752157915, "learning_rate": 4.203338391502276e-06, "loss": 0.5043, "step": 115 }, { "epoch": 0.5184357541899441, "grad_norm": 0.5915094303426857, "learning_rate": 4.195751138088012e-06, "loss": 0.5115, "step": 116 }, { "epoch": 0.5229050279329609, "grad_norm": 0.5079762717076601, "learning_rate": 4.1881638846737485e-06, "loss": 0.5491, "step": 117 }, { "epoch": 0.5273743016759777, "grad_norm": 0.5933329970849119, "learning_rate": 4.180576631259484e-06, "loss": 0.5095, "step": 118 }, { "epoch": 0.5318435754189944, "grad_norm": 0.5053074762991974, "learning_rate": 4.17298937784522e-06, "loss": 0.4881, "step": 119 }, { "epoch": 0.5363128491620112, "grad_norm": 0.6515342691068152, "learning_rate": 4.1654021244309564e-06, "loss": 0.5069, "step": 120 }, { "epoch": 0.5407821229050279, "grad_norm": 0.5774345075195761, "learning_rate": 4.157814871016692e-06, "loss": 0.5007, "step": 121 }, { "epoch": 0.5452513966480447, "grad_norm": 0.6308991657157696, "learning_rate": 4.150227617602428e-06, "loss": 0.5083, "step": 122 }, { "epoch": 0.5497206703910614, "grad_norm": 0.5258198371155504, "learning_rate": 4.142640364188164e-06, "loss": 0.5103, "step": 123 }, { "epoch": 0.5541899441340782, "grad_norm": 0.5629243370057063, "learning_rate": 4.1350531107739e-06, "loss": 0.5016, "step": 124 }, { "epoch": 0.5586592178770949, "grad_norm": 0.5440095922230602, "learning_rate": 4.127465857359636e-06, "loss": 0.491, "step": 125 }, { "epoch": 0.5631284916201117, "grad_norm": 0.5126793634050918, "learning_rate": 4.119878603945372e-06, "loss": 0.5239, "step": 126 }, { "epoch": 0.5675977653631284, "grad_norm": 0.5476299968831635, "learning_rate": 4.112291350531108e-06, "loss": 0.5176, "step": 127 }, { "epoch": 0.5720670391061452, "grad_norm": 0.5570481457317527, "learning_rate": 4.104704097116844e-06, "loss": 0.5084, "step": 128 }, { "epoch": 0.576536312849162, "grad_norm": 0.49480777896006917, "learning_rate": 4.09711684370258e-06, "loss": 0.4966, "step": 129 }, { "epoch": 0.5810055865921788, "grad_norm": 0.5089889598386482, "learning_rate": 4.089529590288316e-06, "loss": 0.5059, "step": 130 }, { "epoch": 0.5854748603351956, "grad_norm": 0.4954780811768282, "learning_rate": 4.081942336874052e-06, "loss": 0.5166, "step": 131 }, { "epoch": 0.5899441340782123, "grad_norm": 0.4809684003541764, "learning_rate": 4.074355083459787e-06, "loss": 0.4857, "step": 132 }, { "epoch": 0.5944134078212291, "grad_norm": 0.5009748980192239, "learning_rate": 4.066767830045524e-06, "loss": 0.523, "step": 133 }, { "epoch": 0.5988826815642458, "grad_norm": 0.48320950023108783, "learning_rate": 4.05918057663126e-06, "loss": 0.4956, "step": 134 }, { "epoch": 0.6033519553072626, "grad_norm": 0.49399181892388616, "learning_rate": 4.051593323216995e-06, "loss": 0.5163, "step": 135 }, { "epoch": 0.6078212290502794, "grad_norm": 0.4821527638840584, "learning_rate": 4.044006069802732e-06, "loss": 0.4981, "step": 136 }, { "epoch": 0.6122905027932961, "grad_norm": 0.5099314085164022, "learning_rate": 4.036418816388468e-06, "loss": 0.5121, "step": 137 }, { "epoch": 0.6167597765363129, "grad_norm": 0.5323495276796993, "learning_rate": 4.028831562974203e-06, "loss": 0.5074, "step": 138 }, { "epoch": 0.6212290502793296, "grad_norm": 0.48851533865778524, "learning_rate": 4.02124430955994e-06, "loss": 0.52, "step": 139 }, { "epoch": 0.6256983240223464, "grad_norm": 0.5281525574471102, "learning_rate": 4.0136570561456756e-06, "loss": 0.5039, "step": 140 }, { "epoch": 0.6301675977653631, "grad_norm": 0.5279406819094602, "learning_rate": 4.006069802731411e-06, "loss": 0.4933, "step": 141 }, { "epoch": 0.6346368715083799, "grad_norm": 0.45822529010842367, "learning_rate": 3.998482549317148e-06, "loss": 0.5151, "step": 142 }, { "epoch": 0.6391061452513966, "grad_norm": 0.5349296018392958, "learning_rate": 3.9908952959028835e-06, "loss": 0.4983, "step": 143 }, { "epoch": 0.6435754189944134, "grad_norm": 0.5316574655817645, "learning_rate": 3.983308042488619e-06, "loss": 0.5344, "step": 144 }, { "epoch": 0.6480446927374302, "grad_norm": 0.5107562456487705, "learning_rate": 3.975720789074356e-06, "loss": 0.52, "step": 145 }, { "epoch": 0.6525139664804469, "grad_norm": 0.5072583855988344, "learning_rate": 3.9681335356600915e-06, "loss": 0.4923, "step": 146 }, { "epoch": 0.6569832402234637, "grad_norm": 0.5836000010606872, "learning_rate": 3.960546282245827e-06, "loss": 0.5295, "step": 147 }, { "epoch": 0.6614525139664804, "grad_norm": 0.49884721978842206, "learning_rate": 3.952959028831564e-06, "loss": 0.5216, "step": 148 }, { "epoch": 0.6659217877094972, "grad_norm": 0.5120991222561042, "learning_rate": 3.945371775417299e-06, "loss": 0.5414, "step": 149 }, { "epoch": 0.6703910614525139, "grad_norm": 0.49838481084559894, "learning_rate": 3.937784522003035e-06, "loss": 0.5381, "step": 150 }, { "epoch": 0.6748603351955307, "grad_norm": 0.5398799415397277, "learning_rate": 3.930197268588772e-06, "loss": 0.4896, "step": 151 }, { "epoch": 0.6793296089385474, "grad_norm": 0.565748454568293, "learning_rate": 3.922610015174507e-06, "loss": 0.4985, "step": 152 }, { "epoch": 0.6837988826815642, "grad_norm": 0.5986782671726579, "learning_rate": 3.915022761760243e-06, "loss": 0.5125, "step": 153 }, { "epoch": 0.6882681564245811, "grad_norm": 0.5290750194980306, "learning_rate": 3.907435508345979e-06, "loss": 0.5078, "step": 154 }, { "epoch": 0.6927374301675978, "grad_norm": 0.5706479438251948, "learning_rate": 3.899848254931715e-06, "loss": 0.5205, "step": 155 }, { "epoch": 0.6972067039106146, "grad_norm": 0.5173864500214489, "learning_rate": 3.892261001517451e-06, "loss": 0.4988, "step": 156 }, { "epoch": 0.7016759776536313, "grad_norm": 0.4920045809108581, "learning_rate": 3.884673748103187e-06, "loss": 0.4954, "step": 157 }, { "epoch": 0.7061452513966481, "grad_norm": 0.5237901072069291, "learning_rate": 3.877086494688923e-06, "loss": 0.5253, "step": 158 }, { "epoch": 0.7106145251396648, "grad_norm": 0.5212795514388354, "learning_rate": 3.869499241274659e-06, "loss": 0.5029, "step": 159 }, { "epoch": 0.7150837988826816, "grad_norm": 0.48658569533086143, "learning_rate": 3.861911987860395e-06, "loss": 0.5023, "step": 160 }, { "epoch": 0.7195530726256983, "grad_norm": 0.5145837815914728, "learning_rate": 3.854324734446131e-06, "loss": 0.5436, "step": 161 }, { "epoch": 0.7240223463687151, "grad_norm": 0.49182827281373437, "learning_rate": 3.846737481031867e-06, "loss": 0.4948, "step": 162 }, { "epoch": 0.7284916201117319, "grad_norm": 0.5113640333759738, "learning_rate": 3.839150227617603e-06, "loss": 0.4778, "step": 163 }, { "epoch": 0.7329608938547486, "grad_norm": 0.47675402744999507, "learning_rate": 3.831562974203339e-06, "loss": 0.5182, "step": 164 }, { "epoch": 0.7374301675977654, "grad_norm": 0.5214515656344721, "learning_rate": 3.823975720789075e-06, "loss": 0.5204, "step": 165 }, { "epoch": 0.7418994413407821, "grad_norm": 0.5130606167326404, "learning_rate": 3.816388467374811e-06, "loss": 0.492, "step": 166 }, { "epoch": 0.7463687150837989, "grad_norm": 0.49678736596959705, "learning_rate": 3.8088012139605467e-06, "loss": 0.4997, "step": 167 }, { "epoch": 0.7508379888268156, "grad_norm": 0.5275415964748987, "learning_rate": 3.801213960546283e-06, "loss": 0.5163, "step": 168 }, { "epoch": 0.7553072625698324, "grad_norm": 0.5323621998896121, "learning_rate": 3.7936267071320185e-06, "loss": 0.4846, "step": 169 }, { "epoch": 0.7597765363128491, "grad_norm": 0.5302675218909635, "learning_rate": 3.7860394537177547e-06, "loss": 0.5193, "step": 170 }, { "epoch": 0.7642458100558659, "grad_norm": 0.5097255453283106, "learning_rate": 3.778452200303491e-06, "loss": 0.5248, "step": 171 }, { "epoch": 0.7687150837988826, "grad_norm": 0.5221084637587756, "learning_rate": 3.7708649468892265e-06, "loss": 0.5217, "step": 172 }, { "epoch": 0.7731843575418994, "grad_norm": 0.5823307367451829, "learning_rate": 3.7632776934749626e-06, "loss": 0.5057, "step": 173 }, { "epoch": 0.7776536312849162, "grad_norm": 0.48586162675121547, "learning_rate": 3.7556904400606987e-06, "loss": 0.5372, "step": 174 }, { "epoch": 0.7821229050279329, "grad_norm": 0.5099958749272927, "learning_rate": 3.748103186646434e-06, "loss": 0.4844, "step": 175 }, { "epoch": 0.7865921787709497, "grad_norm": 0.5221594966770278, "learning_rate": 3.74051593323217e-06, "loss": 0.4968, "step": 176 }, { "epoch": 0.7910614525139665, "grad_norm": 0.5392129841421105, "learning_rate": 3.732928679817906e-06, "loss": 0.4906, "step": 177 }, { "epoch": 0.7955307262569833, "grad_norm": 0.5239629013619267, "learning_rate": 3.725341426403642e-06, "loss": 0.5172, "step": 178 }, { "epoch": 0.8, "grad_norm": 0.537861417175095, "learning_rate": 3.717754172989378e-06, "loss": 0.4914, "step": 179 }, { "epoch": 0.8044692737430168, "grad_norm": 0.5001204419984062, "learning_rate": 3.710166919575114e-06, "loss": 0.5014, "step": 180 }, { "epoch": 0.8089385474860336, "grad_norm": 0.5243758405421232, "learning_rate": 3.70257966616085e-06, "loss": 0.4979, "step": 181 }, { "epoch": 0.8134078212290503, "grad_norm": 0.507560920806033, "learning_rate": 3.6949924127465856e-06, "loss": 0.5003, "step": 182 }, { "epoch": 0.8178770949720671, "grad_norm": 0.5623627396713589, "learning_rate": 3.6874051593323218e-06, "loss": 0.5011, "step": 183 }, { "epoch": 0.8223463687150838, "grad_norm": 0.5225320984203384, "learning_rate": 3.679817905918058e-06, "loss": 0.5117, "step": 184 }, { "epoch": 0.8268156424581006, "grad_norm": 0.5371208831947621, "learning_rate": 3.6722306525037936e-06, "loss": 0.5028, "step": 185 }, { "epoch": 0.8312849162011173, "grad_norm": 0.5412513689241548, "learning_rate": 3.6646433990895297e-06, "loss": 0.5111, "step": 186 }, { "epoch": 0.8357541899441341, "grad_norm": 0.49083593539874787, "learning_rate": 3.657056145675266e-06, "loss": 0.5002, "step": 187 }, { "epoch": 0.8402234636871508, "grad_norm": 0.5144646076400369, "learning_rate": 3.6494688922610015e-06, "loss": 0.4875, "step": 188 }, { "epoch": 0.8446927374301676, "grad_norm": 0.5385458846653849, "learning_rate": 3.6418816388467377e-06, "loss": 0.5072, "step": 189 }, { "epoch": 0.8491620111731844, "grad_norm": 0.48088002744673064, "learning_rate": 3.6342943854324738e-06, "loss": 0.549, "step": 190 }, { "epoch": 0.8536312849162011, "grad_norm": 0.563960197772806, "learning_rate": 3.6267071320182095e-06, "loss": 0.5344, "step": 191 }, { "epoch": 0.8581005586592179, "grad_norm": 0.5235289059912077, "learning_rate": 3.6191198786039456e-06, "loss": 0.5172, "step": 192 }, { "epoch": 0.8625698324022346, "grad_norm": 0.5059904029741168, "learning_rate": 3.6115326251896813e-06, "loss": 0.4908, "step": 193 }, { "epoch": 0.8670391061452514, "grad_norm": 0.507980780742306, "learning_rate": 3.6039453717754174e-06, "loss": 0.4951, "step": 194 }, { "epoch": 0.8715083798882681, "grad_norm": 0.5552145890763519, "learning_rate": 3.5963581183611536e-06, "loss": 0.505, "step": 195 }, { "epoch": 0.8759776536312849, "grad_norm": 0.5379367693029108, "learning_rate": 3.5887708649468893e-06, "loss": 0.52, "step": 196 }, { "epoch": 0.8804469273743016, "grad_norm": 0.5813281708032492, "learning_rate": 3.5811836115326254e-06, "loss": 0.4966, "step": 197 }, { "epoch": 0.8849162011173184, "grad_norm": 0.5138017935214574, "learning_rate": 3.5735963581183615e-06, "loss": 0.495, "step": 198 }, { "epoch": 0.8893854748603351, "grad_norm": 0.5116405158423362, "learning_rate": 3.5660091047040972e-06, "loss": 0.519, "step": 199 }, { "epoch": 0.8938547486033519, "grad_norm": 0.5696283061922035, "learning_rate": 3.5584218512898333e-06, "loss": 0.5002, "step": 200 }, { "epoch": 0.8983240223463688, "grad_norm": 0.5541549121266046, "learning_rate": 3.5508345978755695e-06, "loss": 0.5125, "step": 201 }, { "epoch": 0.9027932960893855, "grad_norm": 0.5017863683272789, "learning_rate": 3.543247344461305e-06, "loss": 0.499, "step": 202 }, { "epoch": 0.9072625698324023, "grad_norm": 0.5776107100347877, "learning_rate": 3.5356600910470413e-06, "loss": 0.4883, "step": 203 }, { "epoch": 0.911731843575419, "grad_norm": 0.5350847562000018, "learning_rate": 3.528072837632777e-06, "loss": 0.5052, "step": 204 }, { "epoch": 0.9162011173184358, "grad_norm": 0.5012614076658582, "learning_rate": 3.520485584218513e-06, "loss": 0.5422, "step": 205 }, { "epoch": 0.9206703910614525, "grad_norm": 0.6107847324258918, "learning_rate": 3.5128983308042493e-06, "loss": 0.5204, "step": 206 }, { "epoch": 0.9251396648044693, "grad_norm": 0.49314942553653257, "learning_rate": 3.505311077389985e-06, "loss": 0.5074, "step": 207 }, { "epoch": 0.929608938547486, "grad_norm": 0.5340833706807174, "learning_rate": 3.497723823975721e-06, "loss": 0.4936, "step": 208 }, { "epoch": 0.9340782122905028, "grad_norm": 0.5342137283430334, "learning_rate": 3.490136570561457e-06, "loss": 0.4945, "step": 209 }, { "epoch": 0.9385474860335196, "grad_norm": 0.5301769352005536, "learning_rate": 3.482549317147193e-06, "loss": 0.4963, "step": 210 }, { "epoch": 0.9430167597765363, "grad_norm": 0.5455545680194068, "learning_rate": 3.474962063732929e-06, "loss": 0.4931, "step": 211 }, { "epoch": 0.9474860335195531, "grad_norm": 0.5282021847913991, "learning_rate": 3.467374810318665e-06, "loss": 0.4894, "step": 212 }, { "epoch": 0.9519553072625698, "grad_norm": 0.5341400870442735, "learning_rate": 3.459787556904401e-06, "loss": 0.5102, "step": 213 }, { "epoch": 0.9564245810055866, "grad_norm": 0.5313633005763995, "learning_rate": 3.452200303490137e-06, "loss": 0.501, "step": 214 }, { "epoch": 0.9608938547486033, "grad_norm": 0.5774555510069799, "learning_rate": 3.4446130500758727e-06, "loss": 0.5021, "step": 215 }, { "epoch": 0.9653631284916201, "grad_norm": 0.521646351973494, "learning_rate": 3.437025796661609e-06, "loss": 0.5071, "step": 216 }, { "epoch": 0.9698324022346368, "grad_norm": 0.5335773786656196, "learning_rate": 3.429438543247345e-06, "loss": 0.4993, "step": 217 }, { "epoch": 0.9743016759776536, "grad_norm": 0.5774515212071393, "learning_rate": 3.4218512898330806e-06, "loss": 0.4802, "step": 218 }, { "epoch": 0.9787709497206704, "grad_norm": 0.5159511449041201, "learning_rate": 3.4142640364188168e-06, "loss": 0.483, "step": 219 }, { "epoch": 0.9832402234636871, "grad_norm": 0.509916149384416, "learning_rate": 3.406676783004553e-06, "loss": 0.4774, "step": 220 }, { "epoch": 0.9877094972067039, "grad_norm": 0.5226659969898425, "learning_rate": 3.3990895295902886e-06, "loss": 0.516, "step": 221 }, { "epoch": 0.9921787709497206, "grad_norm": 0.5632866199425641, "learning_rate": 3.3915022761760247e-06, "loss": 0.5346, "step": 222 }, { "epoch": 0.9966480446927374, "grad_norm": 0.5224220899066443, "learning_rate": 3.383915022761761e-06, "loss": 0.5441, "step": 223 }, { "epoch": 1.0011173184357542, "grad_norm": 0.5965341004746808, "learning_rate": 3.3763277693474965e-06, "loss": 0.512, "step": 224 }, { "epoch": 1.0011173184357542, "eval_loss": 0.5086758732795715, "eval_runtime": 32.322, "eval_samples_per_second": 82.606, "eval_steps_per_second": 5.167, "step": 224 }, { "epoch": 1.005586592178771, "grad_norm": 0.5234315419190534, "learning_rate": 3.3687405159332327e-06, "loss": 0.5054, "step": 225 }, { "epoch": 1.0100558659217878, "grad_norm": 0.5575750854403085, "learning_rate": 3.3611532625189684e-06, "loss": 0.4878, "step": 226 }, { "epoch": 1.0145251396648045, "grad_norm": 0.5253448160628447, "learning_rate": 3.3535660091047045e-06, "loss": 0.5054, "step": 227 }, { "epoch": 1.0189944134078213, "grad_norm": 0.49020025660118194, "learning_rate": 3.3459787556904406e-06, "loss": 0.4818, "step": 228 }, { "epoch": 1.0044692737430168, "grad_norm": 0.6277138162928673, "learning_rate": 3.3383915022761763e-06, "loss": 0.5008, "step": 229 }, { "epoch": 1.0089385474860335, "grad_norm": 0.5276575275655174, "learning_rate": 3.3308042488619125e-06, "loss": 0.4822, "step": 230 }, { "epoch": 1.0134078212290503, "grad_norm": 0.516729892988387, "learning_rate": 3.3232169954476486e-06, "loss": 0.4783, "step": 231 }, { "epoch": 1.017877094972067, "grad_norm": 0.5648907712973277, "learning_rate": 3.3156297420333843e-06, "loss": 0.4927, "step": 232 }, { "epoch": 1.0223463687150838, "grad_norm": 0.5521221146190025, "learning_rate": 3.3080424886191204e-06, "loss": 0.4742, "step": 233 }, { "epoch": 1.0268156424581005, "grad_norm": 0.5288379239842262, "learning_rate": 3.3004552352048565e-06, "loss": 0.4863, "step": 234 }, { "epoch": 1.0312849162011173, "grad_norm": 0.4932863276651226, "learning_rate": 3.2928679817905922e-06, "loss": 0.4613, "step": 235 }, { "epoch": 1.035754189944134, "grad_norm": 0.5069055178884726, "learning_rate": 3.2852807283763284e-06, "loss": 0.4969, "step": 236 }, { "epoch": 1.0402234636871508, "grad_norm": 0.49009866664123986, "learning_rate": 3.277693474962064e-06, "loss": 0.4759, "step": 237 }, { "epoch": 1.0446927374301676, "grad_norm": 0.5220649844244807, "learning_rate": 3.2701062215478e-06, "loss": 0.4654, "step": 238 }, { "epoch": 1.0491620111731843, "grad_norm": 0.5354627141925222, "learning_rate": 3.2625189681335363e-06, "loss": 0.5009, "step": 239 }, { "epoch": 1.053631284916201, "grad_norm": 0.5561818122893151, "learning_rate": 3.254931714719272e-06, "loss": 0.4564, "step": 240 }, { "epoch": 1.0581005586592178, "grad_norm": 0.48719455414113416, "learning_rate": 3.2473444613050077e-06, "loss": 0.4924, "step": 241 }, { "epoch": 1.0625698324022346, "grad_norm": 0.5527905454518657, "learning_rate": 3.2397572078907434e-06, "loss": 0.4856, "step": 242 }, { "epoch": 1.0670391061452513, "grad_norm": 0.5381863443762583, "learning_rate": 3.2321699544764795e-06, "loss": 0.4935, "step": 243 }, { "epoch": 1.071508379888268, "grad_norm": 0.5355360844432822, "learning_rate": 3.2245827010622157e-06, "loss": 0.4858, "step": 244 }, { "epoch": 1.0759776536312848, "grad_norm": 0.5191892602172652, "learning_rate": 3.2169954476479514e-06, "loss": 0.4819, "step": 245 }, { "epoch": 1.0804469273743016, "grad_norm": 0.4952672717234948, "learning_rate": 3.2094081942336875e-06, "loss": 0.5007, "step": 246 }, { "epoch": 1.0849162011173183, "grad_norm": 0.5331734326280092, "learning_rate": 3.201820940819423e-06, "loss": 0.4874, "step": 247 }, { "epoch": 1.089385474860335, "grad_norm": 0.49333273079420603, "learning_rate": 3.1942336874051593e-06, "loss": 0.5045, "step": 248 }, { "epoch": 1.0938547486033519, "grad_norm": 0.5143326582021585, "learning_rate": 3.1866464339908955e-06, "loss": 0.5047, "step": 249 }, { "epoch": 1.0983240223463686, "grad_norm": 0.5027227799706213, "learning_rate": 3.179059180576631e-06, "loss": 0.4803, "step": 250 }, { "epoch": 1.1027932960893856, "grad_norm": 0.5257359502707071, "learning_rate": 3.1714719271623673e-06, "loss": 0.4654, "step": 251 }, { "epoch": 1.1072625698324021, "grad_norm": 0.5244897759431714, "learning_rate": 3.1638846737481034e-06, "loss": 0.5329, "step": 252 }, { "epoch": 1.111731843575419, "grad_norm": 0.5028531015372357, "learning_rate": 3.156297420333839e-06, "loss": 0.4721, "step": 253 }, { "epoch": 1.1162011173184359, "grad_norm": 0.5342121642252061, "learning_rate": 3.1487101669195752e-06, "loss": 0.4962, "step": 254 }, { "epoch": 1.1206703910614526, "grad_norm": 0.49550125400926287, "learning_rate": 3.1411229135053114e-06, "loss": 0.4884, "step": 255 }, { "epoch": 1.1251396648044694, "grad_norm": 0.5724921757672813, "learning_rate": 3.133535660091047e-06, "loss": 0.4676, "step": 256 }, { "epoch": 1.1296089385474861, "grad_norm": 0.5012184887171401, "learning_rate": 3.125948406676783e-06, "loss": 0.48, "step": 257 }, { "epoch": 1.1340782122905029, "grad_norm": 0.5049054849116781, "learning_rate": 3.118361153262519e-06, "loss": 0.494, "step": 258 }, { "epoch": 1.1385474860335196, "grad_norm": 0.5291980859522503, "learning_rate": 3.110773899848255e-06, "loss": 0.4787, "step": 259 }, { "epoch": 1.1430167597765364, "grad_norm": 0.5235318263865767, "learning_rate": 3.103186646433991e-06, "loss": 0.4745, "step": 260 }, { "epoch": 1.1474860335195531, "grad_norm": 0.5067236081675356, "learning_rate": 3.095599393019727e-06, "loss": 0.5863, "step": 261 }, { "epoch": 1.15195530726257, "grad_norm": 0.48763909341918976, "learning_rate": 3.088012139605463e-06, "loss": 0.5264, "step": 262 }, { "epoch": 1.1564245810055866, "grad_norm": 0.5198110315693464, "learning_rate": 3.080424886191199e-06, "loss": 0.4716, "step": 263 }, { "epoch": 1.1608938547486034, "grad_norm": 0.5093655208180958, "learning_rate": 3.072837632776935e-06, "loss": 0.4768, "step": 264 }, { "epoch": 1.1653631284916202, "grad_norm": 0.49489186143964614, "learning_rate": 3.065250379362671e-06, "loss": 0.4765, "step": 265 }, { "epoch": 1.169832402234637, "grad_norm": 0.5209973994944476, "learning_rate": 3.057663125948407e-06, "loss": 0.4869, "step": 266 }, { "epoch": 1.1743016759776537, "grad_norm": 0.4873878645874658, "learning_rate": 3.0500758725341427e-06, "loss": 0.4623, "step": 267 }, { "epoch": 1.1787709497206704, "grad_norm": 0.5074910158328607, "learning_rate": 3.042488619119879e-06, "loss": 0.4765, "step": 268 }, { "epoch": 1.1832402234636872, "grad_norm": 0.5195087180758534, "learning_rate": 3.0349013657056146e-06, "loss": 0.4838, "step": 269 }, { "epoch": 1.187709497206704, "grad_norm": 0.5264228931092361, "learning_rate": 3.0273141122913507e-06, "loss": 0.4638, "step": 270 }, { "epoch": 1.1921787709497207, "grad_norm": 0.4919380409463276, "learning_rate": 3.019726858877087e-06, "loss": 0.4981, "step": 271 }, { "epoch": 1.1966480446927374, "grad_norm": 0.5060903656555795, "learning_rate": 3.0121396054628225e-06, "loss": 0.477, "step": 272 }, { "epoch": 1.2011173184357542, "grad_norm": 0.4938697284936987, "learning_rate": 3.0045523520485587e-06, "loss": 0.5003, "step": 273 }, { "epoch": 1.205586592178771, "grad_norm": 0.5599516302969035, "learning_rate": 2.9969650986342948e-06, "loss": 0.4673, "step": 274 }, { "epoch": 1.2100558659217877, "grad_norm": 0.5126307823030745, "learning_rate": 2.9893778452200305e-06, "loss": 0.4971, "step": 275 }, { "epoch": 1.2145251396648045, "grad_norm": 0.5927333276589611, "learning_rate": 2.9817905918057666e-06, "loss": 0.5006, "step": 276 }, { "epoch": 1.2189944134078212, "grad_norm": 0.5367579038483986, "learning_rate": 2.9742033383915027e-06, "loss": 0.5143, "step": 277 }, { "epoch": 1.223463687150838, "grad_norm": 0.5179699664039845, "learning_rate": 2.9666160849772384e-06, "loss": 0.4733, "step": 278 }, { "epoch": 1.2279329608938547, "grad_norm": 0.5015385250607317, "learning_rate": 2.9590288315629746e-06, "loss": 0.4786, "step": 279 }, { "epoch": 1.2324022346368715, "grad_norm": 0.4925764923779409, "learning_rate": 2.9514415781487103e-06, "loss": 0.472, "step": 280 }, { "epoch": 1.2368715083798882, "grad_norm": 0.5522056674324646, "learning_rate": 2.9438543247344464e-06, "loss": 0.4675, "step": 281 }, { "epoch": 1.241340782122905, "grad_norm": 0.5128909667557014, "learning_rate": 2.9362670713201825e-06, "loss": 0.4778, "step": 282 }, { "epoch": 1.2458100558659218, "grad_norm": 0.49958051066954534, "learning_rate": 2.9286798179059182e-06, "loss": 0.4744, "step": 283 }, { "epoch": 1.2502793296089385, "grad_norm": 0.5186801983796817, "learning_rate": 2.9210925644916543e-06, "loss": 0.47, "step": 284 }, { "epoch": 1.2547486033519553, "grad_norm": 0.49479081933964797, "learning_rate": 2.9135053110773905e-06, "loss": 0.4693, "step": 285 }, { "epoch": 1.259217877094972, "grad_norm": 0.5172924586559385, "learning_rate": 2.905918057663126e-06, "loss": 0.5034, "step": 286 }, { "epoch": 1.2636871508379888, "grad_norm": 0.49588955761543196, "learning_rate": 2.8983308042488623e-06, "loss": 0.4857, "step": 287 }, { "epoch": 1.2681564245810055, "grad_norm": 0.5651670438316321, "learning_rate": 2.8907435508345984e-06, "loss": 0.4845, "step": 288 }, { "epoch": 1.2726256983240223, "grad_norm": 0.47720259500943096, "learning_rate": 2.883156297420334e-06, "loss": 0.4482, "step": 289 }, { "epoch": 1.277094972067039, "grad_norm": 0.5080092332022862, "learning_rate": 2.8755690440060702e-06, "loss": 0.4826, "step": 290 }, { "epoch": 1.2815642458100558, "grad_norm": 0.6016729996144168, "learning_rate": 2.867981790591806e-06, "loss": 0.5592, "step": 291 }, { "epoch": 1.2860335195530725, "grad_norm": 0.5456040494522852, "learning_rate": 2.860394537177542e-06, "loss": 0.4821, "step": 292 }, { "epoch": 1.2905027932960893, "grad_norm": 0.5858331665900632, "learning_rate": 2.852807283763278e-06, "loss": 0.5089, "step": 293 }, { "epoch": 1.2949720670391063, "grad_norm": 0.5895630636847224, "learning_rate": 2.845220030349014e-06, "loss": 0.5101, "step": 294 }, { "epoch": 1.2994413407821228, "grad_norm": 0.505172328049275, "learning_rate": 2.83763277693475e-06, "loss": 0.5008, "step": 295 }, { "epoch": 1.3039106145251398, "grad_norm": 0.5035894899292893, "learning_rate": 2.830045523520486e-06, "loss": 0.4812, "step": 296 }, { "epoch": 1.3083798882681563, "grad_norm": 0.5304264887862729, "learning_rate": 2.822458270106222e-06, "loss": 0.4579, "step": 297 }, { "epoch": 1.3128491620111733, "grad_norm": 0.4742288658363735, "learning_rate": 2.814871016691958e-06, "loss": 0.4801, "step": 298 }, { "epoch": 1.3173184357541898, "grad_norm": 0.5293412371049411, "learning_rate": 2.807283763277694e-06, "loss": 0.4602, "step": 299 }, { "epoch": 1.3217877094972068, "grad_norm": 0.5333831439310788, "learning_rate": 2.79969650986343e-06, "loss": 0.4963, "step": 300 }, { "epoch": 1.3262569832402233, "grad_norm": 0.4947844594727205, "learning_rate": 2.792109256449166e-06, "loss": 0.4781, "step": 301 }, { "epoch": 1.3307262569832403, "grad_norm": 0.5323122787674339, "learning_rate": 2.7845220030349016e-06, "loss": 0.4998, "step": 302 }, { "epoch": 1.3351955307262569, "grad_norm": 0.48615316005162806, "learning_rate": 2.7769347496206378e-06, "loss": 0.466, "step": 303 }, { "epoch": 1.3396648044692738, "grad_norm": 0.5038897355673981, "learning_rate": 2.769347496206374e-06, "loss": 0.531, "step": 304 }, { "epoch": 1.3441340782122906, "grad_norm": 0.47725378544672703, "learning_rate": 2.7617602427921096e-06, "loss": 0.4917, "step": 305 }, { "epoch": 1.3486033519553073, "grad_norm": 0.5241579403675019, "learning_rate": 2.7541729893778457e-06, "loss": 0.4673, "step": 306 }, { "epoch": 1.353072625698324, "grad_norm": 0.5252091486066182, "learning_rate": 2.746585735963581e-06, "loss": 0.467, "step": 307 }, { "epoch": 1.3575418994413408, "grad_norm": 0.4803361684555166, "learning_rate": 2.738998482549317e-06, "loss": 0.5257, "step": 308 }, { "epoch": 1.3620111731843576, "grad_norm": 0.5300934772683465, "learning_rate": 2.7314112291350532e-06, "loss": 0.4889, "step": 309 }, { "epoch": 1.3664804469273744, "grad_norm": 0.5001374114396476, "learning_rate": 2.723823975720789e-06, "loss": 0.4847, "step": 310 }, { "epoch": 1.3709497206703911, "grad_norm": 0.5240875065710657, "learning_rate": 2.716236722306525e-06, "loss": 0.4578, "step": 311 }, { "epoch": 1.3754189944134079, "grad_norm": 0.4714761354385937, "learning_rate": 2.708649468892261e-06, "loss": 0.4947, "step": 312 }, { "epoch": 1.3798882681564246, "grad_norm": 0.5205222669608968, "learning_rate": 2.701062215477997e-06, "loss": 0.5049, "step": 313 }, { "epoch": 1.3843575418994414, "grad_norm": 0.5409150510364881, "learning_rate": 2.693474962063733e-06, "loss": 0.4946, "step": 314 }, { "epoch": 1.3888268156424581, "grad_norm": 0.49473773833387136, "learning_rate": 2.6858877086494687e-06, "loss": 0.5024, "step": 315 }, { "epoch": 1.393296089385475, "grad_norm": 0.5322303350412809, "learning_rate": 2.678300455235205e-06, "loss": 0.5129, "step": 316 }, { "epoch": 1.3977653631284916, "grad_norm": 0.5249345159944246, "learning_rate": 2.670713201820941e-06, "loss": 0.4852, "step": 317 }, { "epoch": 1.4022346368715084, "grad_norm": 0.526125738227021, "learning_rate": 2.6631259484066767e-06, "loss": 0.4755, "step": 318 }, { "epoch": 1.4067039106145252, "grad_norm": 0.5077327993266544, "learning_rate": 2.655538694992413e-06, "loss": 0.4708, "step": 319 }, { "epoch": 1.411173184357542, "grad_norm": 0.509006777595202, "learning_rate": 2.647951441578149e-06, "loss": 0.4574, "step": 320 }, { "epoch": 1.4156424581005587, "grad_norm": 0.5397063984545032, "learning_rate": 2.6403641881638846e-06, "loss": 0.4983, "step": 321 }, { "epoch": 1.4201117318435754, "grad_norm": 0.588039527577832, "learning_rate": 2.6327769347496208e-06, "loss": 0.5063, "step": 322 }, { "epoch": 1.4245810055865922, "grad_norm": 0.5086942941008908, "learning_rate": 2.625189681335357e-06, "loss": 0.5439, "step": 323 }, { "epoch": 1.429050279329609, "grad_norm": 0.5593378373560197, "learning_rate": 2.6176024279210926e-06, "loss": 0.4789, "step": 324 }, { "epoch": 1.4335195530726257, "grad_norm": 0.5473461722884195, "learning_rate": 2.6100151745068287e-06, "loss": 0.4861, "step": 325 }, { "epoch": 1.4379888268156424, "grad_norm": 0.5744717523131797, "learning_rate": 2.6024279210925644e-06, "loss": 0.4818, "step": 326 }, { "epoch": 1.4424581005586592, "grad_norm": 0.5240335872273282, "learning_rate": 2.5948406676783005e-06, "loss": 0.4785, "step": 327 }, { "epoch": 1.446927374301676, "grad_norm": 0.5608475565499077, "learning_rate": 2.5872534142640367e-06, "loss": 0.5042, "step": 328 }, { "epoch": 1.4513966480446927, "grad_norm": 0.5363050759442354, "learning_rate": 2.5796661608497724e-06, "loss": 0.4995, "step": 329 }, { "epoch": 1.4558659217877095, "grad_norm": 0.5027607995293213, "learning_rate": 2.5720789074355085e-06, "loss": 0.4954, "step": 330 }, { "epoch": 1.4603351955307262, "grad_norm": 0.4969954043343912, "learning_rate": 2.5644916540212446e-06, "loss": 0.4882, "step": 331 }, { "epoch": 1.464804469273743, "grad_norm": 0.5718337637259987, "learning_rate": 2.5569044006069803e-06, "loss": 0.4726, "step": 332 }, { "epoch": 1.4692737430167597, "grad_norm": 0.5234285673769393, "learning_rate": 2.5493171471927164e-06, "loss": 0.4697, "step": 333 }, { "epoch": 1.4737430167597765, "grad_norm": 0.5219608223935474, "learning_rate": 2.5417298937784526e-06, "loss": 0.4776, "step": 334 }, { "epoch": 1.4782122905027932, "grad_norm": 0.606768683027719, "learning_rate": 2.5341426403641883e-06, "loss": 0.4756, "step": 335 }, { "epoch": 1.48268156424581, "grad_norm": 0.5432176614167162, "learning_rate": 2.5265553869499244e-06, "loss": 0.4619, "step": 336 }, { "epoch": 1.48268156424581, "eval_loss": 0.5032872557640076, "eval_runtime": 32.2653, "eval_samples_per_second": 82.751, "eval_steps_per_second": 5.176, "step": 336 }, { "epoch": 1.4871508379888267, "grad_norm": 0.5377224565720307, "learning_rate": 2.51896813353566e-06, "loss": 0.4667, "step": 337 }, { "epoch": 1.4916201117318435, "grad_norm": 0.5875842821019925, "learning_rate": 2.5113808801213962e-06, "loss": 0.4873, "step": 338 }, { "epoch": 1.4960893854748603, "grad_norm": 0.5624618850955563, "learning_rate": 2.5037936267071324e-06, "loss": 0.4626, "step": 339 }, { "epoch": 1.500558659217877, "grad_norm": 0.5380665704623208, "learning_rate": 2.496206373292868e-06, "loss": 0.4749, "step": 340 }, { "epoch": 1.505027932960894, "grad_norm": 0.5515048280645092, "learning_rate": 2.488619119878604e-06, "loss": 0.4862, "step": 341 }, { "epoch": 1.5094972067039105, "grad_norm": 0.5541495309875792, "learning_rate": 2.4810318664643403e-06, "loss": 0.4533, "step": 342 }, { "epoch": 1.5139664804469275, "grad_norm": 0.514569610923219, "learning_rate": 2.473444613050076e-06, "loss": 0.4585, "step": 343 }, { "epoch": 1.518435754189944, "grad_norm": 0.5862261680615611, "learning_rate": 2.465857359635812e-06, "loss": 0.4835, "step": 344 }, { "epoch": 1.522905027932961, "grad_norm": 0.536975433650959, "learning_rate": 2.458270106221548e-06, "loss": 0.4638, "step": 345 }, { "epoch": 1.5273743016759775, "grad_norm": 0.5142279881627296, "learning_rate": 2.450682852807284e-06, "loss": 0.4832, "step": 346 }, { "epoch": 1.5318435754189945, "grad_norm": 0.5124062345457664, "learning_rate": 2.44309559939302e-06, "loss": 0.4857, "step": 347 }, { "epoch": 1.536312849162011, "grad_norm": 0.509905421315166, "learning_rate": 2.435508345978756e-06, "loss": 0.4807, "step": 348 }, { "epoch": 1.540782122905028, "grad_norm": 0.5275507665643101, "learning_rate": 2.427921092564492e-06, "loss": 0.4722, "step": 349 }, { "epoch": 1.5452513966480446, "grad_norm": 0.5003951978639241, "learning_rate": 2.420333839150228e-06, "loss": 0.4941, "step": 350 }, { "epoch": 1.5497206703910615, "grad_norm": 0.5251614039487292, "learning_rate": 2.4127465857359637e-06, "loss": 0.5376, "step": 351 }, { "epoch": 1.554189944134078, "grad_norm": 0.5128097582851606, "learning_rate": 2.4051593323217e-06, "loss": 0.5001, "step": 352 }, { "epoch": 1.558659217877095, "grad_norm": 0.4971288144993544, "learning_rate": 2.397572078907436e-06, "loss": 0.4721, "step": 353 }, { "epoch": 1.5631284916201116, "grad_norm": 0.5111491573396423, "learning_rate": 2.3899848254931717e-06, "loss": 0.5192, "step": 354 }, { "epoch": 1.5675977653631286, "grad_norm": 0.5025703950184777, "learning_rate": 2.382397572078908e-06, "loss": 0.4541, "step": 355 }, { "epoch": 1.572067039106145, "grad_norm": 0.5011840454618501, "learning_rate": 2.3748103186646435e-06, "loss": 0.4789, "step": 356 }, { "epoch": 1.576536312849162, "grad_norm": 0.4981690031939279, "learning_rate": 2.3672230652503792e-06, "loss": 0.497, "step": 357 }, { "epoch": 1.5810055865921788, "grad_norm": 0.4994506284814992, "learning_rate": 2.3596358118361154e-06, "loss": 0.4655, "step": 358 }, { "epoch": 1.5854748603351956, "grad_norm": 0.49694019288140645, "learning_rate": 2.3520485584218515e-06, "loss": 0.4883, "step": 359 }, { "epoch": 1.5899441340782123, "grad_norm": 0.5103222760000443, "learning_rate": 2.344461305007587e-06, "loss": 0.4615, "step": 360 }, { "epoch": 1.594413407821229, "grad_norm": 0.4960772434204748, "learning_rate": 2.3368740515933233e-06, "loss": 0.4748, "step": 361 }, { "epoch": 1.5988826815642458, "grad_norm": 0.5372239368130248, "learning_rate": 2.3292867981790594e-06, "loss": 0.461, "step": 362 }, { "epoch": 1.6033519553072626, "grad_norm": 0.5002183877541576, "learning_rate": 2.321699544764795e-06, "loss": 0.4985, "step": 363 }, { "epoch": 1.6078212290502794, "grad_norm": 0.5065612982842012, "learning_rate": 2.3141122913505313e-06, "loss": 0.4846, "step": 364 }, { "epoch": 1.612290502793296, "grad_norm": 0.5294145045424744, "learning_rate": 2.3065250379362674e-06, "loss": 0.4769, "step": 365 }, { "epoch": 1.6167597765363129, "grad_norm": 0.5019842435412675, "learning_rate": 2.298937784522003e-06, "loss": 0.4738, "step": 366 }, { "epoch": 1.6212290502793296, "grad_norm": 0.4973490272845337, "learning_rate": 2.291350531107739e-06, "loss": 0.4773, "step": 367 }, { "epoch": 1.6256983240223464, "grad_norm": 0.4917027064554639, "learning_rate": 2.283763277693475e-06, "loss": 0.4921, "step": 368 }, { "epoch": 1.6301675977653631, "grad_norm": 0.5191932977137537, "learning_rate": 2.276176024279211e-06, "loss": 0.4823, "step": 369 }, { "epoch": 1.6346368715083799, "grad_norm": 0.5130853619632637, "learning_rate": 2.268588770864947e-06, "loss": 0.4631, "step": 370 }, { "epoch": 1.6391061452513966, "grad_norm": 0.5402133951776769, "learning_rate": 2.261001517450683e-06, "loss": 0.456, "step": 371 }, { "epoch": 1.6435754189944134, "grad_norm": 0.5136201214507364, "learning_rate": 2.253414264036419e-06, "loss": 0.4715, "step": 372 }, { "epoch": 1.6480446927374302, "grad_norm": 0.5397223114168315, "learning_rate": 2.245827010622155e-06, "loss": 0.465, "step": 373 }, { "epoch": 1.652513966480447, "grad_norm": 0.5178224930093817, "learning_rate": 2.238239757207891e-06, "loss": 0.4526, "step": 374 }, { "epoch": 1.6569832402234637, "grad_norm": 0.5399618966600667, "learning_rate": 2.230652503793627e-06, "loss": 0.4873, "step": 375 }, { "epoch": 1.6614525139664804, "grad_norm": 0.5106529140250409, "learning_rate": 2.223065250379363e-06, "loss": 0.4771, "step": 376 }, { "epoch": 1.6659217877094972, "grad_norm": 0.5460841160518792, "learning_rate": 2.2154779969650988e-06, "loss": 0.4743, "step": 377 }, { "epoch": 1.670391061452514, "grad_norm": 0.5846706720135237, "learning_rate": 2.207890743550835e-06, "loss": 0.4552, "step": 378 }, { "epoch": 1.6748603351955307, "grad_norm": 0.534699591229056, "learning_rate": 2.2003034901365706e-06, "loss": 0.495, "step": 379 }, { "epoch": 1.6793296089385474, "grad_norm": 0.5377842578832556, "learning_rate": 2.1927162367223067e-06, "loss": 0.4824, "step": 380 }, { "epoch": 1.6837988826815642, "grad_norm": 0.5799148072307362, "learning_rate": 2.185128983308043e-06, "loss": 0.4708, "step": 381 }, { "epoch": 1.6882681564245812, "grad_norm": 0.5076181067753225, "learning_rate": 2.1775417298937786e-06, "loss": 0.4631, "step": 382 }, { "epoch": 1.6927374301675977, "grad_norm": 0.5025103762077108, "learning_rate": 2.1699544764795147e-06, "loss": 0.4676, "step": 383 }, { "epoch": 1.6972067039106147, "grad_norm": 0.48231447315943815, "learning_rate": 2.162367223065251e-06, "loss": 0.4694, "step": 384 }, { "epoch": 1.7016759776536312, "grad_norm": 0.5003437712619672, "learning_rate": 2.1547799696509865e-06, "loss": 0.4607, "step": 385 }, { "epoch": 1.7061452513966482, "grad_norm": 0.5149834201342615, "learning_rate": 2.1471927162367226e-06, "loss": 0.5253, "step": 386 }, { "epoch": 1.7106145251396647, "grad_norm": 0.5406857228194807, "learning_rate": 2.1396054628224588e-06, "loss": 0.5352, "step": 387 }, { "epoch": 1.7150837988826817, "grad_norm": 0.5204717082880796, "learning_rate": 2.1320182094081945e-06, "loss": 0.4936, "step": 388 }, { "epoch": 1.7195530726256982, "grad_norm": 0.5202749448515015, "learning_rate": 2.12443095599393e-06, "loss": 0.4594, "step": 389 }, { "epoch": 1.7240223463687152, "grad_norm": 0.4952910050954741, "learning_rate": 2.1168437025796663e-06, "loss": 0.4793, "step": 390 }, { "epoch": 1.7284916201117317, "grad_norm": 0.48637344511275765, "learning_rate": 2.109256449165402e-06, "loss": 0.4745, "step": 391 }, { "epoch": 1.7329608938547487, "grad_norm": 0.4906271558417189, "learning_rate": 2.101669195751138e-06, "loss": 0.4652, "step": 392 }, { "epoch": 1.7374301675977653, "grad_norm": 0.5090727761430808, "learning_rate": 2.0940819423368742e-06, "loss": 0.5298, "step": 393 }, { "epoch": 1.7418994413407822, "grad_norm": 0.4866596430096002, "learning_rate": 2.08649468892261e-06, "loss": 0.4838, "step": 394 }, { "epoch": 1.7463687150837988, "grad_norm": 0.49259128094078225, "learning_rate": 2.078907435508346e-06, "loss": 0.4737, "step": 395 }, { "epoch": 1.7508379888268157, "grad_norm": 0.5031034927365164, "learning_rate": 2.071320182094082e-06, "loss": 0.4777, "step": 396 }, { "epoch": 1.7553072625698323, "grad_norm": 0.48520021809068, "learning_rate": 2.063732928679818e-06, "loss": 0.454, "step": 397 }, { "epoch": 1.7597765363128492, "grad_norm": 0.5217902917498033, "learning_rate": 2.056145675265554e-06, "loss": 0.5199, "step": 398 }, { "epoch": 1.7642458100558658, "grad_norm": 0.4963706354662438, "learning_rate": 2.04855842185129e-06, "loss": 0.4649, "step": 399 }, { "epoch": 1.7687150837988828, "grad_norm": 0.49552580249683814, "learning_rate": 2.040971168437026e-06, "loss": 0.4887, "step": 400 }, { "epoch": 1.7731843575418993, "grad_norm": 0.49468231483436587, "learning_rate": 2.033383915022762e-06, "loss": 0.4757, "step": 401 }, { "epoch": 1.7776536312849163, "grad_norm": 0.47502769186897137, "learning_rate": 2.0257966616084977e-06, "loss": 0.4846, "step": 402 }, { "epoch": 1.7821229050279328, "grad_norm": 0.5089134108324753, "learning_rate": 2.018209408194234e-06, "loss": 0.4692, "step": 403 }, { "epoch": 1.7865921787709498, "grad_norm": 0.4981731250179096, "learning_rate": 2.01062215477997e-06, "loss": 0.4665, "step": 404 }, { "epoch": 1.7910614525139665, "grad_norm": 0.4690174096990647, "learning_rate": 2.0030349013657056e-06, "loss": 0.4606, "step": 405 }, { "epoch": 1.7955307262569833, "grad_norm": 0.4914411018326139, "learning_rate": 1.9954476479514418e-06, "loss": 0.4847, "step": 406 }, { "epoch": 1.8, "grad_norm": 0.5513335287615446, "learning_rate": 1.987860394537178e-06, "loss": 0.488, "step": 407 }, { "epoch": 1.8044692737430168, "grad_norm": 0.48992847389625854, "learning_rate": 1.9802731411229136e-06, "loss": 0.465, "step": 408 }, { "epoch": 1.8089385474860336, "grad_norm": 0.5185912675971102, "learning_rate": 1.9726858877086497e-06, "loss": 0.4872, "step": 409 }, { "epoch": 1.8134078212290503, "grad_norm": 0.47300994432774746, "learning_rate": 1.965098634294386e-06, "loss": 0.4534, "step": 410 }, { "epoch": 1.817877094972067, "grad_norm": 0.5034275716173061, "learning_rate": 1.9575113808801215e-06, "loss": 0.4668, "step": 411 }, { "epoch": 1.8223463687150838, "grad_norm": 0.46435516708821234, "learning_rate": 1.9499241274658577e-06, "loss": 0.491, "step": 412 }, { "epoch": 1.8268156424581006, "grad_norm": 0.4915994950607221, "learning_rate": 1.9423368740515934e-06, "loss": 0.4793, "step": 413 }, { "epoch": 1.8312849162011173, "grad_norm": 0.49834550842622505, "learning_rate": 1.9347496206373295e-06, "loss": 0.4897, "step": 414 }, { "epoch": 1.835754189944134, "grad_norm": 0.5275123854480922, "learning_rate": 1.9271623672230656e-06, "loss": 0.5167, "step": 415 }, { "epoch": 1.8402234636871508, "grad_norm": 0.501447167095016, "learning_rate": 1.9195751138088013e-06, "loss": 0.5522, "step": 416 }, { "epoch": 1.8446927374301676, "grad_norm": 0.4929358533949542, "learning_rate": 1.9119878603945374e-06, "loss": 0.464, "step": 417 }, { "epoch": 1.8491620111731844, "grad_norm": 0.4753893398726578, "learning_rate": 1.9044006069802734e-06, "loss": 0.4707, "step": 418 }, { "epoch": 1.853631284916201, "grad_norm": 0.49249135995729143, "learning_rate": 1.8968133535660093e-06, "loss": 0.4674, "step": 419 }, { "epoch": 1.8581005586592179, "grad_norm": 0.4959065755228421, "learning_rate": 1.8892261001517454e-06, "loss": 0.4731, "step": 420 }, { "epoch": 1.8625698324022346, "grad_norm": 0.4830053823448673, "learning_rate": 1.8816388467374813e-06, "loss": 0.4688, "step": 421 }, { "epoch": 1.8670391061452514, "grad_norm": 0.5026774544021604, "learning_rate": 1.874051593323217e-06, "loss": 0.4777, "step": 422 }, { "epoch": 1.8715083798882681, "grad_norm": 0.48376669794821286, "learning_rate": 1.866464339908953e-06, "loss": 0.4532, "step": 423 }, { "epoch": 1.8759776536312849, "grad_norm": 0.5036910075540445, "learning_rate": 1.858877086494689e-06, "loss": 0.4749, "step": 424 }, { "epoch": 1.8804469273743016, "grad_norm": 0.464687327497754, "learning_rate": 1.851289833080425e-06, "loss": 0.4582, "step": 425 }, { "epoch": 1.8849162011173184, "grad_norm": 0.4713537790626329, "learning_rate": 1.8437025796661609e-06, "loss": 0.4954, "step": 426 }, { "epoch": 1.8893854748603351, "grad_norm": 0.4964027557908742, "learning_rate": 1.8361153262518968e-06, "loss": 0.4777, "step": 427 }, { "epoch": 1.893854748603352, "grad_norm": 0.5022160749373462, "learning_rate": 1.828528072837633e-06, "loss": 0.4723, "step": 428 }, { "epoch": 1.8983240223463689, "grad_norm": 0.4971783807561644, "learning_rate": 1.8209408194233688e-06, "loss": 0.4619, "step": 429 }, { "epoch": 1.9027932960893854, "grad_norm": 0.5020482529233811, "learning_rate": 1.8133535660091047e-06, "loss": 0.4923, "step": 430 }, { "epoch": 1.9072625698324024, "grad_norm": 0.5027930539648167, "learning_rate": 1.8057663125948407e-06, "loss": 0.4936, "step": 431 }, { "epoch": 1.911731843575419, "grad_norm": 0.5048109425957825, "learning_rate": 1.7981790591805768e-06, "loss": 0.4647, "step": 432 }, { "epoch": 1.916201117318436, "grad_norm": 0.5152625504278344, "learning_rate": 1.7905918057663127e-06, "loss": 0.4962, "step": 433 }, { "epoch": 1.9206703910614524, "grad_norm": 0.5486472960937676, "learning_rate": 1.7830045523520486e-06, "loss": 0.4643, "step": 434 }, { "epoch": 1.9251396648044694, "grad_norm": 0.5074159325604022, "learning_rate": 1.7754172989377847e-06, "loss": 0.4838, "step": 435 }, { "epoch": 1.929608938547486, "grad_norm": 0.487237277239409, "learning_rate": 1.7678300455235207e-06, "loss": 0.4953, "step": 436 }, { "epoch": 1.934078212290503, "grad_norm": 0.5030243156211434, "learning_rate": 1.7602427921092566e-06, "loss": 0.4669, "step": 437 }, { "epoch": 1.9385474860335195, "grad_norm": 0.54024162061195, "learning_rate": 1.7526555386949925e-06, "loss": 0.4645, "step": 438 }, { "epoch": 1.9430167597765364, "grad_norm": 0.4924548318787792, "learning_rate": 1.7450682852807286e-06, "loss": 0.487, "step": 439 }, { "epoch": 1.947486033519553, "grad_norm": 0.5002700607677886, "learning_rate": 1.7374810318664645e-06, "loss": 0.4904, "step": 440 }, { "epoch": 1.95195530726257, "grad_norm": 0.49052978017153487, "learning_rate": 1.7298937784522004e-06, "loss": 0.4721, "step": 441 }, { "epoch": 1.9564245810055865, "grad_norm": 0.5330277427973065, "learning_rate": 1.7223065250379363e-06, "loss": 0.4674, "step": 442 }, { "epoch": 1.9608938547486034, "grad_norm": 0.49644600432443986, "learning_rate": 1.7147192716236725e-06, "loss": 0.4735, "step": 443 }, { "epoch": 1.96536312849162, "grad_norm": 0.4763428511189237, "learning_rate": 1.7071320182094084e-06, "loss": 0.4758, "step": 444 }, { "epoch": 1.969832402234637, "grad_norm": 0.49662051116381534, "learning_rate": 1.6995447647951443e-06, "loss": 0.5025, "step": 445 }, { "epoch": 1.9743016759776535, "grad_norm": 0.5368027446646121, "learning_rate": 1.6919575113808804e-06, "loss": 0.4708, "step": 446 }, { "epoch": 1.9787709497206705, "grad_norm": 0.49270124177132796, "learning_rate": 1.6843702579666163e-06, "loss": 0.4859, "step": 447 }, { "epoch": 1.983240223463687, "grad_norm": 0.4886244120662818, "learning_rate": 1.6767830045523523e-06, "loss": 0.4542, "step": 448 }, { "epoch": 1.983240223463687, "eval_loss": 0.49883514642715454, "eval_runtime": 32.2692, "eval_samples_per_second": 82.741, "eval_steps_per_second": 5.175, "step": 448 }, { "epoch": 1.987709497206704, "grad_norm": 0.48474622370426634, "learning_rate": 1.6691957511380882e-06, "loss": 0.4752, "step": 449 }, { "epoch": 1.9921787709497205, "grad_norm": 0.5089879136002737, "learning_rate": 1.6616084977238243e-06, "loss": 0.4835, "step": 450 }, { "epoch": 1.9966480446927375, "grad_norm": 0.540987114653738, "learning_rate": 1.6540212443095602e-06, "loss": 0.4685, "step": 451 }, { "epoch": 2.001117318435754, "grad_norm": 0.5330263791441335, "learning_rate": 1.6464339908952961e-06, "loss": 0.4729, "step": 452 }, { "epoch": 2.005586592178771, "grad_norm": 0.5025591514510047, "learning_rate": 1.638846737481032e-06, "loss": 0.4739, "step": 453 }, { "epoch": 2.010055865921788, "grad_norm": 0.4938210284492649, "learning_rate": 1.6312594840667682e-06, "loss": 0.4772, "step": 454 }, { "epoch": 2.0145251396648045, "grad_norm": 0.5590423643222178, "learning_rate": 1.6236722306525039e-06, "loss": 0.4651, "step": 455 }, { "epoch": 2.018994413407821, "grad_norm": 0.4994221649154916, "learning_rate": 1.6160849772382398e-06, "loss": 0.4989, "step": 456 }, { "epoch": 2.004469273743017, "grad_norm": 0.5578750520607872, "learning_rate": 1.6084977238239757e-06, "loss": 0.4779, "step": 457 }, { "epoch": 2.0089385474860335, "grad_norm": 0.5277075050354076, "learning_rate": 1.6009104704097116e-06, "loss": 0.462, "step": 458 }, { "epoch": 2.0134078212290505, "grad_norm": 0.5125996961203592, "learning_rate": 1.5933232169954477e-06, "loss": 0.4892, "step": 459 }, { "epoch": 2.017877094972067, "grad_norm": 0.5150882727539035, "learning_rate": 1.5857359635811836e-06, "loss": 0.4913, "step": 460 }, { "epoch": 2.022346368715084, "grad_norm": 0.5395569256138755, "learning_rate": 1.5781487101669196e-06, "loss": 0.4789, "step": 461 }, { "epoch": 2.0268156424581005, "grad_norm": 0.4907655840755226, "learning_rate": 1.5705614567526557e-06, "loss": 0.4699, "step": 462 }, { "epoch": 2.0312849162011175, "grad_norm": 0.508487377683692, "learning_rate": 1.5629742033383916e-06, "loss": 0.4671, "step": 463 }, { "epoch": 2.035754189944134, "grad_norm": 0.5211842893319257, "learning_rate": 1.5553869499241275e-06, "loss": 0.48, "step": 464 }, { "epoch": 2.040223463687151, "grad_norm": 0.5540445804429339, "learning_rate": 1.5477996965098634e-06, "loss": 0.4839, "step": 465 }, { "epoch": 2.0446927374301676, "grad_norm": 0.502144462775385, "learning_rate": 1.5402124430955995e-06, "loss": 0.4991, "step": 466 }, { "epoch": 2.0491620111731845, "grad_norm": 0.4963826920124019, "learning_rate": 1.5326251896813355e-06, "loss": 0.5268, "step": 467 }, { "epoch": 2.053631284916201, "grad_norm": 0.48862357759049696, "learning_rate": 1.5250379362670714e-06, "loss": 0.4572, "step": 468 }, { "epoch": 2.058100558659218, "grad_norm": 0.48705814461293884, "learning_rate": 1.5174506828528073e-06, "loss": 0.4761, "step": 469 }, { "epoch": 2.0625698324022346, "grad_norm": 0.50151248823518, "learning_rate": 1.5098634294385434e-06, "loss": 0.468, "step": 470 }, { "epoch": 2.0670391061452515, "grad_norm": 0.47075309247993236, "learning_rate": 1.5022761760242793e-06, "loss": 0.4496, "step": 471 }, { "epoch": 2.071508379888268, "grad_norm": 0.5055609634042088, "learning_rate": 1.4946889226100152e-06, "loss": 0.4626, "step": 472 }, { "epoch": 2.075977653631285, "grad_norm": 0.4931560723293831, "learning_rate": 1.4871016691957514e-06, "loss": 0.4743, "step": 473 }, { "epoch": 2.0804469273743016, "grad_norm": 0.4970179492434698, "learning_rate": 1.4795144157814873e-06, "loss": 0.4849, "step": 474 }, { "epoch": 2.0849162011173186, "grad_norm": 0.5205249715468657, "learning_rate": 1.4719271623672232e-06, "loss": 0.4647, "step": 475 }, { "epoch": 2.089385474860335, "grad_norm": 0.5106891771853166, "learning_rate": 1.4643399089529591e-06, "loss": 0.4698, "step": 476 }, { "epoch": 2.093854748603352, "grad_norm": 0.5030533511191912, "learning_rate": 1.4567526555386952e-06, "loss": 0.4435, "step": 477 }, { "epoch": 2.0983240223463686, "grad_norm": 0.5267045945145785, "learning_rate": 1.4491654021244311e-06, "loss": 0.4551, "step": 478 }, { "epoch": 2.1027932960893856, "grad_norm": 0.4910735265010368, "learning_rate": 1.441578148710167e-06, "loss": 0.4893, "step": 479 }, { "epoch": 2.107262569832402, "grad_norm": 0.5098674286346987, "learning_rate": 1.433990895295903e-06, "loss": 0.4631, "step": 480 }, { "epoch": 2.111731843575419, "grad_norm": 0.49833407519162715, "learning_rate": 1.426403641881639e-06, "loss": 0.4478, "step": 481 }, { "epoch": 2.1162011173184356, "grad_norm": 0.4968638787102979, "learning_rate": 1.418816388467375e-06, "loss": 0.4661, "step": 482 }, { "epoch": 2.1206703910614526, "grad_norm": 0.5287581344581138, "learning_rate": 1.411229135053111e-06, "loss": 0.487, "step": 483 }, { "epoch": 2.125139664804469, "grad_norm": 0.5169520017940434, "learning_rate": 1.403641881638847e-06, "loss": 0.4641, "step": 484 }, { "epoch": 2.129608938547486, "grad_norm": 0.4889710176907046, "learning_rate": 1.396054628224583e-06, "loss": 0.4657, "step": 485 }, { "epoch": 2.1340782122905027, "grad_norm": 0.5068054201127302, "learning_rate": 1.3884673748103189e-06, "loss": 0.4655, "step": 486 }, { "epoch": 2.1385474860335196, "grad_norm": 0.5227391076748372, "learning_rate": 1.3808801213960548e-06, "loss": 0.4489, "step": 487 }, { "epoch": 2.143016759776536, "grad_norm": 0.4685531353247096, "learning_rate": 1.3732928679817905e-06, "loss": 0.4518, "step": 488 }, { "epoch": 2.147486033519553, "grad_norm": 0.48729624697805496, "learning_rate": 1.3657056145675266e-06, "loss": 0.4539, "step": 489 }, { "epoch": 2.1519553072625697, "grad_norm": 0.49686659002832273, "learning_rate": 1.3581183611532625e-06, "loss": 0.4698, "step": 490 }, { "epoch": 2.1564245810055866, "grad_norm": 0.5165370496352449, "learning_rate": 1.3505311077389985e-06, "loss": 0.5282, "step": 491 }, { "epoch": 2.160893854748603, "grad_norm": 0.49390583090447004, "learning_rate": 1.3429438543247344e-06, "loss": 0.5011, "step": 492 }, { "epoch": 2.16536312849162, "grad_norm": 0.4966083299512843, "learning_rate": 1.3353566009104705e-06, "loss": 0.4888, "step": 493 }, { "epoch": 2.1698324022346367, "grad_norm": 0.47651097727052893, "learning_rate": 1.3277693474962064e-06, "loss": 0.4629, "step": 494 }, { "epoch": 2.1743016759776537, "grad_norm": 0.5171229401967155, "learning_rate": 1.3201820940819423e-06, "loss": 0.4538, "step": 495 }, { "epoch": 2.17877094972067, "grad_norm": 0.5177579051508318, "learning_rate": 1.3125948406676784e-06, "loss": 0.4822, "step": 496 }, { "epoch": 2.183240223463687, "grad_norm": 0.5363730349952008, "learning_rate": 1.3050075872534144e-06, "loss": 0.4879, "step": 497 }, { "epoch": 2.1877094972067037, "grad_norm": 0.49547274684149156, "learning_rate": 1.2974203338391503e-06, "loss": 0.4732, "step": 498 }, { "epoch": 2.1921787709497207, "grad_norm": 0.48560343902497966, "learning_rate": 1.2898330804248862e-06, "loss": 0.4918, "step": 499 }, { "epoch": 2.1966480446927372, "grad_norm": 0.49966297308311575, "learning_rate": 1.2822458270106223e-06, "loss": 0.4618, "step": 500 }, { "epoch": 2.201117318435754, "grad_norm": 0.472067295226221, "learning_rate": 1.2746585735963582e-06, "loss": 0.4953, "step": 501 }, { "epoch": 2.205586592178771, "grad_norm": 0.5241464474412257, "learning_rate": 1.2670713201820941e-06, "loss": 0.4797, "step": 502 }, { "epoch": 2.2100558659217877, "grad_norm": 0.5063960714766558, "learning_rate": 1.25948406676783e-06, "loss": 0.4493, "step": 503 }, { "epoch": 2.2145251396648042, "grad_norm": 0.48407847297667117, "learning_rate": 1.2518968133535662e-06, "loss": 0.4736, "step": 504 }, { "epoch": 2.218994413407821, "grad_norm": 0.4859389089228068, "learning_rate": 1.244309559939302e-06, "loss": 0.4476, "step": 505 }, { "epoch": 2.223463687150838, "grad_norm": 0.49468343574845597, "learning_rate": 1.236722306525038e-06, "loss": 0.4447, "step": 506 }, { "epoch": 2.2279329608938547, "grad_norm": 0.45980997966401754, "learning_rate": 1.229135053110774e-06, "loss": 0.4599, "step": 507 }, { "epoch": 2.2324022346368717, "grad_norm": 0.5135776498149942, "learning_rate": 1.22154779969651e-06, "loss": 0.4852, "step": 508 }, { "epoch": 2.2368715083798882, "grad_norm": 0.48749385873656575, "learning_rate": 1.213960546282246e-06, "loss": 0.5157, "step": 509 }, { "epoch": 2.241340782122905, "grad_norm": 0.48899276420726745, "learning_rate": 1.2063732928679819e-06, "loss": 0.4437, "step": 510 }, { "epoch": 2.2458100558659218, "grad_norm": 0.4918342780724834, "learning_rate": 1.198786039453718e-06, "loss": 0.4974, "step": 511 }, { "epoch": 2.2502793296089387, "grad_norm": 0.4774969997416472, "learning_rate": 1.191198786039454e-06, "loss": 0.4509, "step": 512 }, { "epoch": 2.2547486033519553, "grad_norm": 0.4791749757272505, "learning_rate": 1.1836115326251896e-06, "loss": 0.4599, "step": 513 }, { "epoch": 2.2592178770949722, "grad_norm": 0.5072314853369309, "learning_rate": 1.1760242792109257e-06, "loss": 0.4643, "step": 514 }, { "epoch": 2.2636871508379888, "grad_norm": 0.5004472392134528, "learning_rate": 1.1684370257966617e-06, "loss": 0.4688, "step": 515 }, { "epoch": 2.2681564245810057, "grad_norm": 0.48150392498478517, "learning_rate": 1.1608497723823976e-06, "loss": 0.5069, "step": 516 }, { "epoch": 2.2726256983240223, "grad_norm": 0.4828855445414688, "learning_rate": 1.1532625189681337e-06, "loss": 0.4571, "step": 517 }, { "epoch": 2.2770949720670393, "grad_norm": 0.49789128855450915, "learning_rate": 1.1456752655538696e-06, "loss": 0.4651, "step": 518 }, { "epoch": 2.281564245810056, "grad_norm": 0.49890243849633514, "learning_rate": 1.1380880121396055e-06, "loss": 0.4757, "step": 519 }, { "epoch": 2.2860335195530728, "grad_norm": 0.48825603773254755, "learning_rate": 1.1305007587253414e-06, "loss": 0.4668, "step": 520 }, { "epoch": 2.2905027932960893, "grad_norm": 0.48797823917514366, "learning_rate": 1.1229135053110776e-06, "loss": 0.4576, "step": 521 }, { "epoch": 2.2949720670391063, "grad_norm": 0.4713265522467315, "learning_rate": 1.1153262518968135e-06, "loss": 0.4998, "step": 522 }, { "epoch": 2.299441340782123, "grad_norm": 0.48731093463299924, "learning_rate": 1.1077389984825494e-06, "loss": 0.4583, "step": 523 }, { "epoch": 2.30391061452514, "grad_norm": 0.5356992050391173, "learning_rate": 1.1001517450682853e-06, "loss": 0.4964, "step": 524 }, { "epoch": 2.3083798882681563, "grad_norm": 0.5185855845215128, "learning_rate": 1.0925644916540214e-06, "loss": 0.4456, "step": 525 }, { "epoch": 2.3128491620111733, "grad_norm": 0.5340341729511069, "learning_rate": 1.0849772382397573e-06, "loss": 0.4667, "step": 526 }, { "epoch": 2.31731843575419, "grad_norm": 0.46791114784776194, "learning_rate": 1.0773899848254933e-06, "loss": 0.4907, "step": 527 }, { "epoch": 2.321787709497207, "grad_norm": 0.5224682012273069, "learning_rate": 1.0698027314112294e-06, "loss": 0.4411, "step": 528 }, { "epoch": 2.3262569832402233, "grad_norm": 0.5013175127206561, "learning_rate": 1.062215477996965e-06, "loss": 0.4872, "step": 529 }, { "epoch": 2.3307262569832403, "grad_norm": 0.49567401298066655, "learning_rate": 1.054628224582701e-06, "loss": 0.4547, "step": 530 }, { "epoch": 2.335195530726257, "grad_norm": 0.5092075971225333, "learning_rate": 1.0470409711684371e-06, "loss": 0.4774, "step": 531 }, { "epoch": 2.339664804469274, "grad_norm": 0.5206537811339336, "learning_rate": 1.039453717754173e-06, "loss": 0.4593, "step": 532 }, { "epoch": 2.3441340782122904, "grad_norm": 0.5247188237207532, "learning_rate": 1.031866464339909e-06, "loss": 0.4947, "step": 533 }, { "epoch": 2.3486033519553073, "grad_norm": 0.48511496092549006, "learning_rate": 1.024279210925645e-06, "loss": 0.4675, "step": 534 }, { "epoch": 2.353072625698324, "grad_norm": 0.49031229609343957, "learning_rate": 1.016691957511381e-06, "loss": 0.4654, "step": 535 }, { "epoch": 2.357541899441341, "grad_norm": 0.5022437562865831, "learning_rate": 1.009104704097117e-06, "loss": 0.4881, "step": 536 }, { "epoch": 2.3620111731843574, "grad_norm": 0.48427536854604036, "learning_rate": 1.0015174506828528e-06, "loss": 0.4722, "step": 537 }, { "epoch": 2.3664804469273744, "grad_norm": 0.4951077677256604, "learning_rate": 9.93930197268589e-07, "loss": 0.4816, "step": 538 }, { "epoch": 2.370949720670391, "grad_norm": 0.506971274674449, "learning_rate": 9.863429438543249e-07, "loss": 0.448, "step": 539 }, { "epoch": 2.375418994413408, "grad_norm": 0.4843767211745505, "learning_rate": 9.787556904400608e-07, "loss": 0.4644, "step": 540 }, { "epoch": 2.3798882681564244, "grad_norm": 0.4787312911085378, "learning_rate": 9.711684370257967e-07, "loss": 0.4681, "step": 541 }, { "epoch": 2.3843575418994414, "grad_norm": 0.5040950848521704, "learning_rate": 9.635811836115328e-07, "loss": 0.4595, "step": 542 }, { "epoch": 2.388826815642458, "grad_norm": 0.4742009596031962, "learning_rate": 9.559939301972687e-07, "loss": 0.4542, "step": 543 }, { "epoch": 2.393296089385475, "grad_norm": 0.4819914757091554, "learning_rate": 9.484066767830046e-07, "loss": 0.4818, "step": 544 }, { "epoch": 2.3977653631284914, "grad_norm": 0.4794763925814031, "learning_rate": 9.408194233687407e-07, "loss": 0.4459, "step": 545 }, { "epoch": 2.4022346368715084, "grad_norm": 0.5065706041345039, "learning_rate": 9.332321699544765e-07, "loss": 0.4986, "step": 546 }, { "epoch": 2.4067039106145254, "grad_norm": 0.4918492784353524, "learning_rate": 9.256449165402125e-07, "loss": 0.4618, "step": 547 }, { "epoch": 2.411173184357542, "grad_norm": 0.48410824249150014, "learning_rate": 9.180576631259484e-07, "loss": 0.4358, "step": 548 }, { "epoch": 2.4156424581005584, "grad_norm": 0.5071386174484, "learning_rate": 9.104704097116844e-07, "loss": 0.4522, "step": 549 }, { "epoch": 2.4201117318435754, "grad_norm": 0.5179035085963222, "learning_rate": 9.028831562974203e-07, "loss": 0.4718, "step": 550 }, { "epoch": 2.4245810055865924, "grad_norm": 0.48695002255677194, "learning_rate": 8.952959028831563e-07, "loss": 0.4946, "step": 551 }, { "epoch": 2.429050279329609, "grad_norm": 0.4697467884861594, "learning_rate": 8.877086494688924e-07, "loss": 0.482, "step": 552 }, { "epoch": 2.4335195530726255, "grad_norm": 0.5039869709064255, "learning_rate": 8.801213960546283e-07, "loss": 0.4528, "step": 553 }, { "epoch": 2.4379888268156424, "grad_norm": 0.4956541767737936, "learning_rate": 8.725341426403643e-07, "loss": 0.4416, "step": 554 }, { "epoch": 2.4424581005586594, "grad_norm": 0.5220748437476452, "learning_rate": 8.649468892261002e-07, "loss": 0.4534, "step": 555 }, { "epoch": 2.446927374301676, "grad_norm": 0.5004114462806725, "learning_rate": 8.573596358118362e-07, "loss": 0.4523, "step": 556 }, { "epoch": 2.451396648044693, "grad_norm": 0.5002883358165275, "learning_rate": 8.497723823975721e-07, "loss": 0.4483, "step": 557 }, { "epoch": 2.4558659217877095, "grad_norm": 0.5001315929982186, "learning_rate": 8.421851289833082e-07, "loss": 0.4547, "step": 558 }, { "epoch": 2.4603351955307264, "grad_norm": 0.5141518132214271, "learning_rate": 8.345978755690441e-07, "loss": 0.5064, "step": 559 }, { "epoch": 2.464804469273743, "grad_norm": 0.4960566337702956, "learning_rate": 8.270106221547801e-07, "loss": 0.4473, "step": 560 }, { "epoch": 2.464804469273743, "eval_loss": 0.4975128769874573, "eval_runtime": 32.1892, "eval_samples_per_second": 82.947, "eval_steps_per_second": 5.188, "step": 560 }, { "epoch": 2.46927374301676, "grad_norm": 0.4897198333575608, "learning_rate": 8.19423368740516e-07, "loss": 0.4559, "step": 561 }, { "epoch": 2.4737430167597765, "grad_norm": 0.49985877359817477, "learning_rate": 8.118361153262519e-07, "loss": 0.4651, "step": 562 }, { "epoch": 2.4782122905027935, "grad_norm": 0.49485306049230043, "learning_rate": 8.042488619119878e-07, "loss": 0.4803, "step": 563 }, { "epoch": 2.48268156424581, "grad_norm": 0.5208900276481403, "learning_rate": 7.966616084977239e-07, "loss": 0.4553, "step": 564 }, { "epoch": 2.487150837988827, "grad_norm": 0.48837650054879367, "learning_rate": 7.890743550834598e-07, "loss": 0.4681, "step": 565 }, { "epoch": 2.4916201117318435, "grad_norm": 0.47305516115923485, "learning_rate": 7.814871016691958e-07, "loss": 0.4564, "step": 566 }, { "epoch": 2.4960893854748605, "grad_norm": 0.5049602967209187, "learning_rate": 7.738998482549317e-07, "loss": 0.4632, "step": 567 }, { "epoch": 2.500558659217877, "grad_norm": 0.48547661489581506, "learning_rate": 7.663125948406677e-07, "loss": 0.448, "step": 568 }, { "epoch": 2.505027932960894, "grad_norm": 0.5747686136950829, "learning_rate": 7.587253414264036e-07, "loss": 0.4432, "step": 569 }, { "epoch": 2.5094972067039105, "grad_norm": 0.5499322063415041, "learning_rate": 7.511380880121397e-07, "loss": 0.432, "step": 570 }, { "epoch": 2.5139664804469275, "grad_norm": 0.49909881450641125, "learning_rate": 7.435508345978757e-07, "loss": 0.4325, "step": 571 }, { "epoch": 2.518435754189944, "grad_norm": 0.5137001617086465, "learning_rate": 7.359635811836116e-07, "loss": 0.4374, "step": 572 }, { "epoch": 2.522905027932961, "grad_norm": 0.5043757440405017, "learning_rate": 7.283763277693476e-07, "loss": 0.4751, "step": 573 }, { "epoch": 2.5273743016759775, "grad_norm": 0.4882894387336232, "learning_rate": 7.207890743550835e-07, "loss": 0.4368, "step": 574 }, { "epoch": 2.5318435754189945, "grad_norm": 0.5302357654250439, "learning_rate": 7.132018209408196e-07, "loss": 0.4174, "step": 575 }, { "epoch": 2.536312849162011, "grad_norm": 0.5229623842607946, "learning_rate": 7.056145675265555e-07, "loss": 0.4353, "step": 576 }, { "epoch": 2.540782122905028, "grad_norm": 0.5145290981136341, "learning_rate": 6.980273141122915e-07, "loss": 0.4278, "step": 577 }, { "epoch": 2.5452513966480446, "grad_norm": 0.49830066459201744, "learning_rate": 6.904400606980274e-07, "loss": 0.434, "step": 578 }, { "epoch": 2.5497206703910615, "grad_norm": 0.5220459047427843, "learning_rate": 6.828528072837633e-07, "loss": 0.4375, "step": 579 }, { "epoch": 2.554189944134078, "grad_norm": 0.5111391623585094, "learning_rate": 6.752655538694992e-07, "loss": 0.4295, "step": 580 }, { "epoch": 2.558659217877095, "grad_norm": 0.4944043554942434, "learning_rate": 6.676783004552352e-07, "loss": 0.4205, "step": 581 }, { "epoch": 2.5631284916201116, "grad_norm": 0.49154601956517674, "learning_rate": 6.600910470409712e-07, "loss": 0.4502, "step": 582 }, { "epoch": 2.5675977653631286, "grad_norm": 0.5008348019898189, "learning_rate": 6.525037936267072e-07, "loss": 0.4453, "step": 583 }, { "epoch": 2.572067039106145, "grad_norm": 0.5005572885290496, "learning_rate": 6.449165402124431e-07, "loss": 0.4353, "step": 584 }, { "epoch": 2.576536312849162, "grad_norm": 0.5302967367534124, "learning_rate": 6.373292867981791e-07, "loss": 0.4269, "step": 585 }, { "epoch": 2.5810055865921786, "grad_norm": 0.5050943269077487, "learning_rate": 6.29742033383915e-07, "loss": 0.4372, "step": 586 }, { "epoch": 2.5854748603351956, "grad_norm": 0.5223556534474804, "learning_rate": 6.22154779969651e-07, "loss": 0.445, "step": 587 }, { "epoch": 2.5899441340782126, "grad_norm": 0.4891137574673572, "learning_rate": 6.14567526555387e-07, "loss": 0.4171, "step": 588 }, { "epoch": 2.594413407821229, "grad_norm": 0.5172276124686148, "learning_rate": 6.06980273141123e-07, "loss": 0.4522, "step": 589 }, { "epoch": 2.5988826815642456, "grad_norm": 0.48873823307103154, "learning_rate": 5.99393019726859e-07, "loss": 0.4286, "step": 590 }, { "epoch": 2.6033519553072626, "grad_norm": 0.502230133682332, "learning_rate": 5.918057663125948e-07, "loss": 0.4461, "step": 591 }, { "epoch": 2.6078212290502796, "grad_norm": 0.4898396410998909, "learning_rate": 5.842185128983308e-07, "loss": 0.4289, "step": 592 }, { "epoch": 2.612290502793296, "grad_norm": 0.48422568768513635, "learning_rate": 5.766312594840668e-07, "loss": 0.4423, "step": 593 }, { "epoch": 2.6167597765363126, "grad_norm": 0.5003994120834477, "learning_rate": 5.690440060698028e-07, "loss": 0.437, "step": 594 }, { "epoch": 2.6212290502793296, "grad_norm": 0.5079705228889307, "learning_rate": 5.614567526555388e-07, "loss": 0.4516, "step": 595 }, { "epoch": 2.6256983240223466, "grad_norm": 0.49246631419316356, "learning_rate": 5.538694992412747e-07, "loss": 0.4368, "step": 596 }, { "epoch": 2.630167597765363, "grad_norm": 0.5086922889782935, "learning_rate": 5.462822458270107e-07, "loss": 0.4227, "step": 597 }, { "epoch": 2.6346368715083797, "grad_norm": 0.4780759168718152, "learning_rate": 5.386949924127466e-07, "loss": 0.4462, "step": 598 }, { "epoch": 2.6391061452513966, "grad_norm": 0.464930849547806, "learning_rate": 5.311077389984825e-07, "loss": 0.4334, "step": 599 }, { "epoch": 2.6435754189944136, "grad_norm": 0.5002300455424397, "learning_rate": 5.235204855842186e-07, "loss": 0.4641, "step": 600 }, { "epoch": 2.64804469273743, "grad_norm": 0.4901819784256759, "learning_rate": 5.159332321699545e-07, "loss": 0.4534, "step": 601 }, { "epoch": 2.6525139664804467, "grad_norm": 0.48520187595124403, "learning_rate": 5.083459787556905e-07, "loss": 0.4246, "step": 602 }, { "epoch": 2.6569832402234637, "grad_norm": 0.5093634792454178, "learning_rate": 5.007587253414264e-07, "loss": 0.458, "step": 603 }, { "epoch": 2.6614525139664806, "grad_norm": 0.47795678253705354, "learning_rate": 4.931714719271624e-07, "loss": 0.4535, "step": 604 }, { "epoch": 2.665921787709497, "grad_norm": 0.48340944166993405, "learning_rate": 4.855842185128983e-07, "loss": 0.4721, "step": 605 }, { "epoch": 2.6703910614525137, "grad_norm": 0.47996848455446794, "learning_rate": 4.779969650986344e-07, "loss": 0.4718, "step": 606 }, { "epoch": 2.6748603351955307, "grad_norm": 0.5064110255390989, "learning_rate": 4.7040971168437033e-07, "loss": 0.4232, "step": 607 }, { "epoch": 2.6793296089385477, "grad_norm": 0.4824240302682651, "learning_rate": 4.6282245827010624e-07, "loss": 0.4307, "step": 608 }, { "epoch": 2.683798882681564, "grad_norm": 0.4920800315955025, "learning_rate": 4.552352048558422e-07, "loss": 0.4435, "step": 609 }, { "epoch": 2.688268156424581, "grad_norm": 0.48921504704563673, "learning_rate": 4.476479514415782e-07, "loss": 0.4409, "step": 610 }, { "epoch": 2.6927374301675977, "grad_norm": 0.500357475284234, "learning_rate": 4.4006069802731414e-07, "loss": 0.4512, "step": 611 }, { "epoch": 2.6972067039106147, "grad_norm": 0.4772853981005586, "learning_rate": 4.324734446130501e-07, "loss": 0.4305, "step": 612 }, { "epoch": 2.701675977653631, "grad_norm": 0.5075964338612743, "learning_rate": 4.248861911987861e-07, "loss": 0.4285, "step": 613 }, { "epoch": 2.706145251396648, "grad_norm": 0.4916721748305679, "learning_rate": 4.1729893778452204e-07, "loss": 0.4586, "step": 614 }, { "epoch": 2.7106145251396647, "grad_norm": 0.4746003660659448, "learning_rate": 4.09711684370258e-07, "loss": 0.4385, "step": 615 }, { "epoch": 2.7150837988826817, "grad_norm": 0.49004649736827927, "learning_rate": 4.021244309559939e-07, "loss": 0.4349, "step": 616 }, { "epoch": 2.7195530726256982, "grad_norm": 0.5055169203153188, "learning_rate": 3.945371775417299e-07, "loss": 0.4741, "step": 617 }, { "epoch": 2.724022346368715, "grad_norm": 0.5491416139361283, "learning_rate": 3.8694992412746586e-07, "loss": 0.4285, "step": 618 }, { "epoch": 2.7284916201117317, "grad_norm": 0.4860669923591883, "learning_rate": 3.793626707132018e-07, "loss": 0.4147, "step": 619 }, { "epoch": 2.7329608938547487, "grad_norm": 0.48537127261627433, "learning_rate": 3.7177541729893784e-07, "loss": 0.4538, "step": 620 }, { "epoch": 2.7374301675977653, "grad_norm": 0.47557197343237195, "learning_rate": 3.641881638846738e-07, "loss": 0.454, "step": 621 }, { "epoch": 2.7418994413407822, "grad_norm": 0.4844044051975442, "learning_rate": 3.566009104704098e-07, "loss": 0.4273, "step": 622 }, { "epoch": 2.7463687150837988, "grad_norm": 0.5000172198487197, "learning_rate": 3.4901365705614574e-07, "loss": 0.4315, "step": 623 }, { "epoch": 2.7508379888268157, "grad_norm": 0.4848986778566373, "learning_rate": 3.4142640364188166e-07, "loss": 0.4495, "step": 624 }, { "epoch": 2.7553072625698323, "grad_norm": 0.48082262468066483, "learning_rate": 3.338391502276176e-07, "loss": 0.4185, "step": 625 }, { "epoch": 2.7597765363128492, "grad_norm": 0.4865964770369407, "learning_rate": 3.262518968133536e-07, "loss": 0.4538, "step": 626 }, { "epoch": 2.764245810055866, "grad_norm": 0.4815694373638968, "learning_rate": 3.1866464339908956e-07, "loss": 0.459, "step": 627 }, { "epoch": 2.7687150837988828, "grad_norm": 0.49485301744734556, "learning_rate": 3.110773899848255e-07, "loss": 0.4527, "step": 628 }, { "epoch": 2.7731843575418993, "grad_norm": 0.5214886405175438, "learning_rate": 3.034901365705615e-07, "loss": 0.4393, "step": 629 }, { "epoch": 2.7776536312849163, "grad_norm": 0.4645817613806632, "learning_rate": 2.959028831562974e-07, "loss": 0.4723, "step": 630 }, { "epoch": 2.782122905027933, "grad_norm": 0.4887881586065441, "learning_rate": 2.883156297420334e-07, "loss": 0.418, "step": 631 }, { "epoch": 2.78659217877095, "grad_norm": 0.47562569136752975, "learning_rate": 2.807283763277694e-07, "loss": 0.4347, "step": 632 }, { "epoch": 2.7910614525139668, "grad_norm": 0.4848933584055101, "learning_rate": 2.7314112291350536e-07, "loss": 0.4266, "step": 633 }, { "epoch": 2.7955307262569833, "grad_norm": 0.48829866920710585, "learning_rate": 2.6555386949924127e-07, "loss": 0.4522, "step": 634 }, { "epoch": 2.8, "grad_norm": 0.4776396492507318, "learning_rate": 2.5796661608497724e-07, "loss": 0.4293, "step": 635 }, { "epoch": 2.804469273743017, "grad_norm": 0.48342626581685083, "learning_rate": 2.503793626707132e-07, "loss": 0.437, "step": 636 }, { "epoch": 2.8089385474860338, "grad_norm": 0.47481545268800873, "learning_rate": 2.4279210925644917e-07, "loss": 0.4333, "step": 637 }, { "epoch": 2.8134078212290503, "grad_norm": 0.5105857690012979, "learning_rate": 2.3520485584218516e-07, "loss": 0.4367, "step": 638 }, { "epoch": 2.817877094972067, "grad_norm": 0.49397311695562557, "learning_rate": 2.276176024279211e-07, "loss": 0.4355, "step": 639 }, { "epoch": 2.822346368715084, "grad_norm": 0.4766966322472669, "learning_rate": 2.2003034901365707e-07, "loss": 0.4478, "step": 640 }, { "epoch": 2.826815642458101, "grad_norm": 0.47484165703817144, "learning_rate": 2.1244309559939304e-07, "loss": 0.4366, "step": 641 }, { "epoch": 2.8312849162011173, "grad_norm": 0.47271019258961905, "learning_rate": 2.04855842185129e-07, "loss": 0.4473, "step": 642 }, { "epoch": 2.835754189944134, "grad_norm": 0.4668544421054751, "learning_rate": 1.9726858877086494e-07, "loss": 0.4341, "step": 643 }, { "epoch": 2.840223463687151, "grad_norm": 0.47716986376667936, "learning_rate": 1.896813353566009e-07, "loss": 0.4223, "step": 644 }, { "epoch": 2.844692737430168, "grad_norm": 0.4939991882199546, "learning_rate": 1.820940819423369e-07, "loss": 0.4441, "step": 645 }, { "epoch": 2.8491620111731844, "grad_norm": 0.47705877463455093, "learning_rate": 1.7450682852807287e-07, "loss": 0.485, "step": 646 }, { "epoch": 2.853631284916201, "grad_norm": 0.49941990019846055, "learning_rate": 1.669195751138088e-07, "loss": 0.4701, "step": 647 }, { "epoch": 2.858100558659218, "grad_norm": 0.47625945090740057, "learning_rate": 1.5933232169954478e-07, "loss": 0.454, "step": 648 }, { "epoch": 2.862569832402235, "grad_norm": 0.5028271948555065, "learning_rate": 1.5174506828528074e-07, "loss": 0.4296, "step": 649 }, { "epoch": 2.8670391061452514, "grad_norm": 0.49358352504671604, "learning_rate": 1.441578148710167e-07, "loss": 0.4324, "step": 650 }, { "epoch": 2.871508379888268, "grad_norm": 0.48576223310259714, "learning_rate": 1.3657056145675268e-07, "loss": 0.4424, "step": 651 }, { "epoch": 2.875977653631285, "grad_norm": 0.47971171549572994, "learning_rate": 1.2898330804248862e-07, "loss": 0.4555, "step": 652 }, { "epoch": 2.880446927374302, "grad_norm": 0.49955392754214234, "learning_rate": 1.2139605462822459e-07, "loss": 0.4338, "step": 653 }, { "epoch": 2.8849162011173184, "grad_norm": 0.47232821100099637, "learning_rate": 1.1380880121396055e-07, "loss": 0.4349, "step": 654 }, { "epoch": 2.889385474860335, "grad_norm": 0.4834170945858496, "learning_rate": 1.0622154779969652e-07, "loss": 0.4564, "step": 655 }, { "epoch": 2.893854748603352, "grad_norm": 0.48493984226629705, "learning_rate": 9.863429438543247e-08, "loss": 0.4384, "step": 656 }, { "epoch": 2.898324022346369, "grad_norm": 0.4601302405573049, "learning_rate": 9.104704097116845e-08, "loss": 0.4512, "step": 657 }, { "epoch": 2.9027932960893854, "grad_norm": 0.4768143158802154, "learning_rate": 8.34597875569044e-08, "loss": 0.4354, "step": 658 }, { "epoch": 2.9072625698324024, "grad_norm": 0.49427262358096746, "learning_rate": 7.587253414264037e-08, "loss": 0.4259, "step": 659 }, { "epoch": 2.911731843575419, "grad_norm": 0.4749814243858473, "learning_rate": 6.828528072837634e-08, "loss": 0.4445, "step": 660 }, { "epoch": 2.916201117318436, "grad_norm": 0.48533516583361297, "learning_rate": 6.069802731411229e-08, "loss": 0.4799, "step": 661 }, { "epoch": 2.9206703910614524, "grad_norm": 0.4839142899304728, "learning_rate": 5.311077389984826e-08, "loss": 0.4574, "step": 662 }, { "epoch": 2.9251396648044694, "grad_norm": 0.486543073762813, "learning_rate": 4.5523520485584226e-08, "loss": 0.4446, "step": 663 }, { "epoch": 2.929608938547486, "grad_norm": 0.4980109939724879, "learning_rate": 3.7936267071320186e-08, "loss": 0.4315, "step": 664 }, { "epoch": 2.934078212290503, "grad_norm": 0.48840296818117557, "learning_rate": 3.0349013657056146e-08, "loss": 0.4318, "step": 665 }, { "epoch": 2.9385474860335195, "grad_norm": 0.4903615270099699, "learning_rate": 2.2761760242792113e-08, "loss": 0.4357, "step": 666 }, { "epoch": 2.9430167597765364, "grad_norm": 0.48850430880690693, "learning_rate": 1.5174506828528073e-08, "loss": 0.4334, "step": 667 }, { "epoch": 2.947486033519553, "grad_norm": 0.4773488572878224, "learning_rate": 7.587253414264037e-09, "loss": 0.4281, "step": 668 }, { "epoch": 2.95195530726257, "grad_norm": 0.4760606483764159, "learning_rate": 0.0, "loss": 0.4504, "step": 669 } ], "logging_steps": 1, "max_steps": 669, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 56, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 559881199288320.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }