|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.95195530726257, |
|
"eval_steps": 112, |
|
"global_step": 669, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.004464285714285714, |
|
"grad_norm": 5.096107617240597, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 0.81, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.004464285714285714, |
|
"eval_loss": 0.7919066548347473, |
|
"eval_runtime": 32.2179, |
|
"eval_samples_per_second": 82.873, |
|
"eval_steps_per_second": 5.183, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.008928571428571428, |
|
"grad_norm": 4.9041704402978805, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 0.7839, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.013392857142857142, |
|
"grad_norm": 4.977917588021941, |
|
"learning_rate": 1.5e-06, |
|
"loss": 0.7745, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.017857142857142856, |
|
"grad_norm": 4.7373014121550705, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.7681, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.022321428571428572, |
|
"grad_norm": 4.563297129857777, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.7589, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.026785714285714284, |
|
"grad_norm": 3.5264108754930787, |
|
"learning_rate": 3e-06, |
|
"loss": 0.7629, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.03125, |
|
"grad_norm": 3.4126703272457166, |
|
"learning_rate": 3.5e-06, |
|
"loss": 0.7181, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.03571428571428571, |
|
"grad_norm": 3.6971429493758636, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.667, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.04017857142857143, |
|
"grad_norm": 1.940701767077778, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.6629, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.044642857142857144, |
|
"grad_norm": 12.40862535155878, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6832, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.049107142857142856, |
|
"grad_norm": 2.8235547786373867, |
|
"learning_rate": 4.992447129909366e-06, |
|
"loss": 0.6213, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.05357142857142857, |
|
"grad_norm": 1.734602557718062, |
|
"learning_rate": 4.984894259818732e-06, |
|
"loss": 0.6279, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.05803571428571429, |
|
"grad_norm": 1.3840828717613172, |
|
"learning_rate": 4.977341389728097e-06, |
|
"loss": 0.6134, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0625, |
|
"grad_norm": 1.1262172615918475, |
|
"learning_rate": 4.969788519637463e-06, |
|
"loss": 0.6029, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.06696428571428571, |
|
"grad_norm": 0.9263739512436016, |
|
"learning_rate": 4.962235649546828e-06, |
|
"loss": 0.6067, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.07142857142857142, |
|
"grad_norm": 0.8374584146721384, |
|
"learning_rate": 4.954682779456194e-06, |
|
"loss": 0.6014, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.07589285714285714, |
|
"grad_norm": 0.7757487058764371, |
|
"learning_rate": 4.9471299093655595e-06, |
|
"loss": 0.5906, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.08035714285714286, |
|
"grad_norm": 0.6531787929864147, |
|
"learning_rate": 4.939577039274925e-06, |
|
"loss": 0.5589, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.08482142857142858, |
|
"grad_norm": 0.7223853317191319, |
|
"learning_rate": 4.93202416918429e-06, |
|
"loss": 0.5877, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.08928571428571429, |
|
"grad_norm": 0.6793848952870568, |
|
"learning_rate": 4.924471299093656e-06, |
|
"loss": 0.6075, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.09375, |
|
"grad_norm": 0.65689704768199, |
|
"learning_rate": 4.9169184290030215e-06, |
|
"loss": 0.5659, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.09821428571428571, |
|
"grad_norm": 0.6553424447930819, |
|
"learning_rate": 4.909365558912387e-06, |
|
"loss": 0.5728, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.10267857142857142, |
|
"grad_norm": 0.6095768945247507, |
|
"learning_rate": 4.901812688821753e-06, |
|
"loss": 0.5716, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.10714285714285714, |
|
"grad_norm": 0.5874011280839848, |
|
"learning_rate": 4.894259818731118e-06, |
|
"loss": 0.5736, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.11160714285714286, |
|
"grad_norm": 0.642646780124128, |
|
"learning_rate": 4.8867069486404835e-06, |
|
"loss": 0.5889, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.11607142857142858, |
|
"grad_norm": 0.5585168252549827, |
|
"learning_rate": 4.879154078549849e-06, |
|
"loss": 0.5761, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.12053571428571429, |
|
"grad_norm": 0.4948199084353006, |
|
"learning_rate": 4.871601208459215e-06, |
|
"loss": 0.545, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.125, |
|
"grad_norm": 0.580699538693282, |
|
"learning_rate": 4.864048338368581e-06, |
|
"loss": 0.5874, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.12946428571428573, |
|
"grad_norm": 0.5483036703790811, |
|
"learning_rate": 4.8564954682779455e-06, |
|
"loss": 0.553, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.13392857142857142, |
|
"grad_norm": 0.5207516065535861, |
|
"learning_rate": 4.848942598187312e-06, |
|
"loss": 0.54, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.13839285714285715, |
|
"grad_norm": 0.5453894595978196, |
|
"learning_rate": 4.841389728096677e-06, |
|
"loss": 0.5779, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.14285714285714285, |
|
"grad_norm": 0.5223897707340578, |
|
"learning_rate": 4.833836858006043e-06, |
|
"loss": 0.5518, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.14732142857142858, |
|
"grad_norm": 0.5473812522488812, |
|
"learning_rate": 4.826283987915408e-06, |
|
"loss": 0.5831, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.15178571428571427, |
|
"grad_norm": 0.5437546013329395, |
|
"learning_rate": 4.818731117824774e-06, |
|
"loss": 0.5884, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.15625, |
|
"grad_norm": 0.5305579139277316, |
|
"learning_rate": 4.81117824773414e-06, |
|
"loss": 0.5398, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.16071428571428573, |
|
"grad_norm": 0.5258419287632591, |
|
"learning_rate": 4.803625377643505e-06, |
|
"loss": 0.5593, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.16517857142857142, |
|
"grad_norm": 0.5158851354579528, |
|
"learning_rate": 4.79607250755287e-06, |
|
"loss": 0.5564, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.16964285714285715, |
|
"grad_norm": 0.4824844062486759, |
|
"learning_rate": 4.788519637462236e-06, |
|
"loss": 0.5616, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.17410714285714285, |
|
"grad_norm": 0.5173226727977146, |
|
"learning_rate": 4.780966767371602e-06, |
|
"loss": 0.5674, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.17857142857142858, |
|
"grad_norm": 0.4966629400190002, |
|
"learning_rate": 4.773413897280967e-06, |
|
"loss": 0.5833, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.18303571428571427, |
|
"grad_norm": 0.48186435126437244, |
|
"learning_rate": 4.765861027190333e-06, |
|
"loss": 0.5571, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.1875, |
|
"grad_norm": 0.5029558558326376, |
|
"learning_rate": 4.758308157099698e-06, |
|
"loss": 0.5358, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.19196428571428573, |
|
"grad_norm": 0.5249190794688006, |
|
"learning_rate": 4.750755287009064e-06, |
|
"loss": 0.5542, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.19642857142857142, |
|
"grad_norm": 0.4622817114582695, |
|
"learning_rate": 4.743202416918429e-06, |
|
"loss": 0.54, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.20089285714285715, |
|
"grad_norm": 0.49789824735563454, |
|
"learning_rate": 4.735649546827795e-06, |
|
"loss": 0.5308, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.20535714285714285, |
|
"grad_norm": 0.51199116739647, |
|
"learning_rate": 4.728096676737161e-06, |
|
"loss": 0.5784, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.20982142857142858, |
|
"grad_norm": 0.47152096107780506, |
|
"learning_rate": 4.720543806646526e-06, |
|
"loss": 0.5654, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.21428571428571427, |
|
"grad_norm": 0.5233581267403502, |
|
"learning_rate": 4.712990936555891e-06, |
|
"loss": 0.5727, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.21875, |
|
"grad_norm": 0.517908769266997, |
|
"learning_rate": 4.705438066465257e-06, |
|
"loss": 0.5258, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.22321428571428573, |
|
"grad_norm": 0.5103782222366013, |
|
"learning_rate": 4.697885196374623e-06, |
|
"loss": 0.564, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.22767857142857142, |
|
"grad_norm": 0.5894196264675328, |
|
"learning_rate": 4.6903323262839885e-06, |
|
"loss": 0.5355, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.23214285714285715, |
|
"grad_norm": 0.48136408513935, |
|
"learning_rate": 4.682779456193353e-06, |
|
"loss": 0.5714, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.23660714285714285, |
|
"grad_norm": 0.5996790928794941, |
|
"learning_rate": 4.67522658610272e-06, |
|
"loss": 0.5385, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.24107142857142858, |
|
"grad_norm": 0.5231061994660855, |
|
"learning_rate": 4.667673716012085e-06, |
|
"loss": 0.5429, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.24553571428571427, |
|
"grad_norm": 0.5416056944426403, |
|
"learning_rate": 4.6601208459214505e-06, |
|
"loss": 0.5423, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.5734105721889117, |
|
"learning_rate": 4.652567975830816e-06, |
|
"loss": 0.5756, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.2544642857142857, |
|
"grad_norm": 0.5318541491738474, |
|
"learning_rate": 4.645015105740182e-06, |
|
"loss": 0.5486, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.25892857142857145, |
|
"grad_norm": 0.4839987879630258, |
|
"learning_rate": 4.637462235649548e-06, |
|
"loss": 0.5517, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.26339285714285715, |
|
"grad_norm": 0.5843387907103592, |
|
"learning_rate": 4.6299093655589125e-06, |
|
"loss": 0.5288, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.26785714285714285, |
|
"grad_norm": 0.5496996598474941, |
|
"learning_rate": 4.622356495468278e-06, |
|
"loss": 0.542, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.27232142857142855, |
|
"grad_norm": 0.5192508705264864, |
|
"learning_rate": 4.614803625377644e-06, |
|
"loss": 0.5291, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.2767857142857143, |
|
"grad_norm": 0.5918952697319948, |
|
"learning_rate": 4.60725075528701e-06, |
|
"loss": 0.5229, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.28125, |
|
"grad_norm": 0.5914173604947244, |
|
"learning_rate": 4.5996978851963745e-06, |
|
"loss": 0.5754, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.2857142857142857, |
|
"grad_norm": 0.5298896547584293, |
|
"learning_rate": 4.592145015105741e-06, |
|
"loss": 0.5687, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.29017857142857145, |
|
"grad_norm": 0.5671535875314645, |
|
"learning_rate": 4.584592145015106e-06, |
|
"loss": 0.5513, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.29464285714285715, |
|
"grad_norm": 0.5234306194196137, |
|
"learning_rate": 4.577039274924472e-06, |
|
"loss": 0.5226, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.29910714285714285, |
|
"grad_norm": 0.4959853502285333, |
|
"learning_rate": 4.569486404833837e-06, |
|
"loss": 0.5335, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.30357142857142855, |
|
"grad_norm": 0.47729026424742405, |
|
"learning_rate": 4.561933534743202e-06, |
|
"loss": 0.5248, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.3080357142857143, |
|
"grad_norm": 0.47294297292566195, |
|
"learning_rate": 4.554380664652569e-06, |
|
"loss": 0.5588, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.3125, |
|
"grad_norm": 0.5275786368006364, |
|
"learning_rate": 4.5468277945619336e-06, |
|
"loss": 0.5483, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.3169642857142857, |
|
"grad_norm": 0.5057705930103896, |
|
"learning_rate": 4.539274924471299e-06, |
|
"loss": 0.5236, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.32142857142857145, |
|
"grad_norm": 0.49625696780589473, |
|
"learning_rate": 4.531722054380665e-06, |
|
"loss": 0.5221, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.32589285714285715, |
|
"grad_norm": 0.49548627058675154, |
|
"learning_rate": 4.524169184290031e-06, |
|
"loss": 0.5176, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.33035714285714285, |
|
"grad_norm": 0.5011897814029462, |
|
"learning_rate": 4.516616314199396e-06, |
|
"loss": 0.5431, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.33482142857142855, |
|
"grad_norm": 0.5168007116628185, |
|
"learning_rate": 4.509063444108761e-06, |
|
"loss": 0.5289, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.3392857142857143, |
|
"grad_norm": 0.5199488476439844, |
|
"learning_rate": 4.501510574018128e-06, |
|
"loss": 0.5437, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.34375, |
|
"grad_norm": 0.5776348374651958, |
|
"learning_rate": 4.493957703927493e-06, |
|
"loss": 0.5424, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.3482142857142857, |
|
"grad_norm": 0.5785274375251012, |
|
"learning_rate": 4.486404833836858e-06, |
|
"loss": 0.5252, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.35267857142857145, |
|
"grad_norm": 0.5624252036929408, |
|
"learning_rate": 4.478851963746224e-06, |
|
"loss": 0.5502, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.35714285714285715, |
|
"grad_norm": 0.5630273292563954, |
|
"learning_rate": 4.47129909365559e-06, |
|
"loss": 0.5249, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.36160714285714285, |
|
"grad_norm": 0.5229643469397499, |
|
"learning_rate": 4.463746223564955e-06, |
|
"loss": 0.572, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.36607142857142855, |
|
"grad_norm": 0.5401109169653766, |
|
"learning_rate": 4.45619335347432e-06, |
|
"loss": 0.5174, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.3705357142857143, |
|
"grad_norm": 0.572853251261138, |
|
"learning_rate": 4.448640483383686e-06, |
|
"loss": 0.5336, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.375, |
|
"grad_norm": 0.49833694539628726, |
|
"learning_rate": 4.441087613293052e-06, |
|
"loss": 0.5205, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.3794642857142857, |
|
"grad_norm": 0.5167418841184229, |
|
"learning_rate": 4.4335347432024175e-06, |
|
"loss": 0.5405, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.38392857142857145, |
|
"grad_norm": 0.5243612918158316, |
|
"learning_rate": 4.425981873111782e-06, |
|
"loss": 0.5204, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.38839285714285715, |
|
"grad_norm": 0.5159624173670619, |
|
"learning_rate": 4.418429003021149e-06, |
|
"loss": 0.5315, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.39285714285714285, |
|
"grad_norm": 0.5227918139330733, |
|
"learning_rate": 4.410876132930514e-06, |
|
"loss": 0.5418, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.39732142857142855, |
|
"grad_norm": 0.48045718180033986, |
|
"learning_rate": 4.4033232628398795e-06, |
|
"loss": 0.506, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.4017857142857143, |
|
"grad_norm": 0.5791314708394477, |
|
"learning_rate": 4.395770392749245e-06, |
|
"loss": 0.596, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.40625, |
|
"grad_norm": 0.6076262289549809, |
|
"learning_rate": 4.38821752265861e-06, |
|
"loss": 0.5312, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.4107142857142857, |
|
"grad_norm": 0.4780377894473343, |
|
"learning_rate": 4.380664652567977e-06, |
|
"loss": 0.549, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.41517857142857145, |
|
"grad_norm": 0.5549693296410254, |
|
"learning_rate": 4.3731117824773415e-06, |
|
"loss": 0.5159, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.41964285714285715, |
|
"grad_norm": 0.5097636365128929, |
|
"learning_rate": 4.365558912386707e-06, |
|
"loss": 0.5612, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.42410714285714285, |
|
"grad_norm": 0.5021825140627123, |
|
"learning_rate": 4.358006042296073e-06, |
|
"loss": 0.5653, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.42857142857142855, |
|
"grad_norm": 0.537945728158161, |
|
"learning_rate": 4.350453172205439e-06, |
|
"loss": 0.528, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.4330357142857143, |
|
"grad_norm": 0.4846132130879552, |
|
"learning_rate": 4.342900302114804e-06, |
|
"loss": 0.5395, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.4375, |
|
"grad_norm": 0.5206116049646974, |
|
"learning_rate": 4.335347432024169e-06, |
|
"loss": 0.5445, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.4419642857142857, |
|
"grad_norm": 0.551262765339688, |
|
"learning_rate": 4.327794561933535e-06, |
|
"loss": 0.5346, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.44642857142857145, |
|
"grad_norm": 0.4984265695742978, |
|
"learning_rate": 4.3202416918429006e-06, |
|
"loss": 0.5161, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.45089285714285715, |
|
"grad_norm": 0.49943465819192145, |
|
"learning_rate": 4.312688821752266e-06, |
|
"loss": 0.5164, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.45535714285714285, |
|
"grad_norm": 0.48826983665318036, |
|
"learning_rate": 4.305135951661632e-06, |
|
"loss": 0.552, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.45982142857142855, |
|
"grad_norm": 0.5320671379410685, |
|
"learning_rate": 4.297583081570998e-06, |
|
"loss": 0.5512, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.4642857142857143, |
|
"grad_norm": 0.5027439491810191, |
|
"learning_rate": 4.2900302114803626e-06, |
|
"loss": 0.5728, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.46875, |
|
"grad_norm": 0.4875858103308158, |
|
"learning_rate": 4.282477341389728e-06, |
|
"loss": 0.5306, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.4732142857142857, |
|
"grad_norm": 0.5205285237402679, |
|
"learning_rate": 4.274924471299094e-06, |
|
"loss": 0.5274, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.47767857142857145, |
|
"grad_norm": 0.47059444782816534, |
|
"learning_rate": 4.26737160120846e-06, |
|
"loss": 0.5092, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.48214285714285715, |
|
"grad_norm": 0.4998056822147382, |
|
"learning_rate": 4.259818731117825e-06, |
|
"loss": 0.5481, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.48660714285714285, |
|
"grad_norm": 0.5403989848602658, |
|
"learning_rate": 4.25226586102719e-06, |
|
"loss": 0.518, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.49107142857142855, |
|
"grad_norm": 0.48519680312136604, |
|
"learning_rate": 4.244712990936557e-06, |
|
"loss": 0.5522, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.4955357142857143, |
|
"grad_norm": 0.5655590239975465, |
|
"learning_rate": 4.237160120845922e-06, |
|
"loss": 0.526, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.49558151987556875, |
|
"learning_rate": 4.229607250755287e-06, |
|
"loss": 0.5275, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 0.5346882343292236, |
|
"eval_runtime": 32.2399, |
|
"eval_samples_per_second": 82.817, |
|
"eval_steps_per_second": 5.18, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.5050279329608939, |
|
"grad_norm": 0.5662939933016762, |
|
"learning_rate": 4.218512898330804e-06, |
|
"loss": 0.515, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.5094972067039106, |
|
"grad_norm": 0.5225180760232837, |
|
"learning_rate": 4.2109256449165405e-06, |
|
"loss": 0.5038, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.5139664804469274, |
|
"grad_norm": 0.5377410752157915, |
|
"learning_rate": 4.203338391502276e-06, |
|
"loss": 0.5043, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.5184357541899441, |
|
"grad_norm": 0.5915094303426857, |
|
"learning_rate": 4.195751138088012e-06, |
|
"loss": 0.5115, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.5229050279329609, |
|
"grad_norm": 0.5079762717076601, |
|
"learning_rate": 4.1881638846737485e-06, |
|
"loss": 0.5491, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.5273743016759777, |
|
"grad_norm": 0.5933329970849119, |
|
"learning_rate": 4.180576631259484e-06, |
|
"loss": 0.5095, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.5318435754189944, |
|
"grad_norm": 0.5053074762991974, |
|
"learning_rate": 4.17298937784522e-06, |
|
"loss": 0.4881, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.5363128491620112, |
|
"grad_norm": 0.6515342691068152, |
|
"learning_rate": 4.1654021244309564e-06, |
|
"loss": 0.5069, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5407821229050279, |
|
"grad_norm": 0.5774345075195761, |
|
"learning_rate": 4.157814871016692e-06, |
|
"loss": 0.5007, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.5452513966480447, |
|
"grad_norm": 0.6308991657157696, |
|
"learning_rate": 4.150227617602428e-06, |
|
"loss": 0.5083, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.5497206703910614, |
|
"grad_norm": 0.5258198371155504, |
|
"learning_rate": 4.142640364188164e-06, |
|
"loss": 0.5103, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.5541899441340782, |
|
"grad_norm": 0.5629243370057063, |
|
"learning_rate": 4.1350531107739e-06, |
|
"loss": 0.5016, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.5586592178770949, |
|
"grad_norm": 0.5440095922230602, |
|
"learning_rate": 4.127465857359636e-06, |
|
"loss": 0.491, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.5631284916201117, |
|
"grad_norm": 0.5126793634050918, |
|
"learning_rate": 4.119878603945372e-06, |
|
"loss": 0.5239, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.5675977653631284, |
|
"grad_norm": 0.5476299968831635, |
|
"learning_rate": 4.112291350531108e-06, |
|
"loss": 0.5176, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.5720670391061452, |
|
"grad_norm": 0.5570481457317527, |
|
"learning_rate": 4.104704097116844e-06, |
|
"loss": 0.5084, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.576536312849162, |
|
"grad_norm": 0.49480777896006917, |
|
"learning_rate": 4.09711684370258e-06, |
|
"loss": 0.4966, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.5810055865921788, |
|
"grad_norm": 0.5089889598386482, |
|
"learning_rate": 4.089529590288316e-06, |
|
"loss": 0.5059, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5854748603351956, |
|
"grad_norm": 0.4954780811768282, |
|
"learning_rate": 4.081942336874052e-06, |
|
"loss": 0.5166, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.5899441340782123, |
|
"grad_norm": 0.4809684003541764, |
|
"learning_rate": 4.074355083459787e-06, |
|
"loss": 0.4857, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.5944134078212291, |
|
"grad_norm": 0.5009748980192239, |
|
"learning_rate": 4.066767830045524e-06, |
|
"loss": 0.523, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.5988826815642458, |
|
"grad_norm": 0.48320950023108783, |
|
"learning_rate": 4.05918057663126e-06, |
|
"loss": 0.4956, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.6033519553072626, |
|
"grad_norm": 0.49399181892388616, |
|
"learning_rate": 4.051593323216995e-06, |
|
"loss": 0.5163, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.6078212290502794, |
|
"grad_norm": 0.4821527638840584, |
|
"learning_rate": 4.044006069802732e-06, |
|
"loss": 0.4981, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.6122905027932961, |
|
"grad_norm": 0.5099314085164022, |
|
"learning_rate": 4.036418816388468e-06, |
|
"loss": 0.5121, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.6167597765363129, |
|
"grad_norm": 0.5323495276796993, |
|
"learning_rate": 4.028831562974203e-06, |
|
"loss": 0.5074, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.6212290502793296, |
|
"grad_norm": 0.48851533865778524, |
|
"learning_rate": 4.02124430955994e-06, |
|
"loss": 0.52, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.6256983240223464, |
|
"grad_norm": 0.5281525574471102, |
|
"learning_rate": 4.0136570561456756e-06, |
|
"loss": 0.5039, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6301675977653631, |
|
"grad_norm": 0.5279406819094602, |
|
"learning_rate": 4.006069802731411e-06, |
|
"loss": 0.4933, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.6346368715083799, |
|
"grad_norm": 0.45822529010842367, |
|
"learning_rate": 3.998482549317148e-06, |
|
"loss": 0.5151, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.6391061452513966, |
|
"grad_norm": 0.5349296018392958, |
|
"learning_rate": 3.9908952959028835e-06, |
|
"loss": 0.4983, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.6435754189944134, |
|
"grad_norm": 0.5316574655817645, |
|
"learning_rate": 3.983308042488619e-06, |
|
"loss": 0.5344, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.6480446927374302, |
|
"grad_norm": 0.5107562456487705, |
|
"learning_rate": 3.975720789074356e-06, |
|
"loss": 0.52, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.6525139664804469, |
|
"grad_norm": 0.5072583855988344, |
|
"learning_rate": 3.9681335356600915e-06, |
|
"loss": 0.4923, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.6569832402234637, |
|
"grad_norm": 0.5836000010606872, |
|
"learning_rate": 3.960546282245827e-06, |
|
"loss": 0.5295, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.6614525139664804, |
|
"grad_norm": 0.49884721978842206, |
|
"learning_rate": 3.952959028831564e-06, |
|
"loss": 0.5216, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.6659217877094972, |
|
"grad_norm": 0.5120991222561042, |
|
"learning_rate": 3.945371775417299e-06, |
|
"loss": 0.5414, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.6703910614525139, |
|
"grad_norm": 0.49838481084559894, |
|
"learning_rate": 3.937784522003035e-06, |
|
"loss": 0.5381, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.6748603351955307, |
|
"grad_norm": 0.5398799415397277, |
|
"learning_rate": 3.930197268588772e-06, |
|
"loss": 0.4896, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.6793296089385474, |
|
"grad_norm": 0.565748454568293, |
|
"learning_rate": 3.922610015174507e-06, |
|
"loss": 0.4985, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.6837988826815642, |
|
"grad_norm": 0.5986782671726579, |
|
"learning_rate": 3.915022761760243e-06, |
|
"loss": 0.5125, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.6882681564245811, |
|
"grad_norm": 0.5290750194980306, |
|
"learning_rate": 3.907435508345979e-06, |
|
"loss": 0.5078, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.6927374301675978, |
|
"grad_norm": 0.5706479438251948, |
|
"learning_rate": 3.899848254931715e-06, |
|
"loss": 0.5205, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.6972067039106146, |
|
"grad_norm": 0.5173864500214489, |
|
"learning_rate": 3.892261001517451e-06, |
|
"loss": 0.4988, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.7016759776536313, |
|
"grad_norm": 0.4920045809108581, |
|
"learning_rate": 3.884673748103187e-06, |
|
"loss": 0.4954, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.7061452513966481, |
|
"grad_norm": 0.5237901072069291, |
|
"learning_rate": 3.877086494688923e-06, |
|
"loss": 0.5253, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.7106145251396648, |
|
"grad_norm": 0.5212795514388354, |
|
"learning_rate": 3.869499241274659e-06, |
|
"loss": 0.5029, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.7150837988826816, |
|
"grad_norm": 0.48658569533086143, |
|
"learning_rate": 3.861911987860395e-06, |
|
"loss": 0.5023, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.7195530726256983, |
|
"grad_norm": 0.5145837815914728, |
|
"learning_rate": 3.854324734446131e-06, |
|
"loss": 0.5436, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.7240223463687151, |
|
"grad_norm": 0.49182827281373437, |
|
"learning_rate": 3.846737481031867e-06, |
|
"loss": 0.4948, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.7284916201117319, |
|
"grad_norm": 0.5113640333759738, |
|
"learning_rate": 3.839150227617603e-06, |
|
"loss": 0.4778, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.7329608938547486, |
|
"grad_norm": 0.47675402744999507, |
|
"learning_rate": 3.831562974203339e-06, |
|
"loss": 0.5182, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.7374301675977654, |
|
"grad_norm": 0.5214515656344721, |
|
"learning_rate": 3.823975720789075e-06, |
|
"loss": 0.5204, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.7418994413407821, |
|
"grad_norm": 0.5130606167326404, |
|
"learning_rate": 3.816388467374811e-06, |
|
"loss": 0.492, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.7463687150837989, |
|
"grad_norm": 0.49678736596959705, |
|
"learning_rate": 3.8088012139605467e-06, |
|
"loss": 0.4997, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.7508379888268156, |
|
"grad_norm": 0.5275415964748987, |
|
"learning_rate": 3.801213960546283e-06, |
|
"loss": 0.5163, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.7553072625698324, |
|
"grad_norm": 0.5323621998896121, |
|
"learning_rate": 3.7936267071320185e-06, |
|
"loss": 0.4846, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.7597765363128491, |
|
"grad_norm": 0.5302675218909635, |
|
"learning_rate": 3.7860394537177547e-06, |
|
"loss": 0.5193, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.7642458100558659, |
|
"grad_norm": 0.5097255453283106, |
|
"learning_rate": 3.778452200303491e-06, |
|
"loss": 0.5248, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.7687150837988826, |
|
"grad_norm": 0.5221084637587756, |
|
"learning_rate": 3.7708649468892265e-06, |
|
"loss": 0.5217, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.7731843575418994, |
|
"grad_norm": 0.5823307367451829, |
|
"learning_rate": 3.7632776934749626e-06, |
|
"loss": 0.5057, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.7776536312849162, |
|
"grad_norm": 0.48586162675121547, |
|
"learning_rate": 3.7556904400606987e-06, |
|
"loss": 0.5372, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.7821229050279329, |
|
"grad_norm": 0.5099958749272927, |
|
"learning_rate": 3.748103186646434e-06, |
|
"loss": 0.4844, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.7865921787709497, |
|
"grad_norm": 0.5221594966770278, |
|
"learning_rate": 3.74051593323217e-06, |
|
"loss": 0.4968, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.7910614525139665, |
|
"grad_norm": 0.5392129841421105, |
|
"learning_rate": 3.732928679817906e-06, |
|
"loss": 0.4906, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.7955307262569833, |
|
"grad_norm": 0.5239629013619267, |
|
"learning_rate": 3.725341426403642e-06, |
|
"loss": 0.5172, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.537861417175095, |
|
"learning_rate": 3.717754172989378e-06, |
|
"loss": 0.4914, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.8044692737430168, |
|
"grad_norm": 0.5001204419984062, |
|
"learning_rate": 3.710166919575114e-06, |
|
"loss": 0.5014, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.8089385474860336, |
|
"grad_norm": 0.5243758405421232, |
|
"learning_rate": 3.70257966616085e-06, |
|
"loss": 0.4979, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.8134078212290503, |
|
"grad_norm": 0.507560920806033, |
|
"learning_rate": 3.6949924127465856e-06, |
|
"loss": 0.5003, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.8178770949720671, |
|
"grad_norm": 0.5623627396713589, |
|
"learning_rate": 3.6874051593323218e-06, |
|
"loss": 0.5011, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.8223463687150838, |
|
"grad_norm": 0.5225320984203384, |
|
"learning_rate": 3.679817905918058e-06, |
|
"loss": 0.5117, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.8268156424581006, |
|
"grad_norm": 0.5371208831947621, |
|
"learning_rate": 3.6722306525037936e-06, |
|
"loss": 0.5028, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.8312849162011173, |
|
"grad_norm": 0.5412513689241548, |
|
"learning_rate": 3.6646433990895297e-06, |
|
"loss": 0.5111, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.8357541899441341, |
|
"grad_norm": 0.49083593539874787, |
|
"learning_rate": 3.657056145675266e-06, |
|
"loss": 0.5002, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.8402234636871508, |
|
"grad_norm": 0.5144646076400369, |
|
"learning_rate": 3.6494688922610015e-06, |
|
"loss": 0.4875, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.8446927374301676, |
|
"grad_norm": 0.5385458846653849, |
|
"learning_rate": 3.6418816388467377e-06, |
|
"loss": 0.5072, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.8491620111731844, |
|
"grad_norm": 0.48088002744673064, |
|
"learning_rate": 3.6342943854324738e-06, |
|
"loss": 0.549, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.8536312849162011, |
|
"grad_norm": 0.563960197772806, |
|
"learning_rate": 3.6267071320182095e-06, |
|
"loss": 0.5344, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.8581005586592179, |
|
"grad_norm": 0.5235289059912077, |
|
"learning_rate": 3.6191198786039456e-06, |
|
"loss": 0.5172, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.8625698324022346, |
|
"grad_norm": 0.5059904029741168, |
|
"learning_rate": 3.6115326251896813e-06, |
|
"loss": 0.4908, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.8670391061452514, |
|
"grad_norm": 0.507980780742306, |
|
"learning_rate": 3.6039453717754174e-06, |
|
"loss": 0.4951, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.8715083798882681, |
|
"grad_norm": 0.5552145890763519, |
|
"learning_rate": 3.5963581183611536e-06, |
|
"loss": 0.505, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.8759776536312849, |
|
"grad_norm": 0.5379367693029108, |
|
"learning_rate": 3.5887708649468893e-06, |
|
"loss": 0.52, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.8804469273743016, |
|
"grad_norm": 0.5813281708032492, |
|
"learning_rate": 3.5811836115326254e-06, |
|
"loss": 0.4966, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.8849162011173184, |
|
"grad_norm": 0.5138017935214574, |
|
"learning_rate": 3.5735963581183615e-06, |
|
"loss": 0.495, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.8893854748603351, |
|
"grad_norm": 0.5116405158423362, |
|
"learning_rate": 3.5660091047040972e-06, |
|
"loss": 0.519, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.8938547486033519, |
|
"grad_norm": 0.5696283061922035, |
|
"learning_rate": 3.5584218512898333e-06, |
|
"loss": 0.5002, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.8983240223463688, |
|
"grad_norm": 0.5541549121266046, |
|
"learning_rate": 3.5508345978755695e-06, |
|
"loss": 0.5125, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.9027932960893855, |
|
"grad_norm": 0.5017863683272789, |
|
"learning_rate": 3.543247344461305e-06, |
|
"loss": 0.499, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.9072625698324023, |
|
"grad_norm": 0.5776107100347877, |
|
"learning_rate": 3.5356600910470413e-06, |
|
"loss": 0.4883, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.911731843575419, |
|
"grad_norm": 0.5350847562000018, |
|
"learning_rate": 3.528072837632777e-06, |
|
"loss": 0.5052, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.9162011173184358, |
|
"grad_norm": 0.5012614076658582, |
|
"learning_rate": 3.520485584218513e-06, |
|
"loss": 0.5422, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.9206703910614525, |
|
"grad_norm": 0.6107847324258918, |
|
"learning_rate": 3.5128983308042493e-06, |
|
"loss": 0.5204, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.9251396648044693, |
|
"grad_norm": 0.49314942553653257, |
|
"learning_rate": 3.505311077389985e-06, |
|
"loss": 0.5074, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.929608938547486, |
|
"grad_norm": 0.5340833706807174, |
|
"learning_rate": 3.497723823975721e-06, |
|
"loss": 0.4936, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.9340782122905028, |
|
"grad_norm": 0.5342137283430334, |
|
"learning_rate": 3.490136570561457e-06, |
|
"loss": 0.4945, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.9385474860335196, |
|
"grad_norm": 0.5301769352005536, |
|
"learning_rate": 3.482549317147193e-06, |
|
"loss": 0.4963, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.9430167597765363, |
|
"grad_norm": 0.5455545680194068, |
|
"learning_rate": 3.474962063732929e-06, |
|
"loss": 0.4931, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.9474860335195531, |
|
"grad_norm": 0.5282021847913991, |
|
"learning_rate": 3.467374810318665e-06, |
|
"loss": 0.4894, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.9519553072625698, |
|
"grad_norm": 0.5341400870442735, |
|
"learning_rate": 3.459787556904401e-06, |
|
"loss": 0.5102, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.9564245810055866, |
|
"grad_norm": 0.5313633005763995, |
|
"learning_rate": 3.452200303490137e-06, |
|
"loss": 0.501, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.9608938547486033, |
|
"grad_norm": 0.5774555510069799, |
|
"learning_rate": 3.4446130500758727e-06, |
|
"loss": 0.5021, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.9653631284916201, |
|
"grad_norm": 0.521646351973494, |
|
"learning_rate": 3.437025796661609e-06, |
|
"loss": 0.5071, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.9698324022346368, |
|
"grad_norm": 0.5335773786656196, |
|
"learning_rate": 3.429438543247345e-06, |
|
"loss": 0.4993, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.9743016759776536, |
|
"grad_norm": 0.5774515212071393, |
|
"learning_rate": 3.4218512898330806e-06, |
|
"loss": 0.4802, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.9787709497206704, |
|
"grad_norm": 0.5159511449041201, |
|
"learning_rate": 3.4142640364188168e-06, |
|
"loss": 0.483, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.9832402234636871, |
|
"grad_norm": 0.509916149384416, |
|
"learning_rate": 3.406676783004553e-06, |
|
"loss": 0.4774, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.9877094972067039, |
|
"grad_norm": 0.5226659969898425, |
|
"learning_rate": 3.3990895295902886e-06, |
|
"loss": 0.516, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.9921787709497206, |
|
"grad_norm": 0.5632866199425641, |
|
"learning_rate": 3.3915022761760247e-06, |
|
"loss": 0.5346, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.9966480446927374, |
|
"grad_norm": 0.5224220899066443, |
|
"learning_rate": 3.383915022761761e-06, |
|
"loss": 0.5441, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.0011173184357542, |
|
"grad_norm": 0.5965341004746808, |
|
"learning_rate": 3.3763277693474965e-06, |
|
"loss": 0.512, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.0011173184357542, |
|
"eval_loss": 0.5086758732795715, |
|
"eval_runtime": 32.322, |
|
"eval_samples_per_second": 82.606, |
|
"eval_steps_per_second": 5.167, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.005586592178771, |
|
"grad_norm": 0.5234315419190534, |
|
"learning_rate": 3.3687405159332327e-06, |
|
"loss": 0.5054, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.0100558659217878, |
|
"grad_norm": 0.5575750854403085, |
|
"learning_rate": 3.3611532625189684e-06, |
|
"loss": 0.4878, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.0145251396648045, |
|
"grad_norm": 0.5253448160628447, |
|
"learning_rate": 3.3535660091047045e-06, |
|
"loss": 0.5054, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.0189944134078213, |
|
"grad_norm": 0.49020025660118194, |
|
"learning_rate": 3.3459787556904406e-06, |
|
"loss": 0.4818, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.0044692737430168, |
|
"grad_norm": 0.6277138162928673, |
|
"learning_rate": 3.3383915022761763e-06, |
|
"loss": 0.5008, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.0089385474860335, |
|
"grad_norm": 0.5276575275655174, |
|
"learning_rate": 3.3308042488619125e-06, |
|
"loss": 0.4822, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.0134078212290503, |
|
"grad_norm": 0.516729892988387, |
|
"learning_rate": 3.3232169954476486e-06, |
|
"loss": 0.4783, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.017877094972067, |
|
"grad_norm": 0.5648907712973277, |
|
"learning_rate": 3.3156297420333843e-06, |
|
"loss": 0.4927, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.0223463687150838, |
|
"grad_norm": 0.5521221146190025, |
|
"learning_rate": 3.3080424886191204e-06, |
|
"loss": 0.4742, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.0268156424581005, |
|
"grad_norm": 0.5288379239842262, |
|
"learning_rate": 3.3004552352048565e-06, |
|
"loss": 0.4863, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.0312849162011173, |
|
"grad_norm": 0.4932863276651226, |
|
"learning_rate": 3.2928679817905922e-06, |
|
"loss": 0.4613, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.035754189944134, |
|
"grad_norm": 0.5069055178884726, |
|
"learning_rate": 3.2852807283763284e-06, |
|
"loss": 0.4969, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.0402234636871508, |
|
"grad_norm": 0.49009866664123986, |
|
"learning_rate": 3.277693474962064e-06, |
|
"loss": 0.4759, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.0446927374301676, |
|
"grad_norm": 0.5220649844244807, |
|
"learning_rate": 3.2701062215478e-06, |
|
"loss": 0.4654, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.0491620111731843, |
|
"grad_norm": 0.5354627141925222, |
|
"learning_rate": 3.2625189681335363e-06, |
|
"loss": 0.5009, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.053631284916201, |
|
"grad_norm": 0.5561818122893151, |
|
"learning_rate": 3.254931714719272e-06, |
|
"loss": 0.4564, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.0581005586592178, |
|
"grad_norm": 0.48719455414113416, |
|
"learning_rate": 3.2473444613050077e-06, |
|
"loss": 0.4924, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.0625698324022346, |
|
"grad_norm": 0.5527905454518657, |
|
"learning_rate": 3.2397572078907434e-06, |
|
"loss": 0.4856, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.0670391061452513, |
|
"grad_norm": 0.5381863443762583, |
|
"learning_rate": 3.2321699544764795e-06, |
|
"loss": 0.4935, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.071508379888268, |
|
"grad_norm": 0.5355360844432822, |
|
"learning_rate": 3.2245827010622157e-06, |
|
"loss": 0.4858, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.0759776536312848, |
|
"grad_norm": 0.5191892602172652, |
|
"learning_rate": 3.2169954476479514e-06, |
|
"loss": 0.4819, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.0804469273743016, |
|
"grad_norm": 0.4952672717234948, |
|
"learning_rate": 3.2094081942336875e-06, |
|
"loss": 0.5007, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.0849162011173183, |
|
"grad_norm": 0.5331734326280092, |
|
"learning_rate": 3.201820940819423e-06, |
|
"loss": 0.4874, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.089385474860335, |
|
"grad_norm": 0.49333273079420603, |
|
"learning_rate": 3.1942336874051593e-06, |
|
"loss": 0.5045, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.0938547486033519, |
|
"grad_norm": 0.5143326582021585, |
|
"learning_rate": 3.1866464339908955e-06, |
|
"loss": 0.5047, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.0983240223463686, |
|
"grad_norm": 0.5027227799706213, |
|
"learning_rate": 3.179059180576631e-06, |
|
"loss": 0.4803, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.1027932960893856, |
|
"grad_norm": 0.5257359502707071, |
|
"learning_rate": 3.1714719271623673e-06, |
|
"loss": 0.4654, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 1.1072625698324021, |
|
"grad_norm": 0.5244897759431714, |
|
"learning_rate": 3.1638846737481034e-06, |
|
"loss": 0.5329, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.111731843575419, |
|
"grad_norm": 0.5028531015372357, |
|
"learning_rate": 3.156297420333839e-06, |
|
"loss": 0.4721, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 1.1162011173184359, |
|
"grad_norm": 0.5342121642252061, |
|
"learning_rate": 3.1487101669195752e-06, |
|
"loss": 0.4962, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 1.1206703910614526, |
|
"grad_norm": 0.49550125400926287, |
|
"learning_rate": 3.1411229135053114e-06, |
|
"loss": 0.4884, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 1.1251396648044694, |
|
"grad_norm": 0.5724921757672813, |
|
"learning_rate": 3.133535660091047e-06, |
|
"loss": 0.4676, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 1.1296089385474861, |
|
"grad_norm": 0.5012184887171401, |
|
"learning_rate": 3.125948406676783e-06, |
|
"loss": 0.48, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 1.1340782122905029, |
|
"grad_norm": 0.5049054849116781, |
|
"learning_rate": 3.118361153262519e-06, |
|
"loss": 0.494, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 1.1385474860335196, |
|
"grad_norm": 0.5291980859522503, |
|
"learning_rate": 3.110773899848255e-06, |
|
"loss": 0.4787, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 1.1430167597765364, |
|
"grad_norm": 0.5235318263865767, |
|
"learning_rate": 3.103186646433991e-06, |
|
"loss": 0.4745, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.1474860335195531, |
|
"grad_norm": 0.5067236081675356, |
|
"learning_rate": 3.095599393019727e-06, |
|
"loss": 0.5863, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 1.15195530726257, |
|
"grad_norm": 0.48763909341918976, |
|
"learning_rate": 3.088012139605463e-06, |
|
"loss": 0.5264, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 1.1564245810055866, |
|
"grad_norm": 0.5198110315693464, |
|
"learning_rate": 3.080424886191199e-06, |
|
"loss": 0.4716, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 1.1608938547486034, |
|
"grad_norm": 0.5093655208180958, |
|
"learning_rate": 3.072837632776935e-06, |
|
"loss": 0.4768, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 1.1653631284916202, |
|
"grad_norm": 0.49489186143964614, |
|
"learning_rate": 3.065250379362671e-06, |
|
"loss": 0.4765, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 1.169832402234637, |
|
"grad_norm": 0.5209973994944476, |
|
"learning_rate": 3.057663125948407e-06, |
|
"loss": 0.4869, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 1.1743016759776537, |
|
"grad_norm": 0.4873878645874658, |
|
"learning_rate": 3.0500758725341427e-06, |
|
"loss": 0.4623, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 1.1787709497206704, |
|
"grad_norm": 0.5074910158328607, |
|
"learning_rate": 3.042488619119879e-06, |
|
"loss": 0.4765, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 1.1832402234636872, |
|
"grad_norm": 0.5195087180758534, |
|
"learning_rate": 3.0349013657056146e-06, |
|
"loss": 0.4838, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 1.187709497206704, |
|
"grad_norm": 0.5264228931092361, |
|
"learning_rate": 3.0273141122913507e-06, |
|
"loss": 0.4638, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.1921787709497207, |
|
"grad_norm": 0.4919380409463276, |
|
"learning_rate": 3.019726858877087e-06, |
|
"loss": 0.4981, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 1.1966480446927374, |
|
"grad_norm": 0.5060903656555795, |
|
"learning_rate": 3.0121396054628225e-06, |
|
"loss": 0.477, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 1.2011173184357542, |
|
"grad_norm": 0.4938697284936987, |
|
"learning_rate": 3.0045523520485587e-06, |
|
"loss": 0.5003, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 1.205586592178771, |
|
"grad_norm": 0.5599516302969035, |
|
"learning_rate": 2.9969650986342948e-06, |
|
"loss": 0.4673, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 1.2100558659217877, |
|
"grad_norm": 0.5126307823030745, |
|
"learning_rate": 2.9893778452200305e-06, |
|
"loss": 0.4971, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.2145251396648045, |
|
"grad_norm": 0.5927333276589611, |
|
"learning_rate": 2.9817905918057666e-06, |
|
"loss": 0.5006, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 1.2189944134078212, |
|
"grad_norm": 0.5367579038483986, |
|
"learning_rate": 2.9742033383915027e-06, |
|
"loss": 0.5143, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 1.223463687150838, |
|
"grad_norm": 0.5179699664039845, |
|
"learning_rate": 2.9666160849772384e-06, |
|
"loss": 0.4733, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 1.2279329608938547, |
|
"grad_norm": 0.5015385250607317, |
|
"learning_rate": 2.9590288315629746e-06, |
|
"loss": 0.4786, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 1.2324022346368715, |
|
"grad_norm": 0.4925764923779409, |
|
"learning_rate": 2.9514415781487103e-06, |
|
"loss": 0.472, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.2368715083798882, |
|
"grad_norm": 0.5522056674324646, |
|
"learning_rate": 2.9438543247344464e-06, |
|
"loss": 0.4675, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 1.241340782122905, |
|
"grad_norm": 0.5128909667557014, |
|
"learning_rate": 2.9362670713201825e-06, |
|
"loss": 0.4778, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 1.2458100558659218, |
|
"grad_norm": 0.49958051066954534, |
|
"learning_rate": 2.9286798179059182e-06, |
|
"loss": 0.4744, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 1.2502793296089385, |
|
"grad_norm": 0.5186801983796817, |
|
"learning_rate": 2.9210925644916543e-06, |
|
"loss": 0.47, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 1.2547486033519553, |
|
"grad_norm": 0.49479081933964797, |
|
"learning_rate": 2.9135053110773905e-06, |
|
"loss": 0.4693, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 1.259217877094972, |
|
"grad_norm": 0.5172924586559385, |
|
"learning_rate": 2.905918057663126e-06, |
|
"loss": 0.5034, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 1.2636871508379888, |
|
"grad_norm": 0.49588955761543196, |
|
"learning_rate": 2.8983308042488623e-06, |
|
"loss": 0.4857, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 1.2681564245810055, |
|
"grad_norm": 0.5651670438316321, |
|
"learning_rate": 2.8907435508345984e-06, |
|
"loss": 0.4845, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 1.2726256983240223, |
|
"grad_norm": 0.47720259500943096, |
|
"learning_rate": 2.883156297420334e-06, |
|
"loss": 0.4482, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 1.277094972067039, |
|
"grad_norm": 0.5080092332022862, |
|
"learning_rate": 2.8755690440060702e-06, |
|
"loss": 0.4826, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.2815642458100558, |
|
"grad_norm": 0.6016729996144168, |
|
"learning_rate": 2.867981790591806e-06, |
|
"loss": 0.5592, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 1.2860335195530725, |
|
"grad_norm": 0.5456040494522852, |
|
"learning_rate": 2.860394537177542e-06, |
|
"loss": 0.4821, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 1.2905027932960893, |
|
"grad_norm": 0.5858331665900632, |
|
"learning_rate": 2.852807283763278e-06, |
|
"loss": 0.5089, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 1.2949720670391063, |
|
"grad_norm": 0.5895630636847224, |
|
"learning_rate": 2.845220030349014e-06, |
|
"loss": 0.5101, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 1.2994413407821228, |
|
"grad_norm": 0.505172328049275, |
|
"learning_rate": 2.83763277693475e-06, |
|
"loss": 0.5008, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 1.3039106145251398, |
|
"grad_norm": 0.5035894899292893, |
|
"learning_rate": 2.830045523520486e-06, |
|
"loss": 0.4812, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 1.3083798882681563, |
|
"grad_norm": 0.5304264887862729, |
|
"learning_rate": 2.822458270106222e-06, |
|
"loss": 0.4579, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 1.3128491620111733, |
|
"grad_norm": 0.4742288658363735, |
|
"learning_rate": 2.814871016691958e-06, |
|
"loss": 0.4801, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 1.3173184357541898, |
|
"grad_norm": 0.5293412371049411, |
|
"learning_rate": 2.807283763277694e-06, |
|
"loss": 0.4602, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 1.3217877094972068, |
|
"grad_norm": 0.5333831439310788, |
|
"learning_rate": 2.79969650986343e-06, |
|
"loss": 0.4963, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.3262569832402233, |
|
"grad_norm": 0.4947844594727205, |
|
"learning_rate": 2.792109256449166e-06, |
|
"loss": 0.4781, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 1.3307262569832403, |
|
"grad_norm": 0.5323122787674339, |
|
"learning_rate": 2.7845220030349016e-06, |
|
"loss": 0.4998, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 1.3351955307262569, |
|
"grad_norm": 0.48615316005162806, |
|
"learning_rate": 2.7769347496206378e-06, |
|
"loss": 0.466, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 1.3396648044692738, |
|
"grad_norm": 0.5038897355673981, |
|
"learning_rate": 2.769347496206374e-06, |
|
"loss": 0.531, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 1.3441340782122906, |
|
"grad_norm": 0.47725378544672703, |
|
"learning_rate": 2.7617602427921096e-06, |
|
"loss": 0.4917, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 1.3486033519553073, |
|
"grad_norm": 0.5241579403675019, |
|
"learning_rate": 2.7541729893778457e-06, |
|
"loss": 0.4673, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 1.353072625698324, |
|
"grad_norm": 0.5252091486066182, |
|
"learning_rate": 2.746585735963581e-06, |
|
"loss": 0.467, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 1.3575418994413408, |
|
"grad_norm": 0.4803361684555166, |
|
"learning_rate": 2.738998482549317e-06, |
|
"loss": 0.5257, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 1.3620111731843576, |
|
"grad_norm": 0.5300934772683465, |
|
"learning_rate": 2.7314112291350532e-06, |
|
"loss": 0.4889, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.3664804469273744, |
|
"grad_norm": 0.5001374114396476, |
|
"learning_rate": 2.723823975720789e-06, |
|
"loss": 0.4847, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.3709497206703911, |
|
"grad_norm": 0.5240875065710657, |
|
"learning_rate": 2.716236722306525e-06, |
|
"loss": 0.4578, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 1.3754189944134079, |
|
"grad_norm": 0.4714761354385937, |
|
"learning_rate": 2.708649468892261e-06, |
|
"loss": 0.4947, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.3798882681564246, |
|
"grad_norm": 0.5205222669608968, |
|
"learning_rate": 2.701062215477997e-06, |
|
"loss": 0.5049, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 1.3843575418994414, |
|
"grad_norm": 0.5409150510364881, |
|
"learning_rate": 2.693474962063733e-06, |
|
"loss": 0.4946, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.3888268156424581, |
|
"grad_norm": 0.49473773833387136, |
|
"learning_rate": 2.6858877086494687e-06, |
|
"loss": 0.5024, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.393296089385475, |
|
"grad_norm": 0.5322303350412809, |
|
"learning_rate": 2.678300455235205e-06, |
|
"loss": 0.5129, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 1.3977653631284916, |
|
"grad_norm": 0.5249345159944246, |
|
"learning_rate": 2.670713201820941e-06, |
|
"loss": 0.4852, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.4022346368715084, |
|
"grad_norm": 0.526125738227021, |
|
"learning_rate": 2.6631259484066767e-06, |
|
"loss": 0.4755, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.4067039106145252, |
|
"grad_norm": 0.5077327993266544, |
|
"learning_rate": 2.655538694992413e-06, |
|
"loss": 0.4708, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 1.411173184357542, |
|
"grad_norm": 0.509006777595202, |
|
"learning_rate": 2.647951441578149e-06, |
|
"loss": 0.4574, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.4156424581005587, |
|
"grad_norm": 0.5397063984545032, |
|
"learning_rate": 2.6403641881638846e-06, |
|
"loss": 0.4983, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.4201117318435754, |
|
"grad_norm": 0.588039527577832, |
|
"learning_rate": 2.6327769347496208e-06, |
|
"loss": 0.5063, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 1.4245810055865922, |
|
"grad_norm": 0.5086942941008908, |
|
"learning_rate": 2.625189681335357e-06, |
|
"loss": 0.5439, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 1.429050279329609, |
|
"grad_norm": 0.5593378373560197, |
|
"learning_rate": 2.6176024279210926e-06, |
|
"loss": 0.4789, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.4335195530726257, |
|
"grad_norm": 0.5473461722884195, |
|
"learning_rate": 2.6100151745068287e-06, |
|
"loss": 0.4861, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.4379888268156424, |
|
"grad_norm": 0.5744717523131797, |
|
"learning_rate": 2.6024279210925644e-06, |
|
"loss": 0.4818, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 1.4424581005586592, |
|
"grad_norm": 0.5240335872273282, |
|
"learning_rate": 2.5948406676783005e-06, |
|
"loss": 0.4785, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.446927374301676, |
|
"grad_norm": 0.5608475565499077, |
|
"learning_rate": 2.5872534142640367e-06, |
|
"loss": 0.5042, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 1.4513966480446927, |
|
"grad_norm": 0.5363050759442354, |
|
"learning_rate": 2.5796661608497724e-06, |
|
"loss": 0.4995, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 1.4558659217877095, |
|
"grad_norm": 0.5027607995293213, |
|
"learning_rate": 2.5720789074355085e-06, |
|
"loss": 0.4954, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.4603351955307262, |
|
"grad_norm": 0.4969954043343912, |
|
"learning_rate": 2.5644916540212446e-06, |
|
"loss": 0.4882, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 1.464804469273743, |
|
"grad_norm": 0.5718337637259987, |
|
"learning_rate": 2.5569044006069803e-06, |
|
"loss": 0.4726, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.4692737430167597, |
|
"grad_norm": 0.5234285673769393, |
|
"learning_rate": 2.5493171471927164e-06, |
|
"loss": 0.4697, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.4737430167597765, |
|
"grad_norm": 0.5219608223935474, |
|
"learning_rate": 2.5417298937784526e-06, |
|
"loss": 0.4776, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 1.4782122905027932, |
|
"grad_norm": 0.606768683027719, |
|
"learning_rate": 2.5341426403641883e-06, |
|
"loss": 0.4756, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.48268156424581, |
|
"grad_norm": 0.5432176614167162, |
|
"learning_rate": 2.5265553869499244e-06, |
|
"loss": 0.4619, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.48268156424581, |
|
"eval_loss": 0.5032872557640076, |
|
"eval_runtime": 32.2653, |
|
"eval_samples_per_second": 82.751, |
|
"eval_steps_per_second": 5.176, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.4871508379888267, |
|
"grad_norm": 0.5377224565720307, |
|
"learning_rate": 2.51896813353566e-06, |
|
"loss": 0.4667, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.4916201117318435, |
|
"grad_norm": 0.5875842821019925, |
|
"learning_rate": 2.5113808801213962e-06, |
|
"loss": 0.4873, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.4960893854748603, |
|
"grad_norm": 0.5624618850955563, |
|
"learning_rate": 2.5037936267071324e-06, |
|
"loss": 0.4626, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.500558659217877, |
|
"grad_norm": 0.5380665704623208, |
|
"learning_rate": 2.496206373292868e-06, |
|
"loss": 0.4749, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.505027932960894, |
|
"grad_norm": 0.5515048280645092, |
|
"learning_rate": 2.488619119878604e-06, |
|
"loss": 0.4862, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.5094972067039105, |
|
"grad_norm": 0.5541495309875792, |
|
"learning_rate": 2.4810318664643403e-06, |
|
"loss": 0.4533, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.5139664804469275, |
|
"grad_norm": 0.514569610923219, |
|
"learning_rate": 2.473444613050076e-06, |
|
"loss": 0.4585, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.518435754189944, |
|
"grad_norm": 0.5862261680615611, |
|
"learning_rate": 2.465857359635812e-06, |
|
"loss": 0.4835, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.522905027932961, |
|
"grad_norm": 0.536975433650959, |
|
"learning_rate": 2.458270106221548e-06, |
|
"loss": 0.4638, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.5273743016759775, |
|
"grad_norm": 0.5142279881627296, |
|
"learning_rate": 2.450682852807284e-06, |
|
"loss": 0.4832, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.5318435754189945, |
|
"grad_norm": 0.5124062345457664, |
|
"learning_rate": 2.44309559939302e-06, |
|
"loss": 0.4857, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.536312849162011, |
|
"grad_norm": 0.509905421315166, |
|
"learning_rate": 2.435508345978756e-06, |
|
"loss": 0.4807, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.540782122905028, |
|
"grad_norm": 0.5275507665643101, |
|
"learning_rate": 2.427921092564492e-06, |
|
"loss": 0.4722, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.5452513966480446, |
|
"grad_norm": 0.5003951978639241, |
|
"learning_rate": 2.420333839150228e-06, |
|
"loss": 0.4941, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.5497206703910615, |
|
"grad_norm": 0.5251614039487292, |
|
"learning_rate": 2.4127465857359637e-06, |
|
"loss": 0.5376, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.554189944134078, |
|
"grad_norm": 0.5128097582851606, |
|
"learning_rate": 2.4051593323217e-06, |
|
"loss": 0.5001, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.558659217877095, |
|
"grad_norm": 0.4971288144993544, |
|
"learning_rate": 2.397572078907436e-06, |
|
"loss": 0.4721, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.5631284916201116, |
|
"grad_norm": 0.5111491573396423, |
|
"learning_rate": 2.3899848254931717e-06, |
|
"loss": 0.5192, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.5675977653631286, |
|
"grad_norm": 0.5025703950184777, |
|
"learning_rate": 2.382397572078908e-06, |
|
"loss": 0.4541, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.572067039106145, |
|
"grad_norm": 0.5011840454618501, |
|
"learning_rate": 2.3748103186646435e-06, |
|
"loss": 0.4789, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.576536312849162, |
|
"grad_norm": 0.4981690031939279, |
|
"learning_rate": 2.3672230652503792e-06, |
|
"loss": 0.497, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.5810055865921788, |
|
"grad_norm": 0.4994506284814992, |
|
"learning_rate": 2.3596358118361154e-06, |
|
"loss": 0.4655, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.5854748603351956, |
|
"grad_norm": 0.49694019288140645, |
|
"learning_rate": 2.3520485584218515e-06, |
|
"loss": 0.4883, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.5899441340782123, |
|
"grad_norm": 0.5103222760000443, |
|
"learning_rate": 2.344461305007587e-06, |
|
"loss": 0.4615, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.594413407821229, |
|
"grad_norm": 0.4960772434204748, |
|
"learning_rate": 2.3368740515933233e-06, |
|
"loss": 0.4748, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.5988826815642458, |
|
"grad_norm": 0.5372239368130248, |
|
"learning_rate": 2.3292867981790594e-06, |
|
"loss": 0.461, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.6033519553072626, |
|
"grad_norm": 0.5002183877541576, |
|
"learning_rate": 2.321699544764795e-06, |
|
"loss": 0.4985, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.6078212290502794, |
|
"grad_norm": 0.5065612982842012, |
|
"learning_rate": 2.3141122913505313e-06, |
|
"loss": 0.4846, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.612290502793296, |
|
"grad_norm": 0.5294145045424744, |
|
"learning_rate": 2.3065250379362674e-06, |
|
"loss": 0.4769, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.6167597765363129, |
|
"grad_norm": 0.5019842435412675, |
|
"learning_rate": 2.298937784522003e-06, |
|
"loss": 0.4738, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.6212290502793296, |
|
"grad_norm": 0.4973490272845337, |
|
"learning_rate": 2.291350531107739e-06, |
|
"loss": 0.4773, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.6256983240223464, |
|
"grad_norm": 0.4917027064554639, |
|
"learning_rate": 2.283763277693475e-06, |
|
"loss": 0.4921, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.6301675977653631, |
|
"grad_norm": 0.5191932977137537, |
|
"learning_rate": 2.276176024279211e-06, |
|
"loss": 0.4823, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.6346368715083799, |
|
"grad_norm": 0.5130853619632637, |
|
"learning_rate": 2.268588770864947e-06, |
|
"loss": 0.4631, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.6391061452513966, |
|
"grad_norm": 0.5402133951776769, |
|
"learning_rate": 2.261001517450683e-06, |
|
"loss": 0.456, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.6435754189944134, |
|
"grad_norm": 0.5136201214507364, |
|
"learning_rate": 2.253414264036419e-06, |
|
"loss": 0.4715, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.6480446927374302, |
|
"grad_norm": 0.5397223114168315, |
|
"learning_rate": 2.245827010622155e-06, |
|
"loss": 0.465, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.652513966480447, |
|
"grad_norm": 0.5178224930093817, |
|
"learning_rate": 2.238239757207891e-06, |
|
"loss": 0.4526, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.6569832402234637, |
|
"grad_norm": 0.5399618966600667, |
|
"learning_rate": 2.230652503793627e-06, |
|
"loss": 0.4873, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.6614525139664804, |
|
"grad_norm": 0.5106529140250409, |
|
"learning_rate": 2.223065250379363e-06, |
|
"loss": 0.4771, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.6659217877094972, |
|
"grad_norm": 0.5460841160518792, |
|
"learning_rate": 2.2154779969650988e-06, |
|
"loss": 0.4743, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.670391061452514, |
|
"grad_norm": 0.5846706720135237, |
|
"learning_rate": 2.207890743550835e-06, |
|
"loss": 0.4552, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.6748603351955307, |
|
"grad_norm": 0.534699591229056, |
|
"learning_rate": 2.2003034901365706e-06, |
|
"loss": 0.495, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.6793296089385474, |
|
"grad_norm": 0.5377842578832556, |
|
"learning_rate": 2.1927162367223067e-06, |
|
"loss": 0.4824, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.6837988826815642, |
|
"grad_norm": 0.5799148072307362, |
|
"learning_rate": 2.185128983308043e-06, |
|
"loss": 0.4708, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.6882681564245812, |
|
"grad_norm": 0.5076181067753225, |
|
"learning_rate": 2.1775417298937786e-06, |
|
"loss": 0.4631, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.6927374301675977, |
|
"grad_norm": 0.5025103762077108, |
|
"learning_rate": 2.1699544764795147e-06, |
|
"loss": 0.4676, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.6972067039106147, |
|
"grad_norm": 0.48231447315943815, |
|
"learning_rate": 2.162367223065251e-06, |
|
"loss": 0.4694, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.7016759776536312, |
|
"grad_norm": 0.5003437712619672, |
|
"learning_rate": 2.1547799696509865e-06, |
|
"loss": 0.4607, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.7061452513966482, |
|
"grad_norm": 0.5149834201342615, |
|
"learning_rate": 2.1471927162367226e-06, |
|
"loss": 0.5253, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.7106145251396647, |
|
"grad_norm": 0.5406857228194807, |
|
"learning_rate": 2.1396054628224588e-06, |
|
"loss": 0.5352, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.7150837988826817, |
|
"grad_norm": 0.5204717082880796, |
|
"learning_rate": 2.1320182094081945e-06, |
|
"loss": 0.4936, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.7195530726256982, |
|
"grad_norm": 0.5202749448515015, |
|
"learning_rate": 2.12443095599393e-06, |
|
"loss": 0.4594, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.7240223463687152, |
|
"grad_norm": 0.4952910050954741, |
|
"learning_rate": 2.1168437025796663e-06, |
|
"loss": 0.4793, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.7284916201117317, |
|
"grad_norm": 0.48637344511275765, |
|
"learning_rate": 2.109256449165402e-06, |
|
"loss": 0.4745, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.7329608938547487, |
|
"grad_norm": 0.4906271558417189, |
|
"learning_rate": 2.101669195751138e-06, |
|
"loss": 0.4652, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.7374301675977653, |
|
"grad_norm": 0.5090727761430808, |
|
"learning_rate": 2.0940819423368742e-06, |
|
"loss": 0.5298, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.7418994413407822, |
|
"grad_norm": 0.4866596430096002, |
|
"learning_rate": 2.08649468892261e-06, |
|
"loss": 0.4838, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.7463687150837988, |
|
"grad_norm": 0.49259128094078225, |
|
"learning_rate": 2.078907435508346e-06, |
|
"loss": 0.4737, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.7508379888268157, |
|
"grad_norm": 0.5031034927365164, |
|
"learning_rate": 2.071320182094082e-06, |
|
"loss": 0.4777, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.7553072625698323, |
|
"grad_norm": 0.48520021809068, |
|
"learning_rate": 2.063732928679818e-06, |
|
"loss": 0.454, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.7597765363128492, |
|
"grad_norm": 0.5217902917498033, |
|
"learning_rate": 2.056145675265554e-06, |
|
"loss": 0.5199, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.7642458100558658, |
|
"grad_norm": 0.4963706354662438, |
|
"learning_rate": 2.04855842185129e-06, |
|
"loss": 0.4649, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.7687150837988828, |
|
"grad_norm": 0.49552580249683814, |
|
"learning_rate": 2.040971168437026e-06, |
|
"loss": 0.4887, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.7731843575418993, |
|
"grad_norm": 0.49468231483436587, |
|
"learning_rate": 2.033383915022762e-06, |
|
"loss": 0.4757, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.7776536312849163, |
|
"grad_norm": 0.47502769186897137, |
|
"learning_rate": 2.0257966616084977e-06, |
|
"loss": 0.4846, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.7821229050279328, |
|
"grad_norm": 0.5089134108324753, |
|
"learning_rate": 2.018209408194234e-06, |
|
"loss": 0.4692, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.7865921787709498, |
|
"grad_norm": 0.4981731250179096, |
|
"learning_rate": 2.01062215477997e-06, |
|
"loss": 0.4665, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 1.7910614525139665, |
|
"grad_norm": 0.4690174096990647, |
|
"learning_rate": 2.0030349013657056e-06, |
|
"loss": 0.4606, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.7955307262569833, |
|
"grad_norm": 0.4914411018326139, |
|
"learning_rate": 1.9954476479514418e-06, |
|
"loss": 0.4847, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.5513335287615446, |
|
"learning_rate": 1.987860394537178e-06, |
|
"loss": 0.488, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 1.8044692737430168, |
|
"grad_norm": 0.48992847389625854, |
|
"learning_rate": 1.9802731411229136e-06, |
|
"loss": 0.465, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.8089385474860336, |
|
"grad_norm": 0.5185912675971102, |
|
"learning_rate": 1.9726858877086497e-06, |
|
"loss": 0.4872, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 1.8134078212290503, |
|
"grad_norm": 0.47300994432774746, |
|
"learning_rate": 1.965098634294386e-06, |
|
"loss": 0.4534, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.817877094972067, |
|
"grad_norm": 0.5034275716173061, |
|
"learning_rate": 1.9575113808801215e-06, |
|
"loss": 0.4668, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.8223463687150838, |
|
"grad_norm": 0.46435516708821234, |
|
"learning_rate": 1.9499241274658577e-06, |
|
"loss": 0.491, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.8268156424581006, |
|
"grad_norm": 0.4915994950607221, |
|
"learning_rate": 1.9423368740515934e-06, |
|
"loss": 0.4793, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 1.8312849162011173, |
|
"grad_norm": 0.49834550842622505, |
|
"learning_rate": 1.9347496206373295e-06, |
|
"loss": 0.4897, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.835754189944134, |
|
"grad_norm": 0.5275123854480922, |
|
"learning_rate": 1.9271623672230656e-06, |
|
"loss": 0.5167, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.8402234636871508, |
|
"grad_norm": 0.501447167095016, |
|
"learning_rate": 1.9195751138088013e-06, |
|
"loss": 0.5522, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 1.8446927374301676, |
|
"grad_norm": 0.4929358533949542, |
|
"learning_rate": 1.9119878603945374e-06, |
|
"loss": 0.464, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.8491620111731844, |
|
"grad_norm": 0.4753893398726578, |
|
"learning_rate": 1.9044006069802734e-06, |
|
"loss": 0.4707, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 1.853631284916201, |
|
"grad_norm": 0.49249135995729143, |
|
"learning_rate": 1.8968133535660093e-06, |
|
"loss": 0.4674, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 1.8581005586592179, |
|
"grad_norm": 0.4959065755228421, |
|
"learning_rate": 1.8892261001517454e-06, |
|
"loss": 0.4731, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.8625698324022346, |
|
"grad_norm": 0.4830053823448673, |
|
"learning_rate": 1.8816388467374813e-06, |
|
"loss": 0.4688, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.8670391061452514, |
|
"grad_norm": 0.5026774544021604, |
|
"learning_rate": 1.874051593323217e-06, |
|
"loss": 0.4777, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.8715083798882681, |
|
"grad_norm": 0.48376669794821286, |
|
"learning_rate": 1.866464339908953e-06, |
|
"loss": 0.4532, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.8759776536312849, |
|
"grad_norm": 0.5036910075540445, |
|
"learning_rate": 1.858877086494689e-06, |
|
"loss": 0.4749, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.8804469273743016, |
|
"grad_norm": 0.464687327497754, |
|
"learning_rate": 1.851289833080425e-06, |
|
"loss": 0.4582, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.8849162011173184, |
|
"grad_norm": 0.4713537790626329, |
|
"learning_rate": 1.8437025796661609e-06, |
|
"loss": 0.4954, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.8893854748603351, |
|
"grad_norm": 0.4964027557908742, |
|
"learning_rate": 1.8361153262518968e-06, |
|
"loss": 0.4777, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.893854748603352, |
|
"grad_norm": 0.5022160749373462, |
|
"learning_rate": 1.828528072837633e-06, |
|
"loss": 0.4723, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.8983240223463689, |
|
"grad_norm": 0.4971783807561644, |
|
"learning_rate": 1.8209408194233688e-06, |
|
"loss": 0.4619, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.9027932960893854, |
|
"grad_norm": 0.5020482529233811, |
|
"learning_rate": 1.8133535660091047e-06, |
|
"loss": 0.4923, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.9072625698324024, |
|
"grad_norm": 0.5027930539648167, |
|
"learning_rate": 1.8057663125948407e-06, |
|
"loss": 0.4936, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.911731843575419, |
|
"grad_norm": 0.5048109425957825, |
|
"learning_rate": 1.7981790591805768e-06, |
|
"loss": 0.4647, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.916201117318436, |
|
"grad_norm": 0.5152625504278344, |
|
"learning_rate": 1.7905918057663127e-06, |
|
"loss": 0.4962, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.9206703910614524, |
|
"grad_norm": 0.5486472960937676, |
|
"learning_rate": 1.7830045523520486e-06, |
|
"loss": 0.4643, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.9251396648044694, |
|
"grad_norm": 0.5074159325604022, |
|
"learning_rate": 1.7754172989377847e-06, |
|
"loss": 0.4838, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.929608938547486, |
|
"grad_norm": 0.487237277239409, |
|
"learning_rate": 1.7678300455235207e-06, |
|
"loss": 0.4953, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.934078212290503, |
|
"grad_norm": 0.5030243156211434, |
|
"learning_rate": 1.7602427921092566e-06, |
|
"loss": 0.4669, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.9385474860335195, |
|
"grad_norm": 0.54024162061195, |
|
"learning_rate": 1.7526555386949925e-06, |
|
"loss": 0.4645, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.9430167597765364, |
|
"grad_norm": 0.4924548318787792, |
|
"learning_rate": 1.7450682852807286e-06, |
|
"loss": 0.487, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.947486033519553, |
|
"grad_norm": 0.5002700607677886, |
|
"learning_rate": 1.7374810318664645e-06, |
|
"loss": 0.4904, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.95195530726257, |
|
"grad_norm": 0.49052978017153487, |
|
"learning_rate": 1.7298937784522004e-06, |
|
"loss": 0.4721, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.9564245810055865, |
|
"grad_norm": 0.5330277427973065, |
|
"learning_rate": 1.7223065250379363e-06, |
|
"loss": 0.4674, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.9608938547486034, |
|
"grad_norm": 0.49644600432443986, |
|
"learning_rate": 1.7147192716236725e-06, |
|
"loss": 0.4735, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.96536312849162, |
|
"grad_norm": 0.4763428511189237, |
|
"learning_rate": 1.7071320182094084e-06, |
|
"loss": 0.4758, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.969832402234637, |
|
"grad_norm": 0.49662051116381534, |
|
"learning_rate": 1.6995447647951443e-06, |
|
"loss": 0.5025, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.9743016759776535, |
|
"grad_norm": 0.5368027446646121, |
|
"learning_rate": 1.6919575113808804e-06, |
|
"loss": 0.4708, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.9787709497206705, |
|
"grad_norm": 0.49270124177132796, |
|
"learning_rate": 1.6843702579666163e-06, |
|
"loss": 0.4859, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.983240223463687, |
|
"grad_norm": 0.4886244120662818, |
|
"learning_rate": 1.6767830045523523e-06, |
|
"loss": 0.4542, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.983240223463687, |
|
"eval_loss": 0.49883514642715454, |
|
"eval_runtime": 32.2692, |
|
"eval_samples_per_second": 82.741, |
|
"eval_steps_per_second": 5.175, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.987709497206704, |
|
"grad_norm": 0.48474622370426634, |
|
"learning_rate": 1.6691957511380882e-06, |
|
"loss": 0.4752, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.9921787709497205, |
|
"grad_norm": 0.5089879136002737, |
|
"learning_rate": 1.6616084977238243e-06, |
|
"loss": 0.4835, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.9966480446927375, |
|
"grad_norm": 0.540987114653738, |
|
"learning_rate": 1.6540212443095602e-06, |
|
"loss": 0.4685, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 2.001117318435754, |
|
"grad_norm": 0.5330263791441335, |
|
"learning_rate": 1.6464339908952961e-06, |
|
"loss": 0.4729, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 2.005586592178771, |
|
"grad_norm": 0.5025591514510047, |
|
"learning_rate": 1.638846737481032e-06, |
|
"loss": 0.4739, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 2.010055865921788, |
|
"grad_norm": 0.4938210284492649, |
|
"learning_rate": 1.6312594840667682e-06, |
|
"loss": 0.4772, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 2.0145251396648045, |
|
"grad_norm": 0.5590423643222178, |
|
"learning_rate": 1.6236722306525039e-06, |
|
"loss": 0.4651, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 2.018994413407821, |
|
"grad_norm": 0.4994221649154916, |
|
"learning_rate": 1.6160849772382398e-06, |
|
"loss": 0.4989, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 2.004469273743017, |
|
"grad_norm": 0.5578750520607872, |
|
"learning_rate": 1.6084977238239757e-06, |
|
"loss": 0.4779, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 2.0089385474860335, |
|
"grad_norm": 0.5277075050354076, |
|
"learning_rate": 1.6009104704097116e-06, |
|
"loss": 0.462, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 2.0134078212290505, |
|
"grad_norm": 0.5125996961203592, |
|
"learning_rate": 1.5933232169954477e-06, |
|
"loss": 0.4892, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 2.017877094972067, |
|
"grad_norm": 0.5150882727539035, |
|
"learning_rate": 1.5857359635811836e-06, |
|
"loss": 0.4913, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.022346368715084, |
|
"grad_norm": 0.5395569256138755, |
|
"learning_rate": 1.5781487101669196e-06, |
|
"loss": 0.4789, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 2.0268156424581005, |
|
"grad_norm": 0.4907655840755226, |
|
"learning_rate": 1.5705614567526557e-06, |
|
"loss": 0.4699, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 2.0312849162011175, |
|
"grad_norm": 0.508487377683692, |
|
"learning_rate": 1.5629742033383916e-06, |
|
"loss": 0.4671, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 2.035754189944134, |
|
"grad_norm": 0.5211842893319257, |
|
"learning_rate": 1.5553869499241275e-06, |
|
"loss": 0.48, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 2.040223463687151, |
|
"grad_norm": 0.5540445804429339, |
|
"learning_rate": 1.5477996965098634e-06, |
|
"loss": 0.4839, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 2.0446927374301676, |
|
"grad_norm": 0.502144462775385, |
|
"learning_rate": 1.5402124430955995e-06, |
|
"loss": 0.4991, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 2.0491620111731845, |
|
"grad_norm": 0.4963826920124019, |
|
"learning_rate": 1.5326251896813355e-06, |
|
"loss": 0.5268, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 2.053631284916201, |
|
"grad_norm": 0.48862357759049696, |
|
"learning_rate": 1.5250379362670714e-06, |
|
"loss": 0.4572, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 2.058100558659218, |
|
"grad_norm": 0.48705814461293884, |
|
"learning_rate": 1.5174506828528073e-06, |
|
"loss": 0.4761, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 2.0625698324022346, |
|
"grad_norm": 0.50151248823518, |
|
"learning_rate": 1.5098634294385434e-06, |
|
"loss": 0.468, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.0670391061452515, |
|
"grad_norm": 0.47075309247993236, |
|
"learning_rate": 1.5022761760242793e-06, |
|
"loss": 0.4496, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 2.071508379888268, |
|
"grad_norm": 0.5055609634042088, |
|
"learning_rate": 1.4946889226100152e-06, |
|
"loss": 0.4626, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 2.075977653631285, |
|
"grad_norm": 0.4931560723293831, |
|
"learning_rate": 1.4871016691957514e-06, |
|
"loss": 0.4743, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 2.0804469273743016, |
|
"grad_norm": 0.4970179492434698, |
|
"learning_rate": 1.4795144157814873e-06, |
|
"loss": 0.4849, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 2.0849162011173186, |
|
"grad_norm": 0.5205249715468657, |
|
"learning_rate": 1.4719271623672232e-06, |
|
"loss": 0.4647, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.089385474860335, |
|
"grad_norm": 0.5106891771853166, |
|
"learning_rate": 1.4643399089529591e-06, |
|
"loss": 0.4698, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 2.093854748603352, |
|
"grad_norm": 0.5030533511191912, |
|
"learning_rate": 1.4567526555386952e-06, |
|
"loss": 0.4435, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 2.0983240223463686, |
|
"grad_norm": 0.5267045945145785, |
|
"learning_rate": 1.4491654021244311e-06, |
|
"loss": 0.4551, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 2.1027932960893856, |
|
"grad_norm": 0.4910735265010368, |
|
"learning_rate": 1.441578148710167e-06, |
|
"loss": 0.4893, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 2.107262569832402, |
|
"grad_norm": 0.5098674286346987, |
|
"learning_rate": 1.433990895295903e-06, |
|
"loss": 0.4631, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.111731843575419, |
|
"grad_norm": 0.49833407519162715, |
|
"learning_rate": 1.426403641881639e-06, |
|
"loss": 0.4478, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 2.1162011173184356, |
|
"grad_norm": 0.4968638787102979, |
|
"learning_rate": 1.418816388467375e-06, |
|
"loss": 0.4661, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 2.1206703910614526, |
|
"grad_norm": 0.5287581344581138, |
|
"learning_rate": 1.411229135053111e-06, |
|
"loss": 0.487, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 2.125139664804469, |
|
"grad_norm": 0.5169520017940434, |
|
"learning_rate": 1.403641881638847e-06, |
|
"loss": 0.4641, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 2.129608938547486, |
|
"grad_norm": 0.4889710176907046, |
|
"learning_rate": 1.396054628224583e-06, |
|
"loss": 0.4657, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 2.1340782122905027, |
|
"grad_norm": 0.5068054201127302, |
|
"learning_rate": 1.3884673748103189e-06, |
|
"loss": 0.4655, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 2.1385474860335196, |
|
"grad_norm": 0.5227391076748372, |
|
"learning_rate": 1.3808801213960548e-06, |
|
"loss": 0.4489, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 2.143016759776536, |
|
"grad_norm": 0.4685531353247096, |
|
"learning_rate": 1.3732928679817905e-06, |
|
"loss": 0.4518, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 2.147486033519553, |
|
"grad_norm": 0.48729624697805496, |
|
"learning_rate": 1.3657056145675266e-06, |
|
"loss": 0.4539, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 2.1519553072625697, |
|
"grad_norm": 0.49686659002832273, |
|
"learning_rate": 1.3581183611532625e-06, |
|
"loss": 0.4698, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.1564245810055866, |
|
"grad_norm": 0.5165370496352449, |
|
"learning_rate": 1.3505311077389985e-06, |
|
"loss": 0.5282, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 2.160893854748603, |
|
"grad_norm": 0.49390583090447004, |
|
"learning_rate": 1.3429438543247344e-06, |
|
"loss": 0.5011, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 2.16536312849162, |
|
"grad_norm": 0.4966083299512843, |
|
"learning_rate": 1.3353566009104705e-06, |
|
"loss": 0.4888, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 2.1698324022346367, |
|
"grad_norm": 0.47651097727052893, |
|
"learning_rate": 1.3277693474962064e-06, |
|
"loss": 0.4629, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 2.1743016759776537, |
|
"grad_norm": 0.5171229401967155, |
|
"learning_rate": 1.3201820940819423e-06, |
|
"loss": 0.4538, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 2.17877094972067, |
|
"grad_norm": 0.5177579051508318, |
|
"learning_rate": 1.3125948406676784e-06, |
|
"loss": 0.4822, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 2.183240223463687, |
|
"grad_norm": 0.5363730349952008, |
|
"learning_rate": 1.3050075872534144e-06, |
|
"loss": 0.4879, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 2.1877094972067037, |
|
"grad_norm": 0.49547274684149156, |
|
"learning_rate": 1.2974203338391503e-06, |
|
"loss": 0.4732, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 2.1921787709497207, |
|
"grad_norm": 0.48560343902497966, |
|
"learning_rate": 1.2898330804248862e-06, |
|
"loss": 0.4918, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 2.1966480446927372, |
|
"grad_norm": 0.49966297308311575, |
|
"learning_rate": 1.2822458270106223e-06, |
|
"loss": 0.4618, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.201117318435754, |
|
"grad_norm": 0.472067295226221, |
|
"learning_rate": 1.2746585735963582e-06, |
|
"loss": 0.4953, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 2.205586592178771, |
|
"grad_norm": 0.5241464474412257, |
|
"learning_rate": 1.2670713201820941e-06, |
|
"loss": 0.4797, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 2.2100558659217877, |
|
"grad_norm": 0.5063960714766558, |
|
"learning_rate": 1.25948406676783e-06, |
|
"loss": 0.4493, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 2.2145251396648042, |
|
"grad_norm": 0.48407847297667117, |
|
"learning_rate": 1.2518968133535662e-06, |
|
"loss": 0.4736, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 2.218994413407821, |
|
"grad_norm": 0.4859389089228068, |
|
"learning_rate": 1.244309559939302e-06, |
|
"loss": 0.4476, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 2.223463687150838, |
|
"grad_norm": 0.49468343574845597, |
|
"learning_rate": 1.236722306525038e-06, |
|
"loss": 0.4447, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 2.2279329608938547, |
|
"grad_norm": 0.45980997966401754, |
|
"learning_rate": 1.229135053110774e-06, |
|
"loss": 0.4599, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 2.2324022346368717, |
|
"grad_norm": 0.5135776498149942, |
|
"learning_rate": 1.22154779969651e-06, |
|
"loss": 0.4852, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 2.2368715083798882, |
|
"grad_norm": 0.48749385873656575, |
|
"learning_rate": 1.213960546282246e-06, |
|
"loss": 0.5157, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 2.241340782122905, |
|
"grad_norm": 0.48899276420726745, |
|
"learning_rate": 1.2063732928679819e-06, |
|
"loss": 0.4437, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.2458100558659218, |
|
"grad_norm": 0.4918342780724834, |
|
"learning_rate": 1.198786039453718e-06, |
|
"loss": 0.4974, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 2.2502793296089387, |
|
"grad_norm": 0.4774969997416472, |
|
"learning_rate": 1.191198786039454e-06, |
|
"loss": 0.4509, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 2.2547486033519553, |
|
"grad_norm": 0.4791749757272505, |
|
"learning_rate": 1.1836115326251896e-06, |
|
"loss": 0.4599, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 2.2592178770949722, |
|
"grad_norm": 0.5072314853369309, |
|
"learning_rate": 1.1760242792109257e-06, |
|
"loss": 0.4643, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 2.2636871508379888, |
|
"grad_norm": 0.5004472392134528, |
|
"learning_rate": 1.1684370257966617e-06, |
|
"loss": 0.4688, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 2.2681564245810057, |
|
"grad_norm": 0.48150392498478517, |
|
"learning_rate": 1.1608497723823976e-06, |
|
"loss": 0.5069, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 2.2726256983240223, |
|
"grad_norm": 0.4828855445414688, |
|
"learning_rate": 1.1532625189681337e-06, |
|
"loss": 0.4571, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 2.2770949720670393, |
|
"grad_norm": 0.49789128855450915, |
|
"learning_rate": 1.1456752655538696e-06, |
|
"loss": 0.4651, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 2.281564245810056, |
|
"grad_norm": 0.49890243849633514, |
|
"learning_rate": 1.1380880121396055e-06, |
|
"loss": 0.4757, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 2.2860335195530728, |
|
"grad_norm": 0.48825603773254755, |
|
"learning_rate": 1.1305007587253414e-06, |
|
"loss": 0.4668, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.2905027932960893, |
|
"grad_norm": 0.48797823917514366, |
|
"learning_rate": 1.1229135053110776e-06, |
|
"loss": 0.4576, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 2.2949720670391063, |
|
"grad_norm": 0.4713265522467315, |
|
"learning_rate": 1.1153262518968135e-06, |
|
"loss": 0.4998, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 2.299441340782123, |
|
"grad_norm": 0.48731093463299924, |
|
"learning_rate": 1.1077389984825494e-06, |
|
"loss": 0.4583, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 2.30391061452514, |
|
"grad_norm": 0.5356992050391173, |
|
"learning_rate": 1.1001517450682853e-06, |
|
"loss": 0.4964, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 2.3083798882681563, |
|
"grad_norm": 0.5185855845215128, |
|
"learning_rate": 1.0925644916540214e-06, |
|
"loss": 0.4456, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.3128491620111733, |
|
"grad_norm": 0.5340341729511069, |
|
"learning_rate": 1.0849772382397573e-06, |
|
"loss": 0.4667, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 2.31731843575419, |
|
"grad_norm": 0.46791114784776194, |
|
"learning_rate": 1.0773899848254933e-06, |
|
"loss": 0.4907, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 2.321787709497207, |
|
"grad_norm": 0.5224682012273069, |
|
"learning_rate": 1.0698027314112294e-06, |
|
"loss": 0.4411, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 2.3262569832402233, |
|
"grad_norm": 0.5013175127206561, |
|
"learning_rate": 1.062215477996965e-06, |
|
"loss": 0.4872, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 2.3307262569832403, |
|
"grad_norm": 0.49567401298066655, |
|
"learning_rate": 1.054628224582701e-06, |
|
"loss": 0.4547, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.335195530726257, |
|
"grad_norm": 0.5092075971225333, |
|
"learning_rate": 1.0470409711684371e-06, |
|
"loss": 0.4774, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 2.339664804469274, |
|
"grad_norm": 0.5206537811339336, |
|
"learning_rate": 1.039453717754173e-06, |
|
"loss": 0.4593, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 2.3441340782122904, |
|
"grad_norm": 0.5247188237207532, |
|
"learning_rate": 1.031866464339909e-06, |
|
"loss": 0.4947, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 2.3486033519553073, |
|
"grad_norm": 0.48511496092549006, |
|
"learning_rate": 1.024279210925645e-06, |
|
"loss": 0.4675, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 2.353072625698324, |
|
"grad_norm": 0.49031229609343957, |
|
"learning_rate": 1.016691957511381e-06, |
|
"loss": 0.4654, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 2.357541899441341, |
|
"grad_norm": 0.5022437562865831, |
|
"learning_rate": 1.009104704097117e-06, |
|
"loss": 0.4881, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 2.3620111731843574, |
|
"grad_norm": 0.48427536854604036, |
|
"learning_rate": 1.0015174506828528e-06, |
|
"loss": 0.4722, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 2.3664804469273744, |
|
"grad_norm": 0.4951077677256604, |
|
"learning_rate": 9.93930197268589e-07, |
|
"loss": 0.4816, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 2.370949720670391, |
|
"grad_norm": 0.506971274674449, |
|
"learning_rate": 9.863429438543249e-07, |
|
"loss": 0.448, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 2.375418994413408, |
|
"grad_norm": 0.4843767211745505, |
|
"learning_rate": 9.787556904400608e-07, |
|
"loss": 0.4644, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.3798882681564244, |
|
"grad_norm": 0.4787312911085378, |
|
"learning_rate": 9.711684370257967e-07, |
|
"loss": 0.4681, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 2.3843575418994414, |
|
"grad_norm": 0.5040950848521704, |
|
"learning_rate": 9.635811836115328e-07, |
|
"loss": 0.4595, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 2.388826815642458, |
|
"grad_norm": 0.4742009596031962, |
|
"learning_rate": 9.559939301972687e-07, |
|
"loss": 0.4542, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 2.393296089385475, |
|
"grad_norm": 0.4819914757091554, |
|
"learning_rate": 9.484066767830046e-07, |
|
"loss": 0.4818, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 2.3977653631284914, |
|
"grad_norm": 0.4794763925814031, |
|
"learning_rate": 9.408194233687407e-07, |
|
"loss": 0.4459, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 2.4022346368715084, |
|
"grad_norm": 0.5065706041345039, |
|
"learning_rate": 9.332321699544765e-07, |
|
"loss": 0.4986, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 2.4067039106145254, |
|
"grad_norm": 0.4918492784353524, |
|
"learning_rate": 9.256449165402125e-07, |
|
"loss": 0.4618, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 2.411173184357542, |
|
"grad_norm": 0.48410824249150014, |
|
"learning_rate": 9.180576631259484e-07, |
|
"loss": 0.4358, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 2.4156424581005584, |
|
"grad_norm": 0.5071386174484, |
|
"learning_rate": 9.104704097116844e-07, |
|
"loss": 0.4522, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 2.4201117318435754, |
|
"grad_norm": 0.5179035085963222, |
|
"learning_rate": 9.028831562974203e-07, |
|
"loss": 0.4718, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.4245810055865924, |
|
"grad_norm": 0.48695002255677194, |
|
"learning_rate": 8.952959028831563e-07, |
|
"loss": 0.4946, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 2.429050279329609, |
|
"grad_norm": 0.4697467884861594, |
|
"learning_rate": 8.877086494688924e-07, |
|
"loss": 0.482, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 2.4335195530726255, |
|
"grad_norm": 0.5039869709064255, |
|
"learning_rate": 8.801213960546283e-07, |
|
"loss": 0.4528, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 2.4379888268156424, |
|
"grad_norm": 0.4956541767737936, |
|
"learning_rate": 8.725341426403643e-07, |
|
"loss": 0.4416, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 2.4424581005586594, |
|
"grad_norm": 0.5220748437476452, |
|
"learning_rate": 8.649468892261002e-07, |
|
"loss": 0.4534, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 2.446927374301676, |
|
"grad_norm": 0.5004114462806725, |
|
"learning_rate": 8.573596358118362e-07, |
|
"loss": 0.4523, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 2.451396648044693, |
|
"grad_norm": 0.5002883358165275, |
|
"learning_rate": 8.497723823975721e-07, |
|
"loss": 0.4483, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 2.4558659217877095, |
|
"grad_norm": 0.5001315929982186, |
|
"learning_rate": 8.421851289833082e-07, |
|
"loss": 0.4547, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 2.4603351955307264, |
|
"grad_norm": 0.5141518132214271, |
|
"learning_rate": 8.345978755690441e-07, |
|
"loss": 0.5064, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 2.464804469273743, |
|
"grad_norm": 0.4960566337702956, |
|
"learning_rate": 8.270106221547801e-07, |
|
"loss": 0.4473, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.464804469273743, |
|
"eval_loss": 0.4975128769874573, |
|
"eval_runtime": 32.1892, |
|
"eval_samples_per_second": 82.947, |
|
"eval_steps_per_second": 5.188, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.46927374301676, |
|
"grad_norm": 0.4897198333575608, |
|
"learning_rate": 8.19423368740516e-07, |
|
"loss": 0.4559, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 2.4737430167597765, |
|
"grad_norm": 0.49985877359817477, |
|
"learning_rate": 8.118361153262519e-07, |
|
"loss": 0.4651, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 2.4782122905027935, |
|
"grad_norm": 0.49485306049230043, |
|
"learning_rate": 8.042488619119878e-07, |
|
"loss": 0.4803, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 2.48268156424581, |
|
"grad_norm": 0.5208900276481403, |
|
"learning_rate": 7.966616084977239e-07, |
|
"loss": 0.4553, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 2.487150837988827, |
|
"grad_norm": 0.48837650054879367, |
|
"learning_rate": 7.890743550834598e-07, |
|
"loss": 0.4681, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 2.4916201117318435, |
|
"grad_norm": 0.47305516115923485, |
|
"learning_rate": 7.814871016691958e-07, |
|
"loss": 0.4564, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 2.4960893854748605, |
|
"grad_norm": 0.5049602967209187, |
|
"learning_rate": 7.738998482549317e-07, |
|
"loss": 0.4632, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 2.500558659217877, |
|
"grad_norm": 0.48547661489581506, |
|
"learning_rate": 7.663125948406677e-07, |
|
"loss": 0.448, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 2.505027932960894, |
|
"grad_norm": 0.5747686136950829, |
|
"learning_rate": 7.587253414264036e-07, |
|
"loss": 0.4432, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 2.5094972067039105, |
|
"grad_norm": 0.5499322063415041, |
|
"learning_rate": 7.511380880121397e-07, |
|
"loss": 0.432, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.5139664804469275, |
|
"grad_norm": 0.49909881450641125, |
|
"learning_rate": 7.435508345978757e-07, |
|
"loss": 0.4325, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 2.518435754189944, |
|
"grad_norm": 0.5137001617086465, |
|
"learning_rate": 7.359635811836116e-07, |
|
"loss": 0.4374, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 2.522905027932961, |
|
"grad_norm": 0.5043757440405017, |
|
"learning_rate": 7.283763277693476e-07, |
|
"loss": 0.4751, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 2.5273743016759775, |
|
"grad_norm": 0.4882894387336232, |
|
"learning_rate": 7.207890743550835e-07, |
|
"loss": 0.4368, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 2.5318435754189945, |
|
"grad_norm": 0.5302357654250439, |
|
"learning_rate": 7.132018209408196e-07, |
|
"loss": 0.4174, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.536312849162011, |
|
"grad_norm": 0.5229623842607946, |
|
"learning_rate": 7.056145675265555e-07, |
|
"loss": 0.4353, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 2.540782122905028, |
|
"grad_norm": 0.5145290981136341, |
|
"learning_rate": 6.980273141122915e-07, |
|
"loss": 0.4278, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 2.5452513966480446, |
|
"grad_norm": 0.49830066459201744, |
|
"learning_rate": 6.904400606980274e-07, |
|
"loss": 0.434, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 2.5497206703910615, |
|
"grad_norm": 0.5220459047427843, |
|
"learning_rate": 6.828528072837633e-07, |
|
"loss": 0.4375, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 2.554189944134078, |
|
"grad_norm": 0.5111391623585094, |
|
"learning_rate": 6.752655538694992e-07, |
|
"loss": 0.4295, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.558659217877095, |
|
"grad_norm": 0.4944043554942434, |
|
"learning_rate": 6.676783004552352e-07, |
|
"loss": 0.4205, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 2.5631284916201116, |
|
"grad_norm": 0.49154601956517674, |
|
"learning_rate": 6.600910470409712e-07, |
|
"loss": 0.4502, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 2.5675977653631286, |
|
"grad_norm": 0.5008348019898189, |
|
"learning_rate": 6.525037936267072e-07, |
|
"loss": 0.4453, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 2.572067039106145, |
|
"grad_norm": 0.5005572885290496, |
|
"learning_rate": 6.449165402124431e-07, |
|
"loss": 0.4353, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 2.576536312849162, |
|
"grad_norm": 0.5302967367534124, |
|
"learning_rate": 6.373292867981791e-07, |
|
"loss": 0.4269, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 2.5810055865921786, |
|
"grad_norm": 0.5050943269077487, |
|
"learning_rate": 6.29742033383915e-07, |
|
"loss": 0.4372, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 2.5854748603351956, |
|
"grad_norm": 0.5223556534474804, |
|
"learning_rate": 6.22154779969651e-07, |
|
"loss": 0.445, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 2.5899441340782126, |
|
"grad_norm": 0.4891137574673572, |
|
"learning_rate": 6.14567526555387e-07, |
|
"loss": 0.4171, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 2.594413407821229, |
|
"grad_norm": 0.5172276124686148, |
|
"learning_rate": 6.06980273141123e-07, |
|
"loss": 0.4522, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 2.5988826815642456, |
|
"grad_norm": 0.48873823307103154, |
|
"learning_rate": 5.99393019726859e-07, |
|
"loss": 0.4286, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.6033519553072626, |
|
"grad_norm": 0.502230133682332, |
|
"learning_rate": 5.918057663125948e-07, |
|
"loss": 0.4461, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 2.6078212290502796, |
|
"grad_norm": 0.4898396410998909, |
|
"learning_rate": 5.842185128983308e-07, |
|
"loss": 0.4289, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 2.612290502793296, |
|
"grad_norm": 0.48422568768513635, |
|
"learning_rate": 5.766312594840668e-07, |
|
"loss": 0.4423, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 2.6167597765363126, |
|
"grad_norm": 0.5003994120834477, |
|
"learning_rate": 5.690440060698028e-07, |
|
"loss": 0.437, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 2.6212290502793296, |
|
"grad_norm": 0.5079705228889307, |
|
"learning_rate": 5.614567526555388e-07, |
|
"loss": 0.4516, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 2.6256983240223466, |
|
"grad_norm": 0.49246631419316356, |
|
"learning_rate": 5.538694992412747e-07, |
|
"loss": 0.4368, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 2.630167597765363, |
|
"grad_norm": 0.5086922889782935, |
|
"learning_rate": 5.462822458270107e-07, |
|
"loss": 0.4227, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 2.6346368715083797, |
|
"grad_norm": 0.4780759168718152, |
|
"learning_rate": 5.386949924127466e-07, |
|
"loss": 0.4462, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 2.6391061452513966, |
|
"grad_norm": 0.464930849547806, |
|
"learning_rate": 5.311077389984825e-07, |
|
"loss": 0.4334, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 2.6435754189944136, |
|
"grad_norm": 0.5002300455424397, |
|
"learning_rate": 5.235204855842186e-07, |
|
"loss": 0.4641, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.64804469273743, |
|
"grad_norm": 0.4901819784256759, |
|
"learning_rate": 5.159332321699545e-07, |
|
"loss": 0.4534, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 2.6525139664804467, |
|
"grad_norm": 0.48520187595124403, |
|
"learning_rate": 5.083459787556905e-07, |
|
"loss": 0.4246, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 2.6569832402234637, |
|
"grad_norm": 0.5093634792454178, |
|
"learning_rate": 5.007587253414264e-07, |
|
"loss": 0.458, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 2.6614525139664806, |
|
"grad_norm": 0.47795678253705354, |
|
"learning_rate": 4.931714719271624e-07, |
|
"loss": 0.4535, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 2.665921787709497, |
|
"grad_norm": 0.48340944166993405, |
|
"learning_rate": 4.855842185128983e-07, |
|
"loss": 0.4721, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 2.6703910614525137, |
|
"grad_norm": 0.47996848455446794, |
|
"learning_rate": 4.779969650986344e-07, |
|
"loss": 0.4718, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 2.6748603351955307, |
|
"grad_norm": 0.5064110255390989, |
|
"learning_rate": 4.7040971168437033e-07, |
|
"loss": 0.4232, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 2.6793296089385477, |
|
"grad_norm": 0.4824240302682651, |
|
"learning_rate": 4.6282245827010624e-07, |
|
"loss": 0.4307, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 2.683798882681564, |
|
"grad_norm": 0.4920800315955025, |
|
"learning_rate": 4.552352048558422e-07, |
|
"loss": 0.4435, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 2.688268156424581, |
|
"grad_norm": 0.48921504704563673, |
|
"learning_rate": 4.476479514415782e-07, |
|
"loss": 0.4409, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.6927374301675977, |
|
"grad_norm": 0.500357475284234, |
|
"learning_rate": 4.4006069802731414e-07, |
|
"loss": 0.4512, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 2.6972067039106147, |
|
"grad_norm": 0.4772853981005586, |
|
"learning_rate": 4.324734446130501e-07, |
|
"loss": 0.4305, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 2.701675977653631, |
|
"grad_norm": 0.5075964338612743, |
|
"learning_rate": 4.248861911987861e-07, |
|
"loss": 0.4285, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 2.706145251396648, |
|
"grad_norm": 0.4916721748305679, |
|
"learning_rate": 4.1729893778452204e-07, |
|
"loss": 0.4586, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 2.7106145251396647, |
|
"grad_norm": 0.4746003660659448, |
|
"learning_rate": 4.09711684370258e-07, |
|
"loss": 0.4385, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 2.7150837988826817, |
|
"grad_norm": 0.49004649736827927, |
|
"learning_rate": 4.021244309559939e-07, |
|
"loss": 0.4349, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 2.7195530726256982, |
|
"grad_norm": 0.5055169203153188, |
|
"learning_rate": 3.945371775417299e-07, |
|
"loss": 0.4741, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 2.724022346368715, |
|
"grad_norm": 0.5491416139361283, |
|
"learning_rate": 3.8694992412746586e-07, |
|
"loss": 0.4285, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 2.7284916201117317, |
|
"grad_norm": 0.4860669923591883, |
|
"learning_rate": 3.793626707132018e-07, |
|
"loss": 0.4147, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 2.7329608938547487, |
|
"grad_norm": 0.48537127261627433, |
|
"learning_rate": 3.7177541729893784e-07, |
|
"loss": 0.4538, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.7374301675977653, |
|
"grad_norm": 0.47557197343237195, |
|
"learning_rate": 3.641881638846738e-07, |
|
"loss": 0.454, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 2.7418994413407822, |
|
"grad_norm": 0.4844044051975442, |
|
"learning_rate": 3.566009104704098e-07, |
|
"loss": 0.4273, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 2.7463687150837988, |
|
"grad_norm": 0.5000172198487197, |
|
"learning_rate": 3.4901365705614574e-07, |
|
"loss": 0.4315, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 2.7508379888268157, |
|
"grad_norm": 0.4848986778566373, |
|
"learning_rate": 3.4142640364188166e-07, |
|
"loss": 0.4495, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 2.7553072625698323, |
|
"grad_norm": 0.48082262468066483, |
|
"learning_rate": 3.338391502276176e-07, |
|
"loss": 0.4185, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 2.7597765363128492, |
|
"grad_norm": 0.4865964770369407, |
|
"learning_rate": 3.262518968133536e-07, |
|
"loss": 0.4538, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 2.764245810055866, |
|
"grad_norm": 0.4815694373638968, |
|
"learning_rate": 3.1866464339908956e-07, |
|
"loss": 0.459, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 2.7687150837988828, |
|
"grad_norm": 0.49485301744734556, |
|
"learning_rate": 3.110773899848255e-07, |
|
"loss": 0.4527, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 2.7731843575418993, |
|
"grad_norm": 0.5214886405175438, |
|
"learning_rate": 3.034901365705615e-07, |
|
"loss": 0.4393, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 2.7776536312849163, |
|
"grad_norm": 0.4645817613806632, |
|
"learning_rate": 2.959028831562974e-07, |
|
"loss": 0.4723, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.782122905027933, |
|
"grad_norm": 0.4887881586065441, |
|
"learning_rate": 2.883156297420334e-07, |
|
"loss": 0.418, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 2.78659217877095, |
|
"grad_norm": 0.47562569136752975, |
|
"learning_rate": 2.807283763277694e-07, |
|
"loss": 0.4347, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 2.7910614525139668, |
|
"grad_norm": 0.4848933584055101, |
|
"learning_rate": 2.7314112291350536e-07, |
|
"loss": 0.4266, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 2.7955307262569833, |
|
"grad_norm": 0.48829866920710585, |
|
"learning_rate": 2.6555386949924127e-07, |
|
"loss": 0.4522, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 0.4776396492507318, |
|
"learning_rate": 2.5796661608497724e-07, |
|
"loss": 0.4293, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 2.804469273743017, |
|
"grad_norm": 0.48342626581685083, |
|
"learning_rate": 2.503793626707132e-07, |
|
"loss": 0.437, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 2.8089385474860338, |
|
"grad_norm": 0.47481545268800873, |
|
"learning_rate": 2.4279210925644917e-07, |
|
"loss": 0.4333, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 2.8134078212290503, |
|
"grad_norm": 0.5105857690012979, |
|
"learning_rate": 2.3520485584218516e-07, |
|
"loss": 0.4367, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 2.817877094972067, |
|
"grad_norm": 0.49397311695562557, |
|
"learning_rate": 2.276176024279211e-07, |
|
"loss": 0.4355, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 2.822346368715084, |
|
"grad_norm": 0.4766966322472669, |
|
"learning_rate": 2.2003034901365707e-07, |
|
"loss": 0.4478, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.826815642458101, |
|
"grad_norm": 0.47484165703817144, |
|
"learning_rate": 2.1244309559939304e-07, |
|
"loss": 0.4366, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 2.8312849162011173, |
|
"grad_norm": 0.47271019258961905, |
|
"learning_rate": 2.04855842185129e-07, |
|
"loss": 0.4473, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 2.835754189944134, |
|
"grad_norm": 0.4668544421054751, |
|
"learning_rate": 1.9726858877086494e-07, |
|
"loss": 0.4341, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 2.840223463687151, |
|
"grad_norm": 0.47716986376667936, |
|
"learning_rate": 1.896813353566009e-07, |
|
"loss": 0.4223, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 2.844692737430168, |
|
"grad_norm": 0.4939991882199546, |
|
"learning_rate": 1.820940819423369e-07, |
|
"loss": 0.4441, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 2.8491620111731844, |
|
"grad_norm": 0.47705877463455093, |
|
"learning_rate": 1.7450682852807287e-07, |
|
"loss": 0.485, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 2.853631284916201, |
|
"grad_norm": 0.49941990019846055, |
|
"learning_rate": 1.669195751138088e-07, |
|
"loss": 0.4701, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 2.858100558659218, |
|
"grad_norm": 0.47625945090740057, |
|
"learning_rate": 1.5933232169954478e-07, |
|
"loss": 0.454, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 2.862569832402235, |
|
"grad_norm": 0.5028271948555065, |
|
"learning_rate": 1.5174506828528074e-07, |
|
"loss": 0.4296, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 2.8670391061452514, |
|
"grad_norm": 0.49358352504671604, |
|
"learning_rate": 1.441578148710167e-07, |
|
"loss": 0.4324, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.871508379888268, |
|
"grad_norm": 0.48576223310259714, |
|
"learning_rate": 1.3657056145675268e-07, |
|
"loss": 0.4424, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 2.875977653631285, |
|
"grad_norm": 0.47971171549572994, |
|
"learning_rate": 1.2898330804248862e-07, |
|
"loss": 0.4555, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 2.880446927374302, |
|
"grad_norm": 0.49955392754214234, |
|
"learning_rate": 1.2139605462822459e-07, |
|
"loss": 0.4338, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 2.8849162011173184, |
|
"grad_norm": 0.47232821100099637, |
|
"learning_rate": 1.1380880121396055e-07, |
|
"loss": 0.4349, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 2.889385474860335, |
|
"grad_norm": 0.4834170945858496, |
|
"learning_rate": 1.0622154779969652e-07, |
|
"loss": 0.4564, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 2.893854748603352, |
|
"grad_norm": 0.48493984226629705, |
|
"learning_rate": 9.863429438543247e-08, |
|
"loss": 0.4384, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 2.898324022346369, |
|
"grad_norm": 0.4601302405573049, |
|
"learning_rate": 9.104704097116845e-08, |
|
"loss": 0.4512, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 2.9027932960893854, |
|
"grad_norm": 0.4768143158802154, |
|
"learning_rate": 8.34597875569044e-08, |
|
"loss": 0.4354, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 2.9072625698324024, |
|
"grad_norm": 0.49427262358096746, |
|
"learning_rate": 7.587253414264037e-08, |
|
"loss": 0.4259, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 2.911731843575419, |
|
"grad_norm": 0.4749814243858473, |
|
"learning_rate": 6.828528072837634e-08, |
|
"loss": 0.4445, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.916201117318436, |
|
"grad_norm": 0.48533516583361297, |
|
"learning_rate": 6.069802731411229e-08, |
|
"loss": 0.4799, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 2.9206703910614524, |
|
"grad_norm": 0.4839142899304728, |
|
"learning_rate": 5.311077389984826e-08, |
|
"loss": 0.4574, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 2.9251396648044694, |
|
"grad_norm": 0.486543073762813, |
|
"learning_rate": 4.5523520485584226e-08, |
|
"loss": 0.4446, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 2.929608938547486, |
|
"grad_norm": 0.4980109939724879, |
|
"learning_rate": 3.7936267071320186e-08, |
|
"loss": 0.4315, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 2.934078212290503, |
|
"grad_norm": 0.48840296818117557, |
|
"learning_rate": 3.0349013657056146e-08, |
|
"loss": 0.4318, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 2.9385474860335195, |
|
"grad_norm": 0.4903615270099699, |
|
"learning_rate": 2.2761760242792113e-08, |
|
"loss": 0.4357, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 2.9430167597765364, |
|
"grad_norm": 0.48850430880690693, |
|
"learning_rate": 1.5174506828528073e-08, |
|
"loss": 0.4334, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 2.947486033519553, |
|
"grad_norm": 0.4773488572878224, |
|
"learning_rate": 7.587253414264037e-09, |
|
"loss": 0.4281, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 2.95195530726257, |
|
"grad_norm": 0.4760606483764159, |
|
"learning_rate": 0.0, |
|
"loss": 0.4504, |
|
"step": 669 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 669, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 56, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 559881199288320.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|