function_calling_spectrum_L3.1_8B / trainer_state.json
williambarberjr's picture
Upload trainer_state.json with huggingface_hub
c41bc6e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.95195530726257,
"eval_steps": 112,
"global_step": 669,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004464285714285714,
"grad_norm": 5.096107617240597,
"learning_rate": 5.000000000000001e-07,
"loss": 0.81,
"step": 1
},
{
"epoch": 0.004464285714285714,
"eval_loss": 0.7919066548347473,
"eval_runtime": 32.2179,
"eval_samples_per_second": 82.873,
"eval_steps_per_second": 5.183,
"step": 1
},
{
"epoch": 0.008928571428571428,
"grad_norm": 4.9041704402978805,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.7839,
"step": 2
},
{
"epoch": 0.013392857142857142,
"grad_norm": 4.977917588021941,
"learning_rate": 1.5e-06,
"loss": 0.7745,
"step": 3
},
{
"epoch": 0.017857142857142856,
"grad_norm": 4.7373014121550705,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.7681,
"step": 4
},
{
"epoch": 0.022321428571428572,
"grad_norm": 4.563297129857777,
"learning_rate": 2.5e-06,
"loss": 0.7589,
"step": 5
},
{
"epoch": 0.026785714285714284,
"grad_norm": 3.5264108754930787,
"learning_rate": 3e-06,
"loss": 0.7629,
"step": 6
},
{
"epoch": 0.03125,
"grad_norm": 3.4126703272457166,
"learning_rate": 3.5e-06,
"loss": 0.7181,
"step": 7
},
{
"epoch": 0.03571428571428571,
"grad_norm": 3.6971429493758636,
"learning_rate": 4.000000000000001e-06,
"loss": 0.667,
"step": 8
},
{
"epoch": 0.04017857142857143,
"grad_norm": 1.940701767077778,
"learning_rate": 4.5e-06,
"loss": 0.6629,
"step": 9
},
{
"epoch": 0.044642857142857144,
"grad_norm": 12.40862535155878,
"learning_rate": 5e-06,
"loss": 0.6832,
"step": 10
},
{
"epoch": 0.049107142857142856,
"grad_norm": 2.8235547786373867,
"learning_rate": 4.992447129909366e-06,
"loss": 0.6213,
"step": 11
},
{
"epoch": 0.05357142857142857,
"grad_norm": 1.734602557718062,
"learning_rate": 4.984894259818732e-06,
"loss": 0.6279,
"step": 12
},
{
"epoch": 0.05803571428571429,
"grad_norm": 1.3840828717613172,
"learning_rate": 4.977341389728097e-06,
"loss": 0.6134,
"step": 13
},
{
"epoch": 0.0625,
"grad_norm": 1.1262172615918475,
"learning_rate": 4.969788519637463e-06,
"loss": 0.6029,
"step": 14
},
{
"epoch": 0.06696428571428571,
"grad_norm": 0.9263739512436016,
"learning_rate": 4.962235649546828e-06,
"loss": 0.6067,
"step": 15
},
{
"epoch": 0.07142857142857142,
"grad_norm": 0.8374584146721384,
"learning_rate": 4.954682779456194e-06,
"loss": 0.6014,
"step": 16
},
{
"epoch": 0.07589285714285714,
"grad_norm": 0.7757487058764371,
"learning_rate": 4.9471299093655595e-06,
"loss": 0.5906,
"step": 17
},
{
"epoch": 0.08035714285714286,
"grad_norm": 0.6531787929864147,
"learning_rate": 4.939577039274925e-06,
"loss": 0.5589,
"step": 18
},
{
"epoch": 0.08482142857142858,
"grad_norm": 0.7223853317191319,
"learning_rate": 4.93202416918429e-06,
"loss": 0.5877,
"step": 19
},
{
"epoch": 0.08928571428571429,
"grad_norm": 0.6793848952870568,
"learning_rate": 4.924471299093656e-06,
"loss": 0.6075,
"step": 20
},
{
"epoch": 0.09375,
"grad_norm": 0.65689704768199,
"learning_rate": 4.9169184290030215e-06,
"loss": 0.5659,
"step": 21
},
{
"epoch": 0.09821428571428571,
"grad_norm": 0.6553424447930819,
"learning_rate": 4.909365558912387e-06,
"loss": 0.5728,
"step": 22
},
{
"epoch": 0.10267857142857142,
"grad_norm": 0.6095768945247507,
"learning_rate": 4.901812688821753e-06,
"loss": 0.5716,
"step": 23
},
{
"epoch": 0.10714285714285714,
"grad_norm": 0.5874011280839848,
"learning_rate": 4.894259818731118e-06,
"loss": 0.5736,
"step": 24
},
{
"epoch": 0.11160714285714286,
"grad_norm": 0.642646780124128,
"learning_rate": 4.8867069486404835e-06,
"loss": 0.5889,
"step": 25
},
{
"epoch": 0.11607142857142858,
"grad_norm": 0.5585168252549827,
"learning_rate": 4.879154078549849e-06,
"loss": 0.5761,
"step": 26
},
{
"epoch": 0.12053571428571429,
"grad_norm": 0.4948199084353006,
"learning_rate": 4.871601208459215e-06,
"loss": 0.545,
"step": 27
},
{
"epoch": 0.125,
"grad_norm": 0.580699538693282,
"learning_rate": 4.864048338368581e-06,
"loss": 0.5874,
"step": 28
},
{
"epoch": 0.12946428571428573,
"grad_norm": 0.5483036703790811,
"learning_rate": 4.8564954682779455e-06,
"loss": 0.553,
"step": 29
},
{
"epoch": 0.13392857142857142,
"grad_norm": 0.5207516065535861,
"learning_rate": 4.848942598187312e-06,
"loss": 0.54,
"step": 30
},
{
"epoch": 0.13839285714285715,
"grad_norm": 0.5453894595978196,
"learning_rate": 4.841389728096677e-06,
"loss": 0.5779,
"step": 31
},
{
"epoch": 0.14285714285714285,
"grad_norm": 0.5223897707340578,
"learning_rate": 4.833836858006043e-06,
"loss": 0.5518,
"step": 32
},
{
"epoch": 0.14732142857142858,
"grad_norm": 0.5473812522488812,
"learning_rate": 4.826283987915408e-06,
"loss": 0.5831,
"step": 33
},
{
"epoch": 0.15178571428571427,
"grad_norm": 0.5437546013329395,
"learning_rate": 4.818731117824774e-06,
"loss": 0.5884,
"step": 34
},
{
"epoch": 0.15625,
"grad_norm": 0.5305579139277316,
"learning_rate": 4.81117824773414e-06,
"loss": 0.5398,
"step": 35
},
{
"epoch": 0.16071428571428573,
"grad_norm": 0.5258419287632591,
"learning_rate": 4.803625377643505e-06,
"loss": 0.5593,
"step": 36
},
{
"epoch": 0.16517857142857142,
"grad_norm": 0.5158851354579528,
"learning_rate": 4.79607250755287e-06,
"loss": 0.5564,
"step": 37
},
{
"epoch": 0.16964285714285715,
"grad_norm": 0.4824844062486759,
"learning_rate": 4.788519637462236e-06,
"loss": 0.5616,
"step": 38
},
{
"epoch": 0.17410714285714285,
"grad_norm": 0.5173226727977146,
"learning_rate": 4.780966767371602e-06,
"loss": 0.5674,
"step": 39
},
{
"epoch": 0.17857142857142858,
"grad_norm": 0.4966629400190002,
"learning_rate": 4.773413897280967e-06,
"loss": 0.5833,
"step": 40
},
{
"epoch": 0.18303571428571427,
"grad_norm": 0.48186435126437244,
"learning_rate": 4.765861027190333e-06,
"loss": 0.5571,
"step": 41
},
{
"epoch": 0.1875,
"grad_norm": 0.5029558558326376,
"learning_rate": 4.758308157099698e-06,
"loss": 0.5358,
"step": 42
},
{
"epoch": 0.19196428571428573,
"grad_norm": 0.5249190794688006,
"learning_rate": 4.750755287009064e-06,
"loss": 0.5542,
"step": 43
},
{
"epoch": 0.19642857142857142,
"grad_norm": 0.4622817114582695,
"learning_rate": 4.743202416918429e-06,
"loss": 0.54,
"step": 44
},
{
"epoch": 0.20089285714285715,
"grad_norm": 0.49789824735563454,
"learning_rate": 4.735649546827795e-06,
"loss": 0.5308,
"step": 45
},
{
"epoch": 0.20535714285714285,
"grad_norm": 0.51199116739647,
"learning_rate": 4.728096676737161e-06,
"loss": 0.5784,
"step": 46
},
{
"epoch": 0.20982142857142858,
"grad_norm": 0.47152096107780506,
"learning_rate": 4.720543806646526e-06,
"loss": 0.5654,
"step": 47
},
{
"epoch": 0.21428571428571427,
"grad_norm": 0.5233581267403502,
"learning_rate": 4.712990936555891e-06,
"loss": 0.5727,
"step": 48
},
{
"epoch": 0.21875,
"grad_norm": 0.517908769266997,
"learning_rate": 4.705438066465257e-06,
"loss": 0.5258,
"step": 49
},
{
"epoch": 0.22321428571428573,
"grad_norm": 0.5103782222366013,
"learning_rate": 4.697885196374623e-06,
"loss": 0.564,
"step": 50
},
{
"epoch": 0.22767857142857142,
"grad_norm": 0.5894196264675328,
"learning_rate": 4.6903323262839885e-06,
"loss": 0.5355,
"step": 51
},
{
"epoch": 0.23214285714285715,
"grad_norm": 0.48136408513935,
"learning_rate": 4.682779456193353e-06,
"loss": 0.5714,
"step": 52
},
{
"epoch": 0.23660714285714285,
"grad_norm": 0.5996790928794941,
"learning_rate": 4.67522658610272e-06,
"loss": 0.5385,
"step": 53
},
{
"epoch": 0.24107142857142858,
"grad_norm": 0.5231061994660855,
"learning_rate": 4.667673716012085e-06,
"loss": 0.5429,
"step": 54
},
{
"epoch": 0.24553571428571427,
"grad_norm": 0.5416056944426403,
"learning_rate": 4.6601208459214505e-06,
"loss": 0.5423,
"step": 55
},
{
"epoch": 0.25,
"grad_norm": 0.5734105721889117,
"learning_rate": 4.652567975830816e-06,
"loss": 0.5756,
"step": 56
},
{
"epoch": 0.2544642857142857,
"grad_norm": 0.5318541491738474,
"learning_rate": 4.645015105740182e-06,
"loss": 0.5486,
"step": 57
},
{
"epoch": 0.25892857142857145,
"grad_norm": 0.4839987879630258,
"learning_rate": 4.637462235649548e-06,
"loss": 0.5517,
"step": 58
},
{
"epoch": 0.26339285714285715,
"grad_norm": 0.5843387907103592,
"learning_rate": 4.6299093655589125e-06,
"loss": 0.5288,
"step": 59
},
{
"epoch": 0.26785714285714285,
"grad_norm": 0.5496996598474941,
"learning_rate": 4.622356495468278e-06,
"loss": 0.542,
"step": 60
},
{
"epoch": 0.27232142857142855,
"grad_norm": 0.5192508705264864,
"learning_rate": 4.614803625377644e-06,
"loss": 0.5291,
"step": 61
},
{
"epoch": 0.2767857142857143,
"grad_norm": 0.5918952697319948,
"learning_rate": 4.60725075528701e-06,
"loss": 0.5229,
"step": 62
},
{
"epoch": 0.28125,
"grad_norm": 0.5914173604947244,
"learning_rate": 4.5996978851963745e-06,
"loss": 0.5754,
"step": 63
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.5298896547584293,
"learning_rate": 4.592145015105741e-06,
"loss": 0.5687,
"step": 64
},
{
"epoch": 0.29017857142857145,
"grad_norm": 0.5671535875314645,
"learning_rate": 4.584592145015106e-06,
"loss": 0.5513,
"step": 65
},
{
"epoch": 0.29464285714285715,
"grad_norm": 0.5234306194196137,
"learning_rate": 4.577039274924472e-06,
"loss": 0.5226,
"step": 66
},
{
"epoch": 0.29910714285714285,
"grad_norm": 0.4959853502285333,
"learning_rate": 4.569486404833837e-06,
"loss": 0.5335,
"step": 67
},
{
"epoch": 0.30357142857142855,
"grad_norm": 0.47729026424742405,
"learning_rate": 4.561933534743202e-06,
"loss": 0.5248,
"step": 68
},
{
"epoch": 0.3080357142857143,
"grad_norm": 0.47294297292566195,
"learning_rate": 4.554380664652569e-06,
"loss": 0.5588,
"step": 69
},
{
"epoch": 0.3125,
"grad_norm": 0.5275786368006364,
"learning_rate": 4.5468277945619336e-06,
"loss": 0.5483,
"step": 70
},
{
"epoch": 0.3169642857142857,
"grad_norm": 0.5057705930103896,
"learning_rate": 4.539274924471299e-06,
"loss": 0.5236,
"step": 71
},
{
"epoch": 0.32142857142857145,
"grad_norm": 0.49625696780589473,
"learning_rate": 4.531722054380665e-06,
"loss": 0.5221,
"step": 72
},
{
"epoch": 0.32589285714285715,
"grad_norm": 0.49548627058675154,
"learning_rate": 4.524169184290031e-06,
"loss": 0.5176,
"step": 73
},
{
"epoch": 0.33035714285714285,
"grad_norm": 0.5011897814029462,
"learning_rate": 4.516616314199396e-06,
"loss": 0.5431,
"step": 74
},
{
"epoch": 0.33482142857142855,
"grad_norm": 0.5168007116628185,
"learning_rate": 4.509063444108761e-06,
"loss": 0.5289,
"step": 75
},
{
"epoch": 0.3392857142857143,
"grad_norm": 0.5199488476439844,
"learning_rate": 4.501510574018128e-06,
"loss": 0.5437,
"step": 76
},
{
"epoch": 0.34375,
"grad_norm": 0.5776348374651958,
"learning_rate": 4.493957703927493e-06,
"loss": 0.5424,
"step": 77
},
{
"epoch": 0.3482142857142857,
"grad_norm": 0.5785274375251012,
"learning_rate": 4.486404833836858e-06,
"loss": 0.5252,
"step": 78
},
{
"epoch": 0.35267857142857145,
"grad_norm": 0.5624252036929408,
"learning_rate": 4.478851963746224e-06,
"loss": 0.5502,
"step": 79
},
{
"epoch": 0.35714285714285715,
"grad_norm": 0.5630273292563954,
"learning_rate": 4.47129909365559e-06,
"loss": 0.5249,
"step": 80
},
{
"epoch": 0.36160714285714285,
"grad_norm": 0.5229643469397499,
"learning_rate": 4.463746223564955e-06,
"loss": 0.572,
"step": 81
},
{
"epoch": 0.36607142857142855,
"grad_norm": 0.5401109169653766,
"learning_rate": 4.45619335347432e-06,
"loss": 0.5174,
"step": 82
},
{
"epoch": 0.3705357142857143,
"grad_norm": 0.572853251261138,
"learning_rate": 4.448640483383686e-06,
"loss": 0.5336,
"step": 83
},
{
"epoch": 0.375,
"grad_norm": 0.49833694539628726,
"learning_rate": 4.441087613293052e-06,
"loss": 0.5205,
"step": 84
},
{
"epoch": 0.3794642857142857,
"grad_norm": 0.5167418841184229,
"learning_rate": 4.4335347432024175e-06,
"loss": 0.5405,
"step": 85
},
{
"epoch": 0.38392857142857145,
"grad_norm": 0.5243612918158316,
"learning_rate": 4.425981873111782e-06,
"loss": 0.5204,
"step": 86
},
{
"epoch": 0.38839285714285715,
"grad_norm": 0.5159624173670619,
"learning_rate": 4.418429003021149e-06,
"loss": 0.5315,
"step": 87
},
{
"epoch": 0.39285714285714285,
"grad_norm": 0.5227918139330733,
"learning_rate": 4.410876132930514e-06,
"loss": 0.5418,
"step": 88
},
{
"epoch": 0.39732142857142855,
"grad_norm": 0.48045718180033986,
"learning_rate": 4.4033232628398795e-06,
"loss": 0.506,
"step": 89
},
{
"epoch": 0.4017857142857143,
"grad_norm": 0.5791314708394477,
"learning_rate": 4.395770392749245e-06,
"loss": 0.596,
"step": 90
},
{
"epoch": 0.40625,
"grad_norm": 0.6076262289549809,
"learning_rate": 4.38821752265861e-06,
"loss": 0.5312,
"step": 91
},
{
"epoch": 0.4107142857142857,
"grad_norm": 0.4780377894473343,
"learning_rate": 4.380664652567977e-06,
"loss": 0.549,
"step": 92
},
{
"epoch": 0.41517857142857145,
"grad_norm": 0.5549693296410254,
"learning_rate": 4.3731117824773415e-06,
"loss": 0.5159,
"step": 93
},
{
"epoch": 0.41964285714285715,
"grad_norm": 0.5097636365128929,
"learning_rate": 4.365558912386707e-06,
"loss": 0.5612,
"step": 94
},
{
"epoch": 0.42410714285714285,
"grad_norm": 0.5021825140627123,
"learning_rate": 4.358006042296073e-06,
"loss": 0.5653,
"step": 95
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.537945728158161,
"learning_rate": 4.350453172205439e-06,
"loss": 0.528,
"step": 96
},
{
"epoch": 0.4330357142857143,
"grad_norm": 0.4846132130879552,
"learning_rate": 4.342900302114804e-06,
"loss": 0.5395,
"step": 97
},
{
"epoch": 0.4375,
"grad_norm": 0.5206116049646974,
"learning_rate": 4.335347432024169e-06,
"loss": 0.5445,
"step": 98
},
{
"epoch": 0.4419642857142857,
"grad_norm": 0.551262765339688,
"learning_rate": 4.327794561933535e-06,
"loss": 0.5346,
"step": 99
},
{
"epoch": 0.44642857142857145,
"grad_norm": 0.4984265695742978,
"learning_rate": 4.3202416918429006e-06,
"loss": 0.5161,
"step": 100
},
{
"epoch": 0.45089285714285715,
"grad_norm": 0.49943465819192145,
"learning_rate": 4.312688821752266e-06,
"loss": 0.5164,
"step": 101
},
{
"epoch": 0.45535714285714285,
"grad_norm": 0.48826983665318036,
"learning_rate": 4.305135951661632e-06,
"loss": 0.552,
"step": 102
},
{
"epoch": 0.45982142857142855,
"grad_norm": 0.5320671379410685,
"learning_rate": 4.297583081570998e-06,
"loss": 0.5512,
"step": 103
},
{
"epoch": 0.4642857142857143,
"grad_norm": 0.5027439491810191,
"learning_rate": 4.2900302114803626e-06,
"loss": 0.5728,
"step": 104
},
{
"epoch": 0.46875,
"grad_norm": 0.4875858103308158,
"learning_rate": 4.282477341389728e-06,
"loss": 0.5306,
"step": 105
},
{
"epoch": 0.4732142857142857,
"grad_norm": 0.5205285237402679,
"learning_rate": 4.274924471299094e-06,
"loss": 0.5274,
"step": 106
},
{
"epoch": 0.47767857142857145,
"grad_norm": 0.47059444782816534,
"learning_rate": 4.26737160120846e-06,
"loss": 0.5092,
"step": 107
},
{
"epoch": 0.48214285714285715,
"grad_norm": 0.4998056822147382,
"learning_rate": 4.259818731117825e-06,
"loss": 0.5481,
"step": 108
},
{
"epoch": 0.48660714285714285,
"grad_norm": 0.5403989848602658,
"learning_rate": 4.25226586102719e-06,
"loss": 0.518,
"step": 109
},
{
"epoch": 0.49107142857142855,
"grad_norm": 0.48519680312136604,
"learning_rate": 4.244712990936557e-06,
"loss": 0.5522,
"step": 110
},
{
"epoch": 0.4955357142857143,
"grad_norm": 0.5655590239975465,
"learning_rate": 4.237160120845922e-06,
"loss": 0.526,
"step": 111
},
{
"epoch": 0.5,
"grad_norm": 0.49558151987556875,
"learning_rate": 4.229607250755287e-06,
"loss": 0.5275,
"step": 112
},
{
"epoch": 0.5,
"eval_loss": 0.5346882343292236,
"eval_runtime": 32.2399,
"eval_samples_per_second": 82.817,
"eval_steps_per_second": 5.18,
"step": 112
},
{
"epoch": 0.5050279329608939,
"grad_norm": 0.5662939933016762,
"learning_rate": 4.218512898330804e-06,
"loss": 0.515,
"step": 113
},
{
"epoch": 0.5094972067039106,
"grad_norm": 0.5225180760232837,
"learning_rate": 4.2109256449165405e-06,
"loss": 0.5038,
"step": 114
},
{
"epoch": 0.5139664804469274,
"grad_norm": 0.5377410752157915,
"learning_rate": 4.203338391502276e-06,
"loss": 0.5043,
"step": 115
},
{
"epoch": 0.5184357541899441,
"grad_norm": 0.5915094303426857,
"learning_rate": 4.195751138088012e-06,
"loss": 0.5115,
"step": 116
},
{
"epoch": 0.5229050279329609,
"grad_norm": 0.5079762717076601,
"learning_rate": 4.1881638846737485e-06,
"loss": 0.5491,
"step": 117
},
{
"epoch": 0.5273743016759777,
"grad_norm": 0.5933329970849119,
"learning_rate": 4.180576631259484e-06,
"loss": 0.5095,
"step": 118
},
{
"epoch": 0.5318435754189944,
"grad_norm": 0.5053074762991974,
"learning_rate": 4.17298937784522e-06,
"loss": 0.4881,
"step": 119
},
{
"epoch": 0.5363128491620112,
"grad_norm": 0.6515342691068152,
"learning_rate": 4.1654021244309564e-06,
"loss": 0.5069,
"step": 120
},
{
"epoch": 0.5407821229050279,
"grad_norm": 0.5774345075195761,
"learning_rate": 4.157814871016692e-06,
"loss": 0.5007,
"step": 121
},
{
"epoch": 0.5452513966480447,
"grad_norm": 0.6308991657157696,
"learning_rate": 4.150227617602428e-06,
"loss": 0.5083,
"step": 122
},
{
"epoch": 0.5497206703910614,
"grad_norm": 0.5258198371155504,
"learning_rate": 4.142640364188164e-06,
"loss": 0.5103,
"step": 123
},
{
"epoch": 0.5541899441340782,
"grad_norm": 0.5629243370057063,
"learning_rate": 4.1350531107739e-06,
"loss": 0.5016,
"step": 124
},
{
"epoch": 0.5586592178770949,
"grad_norm": 0.5440095922230602,
"learning_rate": 4.127465857359636e-06,
"loss": 0.491,
"step": 125
},
{
"epoch": 0.5631284916201117,
"grad_norm": 0.5126793634050918,
"learning_rate": 4.119878603945372e-06,
"loss": 0.5239,
"step": 126
},
{
"epoch": 0.5675977653631284,
"grad_norm": 0.5476299968831635,
"learning_rate": 4.112291350531108e-06,
"loss": 0.5176,
"step": 127
},
{
"epoch": 0.5720670391061452,
"grad_norm": 0.5570481457317527,
"learning_rate": 4.104704097116844e-06,
"loss": 0.5084,
"step": 128
},
{
"epoch": 0.576536312849162,
"grad_norm": 0.49480777896006917,
"learning_rate": 4.09711684370258e-06,
"loss": 0.4966,
"step": 129
},
{
"epoch": 0.5810055865921788,
"grad_norm": 0.5089889598386482,
"learning_rate": 4.089529590288316e-06,
"loss": 0.5059,
"step": 130
},
{
"epoch": 0.5854748603351956,
"grad_norm": 0.4954780811768282,
"learning_rate": 4.081942336874052e-06,
"loss": 0.5166,
"step": 131
},
{
"epoch": 0.5899441340782123,
"grad_norm": 0.4809684003541764,
"learning_rate": 4.074355083459787e-06,
"loss": 0.4857,
"step": 132
},
{
"epoch": 0.5944134078212291,
"grad_norm": 0.5009748980192239,
"learning_rate": 4.066767830045524e-06,
"loss": 0.523,
"step": 133
},
{
"epoch": 0.5988826815642458,
"grad_norm": 0.48320950023108783,
"learning_rate": 4.05918057663126e-06,
"loss": 0.4956,
"step": 134
},
{
"epoch": 0.6033519553072626,
"grad_norm": 0.49399181892388616,
"learning_rate": 4.051593323216995e-06,
"loss": 0.5163,
"step": 135
},
{
"epoch": 0.6078212290502794,
"grad_norm": 0.4821527638840584,
"learning_rate": 4.044006069802732e-06,
"loss": 0.4981,
"step": 136
},
{
"epoch": 0.6122905027932961,
"grad_norm": 0.5099314085164022,
"learning_rate": 4.036418816388468e-06,
"loss": 0.5121,
"step": 137
},
{
"epoch": 0.6167597765363129,
"grad_norm": 0.5323495276796993,
"learning_rate": 4.028831562974203e-06,
"loss": 0.5074,
"step": 138
},
{
"epoch": 0.6212290502793296,
"grad_norm": 0.48851533865778524,
"learning_rate": 4.02124430955994e-06,
"loss": 0.52,
"step": 139
},
{
"epoch": 0.6256983240223464,
"grad_norm": 0.5281525574471102,
"learning_rate": 4.0136570561456756e-06,
"loss": 0.5039,
"step": 140
},
{
"epoch": 0.6301675977653631,
"grad_norm": 0.5279406819094602,
"learning_rate": 4.006069802731411e-06,
"loss": 0.4933,
"step": 141
},
{
"epoch": 0.6346368715083799,
"grad_norm": 0.45822529010842367,
"learning_rate": 3.998482549317148e-06,
"loss": 0.5151,
"step": 142
},
{
"epoch": 0.6391061452513966,
"grad_norm": 0.5349296018392958,
"learning_rate": 3.9908952959028835e-06,
"loss": 0.4983,
"step": 143
},
{
"epoch": 0.6435754189944134,
"grad_norm": 0.5316574655817645,
"learning_rate": 3.983308042488619e-06,
"loss": 0.5344,
"step": 144
},
{
"epoch": 0.6480446927374302,
"grad_norm": 0.5107562456487705,
"learning_rate": 3.975720789074356e-06,
"loss": 0.52,
"step": 145
},
{
"epoch": 0.6525139664804469,
"grad_norm": 0.5072583855988344,
"learning_rate": 3.9681335356600915e-06,
"loss": 0.4923,
"step": 146
},
{
"epoch": 0.6569832402234637,
"grad_norm": 0.5836000010606872,
"learning_rate": 3.960546282245827e-06,
"loss": 0.5295,
"step": 147
},
{
"epoch": 0.6614525139664804,
"grad_norm": 0.49884721978842206,
"learning_rate": 3.952959028831564e-06,
"loss": 0.5216,
"step": 148
},
{
"epoch": 0.6659217877094972,
"grad_norm": 0.5120991222561042,
"learning_rate": 3.945371775417299e-06,
"loss": 0.5414,
"step": 149
},
{
"epoch": 0.6703910614525139,
"grad_norm": 0.49838481084559894,
"learning_rate": 3.937784522003035e-06,
"loss": 0.5381,
"step": 150
},
{
"epoch": 0.6748603351955307,
"grad_norm": 0.5398799415397277,
"learning_rate": 3.930197268588772e-06,
"loss": 0.4896,
"step": 151
},
{
"epoch": 0.6793296089385474,
"grad_norm": 0.565748454568293,
"learning_rate": 3.922610015174507e-06,
"loss": 0.4985,
"step": 152
},
{
"epoch": 0.6837988826815642,
"grad_norm": 0.5986782671726579,
"learning_rate": 3.915022761760243e-06,
"loss": 0.5125,
"step": 153
},
{
"epoch": 0.6882681564245811,
"grad_norm": 0.5290750194980306,
"learning_rate": 3.907435508345979e-06,
"loss": 0.5078,
"step": 154
},
{
"epoch": 0.6927374301675978,
"grad_norm": 0.5706479438251948,
"learning_rate": 3.899848254931715e-06,
"loss": 0.5205,
"step": 155
},
{
"epoch": 0.6972067039106146,
"grad_norm": 0.5173864500214489,
"learning_rate": 3.892261001517451e-06,
"loss": 0.4988,
"step": 156
},
{
"epoch": 0.7016759776536313,
"grad_norm": 0.4920045809108581,
"learning_rate": 3.884673748103187e-06,
"loss": 0.4954,
"step": 157
},
{
"epoch": 0.7061452513966481,
"grad_norm": 0.5237901072069291,
"learning_rate": 3.877086494688923e-06,
"loss": 0.5253,
"step": 158
},
{
"epoch": 0.7106145251396648,
"grad_norm": 0.5212795514388354,
"learning_rate": 3.869499241274659e-06,
"loss": 0.5029,
"step": 159
},
{
"epoch": 0.7150837988826816,
"grad_norm": 0.48658569533086143,
"learning_rate": 3.861911987860395e-06,
"loss": 0.5023,
"step": 160
},
{
"epoch": 0.7195530726256983,
"grad_norm": 0.5145837815914728,
"learning_rate": 3.854324734446131e-06,
"loss": 0.5436,
"step": 161
},
{
"epoch": 0.7240223463687151,
"grad_norm": 0.49182827281373437,
"learning_rate": 3.846737481031867e-06,
"loss": 0.4948,
"step": 162
},
{
"epoch": 0.7284916201117319,
"grad_norm": 0.5113640333759738,
"learning_rate": 3.839150227617603e-06,
"loss": 0.4778,
"step": 163
},
{
"epoch": 0.7329608938547486,
"grad_norm": 0.47675402744999507,
"learning_rate": 3.831562974203339e-06,
"loss": 0.5182,
"step": 164
},
{
"epoch": 0.7374301675977654,
"grad_norm": 0.5214515656344721,
"learning_rate": 3.823975720789075e-06,
"loss": 0.5204,
"step": 165
},
{
"epoch": 0.7418994413407821,
"grad_norm": 0.5130606167326404,
"learning_rate": 3.816388467374811e-06,
"loss": 0.492,
"step": 166
},
{
"epoch": 0.7463687150837989,
"grad_norm": 0.49678736596959705,
"learning_rate": 3.8088012139605467e-06,
"loss": 0.4997,
"step": 167
},
{
"epoch": 0.7508379888268156,
"grad_norm": 0.5275415964748987,
"learning_rate": 3.801213960546283e-06,
"loss": 0.5163,
"step": 168
},
{
"epoch": 0.7553072625698324,
"grad_norm": 0.5323621998896121,
"learning_rate": 3.7936267071320185e-06,
"loss": 0.4846,
"step": 169
},
{
"epoch": 0.7597765363128491,
"grad_norm": 0.5302675218909635,
"learning_rate": 3.7860394537177547e-06,
"loss": 0.5193,
"step": 170
},
{
"epoch": 0.7642458100558659,
"grad_norm": 0.5097255453283106,
"learning_rate": 3.778452200303491e-06,
"loss": 0.5248,
"step": 171
},
{
"epoch": 0.7687150837988826,
"grad_norm": 0.5221084637587756,
"learning_rate": 3.7708649468892265e-06,
"loss": 0.5217,
"step": 172
},
{
"epoch": 0.7731843575418994,
"grad_norm": 0.5823307367451829,
"learning_rate": 3.7632776934749626e-06,
"loss": 0.5057,
"step": 173
},
{
"epoch": 0.7776536312849162,
"grad_norm": 0.48586162675121547,
"learning_rate": 3.7556904400606987e-06,
"loss": 0.5372,
"step": 174
},
{
"epoch": 0.7821229050279329,
"grad_norm": 0.5099958749272927,
"learning_rate": 3.748103186646434e-06,
"loss": 0.4844,
"step": 175
},
{
"epoch": 0.7865921787709497,
"grad_norm": 0.5221594966770278,
"learning_rate": 3.74051593323217e-06,
"loss": 0.4968,
"step": 176
},
{
"epoch": 0.7910614525139665,
"grad_norm": 0.5392129841421105,
"learning_rate": 3.732928679817906e-06,
"loss": 0.4906,
"step": 177
},
{
"epoch": 0.7955307262569833,
"grad_norm": 0.5239629013619267,
"learning_rate": 3.725341426403642e-06,
"loss": 0.5172,
"step": 178
},
{
"epoch": 0.8,
"grad_norm": 0.537861417175095,
"learning_rate": 3.717754172989378e-06,
"loss": 0.4914,
"step": 179
},
{
"epoch": 0.8044692737430168,
"grad_norm": 0.5001204419984062,
"learning_rate": 3.710166919575114e-06,
"loss": 0.5014,
"step": 180
},
{
"epoch": 0.8089385474860336,
"grad_norm": 0.5243758405421232,
"learning_rate": 3.70257966616085e-06,
"loss": 0.4979,
"step": 181
},
{
"epoch": 0.8134078212290503,
"grad_norm": 0.507560920806033,
"learning_rate": 3.6949924127465856e-06,
"loss": 0.5003,
"step": 182
},
{
"epoch": 0.8178770949720671,
"grad_norm": 0.5623627396713589,
"learning_rate": 3.6874051593323218e-06,
"loss": 0.5011,
"step": 183
},
{
"epoch": 0.8223463687150838,
"grad_norm": 0.5225320984203384,
"learning_rate": 3.679817905918058e-06,
"loss": 0.5117,
"step": 184
},
{
"epoch": 0.8268156424581006,
"grad_norm": 0.5371208831947621,
"learning_rate": 3.6722306525037936e-06,
"loss": 0.5028,
"step": 185
},
{
"epoch": 0.8312849162011173,
"grad_norm": 0.5412513689241548,
"learning_rate": 3.6646433990895297e-06,
"loss": 0.5111,
"step": 186
},
{
"epoch": 0.8357541899441341,
"grad_norm": 0.49083593539874787,
"learning_rate": 3.657056145675266e-06,
"loss": 0.5002,
"step": 187
},
{
"epoch": 0.8402234636871508,
"grad_norm": 0.5144646076400369,
"learning_rate": 3.6494688922610015e-06,
"loss": 0.4875,
"step": 188
},
{
"epoch": 0.8446927374301676,
"grad_norm": 0.5385458846653849,
"learning_rate": 3.6418816388467377e-06,
"loss": 0.5072,
"step": 189
},
{
"epoch": 0.8491620111731844,
"grad_norm": 0.48088002744673064,
"learning_rate": 3.6342943854324738e-06,
"loss": 0.549,
"step": 190
},
{
"epoch": 0.8536312849162011,
"grad_norm": 0.563960197772806,
"learning_rate": 3.6267071320182095e-06,
"loss": 0.5344,
"step": 191
},
{
"epoch": 0.8581005586592179,
"grad_norm": 0.5235289059912077,
"learning_rate": 3.6191198786039456e-06,
"loss": 0.5172,
"step": 192
},
{
"epoch": 0.8625698324022346,
"grad_norm": 0.5059904029741168,
"learning_rate": 3.6115326251896813e-06,
"loss": 0.4908,
"step": 193
},
{
"epoch": 0.8670391061452514,
"grad_norm": 0.507980780742306,
"learning_rate": 3.6039453717754174e-06,
"loss": 0.4951,
"step": 194
},
{
"epoch": 0.8715083798882681,
"grad_norm": 0.5552145890763519,
"learning_rate": 3.5963581183611536e-06,
"loss": 0.505,
"step": 195
},
{
"epoch": 0.8759776536312849,
"grad_norm": 0.5379367693029108,
"learning_rate": 3.5887708649468893e-06,
"loss": 0.52,
"step": 196
},
{
"epoch": 0.8804469273743016,
"grad_norm": 0.5813281708032492,
"learning_rate": 3.5811836115326254e-06,
"loss": 0.4966,
"step": 197
},
{
"epoch": 0.8849162011173184,
"grad_norm": 0.5138017935214574,
"learning_rate": 3.5735963581183615e-06,
"loss": 0.495,
"step": 198
},
{
"epoch": 0.8893854748603351,
"grad_norm": 0.5116405158423362,
"learning_rate": 3.5660091047040972e-06,
"loss": 0.519,
"step": 199
},
{
"epoch": 0.8938547486033519,
"grad_norm": 0.5696283061922035,
"learning_rate": 3.5584218512898333e-06,
"loss": 0.5002,
"step": 200
},
{
"epoch": 0.8983240223463688,
"grad_norm": 0.5541549121266046,
"learning_rate": 3.5508345978755695e-06,
"loss": 0.5125,
"step": 201
},
{
"epoch": 0.9027932960893855,
"grad_norm": 0.5017863683272789,
"learning_rate": 3.543247344461305e-06,
"loss": 0.499,
"step": 202
},
{
"epoch": 0.9072625698324023,
"grad_norm": 0.5776107100347877,
"learning_rate": 3.5356600910470413e-06,
"loss": 0.4883,
"step": 203
},
{
"epoch": 0.911731843575419,
"grad_norm": 0.5350847562000018,
"learning_rate": 3.528072837632777e-06,
"loss": 0.5052,
"step": 204
},
{
"epoch": 0.9162011173184358,
"grad_norm": 0.5012614076658582,
"learning_rate": 3.520485584218513e-06,
"loss": 0.5422,
"step": 205
},
{
"epoch": 0.9206703910614525,
"grad_norm": 0.6107847324258918,
"learning_rate": 3.5128983308042493e-06,
"loss": 0.5204,
"step": 206
},
{
"epoch": 0.9251396648044693,
"grad_norm": 0.49314942553653257,
"learning_rate": 3.505311077389985e-06,
"loss": 0.5074,
"step": 207
},
{
"epoch": 0.929608938547486,
"grad_norm": 0.5340833706807174,
"learning_rate": 3.497723823975721e-06,
"loss": 0.4936,
"step": 208
},
{
"epoch": 0.9340782122905028,
"grad_norm": 0.5342137283430334,
"learning_rate": 3.490136570561457e-06,
"loss": 0.4945,
"step": 209
},
{
"epoch": 0.9385474860335196,
"grad_norm": 0.5301769352005536,
"learning_rate": 3.482549317147193e-06,
"loss": 0.4963,
"step": 210
},
{
"epoch": 0.9430167597765363,
"grad_norm": 0.5455545680194068,
"learning_rate": 3.474962063732929e-06,
"loss": 0.4931,
"step": 211
},
{
"epoch": 0.9474860335195531,
"grad_norm": 0.5282021847913991,
"learning_rate": 3.467374810318665e-06,
"loss": 0.4894,
"step": 212
},
{
"epoch": 0.9519553072625698,
"grad_norm": 0.5341400870442735,
"learning_rate": 3.459787556904401e-06,
"loss": 0.5102,
"step": 213
},
{
"epoch": 0.9564245810055866,
"grad_norm": 0.5313633005763995,
"learning_rate": 3.452200303490137e-06,
"loss": 0.501,
"step": 214
},
{
"epoch": 0.9608938547486033,
"grad_norm": 0.5774555510069799,
"learning_rate": 3.4446130500758727e-06,
"loss": 0.5021,
"step": 215
},
{
"epoch": 0.9653631284916201,
"grad_norm": 0.521646351973494,
"learning_rate": 3.437025796661609e-06,
"loss": 0.5071,
"step": 216
},
{
"epoch": 0.9698324022346368,
"grad_norm": 0.5335773786656196,
"learning_rate": 3.429438543247345e-06,
"loss": 0.4993,
"step": 217
},
{
"epoch": 0.9743016759776536,
"grad_norm": 0.5774515212071393,
"learning_rate": 3.4218512898330806e-06,
"loss": 0.4802,
"step": 218
},
{
"epoch": 0.9787709497206704,
"grad_norm": 0.5159511449041201,
"learning_rate": 3.4142640364188168e-06,
"loss": 0.483,
"step": 219
},
{
"epoch": 0.9832402234636871,
"grad_norm": 0.509916149384416,
"learning_rate": 3.406676783004553e-06,
"loss": 0.4774,
"step": 220
},
{
"epoch": 0.9877094972067039,
"grad_norm": 0.5226659969898425,
"learning_rate": 3.3990895295902886e-06,
"loss": 0.516,
"step": 221
},
{
"epoch": 0.9921787709497206,
"grad_norm": 0.5632866199425641,
"learning_rate": 3.3915022761760247e-06,
"loss": 0.5346,
"step": 222
},
{
"epoch": 0.9966480446927374,
"grad_norm": 0.5224220899066443,
"learning_rate": 3.383915022761761e-06,
"loss": 0.5441,
"step": 223
},
{
"epoch": 1.0011173184357542,
"grad_norm": 0.5965341004746808,
"learning_rate": 3.3763277693474965e-06,
"loss": 0.512,
"step": 224
},
{
"epoch": 1.0011173184357542,
"eval_loss": 0.5086758732795715,
"eval_runtime": 32.322,
"eval_samples_per_second": 82.606,
"eval_steps_per_second": 5.167,
"step": 224
},
{
"epoch": 1.005586592178771,
"grad_norm": 0.5234315419190534,
"learning_rate": 3.3687405159332327e-06,
"loss": 0.5054,
"step": 225
},
{
"epoch": 1.0100558659217878,
"grad_norm": 0.5575750854403085,
"learning_rate": 3.3611532625189684e-06,
"loss": 0.4878,
"step": 226
},
{
"epoch": 1.0145251396648045,
"grad_norm": 0.5253448160628447,
"learning_rate": 3.3535660091047045e-06,
"loss": 0.5054,
"step": 227
},
{
"epoch": 1.0189944134078213,
"grad_norm": 0.49020025660118194,
"learning_rate": 3.3459787556904406e-06,
"loss": 0.4818,
"step": 228
},
{
"epoch": 1.0044692737430168,
"grad_norm": 0.6277138162928673,
"learning_rate": 3.3383915022761763e-06,
"loss": 0.5008,
"step": 229
},
{
"epoch": 1.0089385474860335,
"grad_norm": 0.5276575275655174,
"learning_rate": 3.3308042488619125e-06,
"loss": 0.4822,
"step": 230
},
{
"epoch": 1.0134078212290503,
"grad_norm": 0.516729892988387,
"learning_rate": 3.3232169954476486e-06,
"loss": 0.4783,
"step": 231
},
{
"epoch": 1.017877094972067,
"grad_norm": 0.5648907712973277,
"learning_rate": 3.3156297420333843e-06,
"loss": 0.4927,
"step": 232
},
{
"epoch": 1.0223463687150838,
"grad_norm": 0.5521221146190025,
"learning_rate": 3.3080424886191204e-06,
"loss": 0.4742,
"step": 233
},
{
"epoch": 1.0268156424581005,
"grad_norm": 0.5288379239842262,
"learning_rate": 3.3004552352048565e-06,
"loss": 0.4863,
"step": 234
},
{
"epoch": 1.0312849162011173,
"grad_norm": 0.4932863276651226,
"learning_rate": 3.2928679817905922e-06,
"loss": 0.4613,
"step": 235
},
{
"epoch": 1.035754189944134,
"grad_norm": 0.5069055178884726,
"learning_rate": 3.2852807283763284e-06,
"loss": 0.4969,
"step": 236
},
{
"epoch": 1.0402234636871508,
"grad_norm": 0.49009866664123986,
"learning_rate": 3.277693474962064e-06,
"loss": 0.4759,
"step": 237
},
{
"epoch": 1.0446927374301676,
"grad_norm": 0.5220649844244807,
"learning_rate": 3.2701062215478e-06,
"loss": 0.4654,
"step": 238
},
{
"epoch": 1.0491620111731843,
"grad_norm": 0.5354627141925222,
"learning_rate": 3.2625189681335363e-06,
"loss": 0.5009,
"step": 239
},
{
"epoch": 1.053631284916201,
"grad_norm": 0.5561818122893151,
"learning_rate": 3.254931714719272e-06,
"loss": 0.4564,
"step": 240
},
{
"epoch": 1.0581005586592178,
"grad_norm": 0.48719455414113416,
"learning_rate": 3.2473444613050077e-06,
"loss": 0.4924,
"step": 241
},
{
"epoch": 1.0625698324022346,
"grad_norm": 0.5527905454518657,
"learning_rate": 3.2397572078907434e-06,
"loss": 0.4856,
"step": 242
},
{
"epoch": 1.0670391061452513,
"grad_norm": 0.5381863443762583,
"learning_rate": 3.2321699544764795e-06,
"loss": 0.4935,
"step": 243
},
{
"epoch": 1.071508379888268,
"grad_norm": 0.5355360844432822,
"learning_rate": 3.2245827010622157e-06,
"loss": 0.4858,
"step": 244
},
{
"epoch": 1.0759776536312848,
"grad_norm": 0.5191892602172652,
"learning_rate": 3.2169954476479514e-06,
"loss": 0.4819,
"step": 245
},
{
"epoch": 1.0804469273743016,
"grad_norm": 0.4952672717234948,
"learning_rate": 3.2094081942336875e-06,
"loss": 0.5007,
"step": 246
},
{
"epoch": 1.0849162011173183,
"grad_norm": 0.5331734326280092,
"learning_rate": 3.201820940819423e-06,
"loss": 0.4874,
"step": 247
},
{
"epoch": 1.089385474860335,
"grad_norm": 0.49333273079420603,
"learning_rate": 3.1942336874051593e-06,
"loss": 0.5045,
"step": 248
},
{
"epoch": 1.0938547486033519,
"grad_norm": 0.5143326582021585,
"learning_rate": 3.1866464339908955e-06,
"loss": 0.5047,
"step": 249
},
{
"epoch": 1.0983240223463686,
"grad_norm": 0.5027227799706213,
"learning_rate": 3.179059180576631e-06,
"loss": 0.4803,
"step": 250
},
{
"epoch": 1.1027932960893856,
"grad_norm": 0.5257359502707071,
"learning_rate": 3.1714719271623673e-06,
"loss": 0.4654,
"step": 251
},
{
"epoch": 1.1072625698324021,
"grad_norm": 0.5244897759431714,
"learning_rate": 3.1638846737481034e-06,
"loss": 0.5329,
"step": 252
},
{
"epoch": 1.111731843575419,
"grad_norm": 0.5028531015372357,
"learning_rate": 3.156297420333839e-06,
"loss": 0.4721,
"step": 253
},
{
"epoch": 1.1162011173184359,
"grad_norm": 0.5342121642252061,
"learning_rate": 3.1487101669195752e-06,
"loss": 0.4962,
"step": 254
},
{
"epoch": 1.1206703910614526,
"grad_norm": 0.49550125400926287,
"learning_rate": 3.1411229135053114e-06,
"loss": 0.4884,
"step": 255
},
{
"epoch": 1.1251396648044694,
"grad_norm": 0.5724921757672813,
"learning_rate": 3.133535660091047e-06,
"loss": 0.4676,
"step": 256
},
{
"epoch": 1.1296089385474861,
"grad_norm": 0.5012184887171401,
"learning_rate": 3.125948406676783e-06,
"loss": 0.48,
"step": 257
},
{
"epoch": 1.1340782122905029,
"grad_norm": 0.5049054849116781,
"learning_rate": 3.118361153262519e-06,
"loss": 0.494,
"step": 258
},
{
"epoch": 1.1385474860335196,
"grad_norm": 0.5291980859522503,
"learning_rate": 3.110773899848255e-06,
"loss": 0.4787,
"step": 259
},
{
"epoch": 1.1430167597765364,
"grad_norm": 0.5235318263865767,
"learning_rate": 3.103186646433991e-06,
"loss": 0.4745,
"step": 260
},
{
"epoch": 1.1474860335195531,
"grad_norm": 0.5067236081675356,
"learning_rate": 3.095599393019727e-06,
"loss": 0.5863,
"step": 261
},
{
"epoch": 1.15195530726257,
"grad_norm": 0.48763909341918976,
"learning_rate": 3.088012139605463e-06,
"loss": 0.5264,
"step": 262
},
{
"epoch": 1.1564245810055866,
"grad_norm": 0.5198110315693464,
"learning_rate": 3.080424886191199e-06,
"loss": 0.4716,
"step": 263
},
{
"epoch": 1.1608938547486034,
"grad_norm": 0.5093655208180958,
"learning_rate": 3.072837632776935e-06,
"loss": 0.4768,
"step": 264
},
{
"epoch": 1.1653631284916202,
"grad_norm": 0.49489186143964614,
"learning_rate": 3.065250379362671e-06,
"loss": 0.4765,
"step": 265
},
{
"epoch": 1.169832402234637,
"grad_norm": 0.5209973994944476,
"learning_rate": 3.057663125948407e-06,
"loss": 0.4869,
"step": 266
},
{
"epoch": 1.1743016759776537,
"grad_norm": 0.4873878645874658,
"learning_rate": 3.0500758725341427e-06,
"loss": 0.4623,
"step": 267
},
{
"epoch": 1.1787709497206704,
"grad_norm": 0.5074910158328607,
"learning_rate": 3.042488619119879e-06,
"loss": 0.4765,
"step": 268
},
{
"epoch": 1.1832402234636872,
"grad_norm": 0.5195087180758534,
"learning_rate": 3.0349013657056146e-06,
"loss": 0.4838,
"step": 269
},
{
"epoch": 1.187709497206704,
"grad_norm": 0.5264228931092361,
"learning_rate": 3.0273141122913507e-06,
"loss": 0.4638,
"step": 270
},
{
"epoch": 1.1921787709497207,
"grad_norm": 0.4919380409463276,
"learning_rate": 3.019726858877087e-06,
"loss": 0.4981,
"step": 271
},
{
"epoch": 1.1966480446927374,
"grad_norm": 0.5060903656555795,
"learning_rate": 3.0121396054628225e-06,
"loss": 0.477,
"step": 272
},
{
"epoch": 1.2011173184357542,
"grad_norm": 0.4938697284936987,
"learning_rate": 3.0045523520485587e-06,
"loss": 0.5003,
"step": 273
},
{
"epoch": 1.205586592178771,
"grad_norm": 0.5599516302969035,
"learning_rate": 2.9969650986342948e-06,
"loss": 0.4673,
"step": 274
},
{
"epoch": 1.2100558659217877,
"grad_norm": 0.5126307823030745,
"learning_rate": 2.9893778452200305e-06,
"loss": 0.4971,
"step": 275
},
{
"epoch": 1.2145251396648045,
"grad_norm": 0.5927333276589611,
"learning_rate": 2.9817905918057666e-06,
"loss": 0.5006,
"step": 276
},
{
"epoch": 1.2189944134078212,
"grad_norm": 0.5367579038483986,
"learning_rate": 2.9742033383915027e-06,
"loss": 0.5143,
"step": 277
},
{
"epoch": 1.223463687150838,
"grad_norm": 0.5179699664039845,
"learning_rate": 2.9666160849772384e-06,
"loss": 0.4733,
"step": 278
},
{
"epoch": 1.2279329608938547,
"grad_norm": 0.5015385250607317,
"learning_rate": 2.9590288315629746e-06,
"loss": 0.4786,
"step": 279
},
{
"epoch": 1.2324022346368715,
"grad_norm": 0.4925764923779409,
"learning_rate": 2.9514415781487103e-06,
"loss": 0.472,
"step": 280
},
{
"epoch": 1.2368715083798882,
"grad_norm": 0.5522056674324646,
"learning_rate": 2.9438543247344464e-06,
"loss": 0.4675,
"step": 281
},
{
"epoch": 1.241340782122905,
"grad_norm": 0.5128909667557014,
"learning_rate": 2.9362670713201825e-06,
"loss": 0.4778,
"step": 282
},
{
"epoch": 1.2458100558659218,
"grad_norm": 0.49958051066954534,
"learning_rate": 2.9286798179059182e-06,
"loss": 0.4744,
"step": 283
},
{
"epoch": 1.2502793296089385,
"grad_norm": 0.5186801983796817,
"learning_rate": 2.9210925644916543e-06,
"loss": 0.47,
"step": 284
},
{
"epoch": 1.2547486033519553,
"grad_norm": 0.49479081933964797,
"learning_rate": 2.9135053110773905e-06,
"loss": 0.4693,
"step": 285
},
{
"epoch": 1.259217877094972,
"grad_norm": 0.5172924586559385,
"learning_rate": 2.905918057663126e-06,
"loss": 0.5034,
"step": 286
},
{
"epoch": 1.2636871508379888,
"grad_norm": 0.49588955761543196,
"learning_rate": 2.8983308042488623e-06,
"loss": 0.4857,
"step": 287
},
{
"epoch": 1.2681564245810055,
"grad_norm": 0.5651670438316321,
"learning_rate": 2.8907435508345984e-06,
"loss": 0.4845,
"step": 288
},
{
"epoch": 1.2726256983240223,
"grad_norm": 0.47720259500943096,
"learning_rate": 2.883156297420334e-06,
"loss": 0.4482,
"step": 289
},
{
"epoch": 1.277094972067039,
"grad_norm": 0.5080092332022862,
"learning_rate": 2.8755690440060702e-06,
"loss": 0.4826,
"step": 290
},
{
"epoch": 1.2815642458100558,
"grad_norm": 0.6016729996144168,
"learning_rate": 2.867981790591806e-06,
"loss": 0.5592,
"step": 291
},
{
"epoch": 1.2860335195530725,
"grad_norm": 0.5456040494522852,
"learning_rate": 2.860394537177542e-06,
"loss": 0.4821,
"step": 292
},
{
"epoch": 1.2905027932960893,
"grad_norm": 0.5858331665900632,
"learning_rate": 2.852807283763278e-06,
"loss": 0.5089,
"step": 293
},
{
"epoch": 1.2949720670391063,
"grad_norm": 0.5895630636847224,
"learning_rate": 2.845220030349014e-06,
"loss": 0.5101,
"step": 294
},
{
"epoch": 1.2994413407821228,
"grad_norm": 0.505172328049275,
"learning_rate": 2.83763277693475e-06,
"loss": 0.5008,
"step": 295
},
{
"epoch": 1.3039106145251398,
"grad_norm": 0.5035894899292893,
"learning_rate": 2.830045523520486e-06,
"loss": 0.4812,
"step": 296
},
{
"epoch": 1.3083798882681563,
"grad_norm": 0.5304264887862729,
"learning_rate": 2.822458270106222e-06,
"loss": 0.4579,
"step": 297
},
{
"epoch": 1.3128491620111733,
"grad_norm": 0.4742288658363735,
"learning_rate": 2.814871016691958e-06,
"loss": 0.4801,
"step": 298
},
{
"epoch": 1.3173184357541898,
"grad_norm": 0.5293412371049411,
"learning_rate": 2.807283763277694e-06,
"loss": 0.4602,
"step": 299
},
{
"epoch": 1.3217877094972068,
"grad_norm": 0.5333831439310788,
"learning_rate": 2.79969650986343e-06,
"loss": 0.4963,
"step": 300
},
{
"epoch": 1.3262569832402233,
"grad_norm": 0.4947844594727205,
"learning_rate": 2.792109256449166e-06,
"loss": 0.4781,
"step": 301
},
{
"epoch": 1.3307262569832403,
"grad_norm": 0.5323122787674339,
"learning_rate": 2.7845220030349016e-06,
"loss": 0.4998,
"step": 302
},
{
"epoch": 1.3351955307262569,
"grad_norm": 0.48615316005162806,
"learning_rate": 2.7769347496206378e-06,
"loss": 0.466,
"step": 303
},
{
"epoch": 1.3396648044692738,
"grad_norm": 0.5038897355673981,
"learning_rate": 2.769347496206374e-06,
"loss": 0.531,
"step": 304
},
{
"epoch": 1.3441340782122906,
"grad_norm": 0.47725378544672703,
"learning_rate": 2.7617602427921096e-06,
"loss": 0.4917,
"step": 305
},
{
"epoch": 1.3486033519553073,
"grad_norm": 0.5241579403675019,
"learning_rate": 2.7541729893778457e-06,
"loss": 0.4673,
"step": 306
},
{
"epoch": 1.353072625698324,
"grad_norm": 0.5252091486066182,
"learning_rate": 2.746585735963581e-06,
"loss": 0.467,
"step": 307
},
{
"epoch": 1.3575418994413408,
"grad_norm": 0.4803361684555166,
"learning_rate": 2.738998482549317e-06,
"loss": 0.5257,
"step": 308
},
{
"epoch": 1.3620111731843576,
"grad_norm": 0.5300934772683465,
"learning_rate": 2.7314112291350532e-06,
"loss": 0.4889,
"step": 309
},
{
"epoch": 1.3664804469273744,
"grad_norm": 0.5001374114396476,
"learning_rate": 2.723823975720789e-06,
"loss": 0.4847,
"step": 310
},
{
"epoch": 1.3709497206703911,
"grad_norm": 0.5240875065710657,
"learning_rate": 2.716236722306525e-06,
"loss": 0.4578,
"step": 311
},
{
"epoch": 1.3754189944134079,
"grad_norm": 0.4714761354385937,
"learning_rate": 2.708649468892261e-06,
"loss": 0.4947,
"step": 312
},
{
"epoch": 1.3798882681564246,
"grad_norm": 0.5205222669608968,
"learning_rate": 2.701062215477997e-06,
"loss": 0.5049,
"step": 313
},
{
"epoch": 1.3843575418994414,
"grad_norm": 0.5409150510364881,
"learning_rate": 2.693474962063733e-06,
"loss": 0.4946,
"step": 314
},
{
"epoch": 1.3888268156424581,
"grad_norm": 0.49473773833387136,
"learning_rate": 2.6858877086494687e-06,
"loss": 0.5024,
"step": 315
},
{
"epoch": 1.393296089385475,
"grad_norm": 0.5322303350412809,
"learning_rate": 2.678300455235205e-06,
"loss": 0.5129,
"step": 316
},
{
"epoch": 1.3977653631284916,
"grad_norm": 0.5249345159944246,
"learning_rate": 2.670713201820941e-06,
"loss": 0.4852,
"step": 317
},
{
"epoch": 1.4022346368715084,
"grad_norm": 0.526125738227021,
"learning_rate": 2.6631259484066767e-06,
"loss": 0.4755,
"step": 318
},
{
"epoch": 1.4067039106145252,
"grad_norm": 0.5077327993266544,
"learning_rate": 2.655538694992413e-06,
"loss": 0.4708,
"step": 319
},
{
"epoch": 1.411173184357542,
"grad_norm": 0.509006777595202,
"learning_rate": 2.647951441578149e-06,
"loss": 0.4574,
"step": 320
},
{
"epoch": 1.4156424581005587,
"grad_norm": 0.5397063984545032,
"learning_rate": 2.6403641881638846e-06,
"loss": 0.4983,
"step": 321
},
{
"epoch": 1.4201117318435754,
"grad_norm": 0.588039527577832,
"learning_rate": 2.6327769347496208e-06,
"loss": 0.5063,
"step": 322
},
{
"epoch": 1.4245810055865922,
"grad_norm": 0.5086942941008908,
"learning_rate": 2.625189681335357e-06,
"loss": 0.5439,
"step": 323
},
{
"epoch": 1.429050279329609,
"grad_norm": 0.5593378373560197,
"learning_rate": 2.6176024279210926e-06,
"loss": 0.4789,
"step": 324
},
{
"epoch": 1.4335195530726257,
"grad_norm": 0.5473461722884195,
"learning_rate": 2.6100151745068287e-06,
"loss": 0.4861,
"step": 325
},
{
"epoch": 1.4379888268156424,
"grad_norm": 0.5744717523131797,
"learning_rate": 2.6024279210925644e-06,
"loss": 0.4818,
"step": 326
},
{
"epoch": 1.4424581005586592,
"grad_norm": 0.5240335872273282,
"learning_rate": 2.5948406676783005e-06,
"loss": 0.4785,
"step": 327
},
{
"epoch": 1.446927374301676,
"grad_norm": 0.5608475565499077,
"learning_rate": 2.5872534142640367e-06,
"loss": 0.5042,
"step": 328
},
{
"epoch": 1.4513966480446927,
"grad_norm": 0.5363050759442354,
"learning_rate": 2.5796661608497724e-06,
"loss": 0.4995,
"step": 329
},
{
"epoch": 1.4558659217877095,
"grad_norm": 0.5027607995293213,
"learning_rate": 2.5720789074355085e-06,
"loss": 0.4954,
"step": 330
},
{
"epoch": 1.4603351955307262,
"grad_norm": 0.4969954043343912,
"learning_rate": 2.5644916540212446e-06,
"loss": 0.4882,
"step": 331
},
{
"epoch": 1.464804469273743,
"grad_norm": 0.5718337637259987,
"learning_rate": 2.5569044006069803e-06,
"loss": 0.4726,
"step": 332
},
{
"epoch": 1.4692737430167597,
"grad_norm": 0.5234285673769393,
"learning_rate": 2.5493171471927164e-06,
"loss": 0.4697,
"step": 333
},
{
"epoch": 1.4737430167597765,
"grad_norm": 0.5219608223935474,
"learning_rate": 2.5417298937784526e-06,
"loss": 0.4776,
"step": 334
},
{
"epoch": 1.4782122905027932,
"grad_norm": 0.606768683027719,
"learning_rate": 2.5341426403641883e-06,
"loss": 0.4756,
"step": 335
},
{
"epoch": 1.48268156424581,
"grad_norm": 0.5432176614167162,
"learning_rate": 2.5265553869499244e-06,
"loss": 0.4619,
"step": 336
},
{
"epoch": 1.48268156424581,
"eval_loss": 0.5032872557640076,
"eval_runtime": 32.2653,
"eval_samples_per_second": 82.751,
"eval_steps_per_second": 5.176,
"step": 336
},
{
"epoch": 1.4871508379888267,
"grad_norm": 0.5377224565720307,
"learning_rate": 2.51896813353566e-06,
"loss": 0.4667,
"step": 337
},
{
"epoch": 1.4916201117318435,
"grad_norm": 0.5875842821019925,
"learning_rate": 2.5113808801213962e-06,
"loss": 0.4873,
"step": 338
},
{
"epoch": 1.4960893854748603,
"grad_norm": 0.5624618850955563,
"learning_rate": 2.5037936267071324e-06,
"loss": 0.4626,
"step": 339
},
{
"epoch": 1.500558659217877,
"grad_norm": 0.5380665704623208,
"learning_rate": 2.496206373292868e-06,
"loss": 0.4749,
"step": 340
},
{
"epoch": 1.505027932960894,
"grad_norm": 0.5515048280645092,
"learning_rate": 2.488619119878604e-06,
"loss": 0.4862,
"step": 341
},
{
"epoch": 1.5094972067039105,
"grad_norm": 0.5541495309875792,
"learning_rate": 2.4810318664643403e-06,
"loss": 0.4533,
"step": 342
},
{
"epoch": 1.5139664804469275,
"grad_norm": 0.514569610923219,
"learning_rate": 2.473444613050076e-06,
"loss": 0.4585,
"step": 343
},
{
"epoch": 1.518435754189944,
"grad_norm": 0.5862261680615611,
"learning_rate": 2.465857359635812e-06,
"loss": 0.4835,
"step": 344
},
{
"epoch": 1.522905027932961,
"grad_norm": 0.536975433650959,
"learning_rate": 2.458270106221548e-06,
"loss": 0.4638,
"step": 345
},
{
"epoch": 1.5273743016759775,
"grad_norm": 0.5142279881627296,
"learning_rate": 2.450682852807284e-06,
"loss": 0.4832,
"step": 346
},
{
"epoch": 1.5318435754189945,
"grad_norm": 0.5124062345457664,
"learning_rate": 2.44309559939302e-06,
"loss": 0.4857,
"step": 347
},
{
"epoch": 1.536312849162011,
"grad_norm": 0.509905421315166,
"learning_rate": 2.435508345978756e-06,
"loss": 0.4807,
"step": 348
},
{
"epoch": 1.540782122905028,
"grad_norm": 0.5275507665643101,
"learning_rate": 2.427921092564492e-06,
"loss": 0.4722,
"step": 349
},
{
"epoch": 1.5452513966480446,
"grad_norm": 0.5003951978639241,
"learning_rate": 2.420333839150228e-06,
"loss": 0.4941,
"step": 350
},
{
"epoch": 1.5497206703910615,
"grad_norm": 0.5251614039487292,
"learning_rate": 2.4127465857359637e-06,
"loss": 0.5376,
"step": 351
},
{
"epoch": 1.554189944134078,
"grad_norm": 0.5128097582851606,
"learning_rate": 2.4051593323217e-06,
"loss": 0.5001,
"step": 352
},
{
"epoch": 1.558659217877095,
"grad_norm": 0.4971288144993544,
"learning_rate": 2.397572078907436e-06,
"loss": 0.4721,
"step": 353
},
{
"epoch": 1.5631284916201116,
"grad_norm": 0.5111491573396423,
"learning_rate": 2.3899848254931717e-06,
"loss": 0.5192,
"step": 354
},
{
"epoch": 1.5675977653631286,
"grad_norm": 0.5025703950184777,
"learning_rate": 2.382397572078908e-06,
"loss": 0.4541,
"step": 355
},
{
"epoch": 1.572067039106145,
"grad_norm": 0.5011840454618501,
"learning_rate": 2.3748103186646435e-06,
"loss": 0.4789,
"step": 356
},
{
"epoch": 1.576536312849162,
"grad_norm": 0.4981690031939279,
"learning_rate": 2.3672230652503792e-06,
"loss": 0.497,
"step": 357
},
{
"epoch": 1.5810055865921788,
"grad_norm": 0.4994506284814992,
"learning_rate": 2.3596358118361154e-06,
"loss": 0.4655,
"step": 358
},
{
"epoch": 1.5854748603351956,
"grad_norm": 0.49694019288140645,
"learning_rate": 2.3520485584218515e-06,
"loss": 0.4883,
"step": 359
},
{
"epoch": 1.5899441340782123,
"grad_norm": 0.5103222760000443,
"learning_rate": 2.344461305007587e-06,
"loss": 0.4615,
"step": 360
},
{
"epoch": 1.594413407821229,
"grad_norm": 0.4960772434204748,
"learning_rate": 2.3368740515933233e-06,
"loss": 0.4748,
"step": 361
},
{
"epoch": 1.5988826815642458,
"grad_norm": 0.5372239368130248,
"learning_rate": 2.3292867981790594e-06,
"loss": 0.461,
"step": 362
},
{
"epoch": 1.6033519553072626,
"grad_norm": 0.5002183877541576,
"learning_rate": 2.321699544764795e-06,
"loss": 0.4985,
"step": 363
},
{
"epoch": 1.6078212290502794,
"grad_norm": 0.5065612982842012,
"learning_rate": 2.3141122913505313e-06,
"loss": 0.4846,
"step": 364
},
{
"epoch": 1.612290502793296,
"grad_norm": 0.5294145045424744,
"learning_rate": 2.3065250379362674e-06,
"loss": 0.4769,
"step": 365
},
{
"epoch": 1.6167597765363129,
"grad_norm": 0.5019842435412675,
"learning_rate": 2.298937784522003e-06,
"loss": 0.4738,
"step": 366
},
{
"epoch": 1.6212290502793296,
"grad_norm": 0.4973490272845337,
"learning_rate": 2.291350531107739e-06,
"loss": 0.4773,
"step": 367
},
{
"epoch": 1.6256983240223464,
"grad_norm": 0.4917027064554639,
"learning_rate": 2.283763277693475e-06,
"loss": 0.4921,
"step": 368
},
{
"epoch": 1.6301675977653631,
"grad_norm": 0.5191932977137537,
"learning_rate": 2.276176024279211e-06,
"loss": 0.4823,
"step": 369
},
{
"epoch": 1.6346368715083799,
"grad_norm": 0.5130853619632637,
"learning_rate": 2.268588770864947e-06,
"loss": 0.4631,
"step": 370
},
{
"epoch": 1.6391061452513966,
"grad_norm": 0.5402133951776769,
"learning_rate": 2.261001517450683e-06,
"loss": 0.456,
"step": 371
},
{
"epoch": 1.6435754189944134,
"grad_norm": 0.5136201214507364,
"learning_rate": 2.253414264036419e-06,
"loss": 0.4715,
"step": 372
},
{
"epoch": 1.6480446927374302,
"grad_norm": 0.5397223114168315,
"learning_rate": 2.245827010622155e-06,
"loss": 0.465,
"step": 373
},
{
"epoch": 1.652513966480447,
"grad_norm": 0.5178224930093817,
"learning_rate": 2.238239757207891e-06,
"loss": 0.4526,
"step": 374
},
{
"epoch": 1.6569832402234637,
"grad_norm": 0.5399618966600667,
"learning_rate": 2.230652503793627e-06,
"loss": 0.4873,
"step": 375
},
{
"epoch": 1.6614525139664804,
"grad_norm": 0.5106529140250409,
"learning_rate": 2.223065250379363e-06,
"loss": 0.4771,
"step": 376
},
{
"epoch": 1.6659217877094972,
"grad_norm": 0.5460841160518792,
"learning_rate": 2.2154779969650988e-06,
"loss": 0.4743,
"step": 377
},
{
"epoch": 1.670391061452514,
"grad_norm": 0.5846706720135237,
"learning_rate": 2.207890743550835e-06,
"loss": 0.4552,
"step": 378
},
{
"epoch": 1.6748603351955307,
"grad_norm": 0.534699591229056,
"learning_rate": 2.2003034901365706e-06,
"loss": 0.495,
"step": 379
},
{
"epoch": 1.6793296089385474,
"grad_norm": 0.5377842578832556,
"learning_rate": 2.1927162367223067e-06,
"loss": 0.4824,
"step": 380
},
{
"epoch": 1.6837988826815642,
"grad_norm": 0.5799148072307362,
"learning_rate": 2.185128983308043e-06,
"loss": 0.4708,
"step": 381
},
{
"epoch": 1.6882681564245812,
"grad_norm": 0.5076181067753225,
"learning_rate": 2.1775417298937786e-06,
"loss": 0.4631,
"step": 382
},
{
"epoch": 1.6927374301675977,
"grad_norm": 0.5025103762077108,
"learning_rate": 2.1699544764795147e-06,
"loss": 0.4676,
"step": 383
},
{
"epoch": 1.6972067039106147,
"grad_norm": 0.48231447315943815,
"learning_rate": 2.162367223065251e-06,
"loss": 0.4694,
"step": 384
},
{
"epoch": 1.7016759776536312,
"grad_norm": 0.5003437712619672,
"learning_rate": 2.1547799696509865e-06,
"loss": 0.4607,
"step": 385
},
{
"epoch": 1.7061452513966482,
"grad_norm": 0.5149834201342615,
"learning_rate": 2.1471927162367226e-06,
"loss": 0.5253,
"step": 386
},
{
"epoch": 1.7106145251396647,
"grad_norm": 0.5406857228194807,
"learning_rate": 2.1396054628224588e-06,
"loss": 0.5352,
"step": 387
},
{
"epoch": 1.7150837988826817,
"grad_norm": 0.5204717082880796,
"learning_rate": 2.1320182094081945e-06,
"loss": 0.4936,
"step": 388
},
{
"epoch": 1.7195530726256982,
"grad_norm": 0.5202749448515015,
"learning_rate": 2.12443095599393e-06,
"loss": 0.4594,
"step": 389
},
{
"epoch": 1.7240223463687152,
"grad_norm": 0.4952910050954741,
"learning_rate": 2.1168437025796663e-06,
"loss": 0.4793,
"step": 390
},
{
"epoch": 1.7284916201117317,
"grad_norm": 0.48637344511275765,
"learning_rate": 2.109256449165402e-06,
"loss": 0.4745,
"step": 391
},
{
"epoch": 1.7329608938547487,
"grad_norm": 0.4906271558417189,
"learning_rate": 2.101669195751138e-06,
"loss": 0.4652,
"step": 392
},
{
"epoch": 1.7374301675977653,
"grad_norm": 0.5090727761430808,
"learning_rate": 2.0940819423368742e-06,
"loss": 0.5298,
"step": 393
},
{
"epoch": 1.7418994413407822,
"grad_norm": 0.4866596430096002,
"learning_rate": 2.08649468892261e-06,
"loss": 0.4838,
"step": 394
},
{
"epoch": 1.7463687150837988,
"grad_norm": 0.49259128094078225,
"learning_rate": 2.078907435508346e-06,
"loss": 0.4737,
"step": 395
},
{
"epoch": 1.7508379888268157,
"grad_norm": 0.5031034927365164,
"learning_rate": 2.071320182094082e-06,
"loss": 0.4777,
"step": 396
},
{
"epoch": 1.7553072625698323,
"grad_norm": 0.48520021809068,
"learning_rate": 2.063732928679818e-06,
"loss": 0.454,
"step": 397
},
{
"epoch": 1.7597765363128492,
"grad_norm": 0.5217902917498033,
"learning_rate": 2.056145675265554e-06,
"loss": 0.5199,
"step": 398
},
{
"epoch": 1.7642458100558658,
"grad_norm": 0.4963706354662438,
"learning_rate": 2.04855842185129e-06,
"loss": 0.4649,
"step": 399
},
{
"epoch": 1.7687150837988828,
"grad_norm": 0.49552580249683814,
"learning_rate": 2.040971168437026e-06,
"loss": 0.4887,
"step": 400
},
{
"epoch": 1.7731843575418993,
"grad_norm": 0.49468231483436587,
"learning_rate": 2.033383915022762e-06,
"loss": 0.4757,
"step": 401
},
{
"epoch": 1.7776536312849163,
"grad_norm": 0.47502769186897137,
"learning_rate": 2.0257966616084977e-06,
"loss": 0.4846,
"step": 402
},
{
"epoch": 1.7821229050279328,
"grad_norm": 0.5089134108324753,
"learning_rate": 2.018209408194234e-06,
"loss": 0.4692,
"step": 403
},
{
"epoch": 1.7865921787709498,
"grad_norm": 0.4981731250179096,
"learning_rate": 2.01062215477997e-06,
"loss": 0.4665,
"step": 404
},
{
"epoch": 1.7910614525139665,
"grad_norm": 0.4690174096990647,
"learning_rate": 2.0030349013657056e-06,
"loss": 0.4606,
"step": 405
},
{
"epoch": 1.7955307262569833,
"grad_norm": 0.4914411018326139,
"learning_rate": 1.9954476479514418e-06,
"loss": 0.4847,
"step": 406
},
{
"epoch": 1.8,
"grad_norm": 0.5513335287615446,
"learning_rate": 1.987860394537178e-06,
"loss": 0.488,
"step": 407
},
{
"epoch": 1.8044692737430168,
"grad_norm": 0.48992847389625854,
"learning_rate": 1.9802731411229136e-06,
"loss": 0.465,
"step": 408
},
{
"epoch": 1.8089385474860336,
"grad_norm": 0.5185912675971102,
"learning_rate": 1.9726858877086497e-06,
"loss": 0.4872,
"step": 409
},
{
"epoch": 1.8134078212290503,
"grad_norm": 0.47300994432774746,
"learning_rate": 1.965098634294386e-06,
"loss": 0.4534,
"step": 410
},
{
"epoch": 1.817877094972067,
"grad_norm": 0.5034275716173061,
"learning_rate": 1.9575113808801215e-06,
"loss": 0.4668,
"step": 411
},
{
"epoch": 1.8223463687150838,
"grad_norm": 0.46435516708821234,
"learning_rate": 1.9499241274658577e-06,
"loss": 0.491,
"step": 412
},
{
"epoch": 1.8268156424581006,
"grad_norm": 0.4915994950607221,
"learning_rate": 1.9423368740515934e-06,
"loss": 0.4793,
"step": 413
},
{
"epoch": 1.8312849162011173,
"grad_norm": 0.49834550842622505,
"learning_rate": 1.9347496206373295e-06,
"loss": 0.4897,
"step": 414
},
{
"epoch": 1.835754189944134,
"grad_norm": 0.5275123854480922,
"learning_rate": 1.9271623672230656e-06,
"loss": 0.5167,
"step": 415
},
{
"epoch": 1.8402234636871508,
"grad_norm": 0.501447167095016,
"learning_rate": 1.9195751138088013e-06,
"loss": 0.5522,
"step": 416
},
{
"epoch": 1.8446927374301676,
"grad_norm": 0.4929358533949542,
"learning_rate": 1.9119878603945374e-06,
"loss": 0.464,
"step": 417
},
{
"epoch": 1.8491620111731844,
"grad_norm": 0.4753893398726578,
"learning_rate": 1.9044006069802734e-06,
"loss": 0.4707,
"step": 418
},
{
"epoch": 1.853631284916201,
"grad_norm": 0.49249135995729143,
"learning_rate": 1.8968133535660093e-06,
"loss": 0.4674,
"step": 419
},
{
"epoch": 1.8581005586592179,
"grad_norm": 0.4959065755228421,
"learning_rate": 1.8892261001517454e-06,
"loss": 0.4731,
"step": 420
},
{
"epoch": 1.8625698324022346,
"grad_norm": 0.4830053823448673,
"learning_rate": 1.8816388467374813e-06,
"loss": 0.4688,
"step": 421
},
{
"epoch": 1.8670391061452514,
"grad_norm": 0.5026774544021604,
"learning_rate": 1.874051593323217e-06,
"loss": 0.4777,
"step": 422
},
{
"epoch": 1.8715083798882681,
"grad_norm": 0.48376669794821286,
"learning_rate": 1.866464339908953e-06,
"loss": 0.4532,
"step": 423
},
{
"epoch": 1.8759776536312849,
"grad_norm": 0.5036910075540445,
"learning_rate": 1.858877086494689e-06,
"loss": 0.4749,
"step": 424
},
{
"epoch": 1.8804469273743016,
"grad_norm": 0.464687327497754,
"learning_rate": 1.851289833080425e-06,
"loss": 0.4582,
"step": 425
},
{
"epoch": 1.8849162011173184,
"grad_norm": 0.4713537790626329,
"learning_rate": 1.8437025796661609e-06,
"loss": 0.4954,
"step": 426
},
{
"epoch": 1.8893854748603351,
"grad_norm": 0.4964027557908742,
"learning_rate": 1.8361153262518968e-06,
"loss": 0.4777,
"step": 427
},
{
"epoch": 1.893854748603352,
"grad_norm": 0.5022160749373462,
"learning_rate": 1.828528072837633e-06,
"loss": 0.4723,
"step": 428
},
{
"epoch": 1.8983240223463689,
"grad_norm": 0.4971783807561644,
"learning_rate": 1.8209408194233688e-06,
"loss": 0.4619,
"step": 429
},
{
"epoch": 1.9027932960893854,
"grad_norm": 0.5020482529233811,
"learning_rate": 1.8133535660091047e-06,
"loss": 0.4923,
"step": 430
},
{
"epoch": 1.9072625698324024,
"grad_norm": 0.5027930539648167,
"learning_rate": 1.8057663125948407e-06,
"loss": 0.4936,
"step": 431
},
{
"epoch": 1.911731843575419,
"grad_norm": 0.5048109425957825,
"learning_rate": 1.7981790591805768e-06,
"loss": 0.4647,
"step": 432
},
{
"epoch": 1.916201117318436,
"grad_norm": 0.5152625504278344,
"learning_rate": 1.7905918057663127e-06,
"loss": 0.4962,
"step": 433
},
{
"epoch": 1.9206703910614524,
"grad_norm": 0.5486472960937676,
"learning_rate": 1.7830045523520486e-06,
"loss": 0.4643,
"step": 434
},
{
"epoch": 1.9251396648044694,
"grad_norm": 0.5074159325604022,
"learning_rate": 1.7754172989377847e-06,
"loss": 0.4838,
"step": 435
},
{
"epoch": 1.929608938547486,
"grad_norm": 0.487237277239409,
"learning_rate": 1.7678300455235207e-06,
"loss": 0.4953,
"step": 436
},
{
"epoch": 1.934078212290503,
"grad_norm": 0.5030243156211434,
"learning_rate": 1.7602427921092566e-06,
"loss": 0.4669,
"step": 437
},
{
"epoch": 1.9385474860335195,
"grad_norm": 0.54024162061195,
"learning_rate": 1.7526555386949925e-06,
"loss": 0.4645,
"step": 438
},
{
"epoch": 1.9430167597765364,
"grad_norm": 0.4924548318787792,
"learning_rate": 1.7450682852807286e-06,
"loss": 0.487,
"step": 439
},
{
"epoch": 1.947486033519553,
"grad_norm": 0.5002700607677886,
"learning_rate": 1.7374810318664645e-06,
"loss": 0.4904,
"step": 440
},
{
"epoch": 1.95195530726257,
"grad_norm": 0.49052978017153487,
"learning_rate": 1.7298937784522004e-06,
"loss": 0.4721,
"step": 441
},
{
"epoch": 1.9564245810055865,
"grad_norm": 0.5330277427973065,
"learning_rate": 1.7223065250379363e-06,
"loss": 0.4674,
"step": 442
},
{
"epoch": 1.9608938547486034,
"grad_norm": 0.49644600432443986,
"learning_rate": 1.7147192716236725e-06,
"loss": 0.4735,
"step": 443
},
{
"epoch": 1.96536312849162,
"grad_norm": 0.4763428511189237,
"learning_rate": 1.7071320182094084e-06,
"loss": 0.4758,
"step": 444
},
{
"epoch": 1.969832402234637,
"grad_norm": 0.49662051116381534,
"learning_rate": 1.6995447647951443e-06,
"loss": 0.5025,
"step": 445
},
{
"epoch": 1.9743016759776535,
"grad_norm": 0.5368027446646121,
"learning_rate": 1.6919575113808804e-06,
"loss": 0.4708,
"step": 446
},
{
"epoch": 1.9787709497206705,
"grad_norm": 0.49270124177132796,
"learning_rate": 1.6843702579666163e-06,
"loss": 0.4859,
"step": 447
},
{
"epoch": 1.983240223463687,
"grad_norm": 0.4886244120662818,
"learning_rate": 1.6767830045523523e-06,
"loss": 0.4542,
"step": 448
},
{
"epoch": 1.983240223463687,
"eval_loss": 0.49883514642715454,
"eval_runtime": 32.2692,
"eval_samples_per_second": 82.741,
"eval_steps_per_second": 5.175,
"step": 448
},
{
"epoch": 1.987709497206704,
"grad_norm": 0.48474622370426634,
"learning_rate": 1.6691957511380882e-06,
"loss": 0.4752,
"step": 449
},
{
"epoch": 1.9921787709497205,
"grad_norm": 0.5089879136002737,
"learning_rate": 1.6616084977238243e-06,
"loss": 0.4835,
"step": 450
},
{
"epoch": 1.9966480446927375,
"grad_norm": 0.540987114653738,
"learning_rate": 1.6540212443095602e-06,
"loss": 0.4685,
"step": 451
},
{
"epoch": 2.001117318435754,
"grad_norm": 0.5330263791441335,
"learning_rate": 1.6464339908952961e-06,
"loss": 0.4729,
"step": 452
},
{
"epoch": 2.005586592178771,
"grad_norm": 0.5025591514510047,
"learning_rate": 1.638846737481032e-06,
"loss": 0.4739,
"step": 453
},
{
"epoch": 2.010055865921788,
"grad_norm": 0.4938210284492649,
"learning_rate": 1.6312594840667682e-06,
"loss": 0.4772,
"step": 454
},
{
"epoch": 2.0145251396648045,
"grad_norm": 0.5590423643222178,
"learning_rate": 1.6236722306525039e-06,
"loss": 0.4651,
"step": 455
},
{
"epoch": 2.018994413407821,
"grad_norm": 0.4994221649154916,
"learning_rate": 1.6160849772382398e-06,
"loss": 0.4989,
"step": 456
},
{
"epoch": 2.004469273743017,
"grad_norm": 0.5578750520607872,
"learning_rate": 1.6084977238239757e-06,
"loss": 0.4779,
"step": 457
},
{
"epoch": 2.0089385474860335,
"grad_norm": 0.5277075050354076,
"learning_rate": 1.6009104704097116e-06,
"loss": 0.462,
"step": 458
},
{
"epoch": 2.0134078212290505,
"grad_norm": 0.5125996961203592,
"learning_rate": 1.5933232169954477e-06,
"loss": 0.4892,
"step": 459
},
{
"epoch": 2.017877094972067,
"grad_norm": 0.5150882727539035,
"learning_rate": 1.5857359635811836e-06,
"loss": 0.4913,
"step": 460
},
{
"epoch": 2.022346368715084,
"grad_norm": 0.5395569256138755,
"learning_rate": 1.5781487101669196e-06,
"loss": 0.4789,
"step": 461
},
{
"epoch": 2.0268156424581005,
"grad_norm": 0.4907655840755226,
"learning_rate": 1.5705614567526557e-06,
"loss": 0.4699,
"step": 462
},
{
"epoch": 2.0312849162011175,
"grad_norm": 0.508487377683692,
"learning_rate": 1.5629742033383916e-06,
"loss": 0.4671,
"step": 463
},
{
"epoch": 2.035754189944134,
"grad_norm": 0.5211842893319257,
"learning_rate": 1.5553869499241275e-06,
"loss": 0.48,
"step": 464
},
{
"epoch": 2.040223463687151,
"grad_norm": 0.5540445804429339,
"learning_rate": 1.5477996965098634e-06,
"loss": 0.4839,
"step": 465
},
{
"epoch": 2.0446927374301676,
"grad_norm": 0.502144462775385,
"learning_rate": 1.5402124430955995e-06,
"loss": 0.4991,
"step": 466
},
{
"epoch": 2.0491620111731845,
"grad_norm": 0.4963826920124019,
"learning_rate": 1.5326251896813355e-06,
"loss": 0.5268,
"step": 467
},
{
"epoch": 2.053631284916201,
"grad_norm": 0.48862357759049696,
"learning_rate": 1.5250379362670714e-06,
"loss": 0.4572,
"step": 468
},
{
"epoch": 2.058100558659218,
"grad_norm": 0.48705814461293884,
"learning_rate": 1.5174506828528073e-06,
"loss": 0.4761,
"step": 469
},
{
"epoch": 2.0625698324022346,
"grad_norm": 0.50151248823518,
"learning_rate": 1.5098634294385434e-06,
"loss": 0.468,
"step": 470
},
{
"epoch": 2.0670391061452515,
"grad_norm": 0.47075309247993236,
"learning_rate": 1.5022761760242793e-06,
"loss": 0.4496,
"step": 471
},
{
"epoch": 2.071508379888268,
"grad_norm": 0.5055609634042088,
"learning_rate": 1.4946889226100152e-06,
"loss": 0.4626,
"step": 472
},
{
"epoch": 2.075977653631285,
"grad_norm": 0.4931560723293831,
"learning_rate": 1.4871016691957514e-06,
"loss": 0.4743,
"step": 473
},
{
"epoch": 2.0804469273743016,
"grad_norm": 0.4970179492434698,
"learning_rate": 1.4795144157814873e-06,
"loss": 0.4849,
"step": 474
},
{
"epoch": 2.0849162011173186,
"grad_norm": 0.5205249715468657,
"learning_rate": 1.4719271623672232e-06,
"loss": 0.4647,
"step": 475
},
{
"epoch": 2.089385474860335,
"grad_norm": 0.5106891771853166,
"learning_rate": 1.4643399089529591e-06,
"loss": 0.4698,
"step": 476
},
{
"epoch": 2.093854748603352,
"grad_norm": 0.5030533511191912,
"learning_rate": 1.4567526555386952e-06,
"loss": 0.4435,
"step": 477
},
{
"epoch": 2.0983240223463686,
"grad_norm": 0.5267045945145785,
"learning_rate": 1.4491654021244311e-06,
"loss": 0.4551,
"step": 478
},
{
"epoch": 2.1027932960893856,
"grad_norm": 0.4910735265010368,
"learning_rate": 1.441578148710167e-06,
"loss": 0.4893,
"step": 479
},
{
"epoch": 2.107262569832402,
"grad_norm": 0.5098674286346987,
"learning_rate": 1.433990895295903e-06,
"loss": 0.4631,
"step": 480
},
{
"epoch": 2.111731843575419,
"grad_norm": 0.49833407519162715,
"learning_rate": 1.426403641881639e-06,
"loss": 0.4478,
"step": 481
},
{
"epoch": 2.1162011173184356,
"grad_norm": 0.4968638787102979,
"learning_rate": 1.418816388467375e-06,
"loss": 0.4661,
"step": 482
},
{
"epoch": 2.1206703910614526,
"grad_norm": 0.5287581344581138,
"learning_rate": 1.411229135053111e-06,
"loss": 0.487,
"step": 483
},
{
"epoch": 2.125139664804469,
"grad_norm": 0.5169520017940434,
"learning_rate": 1.403641881638847e-06,
"loss": 0.4641,
"step": 484
},
{
"epoch": 2.129608938547486,
"grad_norm": 0.4889710176907046,
"learning_rate": 1.396054628224583e-06,
"loss": 0.4657,
"step": 485
},
{
"epoch": 2.1340782122905027,
"grad_norm": 0.5068054201127302,
"learning_rate": 1.3884673748103189e-06,
"loss": 0.4655,
"step": 486
},
{
"epoch": 2.1385474860335196,
"grad_norm": 0.5227391076748372,
"learning_rate": 1.3808801213960548e-06,
"loss": 0.4489,
"step": 487
},
{
"epoch": 2.143016759776536,
"grad_norm": 0.4685531353247096,
"learning_rate": 1.3732928679817905e-06,
"loss": 0.4518,
"step": 488
},
{
"epoch": 2.147486033519553,
"grad_norm": 0.48729624697805496,
"learning_rate": 1.3657056145675266e-06,
"loss": 0.4539,
"step": 489
},
{
"epoch": 2.1519553072625697,
"grad_norm": 0.49686659002832273,
"learning_rate": 1.3581183611532625e-06,
"loss": 0.4698,
"step": 490
},
{
"epoch": 2.1564245810055866,
"grad_norm": 0.5165370496352449,
"learning_rate": 1.3505311077389985e-06,
"loss": 0.5282,
"step": 491
},
{
"epoch": 2.160893854748603,
"grad_norm": 0.49390583090447004,
"learning_rate": 1.3429438543247344e-06,
"loss": 0.5011,
"step": 492
},
{
"epoch": 2.16536312849162,
"grad_norm": 0.4966083299512843,
"learning_rate": 1.3353566009104705e-06,
"loss": 0.4888,
"step": 493
},
{
"epoch": 2.1698324022346367,
"grad_norm": 0.47651097727052893,
"learning_rate": 1.3277693474962064e-06,
"loss": 0.4629,
"step": 494
},
{
"epoch": 2.1743016759776537,
"grad_norm": 0.5171229401967155,
"learning_rate": 1.3201820940819423e-06,
"loss": 0.4538,
"step": 495
},
{
"epoch": 2.17877094972067,
"grad_norm": 0.5177579051508318,
"learning_rate": 1.3125948406676784e-06,
"loss": 0.4822,
"step": 496
},
{
"epoch": 2.183240223463687,
"grad_norm": 0.5363730349952008,
"learning_rate": 1.3050075872534144e-06,
"loss": 0.4879,
"step": 497
},
{
"epoch": 2.1877094972067037,
"grad_norm": 0.49547274684149156,
"learning_rate": 1.2974203338391503e-06,
"loss": 0.4732,
"step": 498
},
{
"epoch": 2.1921787709497207,
"grad_norm": 0.48560343902497966,
"learning_rate": 1.2898330804248862e-06,
"loss": 0.4918,
"step": 499
},
{
"epoch": 2.1966480446927372,
"grad_norm": 0.49966297308311575,
"learning_rate": 1.2822458270106223e-06,
"loss": 0.4618,
"step": 500
},
{
"epoch": 2.201117318435754,
"grad_norm": 0.472067295226221,
"learning_rate": 1.2746585735963582e-06,
"loss": 0.4953,
"step": 501
},
{
"epoch": 2.205586592178771,
"grad_norm": 0.5241464474412257,
"learning_rate": 1.2670713201820941e-06,
"loss": 0.4797,
"step": 502
},
{
"epoch": 2.2100558659217877,
"grad_norm": 0.5063960714766558,
"learning_rate": 1.25948406676783e-06,
"loss": 0.4493,
"step": 503
},
{
"epoch": 2.2145251396648042,
"grad_norm": 0.48407847297667117,
"learning_rate": 1.2518968133535662e-06,
"loss": 0.4736,
"step": 504
},
{
"epoch": 2.218994413407821,
"grad_norm": 0.4859389089228068,
"learning_rate": 1.244309559939302e-06,
"loss": 0.4476,
"step": 505
},
{
"epoch": 2.223463687150838,
"grad_norm": 0.49468343574845597,
"learning_rate": 1.236722306525038e-06,
"loss": 0.4447,
"step": 506
},
{
"epoch": 2.2279329608938547,
"grad_norm": 0.45980997966401754,
"learning_rate": 1.229135053110774e-06,
"loss": 0.4599,
"step": 507
},
{
"epoch": 2.2324022346368717,
"grad_norm": 0.5135776498149942,
"learning_rate": 1.22154779969651e-06,
"loss": 0.4852,
"step": 508
},
{
"epoch": 2.2368715083798882,
"grad_norm": 0.48749385873656575,
"learning_rate": 1.213960546282246e-06,
"loss": 0.5157,
"step": 509
},
{
"epoch": 2.241340782122905,
"grad_norm": 0.48899276420726745,
"learning_rate": 1.2063732928679819e-06,
"loss": 0.4437,
"step": 510
},
{
"epoch": 2.2458100558659218,
"grad_norm": 0.4918342780724834,
"learning_rate": 1.198786039453718e-06,
"loss": 0.4974,
"step": 511
},
{
"epoch": 2.2502793296089387,
"grad_norm": 0.4774969997416472,
"learning_rate": 1.191198786039454e-06,
"loss": 0.4509,
"step": 512
},
{
"epoch": 2.2547486033519553,
"grad_norm": 0.4791749757272505,
"learning_rate": 1.1836115326251896e-06,
"loss": 0.4599,
"step": 513
},
{
"epoch": 2.2592178770949722,
"grad_norm": 0.5072314853369309,
"learning_rate": 1.1760242792109257e-06,
"loss": 0.4643,
"step": 514
},
{
"epoch": 2.2636871508379888,
"grad_norm": 0.5004472392134528,
"learning_rate": 1.1684370257966617e-06,
"loss": 0.4688,
"step": 515
},
{
"epoch": 2.2681564245810057,
"grad_norm": 0.48150392498478517,
"learning_rate": 1.1608497723823976e-06,
"loss": 0.5069,
"step": 516
},
{
"epoch": 2.2726256983240223,
"grad_norm": 0.4828855445414688,
"learning_rate": 1.1532625189681337e-06,
"loss": 0.4571,
"step": 517
},
{
"epoch": 2.2770949720670393,
"grad_norm": 0.49789128855450915,
"learning_rate": 1.1456752655538696e-06,
"loss": 0.4651,
"step": 518
},
{
"epoch": 2.281564245810056,
"grad_norm": 0.49890243849633514,
"learning_rate": 1.1380880121396055e-06,
"loss": 0.4757,
"step": 519
},
{
"epoch": 2.2860335195530728,
"grad_norm": 0.48825603773254755,
"learning_rate": 1.1305007587253414e-06,
"loss": 0.4668,
"step": 520
},
{
"epoch": 2.2905027932960893,
"grad_norm": 0.48797823917514366,
"learning_rate": 1.1229135053110776e-06,
"loss": 0.4576,
"step": 521
},
{
"epoch": 2.2949720670391063,
"grad_norm": 0.4713265522467315,
"learning_rate": 1.1153262518968135e-06,
"loss": 0.4998,
"step": 522
},
{
"epoch": 2.299441340782123,
"grad_norm": 0.48731093463299924,
"learning_rate": 1.1077389984825494e-06,
"loss": 0.4583,
"step": 523
},
{
"epoch": 2.30391061452514,
"grad_norm": 0.5356992050391173,
"learning_rate": 1.1001517450682853e-06,
"loss": 0.4964,
"step": 524
},
{
"epoch": 2.3083798882681563,
"grad_norm": 0.5185855845215128,
"learning_rate": 1.0925644916540214e-06,
"loss": 0.4456,
"step": 525
},
{
"epoch": 2.3128491620111733,
"grad_norm": 0.5340341729511069,
"learning_rate": 1.0849772382397573e-06,
"loss": 0.4667,
"step": 526
},
{
"epoch": 2.31731843575419,
"grad_norm": 0.46791114784776194,
"learning_rate": 1.0773899848254933e-06,
"loss": 0.4907,
"step": 527
},
{
"epoch": 2.321787709497207,
"grad_norm": 0.5224682012273069,
"learning_rate": 1.0698027314112294e-06,
"loss": 0.4411,
"step": 528
},
{
"epoch": 2.3262569832402233,
"grad_norm": 0.5013175127206561,
"learning_rate": 1.062215477996965e-06,
"loss": 0.4872,
"step": 529
},
{
"epoch": 2.3307262569832403,
"grad_norm": 0.49567401298066655,
"learning_rate": 1.054628224582701e-06,
"loss": 0.4547,
"step": 530
},
{
"epoch": 2.335195530726257,
"grad_norm": 0.5092075971225333,
"learning_rate": 1.0470409711684371e-06,
"loss": 0.4774,
"step": 531
},
{
"epoch": 2.339664804469274,
"grad_norm": 0.5206537811339336,
"learning_rate": 1.039453717754173e-06,
"loss": 0.4593,
"step": 532
},
{
"epoch": 2.3441340782122904,
"grad_norm": 0.5247188237207532,
"learning_rate": 1.031866464339909e-06,
"loss": 0.4947,
"step": 533
},
{
"epoch": 2.3486033519553073,
"grad_norm": 0.48511496092549006,
"learning_rate": 1.024279210925645e-06,
"loss": 0.4675,
"step": 534
},
{
"epoch": 2.353072625698324,
"grad_norm": 0.49031229609343957,
"learning_rate": 1.016691957511381e-06,
"loss": 0.4654,
"step": 535
},
{
"epoch": 2.357541899441341,
"grad_norm": 0.5022437562865831,
"learning_rate": 1.009104704097117e-06,
"loss": 0.4881,
"step": 536
},
{
"epoch": 2.3620111731843574,
"grad_norm": 0.48427536854604036,
"learning_rate": 1.0015174506828528e-06,
"loss": 0.4722,
"step": 537
},
{
"epoch": 2.3664804469273744,
"grad_norm": 0.4951077677256604,
"learning_rate": 9.93930197268589e-07,
"loss": 0.4816,
"step": 538
},
{
"epoch": 2.370949720670391,
"grad_norm": 0.506971274674449,
"learning_rate": 9.863429438543249e-07,
"loss": 0.448,
"step": 539
},
{
"epoch": 2.375418994413408,
"grad_norm": 0.4843767211745505,
"learning_rate": 9.787556904400608e-07,
"loss": 0.4644,
"step": 540
},
{
"epoch": 2.3798882681564244,
"grad_norm": 0.4787312911085378,
"learning_rate": 9.711684370257967e-07,
"loss": 0.4681,
"step": 541
},
{
"epoch": 2.3843575418994414,
"grad_norm": 0.5040950848521704,
"learning_rate": 9.635811836115328e-07,
"loss": 0.4595,
"step": 542
},
{
"epoch": 2.388826815642458,
"grad_norm": 0.4742009596031962,
"learning_rate": 9.559939301972687e-07,
"loss": 0.4542,
"step": 543
},
{
"epoch": 2.393296089385475,
"grad_norm": 0.4819914757091554,
"learning_rate": 9.484066767830046e-07,
"loss": 0.4818,
"step": 544
},
{
"epoch": 2.3977653631284914,
"grad_norm": 0.4794763925814031,
"learning_rate": 9.408194233687407e-07,
"loss": 0.4459,
"step": 545
},
{
"epoch": 2.4022346368715084,
"grad_norm": 0.5065706041345039,
"learning_rate": 9.332321699544765e-07,
"loss": 0.4986,
"step": 546
},
{
"epoch": 2.4067039106145254,
"grad_norm": 0.4918492784353524,
"learning_rate": 9.256449165402125e-07,
"loss": 0.4618,
"step": 547
},
{
"epoch": 2.411173184357542,
"grad_norm": 0.48410824249150014,
"learning_rate": 9.180576631259484e-07,
"loss": 0.4358,
"step": 548
},
{
"epoch": 2.4156424581005584,
"grad_norm": 0.5071386174484,
"learning_rate": 9.104704097116844e-07,
"loss": 0.4522,
"step": 549
},
{
"epoch": 2.4201117318435754,
"grad_norm": 0.5179035085963222,
"learning_rate": 9.028831562974203e-07,
"loss": 0.4718,
"step": 550
},
{
"epoch": 2.4245810055865924,
"grad_norm": 0.48695002255677194,
"learning_rate": 8.952959028831563e-07,
"loss": 0.4946,
"step": 551
},
{
"epoch": 2.429050279329609,
"grad_norm": 0.4697467884861594,
"learning_rate": 8.877086494688924e-07,
"loss": 0.482,
"step": 552
},
{
"epoch": 2.4335195530726255,
"grad_norm": 0.5039869709064255,
"learning_rate": 8.801213960546283e-07,
"loss": 0.4528,
"step": 553
},
{
"epoch": 2.4379888268156424,
"grad_norm": 0.4956541767737936,
"learning_rate": 8.725341426403643e-07,
"loss": 0.4416,
"step": 554
},
{
"epoch": 2.4424581005586594,
"grad_norm": 0.5220748437476452,
"learning_rate": 8.649468892261002e-07,
"loss": 0.4534,
"step": 555
},
{
"epoch": 2.446927374301676,
"grad_norm": 0.5004114462806725,
"learning_rate": 8.573596358118362e-07,
"loss": 0.4523,
"step": 556
},
{
"epoch": 2.451396648044693,
"grad_norm": 0.5002883358165275,
"learning_rate": 8.497723823975721e-07,
"loss": 0.4483,
"step": 557
},
{
"epoch": 2.4558659217877095,
"grad_norm": 0.5001315929982186,
"learning_rate": 8.421851289833082e-07,
"loss": 0.4547,
"step": 558
},
{
"epoch": 2.4603351955307264,
"grad_norm": 0.5141518132214271,
"learning_rate": 8.345978755690441e-07,
"loss": 0.5064,
"step": 559
},
{
"epoch": 2.464804469273743,
"grad_norm": 0.4960566337702956,
"learning_rate": 8.270106221547801e-07,
"loss": 0.4473,
"step": 560
},
{
"epoch": 2.464804469273743,
"eval_loss": 0.4975128769874573,
"eval_runtime": 32.1892,
"eval_samples_per_second": 82.947,
"eval_steps_per_second": 5.188,
"step": 560
},
{
"epoch": 2.46927374301676,
"grad_norm": 0.4897198333575608,
"learning_rate": 8.19423368740516e-07,
"loss": 0.4559,
"step": 561
},
{
"epoch": 2.4737430167597765,
"grad_norm": 0.49985877359817477,
"learning_rate": 8.118361153262519e-07,
"loss": 0.4651,
"step": 562
},
{
"epoch": 2.4782122905027935,
"grad_norm": 0.49485306049230043,
"learning_rate": 8.042488619119878e-07,
"loss": 0.4803,
"step": 563
},
{
"epoch": 2.48268156424581,
"grad_norm": 0.5208900276481403,
"learning_rate": 7.966616084977239e-07,
"loss": 0.4553,
"step": 564
},
{
"epoch": 2.487150837988827,
"grad_norm": 0.48837650054879367,
"learning_rate": 7.890743550834598e-07,
"loss": 0.4681,
"step": 565
},
{
"epoch": 2.4916201117318435,
"grad_norm": 0.47305516115923485,
"learning_rate": 7.814871016691958e-07,
"loss": 0.4564,
"step": 566
},
{
"epoch": 2.4960893854748605,
"grad_norm": 0.5049602967209187,
"learning_rate": 7.738998482549317e-07,
"loss": 0.4632,
"step": 567
},
{
"epoch": 2.500558659217877,
"grad_norm": 0.48547661489581506,
"learning_rate": 7.663125948406677e-07,
"loss": 0.448,
"step": 568
},
{
"epoch": 2.505027932960894,
"grad_norm": 0.5747686136950829,
"learning_rate": 7.587253414264036e-07,
"loss": 0.4432,
"step": 569
},
{
"epoch": 2.5094972067039105,
"grad_norm": 0.5499322063415041,
"learning_rate": 7.511380880121397e-07,
"loss": 0.432,
"step": 570
},
{
"epoch": 2.5139664804469275,
"grad_norm": 0.49909881450641125,
"learning_rate": 7.435508345978757e-07,
"loss": 0.4325,
"step": 571
},
{
"epoch": 2.518435754189944,
"grad_norm": 0.5137001617086465,
"learning_rate": 7.359635811836116e-07,
"loss": 0.4374,
"step": 572
},
{
"epoch": 2.522905027932961,
"grad_norm": 0.5043757440405017,
"learning_rate": 7.283763277693476e-07,
"loss": 0.4751,
"step": 573
},
{
"epoch": 2.5273743016759775,
"grad_norm": 0.4882894387336232,
"learning_rate": 7.207890743550835e-07,
"loss": 0.4368,
"step": 574
},
{
"epoch": 2.5318435754189945,
"grad_norm": 0.5302357654250439,
"learning_rate": 7.132018209408196e-07,
"loss": 0.4174,
"step": 575
},
{
"epoch": 2.536312849162011,
"grad_norm": 0.5229623842607946,
"learning_rate": 7.056145675265555e-07,
"loss": 0.4353,
"step": 576
},
{
"epoch": 2.540782122905028,
"grad_norm": 0.5145290981136341,
"learning_rate": 6.980273141122915e-07,
"loss": 0.4278,
"step": 577
},
{
"epoch": 2.5452513966480446,
"grad_norm": 0.49830066459201744,
"learning_rate": 6.904400606980274e-07,
"loss": 0.434,
"step": 578
},
{
"epoch": 2.5497206703910615,
"grad_norm": 0.5220459047427843,
"learning_rate": 6.828528072837633e-07,
"loss": 0.4375,
"step": 579
},
{
"epoch": 2.554189944134078,
"grad_norm": 0.5111391623585094,
"learning_rate": 6.752655538694992e-07,
"loss": 0.4295,
"step": 580
},
{
"epoch": 2.558659217877095,
"grad_norm": 0.4944043554942434,
"learning_rate": 6.676783004552352e-07,
"loss": 0.4205,
"step": 581
},
{
"epoch": 2.5631284916201116,
"grad_norm": 0.49154601956517674,
"learning_rate": 6.600910470409712e-07,
"loss": 0.4502,
"step": 582
},
{
"epoch": 2.5675977653631286,
"grad_norm": 0.5008348019898189,
"learning_rate": 6.525037936267072e-07,
"loss": 0.4453,
"step": 583
},
{
"epoch": 2.572067039106145,
"grad_norm": 0.5005572885290496,
"learning_rate": 6.449165402124431e-07,
"loss": 0.4353,
"step": 584
},
{
"epoch": 2.576536312849162,
"grad_norm": 0.5302967367534124,
"learning_rate": 6.373292867981791e-07,
"loss": 0.4269,
"step": 585
},
{
"epoch": 2.5810055865921786,
"grad_norm": 0.5050943269077487,
"learning_rate": 6.29742033383915e-07,
"loss": 0.4372,
"step": 586
},
{
"epoch": 2.5854748603351956,
"grad_norm": 0.5223556534474804,
"learning_rate": 6.22154779969651e-07,
"loss": 0.445,
"step": 587
},
{
"epoch": 2.5899441340782126,
"grad_norm": 0.4891137574673572,
"learning_rate": 6.14567526555387e-07,
"loss": 0.4171,
"step": 588
},
{
"epoch": 2.594413407821229,
"grad_norm": 0.5172276124686148,
"learning_rate": 6.06980273141123e-07,
"loss": 0.4522,
"step": 589
},
{
"epoch": 2.5988826815642456,
"grad_norm": 0.48873823307103154,
"learning_rate": 5.99393019726859e-07,
"loss": 0.4286,
"step": 590
},
{
"epoch": 2.6033519553072626,
"grad_norm": 0.502230133682332,
"learning_rate": 5.918057663125948e-07,
"loss": 0.4461,
"step": 591
},
{
"epoch": 2.6078212290502796,
"grad_norm": 0.4898396410998909,
"learning_rate": 5.842185128983308e-07,
"loss": 0.4289,
"step": 592
},
{
"epoch": 2.612290502793296,
"grad_norm": 0.48422568768513635,
"learning_rate": 5.766312594840668e-07,
"loss": 0.4423,
"step": 593
},
{
"epoch": 2.6167597765363126,
"grad_norm": 0.5003994120834477,
"learning_rate": 5.690440060698028e-07,
"loss": 0.437,
"step": 594
},
{
"epoch": 2.6212290502793296,
"grad_norm": 0.5079705228889307,
"learning_rate": 5.614567526555388e-07,
"loss": 0.4516,
"step": 595
},
{
"epoch": 2.6256983240223466,
"grad_norm": 0.49246631419316356,
"learning_rate": 5.538694992412747e-07,
"loss": 0.4368,
"step": 596
},
{
"epoch": 2.630167597765363,
"grad_norm": 0.5086922889782935,
"learning_rate": 5.462822458270107e-07,
"loss": 0.4227,
"step": 597
},
{
"epoch": 2.6346368715083797,
"grad_norm": 0.4780759168718152,
"learning_rate": 5.386949924127466e-07,
"loss": 0.4462,
"step": 598
},
{
"epoch": 2.6391061452513966,
"grad_norm": 0.464930849547806,
"learning_rate": 5.311077389984825e-07,
"loss": 0.4334,
"step": 599
},
{
"epoch": 2.6435754189944136,
"grad_norm": 0.5002300455424397,
"learning_rate": 5.235204855842186e-07,
"loss": 0.4641,
"step": 600
},
{
"epoch": 2.64804469273743,
"grad_norm": 0.4901819784256759,
"learning_rate": 5.159332321699545e-07,
"loss": 0.4534,
"step": 601
},
{
"epoch": 2.6525139664804467,
"grad_norm": 0.48520187595124403,
"learning_rate": 5.083459787556905e-07,
"loss": 0.4246,
"step": 602
},
{
"epoch": 2.6569832402234637,
"grad_norm": 0.5093634792454178,
"learning_rate": 5.007587253414264e-07,
"loss": 0.458,
"step": 603
},
{
"epoch": 2.6614525139664806,
"grad_norm": 0.47795678253705354,
"learning_rate": 4.931714719271624e-07,
"loss": 0.4535,
"step": 604
},
{
"epoch": 2.665921787709497,
"grad_norm": 0.48340944166993405,
"learning_rate": 4.855842185128983e-07,
"loss": 0.4721,
"step": 605
},
{
"epoch": 2.6703910614525137,
"grad_norm": 0.47996848455446794,
"learning_rate": 4.779969650986344e-07,
"loss": 0.4718,
"step": 606
},
{
"epoch": 2.6748603351955307,
"grad_norm": 0.5064110255390989,
"learning_rate": 4.7040971168437033e-07,
"loss": 0.4232,
"step": 607
},
{
"epoch": 2.6793296089385477,
"grad_norm": 0.4824240302682651,
"learning_rate": 4.6282245827010624e-07,
"loss": 0.4307,
"step": 608
},
{
"epoch": 2.683798882681564,
"grad_norm": 0.4920800315955025,
"learning_rate": 4.552352048558422e-07,
"loss": 0.4435,
"step": 609
},
{
"epoch": 2.688268156424581,
"grad_norm": 0.48921504704563673,
"learning_rate": 4.476479514415782e-07,
"loss": 0.4409,
"step": 610
},
{
"epoch": 2.6927374301675977,
"grad_norm": 0.500357475284234,
"learning_rate": 4.4006069802731414e-07,
"loss": 0.4512,
"step": 611
},
{
"epoch": 2.6972067039106147,
"grad_norm": 0.4772853981005586,
"learning_rate": 4.324734446130501e-07,
"loss": 0.4305,
"step": 612
},
{
"epoch": 2.701675977653631,
"grad_norm": 0.5075964338612743,
"learning_rate": 4.248861911987861e-07,
"loss": 0.4285,
"step": 613
},
{
"epoch": 2.706145251396648,
"grad_norm": 0.4916721748305679,
"learning_rate": 4.1729893778452204e-07,
"loss": 0.4586,
"step": 614
},
{
"epoch": 2.7106145251396647,
"grad_norm": 0.4746003660659448,
"learning_rate": 4.09711684370258e-07,
"loss": 0.4385,
"step": 615
},
{
"epoch": 2.7150837988826817,
"grad_norm": 0.49004649736827927,
"learning_rate": 4.021244309559939e-07,
"loss": 0.4349,
"step": 616
},
{
"epoch": 2.7195530726256982,
"grad_norm": 0.5055169203153188,
"learning_rate": 3.945371775417299e-07,
"loss": 0.4741,
"step": 617
},
{
"epoch": 2.724022346368715,
"grad_norm": 0.5491416139361283,
"learning_rate": 3.8694992412746586e-07,
"loss": 0.4285,
"step": 618
},
{
"epoch": 2.7284916201117317,
"grad_norm": 0.4860669923591883,
"learning_rate": 3.793626707132018e-07,
"loss": 0.4147,
"step": 619
},
{
"epoch": 2.7329608938547487,
"grad_norm": 0.48537127261627433,
"learning_rate": 3.7177541729893784e-07,
"loss": 0.4538,
"step": 620
},
{
"epoch": 2.7374301675977653,
"grad_norm": 0.47557197343237195,
"learning_rate": 3.641881638846738e-07,
"loss": 0.454,
"step": 621
},
{
"epoch": 2.7418994413407822,
"grad_norm": 0.4844044051975442,
"learning_rate": 3.566009104704098e-07,
"loss": 0.4273,
"step": 622
},
{
"epoch": 2.7463687150837988,
"grad_norm": 0.5000172198487197,
"learning_rate": 3.4901365705614574e-07,
"loss": 0.4315,
"step": 623
},
{
"epoch": 2.7508379888268157,
"grad_norm": 0.4848986778566373,
"learning_rate": 3.4142640364188166e-07,
"loss": 0.4495,
"step": 624
},
{
"epoch": 2.7553072625698323,
"grad_norm": 0.48082262468066483,
"learning_rate": 3.338391502276176e-07,
"loss": 0.4185,
"step": 625
},
{
"epoch": 2.7597765363128492,
"grad_norm": 0.4865964770369407,
"learning_rate": 3.262518968133536e-07,
"loss": 0.4538,
"step": 626
},
{
"epoch": 2.764245810055866,
"grad_norm": 0.4815694373638968,
"learning_rate": 3.1866464339908956e-07,
"loss": 0.459,
"step": 627
},
{
"epoch": 2.7687150837988828,
"grad_norm": 0.49485301744734556,
"learning_rate": 3.110773899848255e-07,
"loss": 0.4527,
"step": 628
},
{
"epoch": 2.7731843575418993,
"grad_norm": 0.5214886405175438,
"learning_rate": 3.034901365705615e-07,
"loss": 0.4393,
"step": 629
},
{
"epoch": 2.7776536312849163,
"grad_norm": 0.4645817613806632,
"learning_rate": 2.959028831562974e-07,
"loss": 0.4723,
"step": 630
},
{
"epoch": 2.782122905027933,
"grad_norm": 0.4887881586065441,
"learning_rate": 2.883156297420334e-07,
"loss": 0.418,
"step": 631
},
{
"epoch": 2.78659217877095,
"grad_norm": 0.47562569136752975,
"learning_rate": 2.807283763277694e-07,
"loss": 0.4347,
"step": 632
},
{
"epoch": 2.7910614525139668,
"grad_norm": 0.4848933584055101,
"learning_rate": 2.7314112291350536e-07,
"loss": 0.4266,
"step": 633
},
{
"epoch": 2.7955307262569833,
"grad_norm": 0.48829866920710585,
"learning_rate": 2.6555386949924127e-07,
"loss": 0.4522,
"step": 634
},
{
"epoch": 2.8,
"grad_norm": 0.4776396492507318,
"learning_rate": 2.5796661608497724e-07,
"loss": 0.4293,
"step": 635
},
{
"epoch": 2.804469273743017,
"grad_norm": 0.48342626581685083,
"learning_rate": 2.503793626707132e-07,
"loss": 0.437,
"step": 636
},
{
"epoch": 2.8089385474860338,
"grad_norm": 0.47481545268800873,
"learning_rate": 2.4279210925644917e-07,
"loss": 0.4333,
"step": 637
},
{
"epoch": 2.8134078212290503,
"grad_norm": 0.5105857690012979,
"learning_rate": 2.3520485584218516e-07,
"loss": 0.4367,
"step": 638
},
{
"epoch": 2.817877094972067,
"grad_norm": 0.49397311695562557,
"learning_rate": 2.276176024279211e-07,
"loss": 0.4355,
"step": 639
},
{
"epoch": 2.822346368715084,
"grad_norm": 0.4766966322472669,
"learning_rate": 2.2003034901365707e-07,
"loss": 0.4478,
"step": 640
},
{
"epoch": 2.826815642458101,
"grad_norm": 0.47484165703817144,
"learning_rate": 2.1244309559939304e-07,
"loss": 0.4366,
"step": 641
},
{
"epoch": 2.8312849162011173,
"grad_norm": 0.47271019258961905,
"learning_rate": 2.04855842185129e-07,
"loss": 0.4473,
"step": 642
},
{
"epoch": 2.835754189944134,
"grad_norm": 0.4668544421054751,
"learning_rate": 1.9726858877086494e-07,
"loss": 0.4341,
"step": 643
},
{
"epoch": 2.840223463687151,
"grad_norm": 0.47716986376667936,
"learning_rate": 1.896813353566009e-07,
"loss": 0.4223,
"step": 644
},
{
"epoch": 2.844692737430168,
"grad_norm": 0.4939991882199546,
"learning_rate": 1.820940819423369e-07,
"loss": 0.4441,
"step": 645
},
{
"epoch": 2.8491620111731844,
"grad_norm": 0.47705877463455093,
"learning_rate": 1.7450682852807287e-07,
"loss": 0.485,
"step": 646
},
{
"epoch": 2.853631284916201,
"grad_norm": 0.49941990019846055,
"learning_rate": 1.669195751138088e-07,
"loss": 0.4701,
"step": 647
},
{
"epoch": 2.858100558659218,
"grad_norm": 0.47625945090740057,
"learning_rate": 1.5933232169954478e-07,
"loss": 0.454,
"step": 648
},
{
"epoch": 2.862569832402235,
"grad_norm": 0.5028271948555065,
"learning_rate": 1.5174506828528074e-07,
"loss": 0.4296,
"step": 649
},
{
"epoch": 2.8670391061452514,
"grad_norm": 0.49358352504671604,
"learning_rate": 1.441578148710167e-07,
"loss": 0.4324,
"step": 650
},
{
"epoch": 2.871508379888268,
"grad_norm": 0.48576223310259714,
"learning_rate": 1.3657056145675268e-07,
"loss": 0.4424,
"step": 651
},
{
"epoch": 2.875977653631285,
"grad_norm": 0.47971171549572994,
"learning_rate": 1.2898330804248862e-07,
"loss": 0.4555,
"step": 652
},
{
"epoch": 2.880446927374302,
"grad_norm": 0.49955392754214234,
"learning_rate": 1.2139605462822459e-07,
"loss": 0.4338,
"step": 653
},
{
"epoch": 2.8849162011173184,
"grad_norm": 0.47232821100099637,
"learning_rate": 1.1380880121396055e-07,
"loss": 0.4349,
"step": 654
},
{
"epoch": 2.889385474860335,
"grad_norm": 0.4834170945858496,
"learning_rate": 1.0622154779969652e-07,
"loss": 0.4564,
"step": 655
},
{
"epoch": 2.893854748603352,
"grad_norm": 0.48493984226629705,
"learning_rate": 9.863429438543247e-08,
"loss": 0.4384,
"step": 656
},
{
"epoch": 2.898324022346369,
"grad_norm": 0.4601302405573049,
"learning_rate": 9.104704097116845e-08,
"loss": 0.4512,
"step": 657
},
{
"epoch": 2.9027932960893854,
"grad_norm": 0.4768143158802154,
"learning_rate": 8.34597875569044e-08,
"loss": 0.4354,
"step": 658
},
{
"epoch": 2.9072625698324024,
"grad_norm": 0.49427262358096746,
"learning_rate": 7.587253414264037e-08,
"loss": 0.4259,
"step": 659
},
{
"epoch": 2.911731843575419,
"grad_norm": 0.4749814243858473,
"learning_rate": 6.828528072837634e-08,
"loss": 0.4445,
"step": 660
},
{
"epoch": 2.916201117318436,
"grad_norm": 0.48533516583361297,
"learning_rate": 6.069802731411229e-08,
"loss": 0.4799,
"step": 661
},
{
"epoch": 2.9206703910614524,
"grad_norm": 0.4839142899304728,
"learning_rate": 5.311077389984826e-08,
"loss": 0.4574,
"step": 662
},
{
"epoch": 2.9251396648044694,
"grad_norm": 0.486543073762813,
"learning_rate": 4.5523520485584226e-08,
"loss": 0.4446,
"step": 663
},
{
"epoch": 2.929608938547486,
"grad_norm": 0.4980109939724879,
"learning_rate": 3.7936267071320186e-08,
"loss": 0.4315,
"step": 664
},
{
"epoch": 2.934078212290503,
"grad_norm": 0.48840296818117557,
"learning_rate": 3.0349013657056146e-08,
"loss": 0.4318,
"step": 665
},
{
"epoch": 2.9385474860335195,
"grad_norm": 0.4903615270099699,
"learning_rate": 2.2761760242792113e-08,
"loss": 0.4357,
"step": 666
},
{
"epoch": 2.9430167597765364,
"grad_norm": 0.48850430880690693,
"learning_rate": 1.5174506828528073e-08,
"loss": 0.4334,
"step": 667
},
{
"epoch": 2.947486033519553,
"grad_norm": 0.4773488572878224,
"learning_rate": 7.587253414264037e-09,
"loss": 0.4281,
"step": 668
},
{
"epoch": 2.95195530726257,
"grad_norm": 0.4760606483764159,
"learning_rate": 0.0,
"loss": 0.4504,
"step": 669
}
],
"logging_steps": 1,
"max_steps": 669,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 56,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 559881199288320.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}