|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 31635, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00948316737790422, |
|
"grad_norm": 3.985076904296875, |
|
"learning_rate": 9.7e-06, |
|
"loss": 2.4371, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01896633475580844, |
|
"grad_norm": 3.8551318645477295, |
|
"learning_rate": 1.97e-05, |
|
"loss": 2.1056, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.02844950213371266, |
|
"grad_norm": 4.302079200744629, |
|
"learning_rate": 2.97e-05, |
|
"loss": 1.9608, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.03793266951161688, |
|
"grad_norm": 3.31756329536438, |
|
"learning_rate": 3.97e-05, |
|
"loss": 1.8338, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.0474158368895211, |
|
"grad_norm": 2.4619405269622803, |
|
"learning_rate": 4.97e-05, |
|
"loss": 1.7855, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.0474158368895211, |
|
"eval_loss": 1.6501274108886719, |
|
"eval_runtime": 72.2019, |
|
"eval_samples_per_second": 129.83, |
|
"eval_steps_per_second": 16.232, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05689900426742532, |
|
"grad_norm": 2.553483724594116, |
|
"learning_rate": 4.9844226754456404e-05, |
|
"loss": 1.7277, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.06638217164532954, |
|
"grad_norm": 2.0428194999694824, |
|
"learning_rate": 4.9683635779669185e-05, |
|
"loss": 1.6971, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.07586533902323377, |
|
"grad_norm": 1.9449608325958252, |
|
"learning_rate": 4.9523044804881966e-05, |
|
"loss": 1.6537, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.08534850640113797, |
|
"grad_norm": 2.5439252853393555, |
|
"learning_rate": 4.9362453830094753e-05, |
|
"loss": 1.6464, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.0948316737790422, |
|
"grad_norm": 2.118544578552246, |
|
"learning_rate": 4.9201862855307534e-05, |
|
"loss": 1.5804, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.0948316737790422, |
|
"eval_loss": 1.5088456869125366, |
|
"eval_runtime": 72.0739, |
|
"eval_samples_per_second": 130.061, |
|
"eval_steps_per_second": 16.261, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.10431484115694642, |
|
"grad_norm": 1.8551363945007324, |
|
"learning_rate": 4.9041271880520315e-05, |
|
"loss": 1.6341, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.11379800853485064, |
|
"grad_norm": 1.9903297424316406, |
|
"learning_rate": 4.88806809057331e-05, |
|
"loss": 1.5718, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.12328117591275486, |
|
"grad_norm": 2.2142210006713867, |
|
"learning_rate": 4.8720089930945884e-05, |
|
"loss": 1.5718, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.1327643432906591, |
|
"grad_norm": 2.2737417221069336, |
|
"learning_rate": 4.8559498956158664e-05, |
|
"loss": 1.5137, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.1422475106685633, |
|
"grad_norm": 2.3361587524414062, |
|
"learning_rate": 4.839890798137145e-05, |
|
"loss": 1.5332, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.1422475106685633, |
|
"eval_loss": 1.4451285600662231, |
|
"eval_runtime": 72.138, |
|
"eval_samples_per_second": 129.945, |
|
"eval_steps_per_second": 16.247, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.15173067804646753, |
|
"grad_norm": 2.335610866546631, |
|
"learning_rate": 4.823831700658423e-05, |
|
"loss": 1.5669, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.16121384542437173, |
|
"grad_norm": 1.811543583869934, |
|
"learning_rate": 4.8077726031797014e-05, |
|
"loss": 1.4985, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.17069701280227595, |
|
"grad_norm": 2.1588528156280518, |
|
"learning_rate": 4.79171350570098e-05, |
|
"loss": 1.4979, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.18018018018018017, |
|
"grad_norm": 1.7643985748291016, |
|
"learning_rate": 4.775654408222258e-05, |
|
"loss": 1.5246, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.1896633475580844, |
|
"grad_norm": 1.9193495512008667, |
|
"learning_rate": 4.759595310743536e-05, |
|
"loss": 1.4915, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.1896633475580844, |
|
"eval_loss": 1.403477430343628, |
|
"eval_runtime": 71.9579, |
|
"eval_samples_per_second": 130.271, |
|
"eval_steps_per_second": 16.287, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.19914651493598862, |
|
"grad_norm": 1.8307377099990845, |
|
"learning_rate": 4.743536213264815e-05, |
|
"loss": 1.5009, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.20862968231389284, |
|
"grad_norm": 1.7923104763031006, |
|
"learning_rate": 4.727477115786093e-05, |
|
"loss": 1.4968, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.21811284969179706, |
|
"grad_norm": 1.925938367843628, |
|
"learning_rate": 4.711418018307371e-05, |
|
"loss": 1.4696, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.22759601706970128, |
|
"grad_norm": 2.106110095977783, |
|
"learning_rate": 4.69535892082865e-05, |
|
"loss": 1.4853, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.2370791844476055, |
|
"grad_norm": 2.345017433166504, |
|
"learning_rate": 4.679299823349928e-05, |
|
"loss": 1.4868, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.2370791844476055, |
|
"eval_loss": 1.3772392272949219, |
|
"eval_runtime": 72.0321, |
|
"eval_samples_per_second": 130.136, |
|
"eval_steps_per_second": 16.271, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.24656235182550973, |
|
"grad_norm": 1.5003846883773804, |
|
"learning_rate": 4.663240725871206e-05, |
|
"loss": 1.4641, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.25604551920341395, |
|
"grad_norm": 1.8472124338150024, |
|
"learning_rate": 4.647181628392485e-05, |
|
"loss": 1.4594, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.2655286865813182, |
|
"grad_norm": 1.8818256855010986, |
|
"learning_rate": 4.631122530913763e-05, |
|
"loss": 1.4547, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.2750118539592224, |
|
"grad_norm": 1.5926233530044556, |
|
"learning_rate": 4.615063433435041e-05, |
|
"loss": 1.4414, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.2844950213371266, |
|
"grad_norm": 1.505327820777893, |
|
"learning_rate": 4.59900433595632e-05, |
|
"loss": 1.4165, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.2844950213371266, |
|
"eval_loss": 1.3518378734588623, |
|
"eval_runtime": 71.9886, |
|
"eval_samples_per_second": 130.215, |
|
"eval_steps_per_second": 16.28, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.29397818871503084, |
|
"grad_norm": 1.77092707157135, |
|
"learning_rate": 4.582945238477598e-05, |
|
"loss": 1.4222, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.30346135609293506, |
|
"grad_norm": 2.265411376953125, |
|
"learning_rate": 4.566886140998876e-05, |
|
"loss": 1.3973, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.3129445234708393, |
|
"grad_norm": 1.4207345247268677, |
|
"learning_rate": 4.550827043520154e-05, |
|
"loss": 1.4423, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.32242769084874345, |
|
"grad_norm": 1.72047758102417, |
|
"learning_rate": 4.534767946041433e-05, |
|
"loss": 1.3939, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.3319108582266477, |
|
"grad_norm": 1.7695670127868652, |
|
"learning_rate": 4.518708848562711e-05, |
|
"loss": 1.3911, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.3319108582266477, |
|
"eval_loss": 1.3347505331039429, |
|
"eval_runtime": 72.0526, |
|
"eval_samples_per_second": 130.099, |
|
"eval_steps_per_second": 16.266, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.3413940256045519, |
|
"grad_norm": 1.93614661693573, |
|
"learning_rate": 4.502649751083989e-05, |
|
"loss": 1.405, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.3508771929824561, |
|
"grad_norm": 1.4412301778793335, |
|
"learning_rate": 4.486590653605268e-05, |
|
"loss": 1.421, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.36036036036036034, |
|
"grad_norm": 1.5761134624481201, |
|
"learning_rate": 4.470531556126546e-05, |
|
"loss": 1.3758, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.36984352773826457, |
|
"grad_norm": 1.7923239469528198, |
|
"learning_rate": 4.454472458647824e-05, |
|
"loss": 1.4087, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.3793266951161688, |
|
"grad_norm": 2.2492587566375732, |
|
"learning_rate": 4.438413361169103e-05, |
|
"loss": 1.3797, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.3793266951161688, |
|
"eval_loss": 1.3214360475540161, |
|
"eval_runtime": 72.0741, |
|
"eval_samples_per_second": 130.061, |
|
"eval_steps_per_second": 16.261, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.388809862494073, |
|
"grad_norm": 1.978060245513916, |
|
"learning_rate": 4.422354263690381e-05, |
|
"loss": 1.4024, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.39829302987197723, |
|
"grad_norm": 1.7838459014892578, |
|
"learning_rate": 4.406295166211659e-05, |
|
"loss": 1.4047, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.40777619724988146, |
|
"grad_norm": 1.682637333869934, |
|
"learning_rate": 4.3902360687329377e-05, |
|
"loss": 1.3709, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.4172593646277857, |
|
"grad_norm": 1.5510674715042114, |
|
"learning_rate": 4.374176971254216e-05, |
|
"loss": 1.4175, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.4267425320056899, |
|
"grad_norm": 1.7401492595672607, |
|
"learning_rate": 4.358117873775494e-05, |
|
"loss": 1.3801, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.4267425320056899, |
|
"eval_loss": 1.3049076795578003, |
|
"eval_runtime": 72.1294, |
|
"eval_samples_per_second": 129.961, |
|
"eval_steps_per_second": 16.249, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.4362256993835941, |
|
"grad_norm": 1.6590989828109741, |
|
"learning_rate": 4.3420587762967726e-05, |
|
"loss": 1.3827, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.44570886676149835, |
|
"grad_norm": 1.5440171957015991, |
|
"learning_rate": 4.325999678818051e-05, |
|
"loss": 1.3617, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.45519203413940257, |
|
"grad_norm": 1.716539978981018, |
|
"learning_rate": 4.309940581339329e-05, |
|
"loss": 1.3463, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.4646752015173068, |
|
"grad_norm": 1.3042521476745605, |
|
"learning_rate": 4.2938814838606075e-05, |
|
"loss": 1.3456, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.474158368895211, |
|
"grad_norm": 1.3467687368392944, |
|
"learning_rate": 4.2778223863818856e-05, |
|
"loss": 1.3559, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.474158368895211, |
|
"eval_loss": 1.2918757200241089, |
|
"eval_runtime": 72.0072, |
|
"eval_samples_per_second": 130.181, |
|
"eval_steps_per_second": 16.276, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.48364153627311524, |
|
"grad_norm": 1.3807010650634766, |
|
"learning_rate": 4.261763288903164e-05, |
|
"loss": 1.3507, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.49312470365101946, |
|
"grad_norm": 1.3885177373886108, |
|
"learning_rate": 4.2457041914244425e-05, |
|
"loss": 1.3552, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.5026078710289237, |
|
"grad_norm": 1.2807698249816895, |
|
"learning_rate": 4.2296450939457205e-05, |
|
"loss": 1.3642, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.5120910384068279, |
|
"grad_norm": 1.4009428024291992, |
|
"learning_rate": 4.2135859964669986e-05, |
|
"loss": 1.3781, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.5215742057847321, |
|
"grad_norm": 1.3763035535812378, |
|
"learning_rate": 4.1975268989882774e-05, |
|
"loss": 1.3717, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.5215742057847321, |
|
"eval_loss": 1.280537724494934, |
|
"eval_runtime": 72.1115, |
|
"eval_samples_per_second": 129.993, |
|
"eval_steps_per_second": 16.253, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.5310573731626363, |
|
"grad_norm": 1.5511786937713623, |
|
"learning_rate": 4.1814678015095555e-05, |
|
"loss": 1.3502, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.5405405405405406, |
|
"grad_norm": 1.4995437860488892, |
|
"learning_rate": 4.1654087040308336e-05, |
|
"loss": 1.3599, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.5500237079184448, |
|
"grad_norm": 1.3496274948120117, |
|
"learning_rate": 4.149349606552112e-05, |
|
"loss": 1.3421, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.559506875296349, |
|
"grad_norm": 1.3634631633758545, |
|
"learning_rate": 4.1332905090733904e-05, |
|
"loss": 1.3617, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.5689900426742532, |
|
"grad_norm": 1.5579423904418945, |
|
"learning_rate": 4.1172314115946685e-05, |
|
"loss": 1.3604, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.5689900426742532, |
|
"eval_loss": 1.2698478698730469, |
|
"eval_runtime": 72.1231, |
|
"eval_samples_per_second": 129.972, |
|
"eval_steps_per_second": 16.25, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.5784732100521575, |
|
"grad_norm": 1.380241870880127, |
|
"learning_rate": 4.101332905090734e-05, |
|
"loss": 1.3379, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.5879563774300617, |
|
"grad_norm": 1.764551043510437, |
|
"learning_rate": 4.085273807612012e-05, |
|
"loss": 1.3208, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.5974395448079659, |
|
"grad_norm": 1.627012848854065, |
|
"learning_rate": 4.069214710133291e-05, |
|
"loss": 1.3448, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.6069227121858701, |
|
"grad_norm": 1.539115071296692, |
|
"learning_rate": 4.053155612654569e-05, |
|
"loss": 1.3422, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.6164058795637744, |
|
"grad_norm": 1.4698444604873657, |
|
"learning_rate": 4.037257106150635e-05, |
|
"loss": 1.3264, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.6164058795637744, |
|
"eval_loss": 1.259299635887146, |
|
"eval_runtime": 72.1176, |
|
"eval_samples_per_second": 129.982, |
|
"eval_steps_per_second": 16.251, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.6258890469416786, |
|
"grad_norm": 1.8150815963745117, |
|
"learning_rate": 4.021198008671913e-05, |
|
"loss": 1.3262, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.6353722143195828, |
|
"grad_norm": 1.4278889894485474, |
|
"learning_rate": 4.005138911193191e-05, |
|
"loss": 1.334, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.6448553816974869, |
|
"grad_norm": 1.4713215827941895, |
|
"learning_rate": 3.98907981371447e-05, |
|
"loss": 1.2924, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.6543385490753911, |
|
"grad_norm": 1.626541018486023, |
|
"learning_rate": 3.9731813072105354e-05, |
|
"loss": 1.3057, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.6638217164532954, |
|
"grad_norm": 1.7835373878479004, |
|
"learning_rate": 3.9571222097318134e-05, |
|
"loss": 1.328, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.6638217164532954, |
|
"eval_loss": 1.252388834953308, |
|
"eval_runtime": 72.2427, |
|
"eval_samples_per_second": 129.757, |
|
"eval_steps_per_second": 16.223, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.6733048838311996, |
|
"grad_norm": 1.8675563335418701, |
|
"learning_rate": 3.9410631122530915e-05, |
|
"loss": 1.322, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.6827880512091038, |
|
"grad_norm": 1.5719430446624756, |
|
"learning_rate": 3.92500401477437e-05, |
|
"loss": 1.3464, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.692271218587008, |
|
"grad_norm": 1.5038641691207886, |
|
"learning_rate": 3.9089449172956484e-05, |
|
"loss": 1.3315, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.7017543859649122, |
|
"grad_norm": 1.777970314025879, |
|
"learning_rate": 3.8928858198169265e-05, |
|
"loss": 1.3549, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.7112375533428165, |
|
"grad_norm": 1.8796472549438477, |
|
"learning_rate": 3.8768267223382045e-05, |
|
"loss": 1.2907, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.7112375533428165, |
|
"eval_loss": 1.2450358867645264, |
|
"eval_runtime": 72.1657, |
|
"eval_samples_per_second": 129.895, |
|
"eval_steps_per_second": 16.24, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.7207207207207207, |
|
"grad_norm": 1.7477796077728271, |
|
"learning_rate": 3.860767624859483e-05, |
|
"loss": 1.3196, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.7302038880986249, |
|
"grad_norm": 1.6598505973815918, |
|
"learning_rate": 3.8447085273807614e-05, |
|
"loss": 1.2799, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.7396870554765291, |
|
"grad_norm": 1.7319283485412598, |
|
"learning_rate": 3.8286494299020395e-05, |
|
"loss": 1.3354, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.7491702228544334, |
|
"grad_norm": 1.847347617149353, |
|
"learning_rate": 3.812590332423318e-05, |
|
"loss": 1.3034, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.7586533902323376, |
|
"grad_norm": 1.6584995985031128, |
|
"learning_rate": 3.796531234944596e-05, |
|
"loss": 1.3092, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.7586533902323376, |
|
"eval_loss": 1.2385543584823608, |
|
"eval_runtime": 72.1594, |
|
"eval_samples_per_second": 129.907, |
|
"eval_steps_per_second": 16.242, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.7681365576102418, |
|
"grad_norm": 1.581036925315857, |
|
"learning_rate": 3.7804721374658744e-05, |
|
"loss": 1.3064, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.777619724988146, |
|
"grad_norm": 1.6824501752853394, |
|
"learning_rate": 3.764413039987153e-05, |
|
"loss": 1.3039, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.7871028923660502, |
|
"grad_norm": 1.4804019927978516, |
|
"learning_rate": 3.748353942508431e-05, |
|
"loss": 1.2774, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.7965860597439545, |
|
"grad_norm": 1.5401322841644287, |
|
"learning_rate": 3.732294845029709e-05, |
|
"loss": 1.3042, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.8060692271218587, |
|
"grad_norm": 1.9226937294006348, |
|
"learning_rate": 3.716235747550988e-05, |
|
"loss": 1.3186, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.8060692271218587, |
|
"eval_loss": 1.2315117120742798, |
|
"eval_runtime": 72.0639, |
|
"eval_samples_per_second": 130.079, |
|
"eval_steps_per_second": 16.263, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.8155523944997629, |
|
"grad_norm": 1.3993178606033325, |
|
"learning_rate": 3.700176650072266e-05, |
|
"loss": 1.3074, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.8250355618776671, |
|
"grad_norm": 1.6044120788574219, |
|
"learning_rate": 3.684117552593544e-05, |
|
"loss": 1.2681, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.8345187292555714, |
|
"grad_norm": 1.6285070180892944, |
|
"learning_rate": 3.668058455114823e-05, |
|
"loss": 1.3198, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.8440018966334756, |
|
"grad_norm": 2.002086639404297, |
|
"learning_rate": 3.651999357636101e-05, |
|
"loss": 1.3227, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.8534850640113798, |
|
"grad_norm": 1.5941271781921387, |
|
"learning_rate": 3.635940260157379e-05, |
|
"loss": 1.2914, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.8534850640113798, |
|
"eval_loss": 1.2264697551727295, |
|
"eval_runtime": 72.0482, |
|
"eval_samples_per_second": 130.107, |
|
"eval_steps_per_second": 16.267, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.862968231389284, |
|
"grad_norm": 1.5721193552017212, |
|
"learning_rate": 3.619881162678658e-05, |
|
"loss": 1.3268, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.8724513987671882, |
|
"grad_norm": 1.7066916227340698, |
|
"learning_rate": 3.603822065199936e-05, |
|
"loss": 1.2845, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.8819345661450925, |
|
"grad_norm": 1.5683172941207886, |
|
"learning_rate": 3.587762967721214e-05, |
|
"loss": 1.2779, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.8914177335229967, |
|
"grad_norm": 1.7200586795806885, |
|
"learning_rate": 3.571703870242493e-05, |
|
"loss": 1.3161, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.9009009009009009, |
|
"grad_norm": 1.4963386058807373, |
|
"learning_rate": 3.555644772763771e-05, |
|
"loss": 1.2668, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.9009009009009009, |
|
"eval_loss": 1.2190866470336914, |
|
"eval_runtime": 72.0991, |
|
"eval_samples_per_second": 130.015, |
|
"eval_steps_per_second": 16.255, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.9103840682788051, |
|
"grad_norm": 1.5414083003997803, |
|
"learning_rate": 3.539585675285049e-05, |
|
"loss": 1.3185, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.9198672356567094, |
|
"grad_norm": 1.46302330493927, |
|
"learning_rate": 3.523526577806328e-05, |
|
"loss": 1.2485, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.9293504030346136, |
|
"grad_norm": 1.4815856218338013, |
|
"learning_rate": 3.507467480327606e-05, |
|
"loss": 1.2912, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.9388335704125178, |
|
"grad_norm": 1.5166754722595215, |
|
"learning_rate": 3.491408382848884e-05, |
|
"loss": 1.2722, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.948316737790422, |
|
"grad_norm": 1.9628846645355225, |
|
"learning_rate": 3.475349285370163e-05, |
|
"loss": 1.2538, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.948316737790422, |
|
"eval_loss": 1.2150416374206543, |
|
"eval_runtime": 72.1513, |
|
"eval_samples_per_second": 129.921, |
|
"eval_steps_per_second": 16.244, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.9577999051683262, |
|
"grad_norm": 1.6791901588439941, |
|
"learning_rate": 3.459290187891441e-05, |
|
"loss": 1.2624, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.9672830725462305, |
|
"grad_norm": 1.5026668310165405, |
|
"learning_rate": 3.443231090412719e-05, |
|
"loss": 1.2696, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.9767662399241347, |
|
"grad_norm": 1.176558017730713, |
|
"learning_rate": 3.427171992933998e-05, |
|
"loss": 1.29, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.9862494073020389, |
|
"grad_norm": 1.5698468685150146, |
|
"learning_rate": 3.411112895455276e-05, |
|
"loss": 1.2874, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.9957325746799431, |
|
"grad_norm": 1.4970085620880127, |
|
"learning_rate": 3.395053797976554e-05, |
|
"loss": 1.2874, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.9957325746799431, |
|
"eval_loss": 1.2110899686813354, |
|
"eval_runtime": 72.0475, |
|
"eval_samples_per_second": 130.109, |
|
"eval_steps_per_second": 16.267, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.0052157420578474, |
|
"grad_norm": 1.284839391708374, |
|
"learning_rate": 3.3789947004978326e-05, |
|
"loss": 1.2793, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 1.0146989094357515, |
|
"grad_norm": 1.680851697921753, |
|
"learning_rate": 3.362935603019111e-05, |
|
"loss": 1.2487, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 1.0241820768136558, |
|
"grad_norm": 1.659610629081726, |
|
"learning_rate": 3.346876505540389e-05, |
|
"loss": 1.2454, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 1.03366524419156, |
|
"grad_norm": 1.6641312837600708, |
|
"learning_rate": 3.330817408061667e-05, |
|
"loss": 1.2323, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 1.0431484115694643, |
|
"grad_norm": 1.481063723564148, |
|
"learning_rate": 3.3147583105829456e-05, |
|
"loss": 1.2646, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.0431484115694643, |
|
"eval_loss": 1.2060637474060059, |
|
"eval_runtime": 71.9819, |
|
"eval_samples_per_second": 130.227, |
|
"eval_steps_per_second": 16.282, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"grad_norm": 1.699491024017334, |
|
"learning_rate": 3.298699213104224e-05, |
|
"loss": 1.2828, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 1.0621147463252727, |
|
"grad_norm": 2.0708415508270264, |
|
"learning_rate": 3.282640115625502e-05, |
|
"loss": 1.2648, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 1.0715979137031768, |
|
"grad_norm": 1.4921772480010986, |
|
"learning_rate": 3.266741609121567e-05, |
|
"loss": 1.2611, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 1.0810810810810811, |
|
"grad_norm": 1.744384765625, |
|
"learning_rate": 3.250682511642846e-05, |
|
"loss": 1.2435, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 1.0905642484589853, |
|
"grad_norm": 1.1988921165466309, |
|
"learning_rate": 3.234623414164124e-05, |
|
"loss": 1.2525, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.0905642484589853, |
|
"eval_loss": 1.2018728256225586, |
|
"eval_runtime": 71.9385, |
|
"eval_samples_per_second": 130.306, |
|
"eval_steps_per_second": 16.292, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.1000474158368896, |
|
"grad_norm": 1.5618336200714111, |
|
"learning_rate": 3.218564316685402e-05, |
|
"loss": 1.2387, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 1.1095305832147937, |
|
"grad_norm": 1.512651801109314, |
|
"learning_rate": 3.202505219206681e-05, |
|
"loss": 1.2507, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 1.119013750592698, |
|
"grad_norm": 2.1945042610168457, |
|
"learning_rate": 3.186446121727959e-05, |
|
"loss": 1.2316, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 1.1284969179706021, |
|
"grad_norm": 1.3046265840530396, |
|
"learning_rate": 3.170387024249237e-05, |
|
"loss": 1.2352, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 1.1379800853485065, |
|
"grad_norm": 1.5922869443893433, |
|
"learning_rate": 3.154327926770516e-05, |
|
"loss": 1.2361, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.1379800853485065, |
|
"eval_loss": 1.1982355117797852, |
|
"eval_runtime": 72.0166, |
|
"eval_samples_per_second": 130.164, |
|
"eval_steps_per_second": 16.274, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.1474632527264106, |
|
"grad_norm": 1.2342475652694702, |
|
"learning_rate": 3.138268829291794e-05, |
|
"loss": 1.2318, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 1.156946420104315, |
|
"grad_norm": 1.630129337310791, |
|
"learning_rate": 3.122209731813072e-05, |
|
"loss": 1.2185, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 1.166429587482219, |
|
"grad_norm": 1.4030356407165527, |
|
"learning_rate": 3.106150634334351e-05, |
|
"loss": 1.2635, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 1.1759127548601234, |
|
"grad_norm": 1.372003436088562, |
|
"learning_rate": 3.090091536855629e-05, |
|
"loss": 1.2131, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 1.1853959222380275, |
|
"grad_norm": 1.1380951404571533, |
|
"learning_rate": 3.074032439376907e-05, |
|
"loss": 1.2553, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.1853959222380275, |
|
"eval_loss": 1.1942973136901855, |
|
"eval_runtime": 71.9892, |
|
"eval_samples_per_second": 130.214, |
|
"eval_steps_per_second": 16.28, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.1948790896159318, |
|
"grad_norm": 1.8760716915130615, |
|
"learning_rate": 3.057973341898186e-05, |
|
"loss": 1.2479, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 1.204362256993836, |
|
"grad_norm": 1.7070045471191406, |
|
"learning_rate": 3.0419142444194638e-05, |
|
"loss": 1.2283, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 1.2138454243717403, |
|
"grad_norm": 1.6677838563919067, |
|
"learning_rate": 3.025855146940742e-05, |
|
"loss": 1.2527, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 1.2233285917496444, |
|
"grad_norm": 1.5015747547149658, |
|
"learning_rate": 3.0097960494620203e-05, |
|
"loss": 1.2402, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 1.2328117591275487, |
|
"grad_norm": 1.613587737083435, |
|
"learning_rate": 2.9937369519832987e-05, |
|
"loss": 1.2288, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.2328117591275487, |
|
"eval_loss": 1.1904593706130981, |
|
"eval_runtime": 72.0827, |
|
"eval_samples_per_second": 130.045, |
|
"eval_steps_per_second": 16.259, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.2422949265054528, |
|
"grad_norm": 1.7170720100402832, |
|
"learning_rate": 2.9776778545045768e-05, |
|
"loss": 1.2199, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 1.251778093883357, |
|
"grad_norm": 1.3260998725891113, |
|
"learning_rate": 2.9616187570258552e-05, |
|
"loss": 1.2575, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 1.2612612612612613, |
|
"grad_norm": 1.450626254081726, |
|
"learning_rate": 2.9455596595471337e-05, |
|
"loss": 1.2267, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 1.2707444286391656, |
|
"grad_norm": 1.51180899143219, |
|
"learning_rate": 2.9295005620684118e-05, |
|
"loss": 1.2546, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 1.2802275960170697, |
|
"grad_norm": 1.846704125404358, |
|
"learning_rate": 2.9134414645896902e-05, |
|
"loss": 1.2216, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.2802275960170697, |
|
"eval_loss": 1.1853208541870117, |
|
"eval_runtime": 72.0024, |
|
"eval_samples_per_second": 130.19, |
|
"eval_steps_per_second": 16.277, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.2897107633949738, |
|
"grad_norm": 1.5088779926300049, |
|
"learning_rate": 2.8973823671109686e-05, |
|
"loss": 1.2028, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 1.2991939307728781, |
|
"grad_norm": 1.2047330141067505, |
|
"learning_rate": 2.8813232696322467e-05, |
|
"loss": 1.2326, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 1.3086770981507825, |
|
"grad_norm": 1.6895666122436523, |
|
"learning_rate": 2.865264172153525e-05, |
|
"loss": 1.2032, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 1.3181602655286866, |
|
"grad_norm": 1.3885574340820312, |
|
"learning_rate": 2.8492050746748032e-05, |
|
"loss": 1.2438, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 1.3276434329065907, |
|
"grad_norm": 1.5129587650299072, |
|
"learning_rate": 2.8331459771960816e-05, |
|
"loss": 1.2099, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.3276434329065907, |
|
"eval_loss": 1.1841365098953247, |
|
"eval_runtime": 72.0289, |
|
"eval_samples_per_second": 130.142, |
|
"eval_steps_per_second": 16.271, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.337126600284495, |
|
"grad_norm": 1.5244189500808716, |
|
"learning_rate": 2.81708687971736e-05, |
|
"loss": 1.2528, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 1.3466097676623994, |
|
"grad_norm": 1.6656090021133423, |
|
"learning_rate": 2.801027782238638e-05, |
|
"loss": 1.2437, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 1.3560929350403035, |
|
"grad_norm": 1.6365015506744385, |
|
"learning_rate": 2.7849686847599165e-05, |
|
"loss": 1.2481, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 1.3655761024182076, |
|
"grad_norm": 1.729038953781128, |
|
"learning_rate": 2.768909587281195e-05, |
|
"loss": 1.2363, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 1.375059269796112, |
|
"grad_norm": 1.663041114807129, |
|
"learning_rate": 2.752850489802473e-05, |
|
"loss": 1.2371, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.375059269796112, |
|
"eval_loss": 1.1793495416641235, |
|
"eval_runtime": 72.0339, |
|
"eval_samples_per_second": 130.133, |
|
"eval_steps_per_second": 16.27, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.384542437174016, |
|
"grad_norm": 1.5626816749572754, |
|
"learning_rate": 2.7367913923237515e-05, |
|
"loss": 1.2287, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 1.3940256045519204, |
|
"grad_norm": 1.2476764917373657, |
|
"learning_rate": 2.72073229484503e-05, |
|
"loss": 1.2129, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 1.4035087719298245, |
|
"grad_norm": 1.4796671867370605, |
|
"learning_rate": 2.704673197366308e-05, |
|
"loss": 1.2143, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 1.4129919393077288, |
|
"grad_norm": 1.8260607719421387, |
|
"learning_rate": 2.6886140998875864e-05, |
|
"loss": 1.2411, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 1.422475106685633, |
|
"grad_norm": 1.6393589973449707, |
|
"learning_rate": 2.6725550024088648e-05, |
|
"loss": 1.2128, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.422475106685633, |
|
"eval_loss": 1.1766639947891235, |
|
"eval_runtime": 72.0436, |
|
"eval_samples_per_second": 130.116, |
|
"eval_steps_per_second": 16.268, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.4319582740635373, |
|
"grad_norm": 1.2327754497528076, |
|
"learning_rate": 2.656495904930143e-05, |
|
"loss": 1.2218, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 1.4414414414414414, |
|
"grad_norm": 1.4845291376113892, |
|
"learning_rate": 2.6405973984262084e-05, |
|
"loss": 1.2158, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 1.4509246088193457, |
|
"grad_norm": 1.5115349292755127, |
|
"learning_rate": 2.6245383009474868e-05, |
|
"loss": 1.2597, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 1.4604077761972498, |
|
"grad_norm": 1.2558484077453613, |
|
"learning_rate": 2.608479203468765e-05, |
|
"loss": 1.2293, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 1.4698909435751542, |
|
"grad_norm": 1.412372350692749, |
|
"learning_rate": 2.5924201059900433e-05, |
|
"loss": 1.2078, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.4698909435751542, |
|
"eval_loss": 1.175757646560669, |
|
"eval_runtime": 72.1719, |
|
"eval_samples_per_second": 129.884, |
|
"eval_steps_per_second": 16.239, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.4793741109530583, |
|
"grad_norm": 1.1586443185806274, |
|
"learning_rate": 2.5763610085113217e-05, |
|
"loss": 1.2167, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 1.4888572783309626, |
|
"grad_norm": 1.535499095916748, |
|
"learning_rate": 2.5603019110325998e-05, |
|
"loss": 1.2177, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 1.4983404457088667, |
|
"grad_norm": 1.3925201892852783, |
|
"learning_rate": 2.5442428135538782e-05, |
|
"loss": 1.2089, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 1.5078236130867708, |
|
"grad_norm": 1.239797592163086, |
|
"learning_rate": 2.5281837160751563e-05, |
|
"loss": 1.2183, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 1.5173067804646752, |
|
"grad_norm": 1.4727925062179565, |
|
"learning_rate": 2.5121246185964347e-05, |
|
"loss": 1.2382, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.5173067804646752, |
|
"eval_loss": 1.1705734729766846, |
|
"eval_runtime": 72.2315, |
|
"eval_samples_per_second": 129.777, |
|
"eval_steps_per_second": 16.226, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.5267899478425795, |
|
"grad_norm": 1.9122114181518555, |
|
"learning_rate": 2.4960655211177135e-05, |
|
"loss": 1.2062, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 1.5362731152204836, |
|
"grad_norm": 1.705417275428772, |
|
"learning_rate": 2.4800064236389916e-05, |
|
"loss": 1.2002, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 1.5457562825983877, |
|
"grad_norm": 1.4141908884048462, |
|
"learning_rate": 2.46394732616027e-05, |
|
"loss": 1.2323, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 1.555239449976292, |
|
"grad_norm": 2.050583839416504, |
|
"learning_rate": 2.4478882286815484e-05, |
|
"loss": 1.2145, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 1.5647226173541964, |
|
"grad_norm": 1.495006799697876, |
|
"learning_rate": 2.4318291312028265e-05, |
|
"loss": 1.2041, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.5647226173541964, |
|
"eval_loss": 1.1694616079330444, |
|
"eval_runtime": 71.9712, |
|
"eval_samples_per_second": 130.247, |
|
"eval_steps_per_second": 16.284, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.5742057847321005, |
|
"grad_norm": 1.4379011392593384, |
|
"learning_rate": 2.415770033724105e-05, |
|
"loss": 1.2045, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 1.5836889521100046, |
|
"grad_norm": 1.6558938026428223, |
|
"learning_rate": 2.399710936245383e-05, |
|
"loss": 1.2234, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 1.593172119487909, |
|
"grad_norm": 1.6931570768356323, |
|
"learning_rate": 2.3836518387666614e-05, |
|
"loss": 1.2061, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 1.6026552868658133, |
|
"grad_norm": 1.445521593093872, |
|
"learning_rate": 2.36759274128794e-05, |
|
"loss": 1.2243, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 1.6121384542437174, |
|
"grad_norm": 1.4067689180374146, |
|
"learning_rate": 2.351533643809218e-05, |
|
"loss": 1.2154, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.6121384542437174, |
|
"eval_loss": 1.1659753322601318, |
|
"eval_runtime": 72.1888, |
|
"eval_samples_per_second": 129.854, |
|
"eval_steps_per_second": 16.235, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.6216216216216215, |
|
"grad_norm": 1.0550585985183716, |
|
"learning_rate": 2.3354745463304964e-05, |
|
"loss": 1.2333, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 1.6311047889995258, |
|
"grad_norm": 1.5547784566879272, |
|
"learning_rate": 2.3194154488517748e-05, |
|
"loss": 1.2088, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 1.6405879563774302, |
|
"grad_norm": 2.006110191345215, |
|
"learning_rate": 2.303356351373053e-05, |
|
"loss": 1.1881, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 1.6500711237553343, |
|
"grad_norm": 1.6522830724716187, |
|
"learning_rate": 2.2872972538943313e-05, |
|
"loss": 1.2158, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 1.6595542911332384, |
|
"grad_norm": 1.2928231954574585, |
|
"learning_rate": 2.2712381564156097e-05, |
|
"loss": 1.2303, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.6595542911332384, |
|
"eval_loss": 1.1643718481063843, |
|
"eval_runtime": 72.2381, |
|
"eval_samples_per_second": 129.765, |
|
"eval_steps_per_second": 16.224, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.6690374585111427, |
|
"grad_norm": 1.38106107711792, |
|
"learning_rate": 2.2551790589368878e-05, |
|
"loss": 1.1969, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 1.678520625889047, |
|
"grad_norm": 1.3726710081100464, |
|
"learning_rate": 2.2391199614581662e-05, |
|
"loss": 1.2122, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 1.6880037932669512, |
|
"grad_norm": 1.2017816305160522, |
|
"learning_rate": 2.2230608639794447e-05, |
|
"loss": 1.2331, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 1.6974869606448553, |
|
"grad_norm": 1.329315423965454, |
|
"learning_rate": 2.2070017665007227e-05, |
|
"loss": 1.2339, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 1.7069701280227596, |
|
"grad_norm": 1.5352445840835571, |
|
"learning_rate": 2.190942669022001e-05, |
|
"loss": 1.2429, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.7069701280227596, |
|
"eval_loss": 1.1619985103607178, |
|
"eval_runtime": 72.1286, |
|
"eval_samples_per_second": 129.962, |
|
"eval_steps_per_second": 16.249, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.716453295400664, |
|
"grad_norm": 1.5836015939712524, |
|
"learning_rate": 2.1748835715432796e-05, |
|
"loss": 1.1925, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 1.725936462778568, |
|
"grad_norm": 1.7755178213119507, |
|
"learning_rate": 2.1588244740645577e-05, |
|
"loss": 1.2146, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 1.7354196301564722, |
|
"grad_norm": 1.3868217468261719, |
|
"learning_rate": 2.142765376585836e-05, |
|
"loss": 1.2082, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 1.7449027975343765, |
|
"grad_norm": 1.320333480834961, |
|
"learning_rate": 2.1267062791071142e-05, |
|
"loss": 1.213, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 1.7543859649122808, |
|
"grad_norm": 1.5032850503921509, |
|
"learning_rate": 2.1106471816283926e-05, |
|
"loss": 1.2048, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.7543859649122808, |
|
"eval_loss": 1.1578137874603271, |
|
"eval_runtime": 72.0841, |
|
"eval_samples_per_second": 130.043, |
|
"eval_steps_per_second": 16.259, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.763869132290185, |
|
"grad_norm": 1.5423904657363892, |
|
"learning_rate": 2.094588084149671e-05, |
|
"loss": 1.2282, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 1.773352299668089, |
|
"grad_norm": 1.439765453338623, |
|
"learning_rate": 2.078528986670949e-05, |
|
"loss": 1.2171, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 1.7828354670459934, |
|
"grad_norm": 1.573088526725769, |
|
"learning_rate": 2.0624698891922275e-05, |
|
"loss": 1.2149, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 1.7923186344238977, |
|
"grad_norm": 1.4882514476776123, |
|
"learning_rate": 2.046410791713506e-05, |
|
"loss": 1.2278, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 1.8018018018018018, |
|
"grad_norm": 1.9028195142745972, |
|
"learning_rate": 2.030351694234784e-05, |
|
"loss": 1.2247, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.8018018018018018, |
|
"eval_loss": 1.157362937927246, |
|
"eval_runtime": 72.1036, |
|
"eval_samples_per_second": 130.007, |
|
"eval_steps_per_second": 16.254, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.811284969179706, |
|
"grad_norm": 1.289600133895874, |
|
"learning_rate": 2.0142925967560625e-05, |
|
"loss": 1.215, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 1.8207681365576103, |
|
"grad_norm": 1.4183131456375122, |
|
"learning_rate": 1.998233499277341e-05, |
|
"loss": 1.2284, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 1.8302513039355146, |
|
"grad_norm": 1.235146403312683, |
|
"learning_rate": 1.982174401798619e-05, |
|
"loss": 1.2067, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 1.8397344713134187, |
|
"grad_norm": 1.486122488975525, |
|
"learning_rate": 1.9661153043198974e-05, |
|
"loss": 1.183, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 1.8492176386913228, |
|
"grad_norm": 1.4615782499313354, |
|
"learning_rate": 1.9500562068411758e-05, |
|
"loss": 1.1847, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.8492176386913228, |
|
"eval_loss": 1.1544617414474487, |
|
"eval_runtime": 72.1411, |
|
"eval_samples_per_second": 129.94, |
|
"eval_steps_per_second": 16.246, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.8587008060692272, |
|
"grad_norm": 1.3062597513198853, |
|
"learning_rate": 1.933997109362454e-05, |
|
"loss": 1.1998, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 1.8681839734471315, |
|
"grad_norm": 1.7676483392715454, |
|
"learning_rate": 1.9180986028585193e-05, |
|
"loss": 1.1985, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 1.8776671408250356, |
|
"grad_norm": 1.55678129196167, |
|
"learning_rate": 1.9020395053797978e-05, |
|
"loss": 1.2155, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 1.8871503082029397, |
|
"grad_norm": 1.2260453701019287, |
|
"learning_rate": 1.885980407901076e-05, |
|
"loss": 1.2282, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 1.896633475580844, |
|
"grad_norm": 1.6828114986419678, |
|
"learning_rate": 1.8699213104223543e-05, |
|
"loss": 1.2183, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.896633475580844, |
|
"eval_loss": 1.1521168947219849, |
|
"eval_runtime": 72.1018, |
|
"eval_samples_per_second": 130.011, |
|
"eval_steps_per_second": 16.255, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.9061166429587484, |
|
"grad_norm": 1.6691786050796509, |
|
"learning_rate": 1.8538622129436327e-05, |
|
"loss": 1.1651, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 1.9155998103366523, |
|
"grad_norm": 1.4728951454162598, |
|
"learning_rate": 1.8378031154649108e-05, |
|
"loss": 1.2022, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 1.9250829777145566, |
|
"grad_norm": 1.6341995000839233, |
|
"learning_rate": 1.8217440179861892e-05, |
|
"loss": 1.1777, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 1.934566145092461, |
|
"grad_norm": 1.4492669105529785, |
|
"learning_rate": 1.8056849205074676e-05, |
|
"loss": 1.2081, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 1.944049312470365, |
|
"grad_norm": 1.6642097234725952, |
|
"learning_rate": 1.7896258230287457e-05, |
|
"loss": 1.1848, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.944049312470365, |
|
"eval_loss": 1.150140404701233, |
|
"eval_runtime": 72.0779, |
|
"eval_samples_per_second": 130.054, |
|
"eval_steps_per_second": 16.26, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.9535324798482692, |
|
"grad_norm": 1.8986822366714478, |
|
"learning_rate": 1.773566725550024e-05, |
|
"loss": 1.2223, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 1.9630156472261735, |
|
"grad_norm": 1.390931248664856, |
|
"learning_rate": 1.7575076280713022e-05, |
|
"loss": 1.2068, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 1.9724988146040778, |
|
"grad_norm": 1.3856289386749268, |
|
"learning_rate": 1.7414485305925806e-05, |
|
"loss": 1.1828, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 1.981981981981982, |
|
"grad_norm": 1.2241305112838745, |
|
"learning_rate": 1.725389433113859e-05, |
|
"loss": 1.1938, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 1.991465149359886, |
|
"grad_norm": 1.5855077505111694, |
|
"learning_rate": 1.709330335635137e-05, |
|
"loss": 1.206, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.991465149359886, |
|
"eval_loss": 1.1497843265533447, |
|
"eval_runtime": 72.1674, |
|
"eval_samples_per_second": 129.893, |
|
"eval_steps_per_second": 16.24, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 2.0009483167377904, |
|
"grad_norm": 2.0832741260528564, |
|
"learning_rate": 1.6932712381564156e-05, |
|
"loss": 1.1805, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 2.0104314841156947, |
|
"grad_norm": 1.893350601196289, |
|
"learning_rate": 1.677212140677694e-05, |
|
"loss": 1.1757, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 2.019914651493599, |
|
"grad_norm": 1.346118688583374, |
|
"learning_rate": 1.661153043198972e-05, |
|
"loss": 1.1938, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 2.029397818871503, |
|
"grad_norm": 1.658034086227417, |
|
"learning_rate": 1.6450939457202505e-05, |
|
"loss": 1.1773, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 2.0388809862494073, |
|
"grad_norm": 1.4759783744812012, |
|
"learning_rate": 1.629034848241529e-05, |
|
"loss": 1.1735, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 2.0388809862494073, |
|
"eval_loss": 1.1474945545196533, |
|
"eval_runtime": 71.9179, |
|
"eval_samples_per_second": 130.343, |
|
"eval_steps_per_second": 16.296, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 2.0483641536273116, |
|
"grad_norm": 1.2887206077575684, |
|
"learning_rate": 1.612975750762807e-05, |
|
"loss": 1.1701, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 2.057847321005216, |
|
"grad_norm": 1.552646279335022, |
|
"learning_rate": 1.5969166532840854e-05, |
|
"loss": 1.1734, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 2.06733048838312, |
|
"grad_norm": 1.6683566570281982, |
|
"learning_rate": 1.581018146780151e-05, |
|
"loss": 1.1883, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 2.076813655761024, |
|
"grad_norm": 1.4613324403762817, |
|
"learning_rate": 1.5649590493014293e-05, |
|
"loss": 1.1845, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 2.0862968231389285, |
|
"grad_norm": 1.5622040033340454, |
|
"learning_rate": 1.5488999518227077e-05, |
|
"loss": 1.1584, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.0862968231389285, |
|
"eval_loss": 1.1467849016189575, |
|
"eval_runtime": 72.0497, |
|
"eval_samples_per_second": 130.105, |
|
"eval_steps_per_second": 16.267, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.095779990516833, |
|
"grad_norm": 1.721030831336975, |
|
"learning_rate": 1.5328408543439858e-05, |
|
"loss": 1.2018, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 2.1052631578947367, |
|
"grad_norm": 1.3872593641281128, |
|
"learning_rate": 1.5167817568652642e-05, |
|
"loss": 1.1659, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 2.114746325272641, |
|
"grad_norm": 1.655704140663147, |
|
"learning_rate": 1.5007226593865425e-05, |
|
"loss": 1.1503, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 2.1242294926505454, |
|
"grad_norm": 1.5672900676727295, |
|
"learning_rate": 1.4848241528826081e-05, |
|
"loss": 1.1879, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 2.1337126600284497, |
|
"grad_norm": 1.6815894842147827, |
|
"learning_rate": 1.4687650554038865e-05, |
|
"loss": 1.1719, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.1337126600284497, |
|
"eval_loss": 1.1450951099395752, |
|
"eval_runtime": 72.1598, |
|
"eval_samples_per_second": 129.906, |
|
"eval_steps_per_second": 16.242, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.1431958274063536, |
|
"grad_norm": 1.040648102760315, |
|
"learning_rate": 1.4527059579251648e-05, |
|
"loss": 1.1629, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 2.152678994784258, |
|
"grad_norm": 1.5001453161239624, |
|
"learning_rate": 1.436646860446443e-05, |
|
"loss": 1.1796, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 2.1621621621621623, |
|
"grad_norm": 1.7325968742370605, |
|
"learning_rate": 1.4205877629677215e-05, |
|
"loss": 1.1757, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 2.171645329540066, |
|
"grad_norm": 1.7485188245773315, |
|
"learning_rate": 1.4045286654889997e-05, |
|
"loss": 1.1485, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 2.1811284969179705, |
|
"grad_norm": 1.4972156286239624, |
|
"learning_rate": 1.388469568010278e-05, |
|
"loss": 1.1667, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 2.1811284969179705, |
|
"eval_loss": 1.144049048423767, |
|
"eval_runtime": 72.1218, |
|
"eval_samples_per_second": 129.975, |
|
"eval_steps_per_second": 16.25, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 2.190611664295875, |
|
"grad_norm": 1.2919082641601562, |
|
"learning_rate": 1.3724104705315564e-05, |
|
"loss": 1.1764, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 2.200094831673779, |
|
"grad_norm": 1.6442806720733643, |
|
"learning_rate": 1.3563513730528346e-05, |
|
"loss": 1.174, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 2.209577999051683, |
|
"grad_norm": 1.480901837348938, |
|
"learning_rate": 1.3402922755741129e-05, |
|
"loss": 1.1666, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 2.2190611664295874, |
|
"grad_norm": 1.6193006038665771, |
|
"learning_rate": 1.3242331780953911e-05, |
|
"loss": 1.1975, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 2.2285443338074917, |
|
"grad_norm": 1.2970917224884033, |
|
"learning_rate": 1.3081740806166696e-05, |
|
"loss": 1.1579, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 2.2285443338074917, |
|
"eval_loss": 1.1433159112930298, |
|
"eval_runtime": 72.0832, |
|
"eval_samples_per_second": 130.044, |
|
"eval_steps_per_second": 16.259, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 2.238027501185396, |
|
"grad_norm": 1.4054538011550903, |
|
"learning_rate": 1.2921149831379478e-05, |
|
"loss": 1.1779, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 2.2475106685633, |
|
"grad_norm": 1.5161010026931763, |
|
"learning_rate": 1.276055885659226e-05, |
|
"loss": 1.1709, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 2.2569938359412043, |
|
"grad_norm": 2.040818929672241, |
|
"learning_rate": 1.2599967881805045e-05, |
|
"loss": 1.1692, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 2.2664770033191086, |
|
"grad_norm": 1.3812401294708252, |
|
"learning_rate": 1.2439376907017826e-05, |
|
"loss": 1.1733, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 2.275960170697013, |
|
"grad_norm": 2.113886833190918, |
|
"learning_rate": 1.2278785932230608e-05, |
|
"loss": 1.1682, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.275960170697013, |
|
"eval_loss": 1.1404303312301636, |
|
"eval_runtime": 72.1649, |
|
"eval_samples_per_second": 129.897, |
|
"eval_steps_per_second": 16.241, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.285443338074917, |
|
"grad_norm": 1.3256770372390747, |
|
"learning_rate": 1.2118194957443393e-05, |
|
"loss": 1.1847, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 2.294926505452821, |
|
"grad_norm": 1.4699623584747314, |
|
"learning_rate": 1.1957603982656175e-05, |
|
"loss": 1.1576, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 2.3044096728307255, |
|
"grad_norm": 1.5492583513259888, |
|
"learning_rate": 1.1797013007868958e-05, |
|
"loss": 1.1532, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 2.31389284020863, |
|
"grad_norm": 1.409488558769226, |
|
"learning_rate": 1.1636422033081742e-05, |
|
"loss": 1.1626, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 2.3233760075865337, |
|
"grad_norm": 1.642247200012207, |
|
"learning_rate": 1.1475831058294524e-05, |
|
"loss": 1.1943, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 2.3233760075865337, |
|
"eval_loss": 1.139186978340149, |
|
"eval_runtime": 72.1131, |
|
"eval_samples_per_second": 129.99, |
|
"eval_steps_per_second": 16.252, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 2.332859174964438, |
|
"grad_norm": 1.4776501655578613, |
|
"learning_rate": 1.1315240083507307e-05, |
|
"loss": 1.1566, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 2.3423423423423424, |
|
"grad_norm": 1.475188136100769, |
|
"learning_rate": 1.115464910872009e-05, |
|
"loss": 1.1743, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 2.3518255097202467, |
|
"grad_norm": 1.48451828956604, |
|
"learning_rate": 1.0994058133932874e-05, |
|
"loss": 1.1539, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 2.3613086770981506, |
|
"grad_norm": 1.4650864601135254, |
|
"learning_rate": 1.0833467159145656e-05, |
|
"loss": 1.2073, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 2.370791844476055, |
|
"grad_norm": 1.71983003616333, |
|
"learning_rate": 1.0672876184358439e-05, |
|
"loss": 1.2021, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.370791844476055, |
|
"eval_loss": 1.1377766132354736, |
|
"eval_runtime": 71.9749, |
|
"eval_samples_per_second": 130.24, |
|
"eval_steps_per_second": 16.283, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.3802750118539593, |
|
"grad_norm": 1.3838121891021729, |
|
"learning_rate": 1.0512285209571223e-05, |
|
"loss": 1.1791, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 2.3897581792318636, |
|
"grad_norm": 1.8836325407028198, |
|
"learning_rate": 1.0351694234784006e-05, |
|
"loss": 1.1834, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 2.3992413466097675, |
|
"grad_norm": 1.3679293394088745, |
|
"learning_rate": 1.0191103259996788e-05, |
|
"loss": 1.183, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 2.408724513987672, |
|
"grad_norm": 1.5593743324279785, |
|
"learning_rate": 1.003051228520957e-05, |
|
"loss": 1.1703, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 2.418207681365576, |
|
"grad_norm": 1.4257512092590332, |
|
"learning_rate": 9.869921310422355e-06, |
|
"loss": 1.172, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 2.418207681365576, |
|
"eval_loss": 1.1378742456436157, |
|
"eval_runtime": 72.0363, |
|
"eval_samples_per_second": 130.129, |
|
"eval_steps_per_second": 16.27, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 2.4276908487434805, |
|
"grad_norm": 1.771941065788269, |
|
"learning_rate": 9.709330335635137e-06, |
|
"loss": 1.1676, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 2.4371740161213844, |
|
"grad_norm": 1.7247157096862793, |
|
"learning_rate": 9.54873936084792e-06, |
|
"loss": 1.1753, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 2.4466571834992887, |
|
"grad_norm": 1.5509614944458008, |
|
"learning_rate": 9.388148386060704e-06, |
|
"loss": 1.1705, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 2.456140350877193, |
|
"grad_norm": 1.8205307722091675, |
|
"learning_rate": 9.227557411273487e-06, |
|
"loss": 1.1938, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 2.4656235182550974, |
|
"grad_norm": 1.501631498336792, |
|
"learning_rate": 9.06696643648627e-06, |
|
"loss": 1.1737, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.4656235182550974, |
|
"eval_loss": 1.1362242698669434, |
|
"eval_runtime": 71.9608, |
|
"eval_samples_per_second": 130.265, |
|
"eval_steps_per_second": 16.287, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.4751066856330013, |
|
"grad_norm": 1.4233213663101196, |
|
"learning_rate": 8.906375461699054e-06, |
|
"loss": 1.1728, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 2.4845898530109056, |
|
"grad_norm": 1.597785472869873, |
|
"learning_rate": 8.745784486911836e-06, |
|
"loss": 1.1559, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 2.49407302038881, |
|
"grad_norm": 1.2396786212921143, |
|
"learning_rate": 8.585193512124619e-06, |
|
"loss": 1.1645, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 2.503556187766714, |
|
"grad_norm": 1.643211841583252, |
|
"learning_rate": 8.424602537337401e-06, |
|
"loss": 1.1948, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 2.513039355144618, |
|
"grad_norm": 1.688436508178711, |
|
"learning_rate": 8.264011562550185e-06, |
|
"loss": 1.1875, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.513039355144618, |
|
"eval_loss": 1.134669303894043, |
|
"eval_runtime": 72.1082, |
|
"eval_samples_per_second": 129.999, |
|
"eval_steps_per_second": 16.253, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.5225225225225225, |
|
"grad_norm": 1.6127384901046753, |
|
"learning_rate": 8.103420587762968e-06, |
|
"loss": 1.1657, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 2.532005689900427, |
|
"grad_norm": 2.12892484664917, |
|
"learning_rate": 7.944435522723622e-06, |
|
"loss": 1.1636, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 2.541488857278331, |
|
"grad_norm": 1.173686146736145, |
|
"learning_rate": 7.783844547936407e-06, |
|
"loss": 1.1866, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 2.550972024656235, |
|
"grad_norm": 1.4527802467346191, |
|
"learning_rate": 7.623253573149189e-06, |
|
"loss": 1.1755, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 2.5604551920341394, |
|
"grad_norm": 1.6228667497634888, |
|
"learning_rate": 7.462662598361972e-06, |
|
"loss": 1.1427, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.5604551920341394, |
|
"eval_loss": 1.134996771812439, |
|
"eval_runtime": 72.1853, |
|
"eval_samples_per_second": 129.86, |
|
"eval_steps_per_second": 16.236, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.5699383594120437, |
|
"grad_norm": 1.5179518461227417, |
|
"learning_rate": 7.302071623574755e-06, |
|
"loss": 1.1496, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 2.5794215267899476, |
|
"grad_norm": 1.2633978128433228, |
|
"learning_rate": 7.141480648787538e-06, |
|
"loss": 1.1633, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 2.588904694167852, |
|
"grad_norm": 1.3050264120101929, |
|
"learning_rate": 6.980889674000321e-06, |
|
"loss": 1.1614, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 2.5983878615457563, |
|
"grad_norm": 1.432268500328064, |
|
"learning_rate": 6.820298699213104e-06, |
|
"loss": 1.1684, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 2.6078710289236606, |
|
"grad_norm": 1.6904171705245972, |
|
"learning_rate": 6.659707724425887e-06, |
|
"loss": 1.1673, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.6078710289236606, |
|
"eval_loss": 1.1333271265029907, |
|
"eval_runtime": 72.187, |
|
"eval_samples_per_second": 129.857, |
|
"eval_steps_per_second": 16.236, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.617354196301565, |
|
"grad_norm": 1.2229042053222656, |
|
"learning_rate": 6.49911674963867e-06, |
|
"loss": 1.1793, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 2.626837363679469, |
|
"grad_norm": 1.7409764528274536, |
|
"learning_rate": 6.338525774851453e-06, |
|
"loss": 1.1963, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 2.636320531057373, |
|
"grad_norm": 1.4706058502197266, |
|
"learning_rate": 6.177934800064237e-06, |
|
"loss": 1.1836, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 2.6458036984352775, |
|
"grad_norm": 1.3871138095855713, |
|
"learning_rate": 6.01734382527702e-06, |
|
"loss": 1.1669, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 2.6552868658131814, |
|
"grad_norm": 1.5841022729873657, |
|
"learning_rate": 5.856752850489803e-06, |
|
"loss": 1.1765, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.6552868658131814, |
|
"eval_loss": 1.1325418949127197, |
|
"eval_runtime": 72.1699, |
|
"eval_samples_per_second": 129.888, |
|
"eval_steps_per_second": 16.239, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.6647700331910857, |
|
"grad_norm": 1.2488940954208374, |
|
"learning_rate": 5.696161875702586e-06, |
|
"loss": 1.1581, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 2.67425320056899, |
|
"grad_norm": 1.633123517036438, |
|
"learning_rate": 5.535570900915369e-06, |
|
"loss": 1.1829, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 2.6837363679468944, |
|
"grad_norm": 1.558030366897583, |
|
"learning_rate": 5.374979926128152e-06, |
|
"loss": 1.1816, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 2.6932195353247987, |
|
"grad_norm": 1.5178041458129883, |
|
"learning_rate": 5.214388951340935e-06, |
|
"loss": 1.1789, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 2.7027027027027026, |
|
"grad_norm": 1.8317012786865234, |
|
"learning_rate": 5.053797976553718e-06, |
|
"loss": 1.1612, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.7027027027027026, |
|
"eval_loss": 1.1320453882217407, |
|
"eval_runtime": 72.3445, |
|
"eval_samples_per_second": 129.575, |
|
"eval_steps_per_second": 16.2, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.712185870080607, |
|
"grad_norm": 1.4248275756835938, |
|
"learning_rate": 4.893207001766502e-06, |
|
"loss": 1.1583, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 2.7216690374585113, |
|
"grad_norm": 1.3696835041046143, |
|
"learning_rate": 4.732616026979284e-06, |
|
"loss": 1.1302, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 2.731152204836415, |
|
"grad_norm": 1.4212887287139893, |
|
"learning_rate": 4.5720250521920675e-06, |
|
"loss": 1.1396, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 2.7406353722143195, |
|
"grad_norm": 1.6230417490005493, |
|
"learning_rate": 4.41143407740485e-06, |
|
"loss": 1.167, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 2.750118539592224, |
|
"grad_norm": 1.4556254148483276, |
|
"learning_rate": 4.252449012365505e-06, |
|
"loss": 1.2229, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.750118539592224, |
|
"eval_loss": 1.1307094097137451, |
|
"eval_runtime": 72.2019, |
|
"eval_samples_per_second": 129.83, |
|
"eval_steps_per_second": 16.232, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.759601706970128, |
|
"grad_norm": 1.399604082107544, |
|
"learning_rate": 4.091858037578288e-06, |
|
"loss": 1.183, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 2.769084874348032, |
|
"grad_norm": 1.3562369346618652, |
|
"learning_rate": 3.931267062791071e-06, |
|
"loss": 1.1729, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 2.7785680417259364, |
|
"grad_norm": 1.4427545070648193, |
|
"learning_rate": 3.7706760880038542e-06, |
|
"loss": 1.1636, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 2.7880512091038407, |
|
"grad_norm": 1.6153539419174194, |
|
"learning_rate": 3.610085113216637e-06, |
|
"loss": 1.1608, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 2.797534376481745, |
|
"grad_norm": 1.553841233253479, |
|
"learning_rate": 3.44949413842942e-06, |
|
"loss": 1.1727, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.797534376481745, |
|
"eval_loss": 1.1305798292160034, |
|
"eval_runtime": 72.4369, |
|
"eval_samples_per_second": 129.409, |
|
"eval_steps_per_second": 16.18, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.807017543859649, |
|
"grad_norm": 1.4503796100616455, |
|
"learning_rate": 3.2889031636422036e-06, |
|
"loss": 1.1533, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 2.8165007112375533, |
|
"grad_norm": 2.3234095573425293, |
|
"learning_rate": 3.1283121888549865e-06, |
|
"loss": 1.1849, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 2.8259838786154576, |
|
"grad_norm": 1.6692347526550293, |
|
"learning_rate": 2.9677212140677695e-06, |
|
"loss": 1.1629, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 2.8354670459933615, |
|
"grad_norm": 1.6683822870254517, |
|
"learning_rate": 2.8071302392805524e-06, |
|
"loss": 1.1584, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 2.844950213371266, |
|
"grad_norm": 1.371102213859558, |
|
"learning_rate": 2.6465392644933354e-06, |
|
"loss": 1.1208, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.844950213371266, |
|
"eval_loss": 1.1299171447753906, |
|
"eval_runtime": 72.1885, |
|
"eval_samples_per_second": 129.854, |
|
"eval_steps_per_second": 16.235, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.85443338074917, |
|
"grad_norm": 1.9285227060317993, |
|
"learning_rate": 2.4859482897061184e-06, |
|
"loss": 1.1871, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 2.8639165481270745, |
|
"grad_norm": 1.5394768714904785, |
|
"learning_rate": 2.3253573149189017e-06, |
|
"loss": 1.1786, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 2.873399715504979, |
|
"grad_norm": 1.606779932975769, |
|
"learning_rate": 2.1647663401316847e-06, |
|
"loss": 1.181, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 2.8828828828828827, |
|
"grad_norm": 1.6637898683547974, |
|
"learning_rate": 2.0041753653444677e-06, |
|
"loss": 1.1435, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 2.892366050260787, |
|
"grad_norm": 1.4190491437911987, |
|
"learning_rate": 1.8435843905572506e-06, |
|
"loss": 1.158, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.892366050260787, |
|
"eval_loss": 1.129961371421814, |
|
"eval_runtime": 72.2984, |
|
"eval_samples_per_second": 129.657, |
|
"eval_steps_per_second": 16.211, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.9018492176386914, |
|
"grad_norm": 1.3839406967163086, |
|
"learning_rate": 1.6829934157700338e-06, |
|
"loss": 1.1716, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 2.9113323850165953, |
|
"grad_norm": 1.2562811374664307, |
|
"learning_rate": 1.5224024409828168e-06, |
|
"loss": 1.1466, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 2.9208155523944996, |
|
"grad_norm": 1.4180203676223755, |
|
"learning_rate": 1.3618114661955997e-06, |
|
"loss": 1.1405, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 2.930298719772404, |
|
"grad_norm": 1.7891360521316528, |
|
"learning_rate": 1.2012204914083829e-06, |
|
"loss": 1.1591, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 2.9397818871503083, |
|
"grad_norm": 1.7551426887512207, |
|
"learning_rate": 1.0406295166211659e-06, |
|
"loss": 1.1833, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.9397818871503083, |
|
"eval_loss": 1.129394292831421, |
|
"eval_runtime": 72.2972, |
|
"eval_samples_per_second": 129.659, |
|
"eval_steps_per_second": 16.211, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.9492650545282126, |
|
"grad_norm": 1.4321238994598389, |
|
"learning_rate": 8.800385418339489e-07, |
|
"loss": 1.1879, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 2.9587482219061165, |
|
"grad_norm": 1.732853651046753, |
|
"learning_rate": 7.210534767946041e-07, |
|
"loss": 1.1682, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 2.968231389284021, |
|
"grad_norm": 1.473656415939331, |
|
"learning_rate": 5.604625020073872e-07, |
|
"loss": 1.1708, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 2.977714556661925, |
|
"grad_norm": 1.2021667957305908, |
|
"learning_rate": 3.998715272201702e-07, |
|
"loss": 1.1679, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 2.987197724039829, |
|
"grad_norm": 1.4972681999206543, |
|
"learning_rate": 2.4088646218082545e-07, |
|
"loss": 1.1678, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.987197724039829, |
|
"eval_loss": 1.129324197769165, |
|
"eval_runtime": 72.2219, |
|
"eval_samples_per_second": 129.794, |
|
"eval_steps_per_second": 16.228, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.9966808914177334, |
|
"grad_norm": 1.7410774230957031, |
|
"learning_rate": 8.029548739360848e-08, |
|
"loss": 1.1645, |
|
"step": 31600 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 31635, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.32254007164928e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|