{ "best_metric": 2.8705949783325195, "best_model_checkpoint": "miner_id_24/checkpoint-500", "epoch": 0.4317789291882556, "eval_steps": 50, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008635578583765112, "eval_loss": 4.056758880615234, "eval_runtime": 6.4154, "eval_samples_per_second": 76.067, "eval_steps_per_second": 19.017, "step": 1 }, { "epoch": 0.008635578583765112, "grad_norm": 109.97903442382812, "learning_rate": 4.0400000000000006e-05, "loss": 7.1856, "step": 10 }, { "epoch": 0.017271157167530225, "grad_norm": 128.83065795898438, "learning_rate": 8.080000000000001e-05, "loss": 7.2879, "step": 20 }, { "epoch": 0.025906735751295335, "grad_norm": 156.83480834960938, "learning_rate": 0.00012119999999999999, "loss": 6.7848, "step": 30 }, { "epoch": 0.03454231433506045, "grad_norm": 94.2466812133789, "learning_rate": 0.00016160000000000002, "loss": 6.6317, "step": 40 }, { "epoch": 0.04317789291882556, "grad_norm": 69.40169525146484, "learning_rate": 0.000202, "loss": 7.2513, "step": 50 }, { "epoch": 0.04317789291882556, "eval_loss": 3.4254961013793945, "eval_runtime": 6.4823, "eval_samples_per_second": 75.282, "eval_steps_per_second": 18.82, "step": 50 }, { "epoch": 0.05181347150259067, "grad_norm": 69.7632064819336, "learning_rate": 0.00020175396907624226, "loss": 6.6028, "step": 60 }, { "epoch": 0.06044905008635579, "grad_norm": 122.16877746582031, "learning_rate": 0.0002010170749428986, "loss": 5.9878, "step": 70 }, { "epoch": 0.0690846286701209, "grad_norm": 137.6359100341797, "learning_rate": 0.00019979290767411438, "loss": 5.9063, "step": 80 }, { "epoch": 0.07772020725388601, "grad_norm": 351.2722473144531, "learning_rate": 0.0001980874312897702, "loss": 6.7379, "step": 90 }, { "epoch": 0.08635578583765112, "grad_norm": 46.962703704833984, "learning_rate": 0.00019590895469937675, "loss": 7.5918, "step": 100 }, { "epoch": 0.08635578583765112, "eval_loss": 3.554054021835327, "eval_runtime": 6.3322, "eval_samples_per_second": 77.067, "eval_steps_per_second": 19.267, "step": 100 }, { "epoch": 0.09499136442141623, "grad_norm": 75.2490234375, "learning_rate": 0.0001932680912219027, "loss": 6.5258, "step": 110 }, { "epoch": 0.10362694300518134, "grad_norm": 60.61024856567383, "learning_rate": 0.00019017770687875164, "loss": 6.6532, "step": 120 }, { "epoch": 0.11226252158894647, "grad_norm": 64.55341339111328, "learning_rate": 0.000186652857711799, "loss": 6.1697, "step": 130 }, { "epoch": 0.12089810017271158, "grad_norm": 98.4996566772461, "learning_rate": 0.00018271071643186968, "loss": 6.6974, "step": 140 }, { "epoch": 0.12953367875647667, "grad_norm": 34.625057220458984, "learning_rate": 0.00017837048875501678, "loss": 6.6976, "step": 150 }, { "epoch": 0.12953367875647667, "eval_loss": 3.2074270248413086, "eval_runtime": 6.4564, "eval_samples_per_second": 75.583, "eval_steps_per_second": 18.896, "step": 150 }, { "epoch": 0.1381692573402418, "grad_norm": 51.142398834228516, "learning_rate": 0.00017365331983420376, "loss": 6.4538, "step": 160 }, { "epoch": 0.14680483592400692, "grad_norm": 60.982696533203125, "learning_rate": 0.0001685821912422447, "loss": 6.1183, "step": 170 }, { "epoch": 0.15544041450777202, "grad_norm": 90.77964782714844, "learning_rate": 0.00016318180900789148, "loss": 6.2612, "step": 180 }, { "epoch": 0.16407599309153714, "grad_norm": 37.461490631103516, "learning_rate": 0.00015747848325054544, "loss": 6.2806, "step": 190 }, { "epoch": 0.17271157167530224, "grad_norm": 57.53730392456055, "learning_rate": 0.0001515, "loss": 6.4748, "step": 200 }, { "epoch": 0.17271157167530224, "eval_loss": 3.1944408416748047, "eval_runtime": 6.5289, "eval_samples_per_second": 74.744, "eval_steps_per_second": 18.686, "step": 200 }, { "epoch": 0.18134715025906736, "grad_norm": 26.914474487304688, "learning_rate": 0.00014527548582569683, "loss": 6.1732, "step": 210 }, { "epoch": 0.18998272884283246, "grad_norm": 42.907127380371094, "learning_rate": 0.00013883526593500714, "loss": 6.1321, "step": 220 }, { "epoch": 0.19861830742659758, "grad_norm": 61.954986572265625, "learning_rate": 0.0001322107164318697, "loss": 6.293, "step": 230 }, { "epoch": 0.20725388601036268, "grad_norm": 41.956809997558594, "learning_rate": 0.00012543411145556643, "loss": 6.3759, "step": 240 }, { "epoch": 0.2158894645941278, "grad_norm": 58.70474624633789, "learning_rate": 0.00011853846594435998, "loss": 6.4753, "step": 250 }, { "epoch": 0.2158894645941278, "eval_loss": 3.325136184692383, "eval_runtime": 6.5343, "eval_samples_per_second": 74.683, "eval_steps_per_second": 18.671, "step": 250 }, { "epoch": 0.22452504317789293, "grad_norm": 37.147239685058594, "learning_rate": 0.00011155737479003301, "loss": 6.4968, "step": 260 }, { "epoch": 0.23316062176165803, "grad_norm": 28.960235595703125, "learning_rate": 0.00010452484916695262, "loss": 6.153, "step": 270 }, { "epoch": 0.24179620034542315, "grad_norm": 50.646121978759766, "learning_rate": 9.747515083304742e-05, "loss": 6.4405, "step": 280 }, { "epoch": 0.2504317789291883, "grad_norm": 61.160247802734375, "learning_rate": 9.044262520996702e-05, "loss": 5.6587, "step": 290 }, { "epoch": 0.25906735751295334, "grad_norm": 126.3609848022461, "learning_rate": 8.346153405564004e-05, "loss": 6.0805, "step": 300 }, { "epoch": 0.25906735751295334, "eval_loss": 3.0865094661712646, "eval_runtime": 6.5059, "eval_samples_per_second": 75.008, "eval_steps_per_second": 18.752, "step": 300 }, { "epoch": 0.26770293609671847, "grad_norm": 107.83584594726562, "learning_rate": 7.656588854443357e-05, "loss": 6.5431, "step": 310 }, { "epoch": 0.2763385146804836, "grad_norm": 63.63986587524414, "learning_rate": 6.978928356813031e-05, "loss": 6.3199, "step": 320 }, { "epoch": 0.2849740932642487, "grad_norm": 61.55841064453125, "learning_rate": 6.316473406499288e-05, "loss": 6.3271, "step": 330 }, { "epoch": 0.29360967184801384, "grad_norm": 59.812896728515625, "learning_rate": 5.672451417430317e-05, "loss": 6.2607, "step": 340 }, { "epoch": 0.3022452504317789, "grad_norm": 81.45887756347656, "learning_rate": 5.050000000000002e-05, "loss": 6.3402, "step": 350 }, { "epoch": 0.3022452504317789, "eval_loss": 3.0198187828063965, "eval_runtime": 6.5239, "eval_samples_per_second": 74.802, "eval_steps_per_second": 18.7, "step": 350 }, { "epoch": 0.31088082901554404, "grad_norm": 37.4752082824707, "learning_rate": 4.452151674945458e-05, "loss": 6.201, "step": 360 }, { "epoch": 0.31951640759930916, "grad_norm": 46.92196273803711, "learning_rate": 3.8818190992108515e-05, "loss": 5.6539, "step": 370 }, { "epoch": 0.3281519861830743, "grad_norm": 111.72332763671875, "learning_rate": 3.3417808757755355e-05, "loss": 6.1284, "step": 380 }, { "epoch": 0.33678756476683935, "grad_norm": 140.5625, "learning_rate": 2.8346680165796253e-05, "loss": 6.0232, "step": 390 }, { "epoch": 0.3454231433506045, "grad_norm": 49.17171096801758, "learning_rate": 2.362951124498323e-05, "loss": 6.0123, "step": 400 }, { "epoch": 0.3454231433506045, "eval_loss": 2.9118363857269287, "eval_runtime": 6.5191, "eval_samples_per_second": 74.857, "eval_steps_per_second": 18.714, "step": 400 }, { "epoch": 0.3540587219343696, "grad_norm": 94.65570068359375, "learning_rate": 1.928928356813032e-05, "loss": 5.9852, "step": 410 }, { "epoch": 0.3626943005181347, "grad_norm": 134.4743194580078, "learning_rate": 1.5347142288200977e-05, "loss": 5.8302, "step": 420 }, { "epoch": 0.37132987910189985, "grad_norm": 459.3276672363281, "learning_rate": 1.1822293121248375e-05, "loss": 5.3816, "step": 430 }, { "epoch": 0.3799654576856649, "grad_norm": 118.11572265625, "learning_rate": 8.731908778097302e-06, "loss": 5.7221, "step": 440 }, { "epoch": 0.38860103626943004, "grad_norm": 46.101783752441406, "learning_rate": 6.09104530062326e-06, "loss": 5.6217, "step": 450 }, { "epoch": 0.38860103626943004, "eval_loss": 2.882382869720459, "eval_runtime": 6.5287, "eval_samples_per_second": 74.746, "eval_steps_per_second": 18.687, "step": 450 }, { "epoch": 0.39723661485319517, "grad_norm": 48.339908599853516, "learning_rate": 3.912568710229791e-06, "loss": 6.0256, "step": 460 }, { "epoch": 0.4058721934369603, "grad_norm": 83.55408477783203, "learning_rate": 2.2070923258856255e-06, "loss": 5.9578, "step": 470 }, { "epoch": 0.41450777202072536, "grad_norm": 55.561649322509766, "learning_rate": 9.829250571013935e-07, "loss": 5.9324, "step": 480 }, { "epoch": 0.4231433506044905, "grad_norm": 66.6230239868164, "learning_rate": 2.4603092375775605e-07, "loss": 5.6569, "step": 490 }, { "epoch": 0.4317789291882556, "grad_norm": 116.90052032470703, "learning_rate": 0.0, "loss": 6.1575, "step": 500 }, { "epoch": 0.4317789291882556, "eval_loss": 2.8705949783325195, "eval_runtime": 6.5205, "eval_samples_per_second": 74.841, "eval_steps_per_second": 18.71, "step": 500 } ], "logging_steps": 10, "max_steps": 500, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4657011228672000.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }