{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 888888, "global_step": 308, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.012987012987012988, "grad_norm": 30.409713806756372, "learning_rate": 4.999479820473016e-07, "logits/chosen": -1.1957358121871948, "logits/rejected": -1.1989657878875732, "logps/chosen": -236.6458282470703, "logps/rejected": -274.4889221191406, "loss": 0.7214, "rewards/accuracies": 0.5625, "rewards/chosen": 0.9376516938209534, "rewards/margins": 0.12735898792743683, "rewards/rejected": 0.810292661190033, "step": 2 }, { "epoch": 0.025974025974025976, "grad_norm": 32.07895252539544, "learning_rate": 4.997919498361457e-07, "logits/chosen": -1.2326594591140747, "logits/rejected": -1.2664812803268433, "logps/chosen": -257.2532958984375, "logps/rejected": -252.87025451660156, "loss": 0.727, "rewards/accuracies": 0.59375, "rewards/chosen": 1.142688512802124, "rewards/margins": 0.25847145915031433, "rewards/rejected": 0.8842170238494873, "step": 4 }, { "epoch": 0.03896103896103896, "grad_norm": 31.29868726835709, "learning_rate": 4.995319682983417e-07, "logits/chosen": -1.183756947517395, "logits/rejected": -1.1092337369918823, "logps/chosen": -191.75558471679688, "logps/rejected": -230.58950805664062, "loss": 0.7031, "rewards/accuracies": 0.4375, "rewards/chosen": 0.7358960509300232, "rewards/margins": -0.0961524099111557, "rewards/rejected": 0.8320484757423401, "step": 6 }, { "epoch": 0.05194805194805195, "grad_norm": 32.01008811759358, "learning_rate": 4.991681456235483e-07, "logits/chosen": -1.2155265808105469, "logits/rejected": -1.3162301778793335, "logps/chosen": -202.1005859375, "logps/rejected": -232.53140258789062, "loss": 0.6947, "rewards/accuracies": 0.65625, "rewards/chosen": 1.0601692199707031, "rewards/margins": 0.3322027921676636, "rewards/rejected": 0.72796630859375, "step": 8 }, { "epoch": 0.06493506493506493, "grad_norm": 26.584702872904153, "learning_rate": 4.98700633214251e-07, "logits/chosen": -1.08168363571167, "logits/rejected": -1.21769118309021, "logps/chosen": -217.7674560546875, "logps/rejected": -195.05032348632812, "loss": 0.6823, "rewards/accuracies": 0.5, "rewards/chosen": 0.8856821060180664, "rewards/margins": 0.06800435483455658, "rewards/rejected": 0.8176776170730591, "step": 10 }, { "epoch": 0.07792207792207792, "grad_norm": 31.831697368646147, "learning_rate": 4.981296256227569e-07, "logits/chosen": -0.912403404712677, "logits/rejected": -0.9036965370178223, "logps/chosen": -204.28919982910156, "logps/rejected": -211.12686157226562, "loss": 0.7049, "rewards/accuracies": 0.53125, "rewards/chosen": 0.849972665309906, "rewards/margins": 0.017677173018455505, "rewards/rejected": 0.8322954177856445, "step": 12 }, { "epoch": 0.09090909090909091, "grad_norm": 29.9192585773342, "learning_rate": 4.974553604702332e-07, "logits/chosen": -1.0211938619613647, "logits/rejected": -1.0898373126983643, "logps/chosen": -166.66610717773438, "logps/rejected": -154.6410675048828, "loss": 0.6707, "rewards/accuracies": 0.6875, "rewards/chosen": 1.0837771892547607, "rewards/margins": 0.4414757192134857, "rewards/rejected": 0.6423014402389526, "step": 14 }, { "epoch": 0.1038961038961039, "grad_norm": 29.446896480009652, "learning_rate": 4.966781183478222e-07, "logits/chosen": -1.1182293891906738, "logits/rejected": -1.116457223892212, "logps/chosen": -201.10256958007812, "logps/rejected": -219.7225341796875, "loss": 0.6796, "rewards/accuracies": 0.625, "rewards/chosen": 0.78659588098526, "rewards/margins": 0.23196566104888916, "rewards/rejected": 0.5546301603317261, "step": 16 }, { "epoch": 0.11688311688311688, "grad_norm": 30.21023973321977, "learning_rate": 4.957982226998757e-07, "logits/chosen": -1.31354820728302, "logits/rejected": -1.2536346912384033, "logps/chosen": -265.8379821777344, "logps/rejected": -281.1967468261719, "loss": 0.7288, "rewards/accuracies": 0.59375, "rewards/chosen": 1.1193804740905762, "rewards/margins": 0.1973106861114502, "rewards/rejected": 0.922069787979126, "step": 18 }, { "epoch": 0.12987012987012986, "grad_norm": 30.94898503765605, "learning_rate": 4.948160396893553e-07, "logits/chosen": -1.2565226554870605, "logits/rejected": -1.15552818775177, "logps/chosen": -174.21267700195312, "logps/rejected": -179.50820922851562, "loss": 0.6927, "rewards/accuracies": 0.59375, "rewards/chosen": 0.931135892868042, "rewards/margins": 0.1621209681034088, "rewards/rejected": 0.7690149545669556, "step": 20 }, { "epoch": 0.14285714285714285, "grad_norm": 29.667041326907114, "learning_rate": 4.937319780454559e-07, "logits/chosen": -1.3007498979568481, "logits/rejected": -1.1949548721313477, "logps/chosen": -230.96115112304688, "logps/rejected": -248.40798950195312, "loss": 0.7216, "rewards/accuracies": 0.34375, "rewards/chosen": 0.7717787027359009, "rewards/margins": -0.18773330748081207, "rewards/rejected": 0.9595120549201965, "step": 22 }, { "epoch": 0.15584415584415584, "grad_norm": 33.222954526021944, "learning_rate": 4.925464888935161e-07, "logits/chosen": -1.1974416971206665, "logits/rejected": -1.2200247049331665, "logps/chosen": -198.02139282226562, "logps/rejected": -242.11419677734375, "loss": 0.7295, "rewards/accuracies": 0.5625, "rewards/chosen": 0.8105030655860901, "rewards/margins": 0.041422925889492035, "rewards/rejected": 0.7690802216529846, "step": 24 }, { "epoch": 0.16883116883116883, "grad_norm": 31.901066989434995, "learning_rate": 4.912600655672849e-07, "logits/chosen": -1.240085482597351, "logits/rejected": -1.2631834745407104, "logps/chosen": -264.38067626953125, "logps/rejected": -317.8161926269531, "loss": 0.6988, "rewards/accuracies": 0.53125, "rewards/chosen": 1.0404579639434814, "rewards/margins": 0.05802140012383461, "rewards/rejected": 0.9824365377426147, "step": 26 }, { "epoch": 0.18181818181818182, "grad_norm": 34.21159593307014, "learning_rate": 4.898732434036243e-07, "logits/chosen": -1.189288854598999, "logits/rejected": -1.048648476600647, "logps/chosen": -262.818115234375, "logps/rejected": -301.31927490234375, "loss": 0.7616, "rewards/accuracies": 0.4375, "rewards/chosen": 0.8156523108482361, "rewards/margins": -0.08142556995153427, "rewards/rejected": 0.8970779180526733, "step": 28 }, { "epoch": 0.19480519480519481, "grad_norm": 32.29301620044956, "learning_rate": 4.883865995197318e-07, "logits/chosen": -1.1413640975952148, "logits/rejected": -1.2122159004211426, "logps/chosen": -217.71823120117188, "logps/rejected": -221.3170166015625, "loss": 0.7261, "rewards/accuracies": 0.78125, "rewards/chosen": 1.1059789657592773, "rewards/margins": 0.40076321363449097, "rewards/rejected": 0.7052158713340759, "step": 30 }, { "epoch": 0.2077922077922078, "grad_norm": 30.16619750369487, "learning_rate": 4.868007525729775e-07, "logits/chosen": -1.36598801612854, "logits/rejected": -1.3407963514328003, "logps/chosen": -239.17111206054688, "logps/rejected": -226.31658935546875, "loss": 0.6825, "rewards/accuracies": 0.5625, "rewards/chosen": 0.7918877005577087, "rewards/margins": 0.05119353532791138, "rewards/rejected": 0.7406941056251526, "step": 32 }, { "epoch": 0.22077922077922077, "grad_norm": 30.605215606478595, "learning_rate": 4.851163625034529e-07, "logits/chosen": -1.1574623584747314, "logits/rejected": -1.193921685218811, "logps/chosen": -230.02108764648438, "logps/rejected": -244.59841918945312, "loss": 0.6851, "rewards/accuracies": 0.65625, "rewards/chosen": 1.080081820487976, "rewards/margins": 0.3703171908855438, "rewards/rejected": 0.7097645998001099, "step": 34 }, { "epoch": 0.23376623376623376, "grad_norm": 31.802179808840666, "learning_rate": 4.833341302593417e-07, "logits/chosen": -1.159416913986206, "logits/rejected": -1.1420115232467651, "logps/chosen": -140.0688018798828, "logps/rejected": -152.40975952148438, "loss": 0.6732, "rewards/accuracies": 0.53125, "rewards/chosen": 0.8693720102310181, "rewards/margins": 0.11663462221622467, "rewards/rejected": 0.7527374625205994, "step": 36 }, { "epoch": 0.24675324675324675, "grad_norm": 30.94540714928562, "learning_rate": 4.814547975052244e-07, "logits/chosen": -1.3302018642425537, "logits/rejected": -1.3434113264083862, "logps/chosen": -191.34698486328125, "logps/rejected": -196.69509887695312, "loss": 0.6714, "rewards/accuracies": 0.46875, "rewards/chosen": 0.7995274662971497, "rewards/margins": 0.047811880707740784, "rewards/rejected": 0.7517155408859253, "step": 38 }, { "epoch": 0.2597402597402597, "grad_norm": 29.865098693610847, "learning_rate": 4.794791463134399e-07, "logits/chosen": -1.2271629571914673, "logits/rejected": -1.3550831079483032, "logps/chosen": -226.42410278320312, "logps/rejected": -214.7489013671875, "loss": 0.6975, "rewards/accuracies": 0.71875, "rewards/chosen": 0.8646127581596375, "rewards/margins": 0.22612789273262024, "rewards/rejected": 0.6384848356246948, "step": 40 }, { "epoch": 0.2727272727272727, "grad_norm": 30.43675432858084, "learning_rate": 4.774079988386296e-07, "logits/chosen": -1.1719427108764648, "logits/rejected": -1.2684495449066162, "logps/chosen": -208.20394897460938, "logps/rejected": -235.50634765625, "loss": 0.6855, "rewards/accuracies": 0.6875, "rewards/chosen": 1.0848238468170166, "rewards/margins": 0.3408503830432892, "rewards/rejected": 0.7439733743667603, "step": 42 }, { "epoch": 0.2857142857142857, "grad_norm": 31.511971354878987, "learning_rate": 4.752422169756047e-07, "logits/chosen": -1.2040029764175415, "logits/rejected": -1.2218215465545654, "logps/chosen": -259.76959228515625, "logps/rejected": -279.75946044921875, "loss": 0.6772, "rewards/accuracies": 0.59375, "rewards/chosen": 0.9812463521957397, "rewards/margins": 0.2615824341773987, "rewards/rejected": 0.7196639776229858, "step": 44 }, { "epoch": 0.2987012987012987, "grad_norm": 35.50720072355199, "learning_rate": 4.729827020006735e-07, "logits/chosen": -1.2218239307403564, "logits/rejected": -1.2937262058258057, "logps/chosen": -239.9827880859375, "logps/rejected": -250.3217315673828, "loss": 0.7242, "rewards/accuracies": 0.5625, "rewards/chosen": 0.8138805627822876, "rewards/margins": 0.09423254430294037, "rewards/rejected": 0.719648003578186, "step": 46 }, { "epoch": 0.3116883116883117, "grad_norm": 32.207699613934025, "learning_rate": 4.706303941965803e-07, "logits/chosen": -1.174890398979187, "logits/rejected": -1.1500227451324463, "logps/chosen": -257.21533203125, "logps/rejected": -302.0556945800781, "loss": 0.6856, "rewards/accuracies": 0.6875, "rewards/chosen": 1.0188751220703125, "rewards/margins": 0.20580947399139404, "rewards/rejected": 0.8130655884742737, "step": 48 }, { "epoch": 0.3246753246753247, "grad_norm": 31.074659437805654, "learning_rate": 4.68186272461214e-07, "logits/chosen": -1.2761106491088867, "logits/rejected": -1.307037353515625, "logps/chosen": -228.11666870117188, "logps/rejected": -254.00048828125, "loss": 0.6711, "rewards/accuracies": 0.71875, "rewards/chosen": 0.9066404104232788, "rewards/margins": 0.32165050506591797, "rewards/rejected": 0.5849898457527161, "step": 50 }, { "epoch": 0.33766233766233766, "grad_norm": 31.027447722981428, "learning_rate": 4.656513539002451e-07, "logits/chosen": -1.2342908382415771, "logits/rejected": -1.2500003576278687, "logps/chosen": -182.6434783935547, "logps/rejected": -210.83444213867188, "loss": 0.6867, "rewards/accuracies": 0.53125, "rewards/chosen": 0.8957956433296204, "rewards/margins": 0.21347832679748535, "rewards/rejected": 0.6823172569274902, "step": 52 }, { "epoch": 0.35064935064935066, "grad_norm": 28.860294980547383, "learning_rate": 4.6302669340386415e-07, "logits/chosen": -1.2332104444503784, "logits/rejected": -1.1100932359695435, "logps/chosen": -194.90341186523438, "logps/rejected": -220.88478088378906, "loss": 0.692, "rewards/accuracies": 0.59375, "rewards/chosen": 0.7950139045715332, "rewards/margins": 0.09959140419960022, "rewards/rejected": 0.6954225897789001, "step": 54 }, { "epoch": 0.36363636363636365, "grad_norm": 37.00612209924237, "learning_rate": 4.603133832077953e-07, "logits/chosen": -1.2518322467803955, "logits/rejected": -1.272511601448059, "logps/chosen": -219.75762939453125, "logps/rejected": -225.9881591796875, "loss": 0.7341, "rewards/accuracies": 0.5625, "rewards/chosen": 0.8070573806762695, "rewards/margins": 0.15917816758155823, "rewards/rejected": 0.6478793025016785, "step": 56 }, { "epoch": 0.37662337662337664, "grad_norm": 34.181154948444195, "learning_rate": 4.575125524387701e-07, "logits/chosen": -1.0896148681640625, "logits/rejected": -1.0761992931365967, "logps/chosen": -246.28172302246094, "logps/rejected": -231.1036376953125, "loss": 0.6999, "rewards/accuracies": 0.5, "rewards/chosen": 0.6937105059623718, "rewards/margins": 0.09433312714099884, "rewards/rejected": 0.5993773937225342, "step": 58 }, { "epoch": 0.38961038961038963, "grad_norm": 33.21940371087566, "learning_rate": 4.5462536664464836e-07, "logits/chosen": -1.2278883457183838, "logits/rejected": -1.2770620584487915, "logps/chosen": -325.75714111328125, "logps/rejected": -309.88836669921875, "loss": 0.685, "rewards/accuracies": 0.65625, "rewards/chosen": 0.9801526069641113, "rewards/margins": 0.22209185361862183, "rewards/rejected": 0.7580606937408447, "step": 60 }, { "epoch": 0.4025974025974026, "grad_norm": 29.951734183402802, "learning_rate": 4.516530273093825e-07, "logits/chosen": -1.3827568292617798, "logits/rejected": -1.3074092864990234, "logps/chosen": -226.56942749023438, "logps/rejected": -261.1898498535156, "loss": 0.6715, "rewards/accuracies": 0.5625, "rewards/chosen": 0.8671592473983765, "rewards/margins": 0.1708277314901352, "rewards/rejected": 0.6963315010070801, "step": 62 }, { "epoch": 0.4155844155844156, "grad_norm": 35.00256057592286, "learning_rate": 4.485967713530281e-07, "logits/chosen": -1.2592735290527344, "logits/rejected": -1.276845097541809, "logps/chosen": -280.13421630859375, "logps/rejected": -272.318359375, "loss": 0.6797, "rewards/accuracies": 0.6875, "rewards/chosen": 0.9574460983276367, "rewards/margins": 0.2645747661590576, "rewards/rejected": 0.6928713321685791, "step": 64 }, { "epoch": 0.42857142857142855, "grad_norm": 130.44143506253818, "learning_rate": 4.4545787061700746e-07, "logits/chosen": -1.201524019241333, "logits/rejected": -1.2254564762115479, "logps/chosen": -222.44644165039062, "logps/rejected": -215.9447021484375, "loss": 0.6865, "rewards/accuracies": 0.59375, "rewards/chosen": 0.8678699731826782, "rewards/margins": 0.32531124353408813, "rewards/rejected": 0.5425587296485901, "step": 66 }, { "epoch": 0.44155844155844154, "grad_norm": 30.886647455965967, "learning_rate": 4.422376313348405e-07, "logits/chosen": -1.2102560997009277, "logits/rejected": -1.1355280876159668, "logps/chosen": -211.4716796875, "logps/rejected": -246.92666625976562, "loss": 0.6728, "rewards/accuracies": 0.75, "rewards/chosen": 0.9252184629440308, "rewards/margins": 0.4986411929130554, "rewards/rejected": 0.4265773296356201, "step": 68 }, { "epoch": 0.45454545454545453, "grad_norm": 38.68657958814148, "learning_rate": 4.3893739358856455e-07, "logits/chosen": -1.0511149168014526, "logits/rejected": -1.1264652013778687, "logps/chosen": -193.53610229492188, "logps/rejected": -218.5620880126953, "loss": 0.701, "rewards/accuracies": 0.5625, "rewards/chosen": 0.7159083485603333, "rewards/margins": 0.046296779066324234, "rewards/rejected": 0.6696116328239441, "step": 70 }, { "epoch": 0.4675324675324675, "grad_norm": 30.130370931610912, "learning_rate": 4.355585307510675e-07, "logits/chosen": -1.246776819229126, "logits/rejected": -1.2227742671966553, "logps/chosen": -159.0496826171875, "logps/rejected": -173.915771484375, "loss": 0.6643, "rewards/accuracies": 0.5, "rewards/chosen": 0.7583217620849609, "rewards/margins": 0.24899302423000336, "rewards/rejected": 0.5093286633491516, "step": 72 }, { "epoch": 0.4805194805194805, "grad_norm": 33.33819375543502, "learning_rate": 4.3210244891456725e-07, "logits/chosen": -1.3857276439666748, "logits/rejected": -1.2871744632720947, "logps/chosen": -237.01812744140625, "logps/rejected": -222.86903381347656, "loss": 0.6923, "rewards/accuracies": 0.53125, "rewards/chosen": 0.6103696227073669, "rewards/margins": 0.10617219656705856, "rewards/rejected": 0.5041974186897278, "step": 74 }, { "epoch": 0.4935064935064935, "grad_norm": 31.374872695083415, "learning_rate": 4.2857058630547586e-07, "logits/chosen": -1.1448109149932861, "logits/rejected": -1.0983011722564697, "logps/chosen": -251.7313690185547, "logps/rejected": -263.6651306152344, "loss": 0.6743, "rewards/accuracies": 0.4375, "rewards/chosen": 0.733715295791626, "rewards/margins": 0.12528231739997864, "rewards/rejected": 0.608432948589325, "step": 76 }, { "epoch": 0.5064935064935064, "grad_norm": 29.77623703163317, "learning_rate": 4.2496441268589047e-07, "logits/chosen": -1.0607296228408813, "logits/rejected": -1.2361115217208862, "logps/chosen": -208.2403564453125, "logps/rejected": -205.75625610351562, "loss": 0.6756, "rewards/accuracies": 0.71875, "rewards/chosen": 0.9102213382720947, "rewards/margins": 0.41290420293807983, "rewards/rejected": 0.4973170757293701, "step": 78 }, { "epoch": 0.5194805194805194, "grad_norm": 29.02788697480319, "learning_rate": 4.2128542874196107e-07, "logits/chosen": -1.0309326648712158, "logits/rejected": -1.143478274345398, "logps/chosen": -173.32151794433594, "logps/rejected": -161.68482971191406, "loss": 0.6691, "rewards/accuracies": 0.5625, "rewards/chosen": 0.7697413563728333, "rewards/margins": 0.14579711854457855, "rewards/rejected": 0.6239442229270935, "step": 80 }, { "epoch": 0.5324675324675324, "grad_norm": 29.784869306609654, "learning_rate": 4.1753516545938986e-07, "logits/chosen": -1.2103991508483887, "logits/rejected": -1.3010125160217285, "logps/chosen": -190.08010864257812, "logps/rejected": -182.82606506347656, "loss": 0.6618, "rewards/accuracies": 0.40625, "rewards/chosen": 0.6524017453193665, "rewards/margins": 0.07943220436573029, "rewards/rejected": 0.5729695558547974, "step": 82 }, { "epoch": 0.5454545454545454, "grad_norm": 32.96758632787842, "learning_rate": 4.137151834863213e-07, "logits/chosen": -1.1179546117782593, "logits/rejected": -1.1667120456695557, "logps/chosen": -252.30087280273438, "logps/rejected": -258.8082275390625, "loss": 0.6678, "rewards/accuracies": 0.59375, "rewards/chosen": 0.7305130958557129, "rewards/margins": 0.33587121963500977, "rewards/rejected": 0.39464184641838074, "step": 84 }, { "epoch": 0.5584415584415584, "grad_norm": 29.473194289500984, "learning_rate": 4.098270724838879e-07, "logits/chosen": -1.2531524896621704, "logits/rejected": -1.2107303142547607, "logps/chosen": -257.7474365234375, "logps/rejected": -297.6868591308594, "loss": 0.6752, "rewards/accuracies": 0.65625, "rewards/chosen": 0.9224438667297363, "rewards/margins": 0.45967769622802734, "rewards/rejected": 0.4627661406993866, "step": 86 }, { "epoch": 0.5714285714285714, "grad_norm": 25.375543282285655, "learning_rate": 4.058724504646834e-07, "logits/chosen": -1.212619662284851, "logits/rejected": -1.218203067779541, "logps/chosen": -242.98947143554688, "logps/rejected": -272.5478515625, "loss": 0.629, "rewards/accuracies": 0.65625, "rewards/chosen": 0.7270645499229431, "rewards/margins": 0.3879895508289337, "rewards/rejected": 0.33907490968704224, "step": 88 }, { "epoch": 0.5844155844155844, "grad_norm": 33.30211893834397, "learning_rate": 4.018529631194369e-07, "logits/chosen": -1.3774781227111816, "logits/rejected": -1.4289326667785645, "logps/chosen": -341.9387512207031, "logps/rejected": -348.09283447265625, "loss": 0.6743, "rewards/accuracies": 0.59375, "rewards/chosen": 0.8079380989074707, "rewards/margins": 0.1343669593334198, "rewards/rejected": 0.6735711097717285, "step": 90 }, { "epoch": 0.5974025974025974, "grad_norm": 33.04398574915242, "learning_rate": 3.9777028313216913e-07, "logits/chosen": -1.1324537992477417, "logits/rejected": -1.2160053253173828, "logps/chosen": -216.34042358398438, "logps/rejected": -239.47853088378906, "loss": 0.6886, "rewards/accuracies": 0.5625, "rewards/chosen": 0.5493214726448059, "rewards/margins": 0.09607386589050293, "rewards/rejected": 0.4532475769519806, "step": 92 }, { "epoch": 0.6103896103896104, "grad_norm": 27.7423356377416, "learning_rate": 3.9362610948411584e-07, "logits/chosen": -1.273106336593628, "logits/rejected": -1.278911828994751, "logps/chosen": -298.0914306640625, "logps/rejected": -303.9120788574219, "loss": 0.6527, "rewards/accuracies": 0.5, "rewards/chosen": 0.9156701564788818, "rewards/margins": 0.36599063873291016, "rewards/rejected": 0.5496795177459717, "step": 94 }, { "epoch": 0.6233766233766234, "grad_norm": 28.666974084256797, "learning_rate": 3.8942216674670737e-07, "logits/chosen": -1.1729968786239624, "logits/rejected": -1.2594560384750366, "logps/chosen": -201.3909454345703, "logps/rejected": -202.02003479003906, "loss": 0.6969, "rewards/accuracies": 0.4375, "rewards/chosen": 0.5368354320526123, "rewards/margins": 0.11352111399173737, "rewards/rejected": 0.4233143925666809, "step": 96 }, { "epoch": 0.6363636363636364, "grad_norm": 31.936323967340297, "learning_rate": 3.851602043638994e-07, "logits/chosen": -1.2385543584823608, "logits/rejected": -1.3384761810302734, "logps/chosen": -247.0762939453125, "logps/rejected": -246.0072021484375, "loss": 0.7103, "rewards/accuracies": 0.625, "rewards/chosen": 0.6583046317100525, "rewards/margins": 0.3179153501987457, "rewards/rejected": 0.3403893709182739, "step": 98 }, { "epoch": 0.6493506493506493, "grad_norm": 32.627149716732674, "learning_rate": 3.80841995924153e-07, "logits/chosen": -1.2925565242767334, "logits/rejected": -1.22373628616333, "logps/chosen": -183.79818725585938, "logps/rejected": -207.68206787109375, "loss": 0.6958, "rewards/accuracies": 0.53125, "rewards/chosen": 0.48615562915802, "rewards/margins": 0.16806553304195404, "rewards/rejected": 0.3180900812149048, "step": 100 }, { "epoch": 0.6623376623376623, "grad_norm": 35.550165524685106, "learning_rate": 3.7646933842236707e-07, "logits/chosen": -1.3273155689239502, "logits/rejected": -1.2905724048614502, "logps/chosen": -210.8981475830078, "logps/rejected": -253.29486083984375, "loss": 0.674, "rewards/accuracies": 0.59375, "rewards/chosen": 0.7722782492637634, "rewards/margins": 0.2569873034954071, "rewards/rejected": 0.5152909755706787, "step": 102 }, { "epoch": 0.6753246753246753, "grad_norm": 37.08342788520365, "learning_rate": 3.720440515120703e-07, "logits/chosen": -1.0195050239562988, "logits/rejected": -1.0432817935943604, "logps/chosen": -190.58543395996094, "logps/rejected": -212.59231567382812, "loss": 0.6802, "rewards/accuracies": 0.5625, "rewards/chosen": 0.7294161915779114, "rewards/margins": 0.22578200697898865, "rewards/rejected": 0.5036342144012451, "step": 104 }, { "epoch": 0.6883116883116883, "grad_norm": 31.103871645303624, "learning_rate": 3.6756797674818417e-07, "logits/chosen": -1.1702743768692017, "logits/rejected": -1.2568817138671875, "logps/chosen": -288.2113037109375, "logps/rejected": -276.70550537109375, "loss": 0.667, "rewards/accuracies": 0.625, "rewards/chosen": 0.7896406650543213, "rewards/margins": 0.35757139325141907, "rewards/rejected": 0.43206924200057983, "step": 106 }, { "epoch": 0.7012987012987013, "grad_norm": 33.27213712499375, "learning_rate": 3.630429768206714e-07, "logits/chosen": -1.1950292587280273, "logits/rejected": -1.2156962156295776, "logps/chosen": -260.03192138671875, "logps/rejected": -280.8499450683594, "loss": 0.6698, "rewards/accuracies": 0.625, "rewards/chosen": 0.6720072627067566, "rewards/margins": 0.1990586519241333, "rewards/rejected": 0.4729485809803009, "step": 108 }, { "epoch": 0.7142857142857143, "grad_norm": 34.347475898145326, "learning_rate": 3.584709347793895e-07, "logits/chosen": -1.2583609819412231, "logits/rejected": -1.2271919250488281, "logps/chosen": -218.08056640625, "logps/rejected": -265.71612548828125, "loss": 0.6771, "rewards/accuracies": 0.6875, "rewards/chosen": 0.7517715692520142, "rewards/margins": 0.31436362862586975, "rewards/rejected": 0.4374079704284668, "step": 110 }, { "epoch": 0.7272727272727273, "grad_norm": 28.787797168356807, "learning_rate": 3.5385375325047163e-07, "logits/chosen": -1.301234245300293, "logits/rejected": -1.3219377994537354, "logps/chosen": -198.1061553955078, "logps/rejected": -215.04071044921875, "loss": 0.6702, "rewards/accuracies": 0.53125, "rewards/chosen": 0.6467806696891785, "rewards/margins": 0.201984703540802, "rewards/rejected": 0.44479596614837646, "step": 112 }, { "epoch": 0.7402597402597403, "grad_norm": 31.164777041001017, "learning_rate": 3.491933536445606e-07, "logits/chosen": -1.3274401426315308, "logits/rejected": -1.2342091798782349, "logps/chosen": -153.36151123046875, "logps/rejected": -195.70001220703125, "loss": 0.6501, "rewards/accuracies": 0.71875, "rewards/chosen": 0.6609562635421753, "rewards/margins": 0.36767348647117615, "rewards/rejected": 0.2932826578617096, "step": 114 }, { "epoch": 0.7532467532467533, "grad_norm": 28.597437073725573, "learning_rate": 3.4449167535722664e-07, "logits/chosen": -1.0330864191055298, "logits/rejected": -1.1435695886611938, "logps/chosen": -273.2784423828125, "logps/rejected": -281.6010437011719, "loss": 0.6587, "rewards/accuracies": 0.71875, "rewards/chosen": 0.6794790029525757, "rewards/margins": 0.4547131061553955, "rewards/rejected": 0.22476595640182495, "step": 116 }, { "epoch": 0.7662337662337663, "grad_norm": 32.62513636994299, "learning_rate": 3.3975067496189963e-07, "logits/chosen": -1.2125685214996338, "logits/rejected": -1.3594530820846558, "logps/chosen": -317.5354919433594, "logps/rejected": -315.6915588378906, "loss": 0.65, "rewards/accuracies": 0.875, "rewards/chosen": 0.8727169036865234, "rewards/margins": 0.7781965136528015, "rewards/rejected": 0.09452031552791595, "step": 118 }, { "epoch": 0.7792207792207793, "grad_norm": 32.86814732954617, "learning_rate": 3.349723253956541e-07, "logits/chosen": -1.0453016757965088, "logits/rejected": -1.1661120653152466, "logps/chosen": -145.11253356933594, "logps/rejected": -178.77328491210938, "loss": 0.6635, "rewards/accuracies": 0.625, "rewards/chosen": 0.7022942900657654, "rewards/margins": 0.42999541759490967, "rewards/rejected": 0.2722988426685333, "step": 120 }, { "epoch": 0.7922077922077922, "grad_norm": 31.986096882532095, "learning_rate": 3.3015861513818383e-07, "logits/chosen": -1.1393004655838013, "logits/rejected": -1.071953296661377, "logps/chosen": -257.8083801269531, "logps/rejected": -292.0323791503906, "loss": 0.66, "rewards/accuracies": 0.65625, "rewards/chosen": 0.6557089686393738, "rewards/margins": 0.2878965139389038, "rewards/rejected": 0.3678124248981476, "step": 122 }, { "epoch": 0.8051948051948052, "grad_norm": 32.75329554864317, "learning_rate": 3.2531154738430853e-07, "logits/chosen": -1.2495208978652954, "logits/rejected": -1.1901376247406006, "logps/chosen": -277.14208984375, "logps/rejected": -308.6678161621094, "loss": 0.6484, "rewards/accuracies": 0.59375, "rewards/chosen": 0.6612053513526917, "rewards/margins": 0.33286944031715393, "rewards/rejected": 0.32833582162857056, "step": 124 }, { "epoch": 0.8181818181818182, "grad_norm": 35.313227849390415, "learning_rate": 3.204331392103574e-07, "logits/chosen": -1.0496852397918701, "logits/rejected": -1.0938959121704102, "logps/chosen": -183.056884765625, "logps/rejected": -218.71853637695312, "loss": 0.6607, "rewards/accuracies": 0.625, "rewards/chosen": 0.7791882157325745, "rewards/margins": 0.2896581292152405, "rewards/rejected": 0.489530086517334, "step": 126 }, { "epoch": 0.8311688311688312, "grad_norm": 32.66165401072115, "learning_rate": 3.155254207347755e-07, "logits/chosen": -1.154829740524292, "logits/rejected": -1.0259711742401123, "logps/chosen": -193.2096710205078, "logps/rejected": -234.06759643554688, "loss": 0.6539, "rewards/accuracies": 0.59375, "rewards/chosen": 0.7302087545394897, "rewards/margins": 0.37150081992149353, "rewards/rejected": 0.3587079346179962, "step": 128 }, { "epoch": 0.8441558441558441, "grad_norm": 34.704917232369, "learning_rate": 3.1059043427330314e-07, "logits/chosen": -1.179085373878479, "logits/rejected": -1.1466186046600342, "logps/chosen": -290.7761535644531, "logps/rejected": -247.35595703125, "loss": 0.689, "rewards/accuracies": 0.4375, "rewards/chosen": 0.7207962274551392, "rewards/margins": 0.21811310946941376, "rewards/rejected": 0.5026831030845642, "step": 130 }, { "epoch": 0.8571428571428571, "grad_norm": 31.672800837060226, "learning_rate": 3.056302334890786e-07, "logits/chosen": -1.0660974979400635, "logits/rejected": -1.1430083513259888, "logps/chosen": -186.87071228027344, "logps/rejected": -191.1336669921875, "loss": 0.6728, "rewards/accuracies": 0.625, "rewards/chosen": 0.577528178691864, "rewards/margins": 0.338672012090683, "rewards/rejected": 0.23885619640350342, "step": 132 }, { "epoch": 0.8701298701298701, "grad_norm": 31.658738700878413, "learning_rate": 3.0064688253802024e-07, "logits/chosen": -1.0814266204833984, "logits/rejected": -1.141867756843567, "logps/chosen": -219.94424438476562, "logps/rejected": -249.24383544921875, "loss": 0.6633, "rewards/accuracies": 0.53125, "rewards/chosen": 0.6220312118530273, "rewards/margins": 0.4069012999534607, "rewards/rejected": 0.21512994170188904, "step": 134 }, { "epoch": 0.8831168831168831, "grad_norm": 33.16970393800331, "learning_rate": 2.956424552098404e-07, "logits/chosen": -1.313258171081543, "logits/rejected": -1.3151941299438477, "logps/chosen": -300.03961181640625, "logps/rejected": -340.99566650390625, "loss": 0.6998, "rewards/accuracies": 0.625, "rewards/chosen": 0.7123932242393494, "rewards/margins": 0.30548784136772156, "rewards/rejected": 0.4069053828716278, "step": 136 }, { "epoch": 0.8961038961038961, "grad_norm": 27.93166457525636, "learning_rate": 2.9061903406505153e-07, "logits/chosen": -1.141364574432373, "logits/rejected": -1.0247657299041748, "logps/chosen": -200.64974975585938, "logps/rejected": -197.97593688964844, "loss": 0.6424, "rewards/accuracies": 0.5625, "rewards/chosen": 0.630258321762085, "rewards/margins": 0.3348933160305023, "rewards/rejected": 0.29536497592926025, "step": 138 }, { "epoch": 0.9090909090909091, "grad_norm": 28.870063032351457, "learning_rate": 2.8557870956832133e-07, "logits/chosen": -1.274742603302002, "logits/rejected": -1.2089191675186157, "logps/chosen": -197.0199737548828, "logps/rejected": -220.10044860839844, "loss": 0.642, "rewards/accuracies": 0.71875, "rewards/chosen": 0.6193188428878784, "rewards/margins": 0.4841911494731903, "rewards/rejected": 0.1351277232170105, "step": 140 }, { "epoch": 0.922077922077922, "grad_norm": 34.83723284802987, "learning_rate": 2.8052357921854e-07, "logits/chosen": -1.0989036560058594, "logits/rejected": -1.2587521076202393, "logps/chosen": -167.8390350341797, "logps/rejected": -161.0415496826172, "loss": 0.6461, "rewards/accuracies": 0.625, "rewards/chosen": 0.7375519275665283, "rewards/margins": 0.4270685017108917, "rewards/rejected": 0.3104833960533142, "step": 142 }, { "epoch": 0.935064935064935, "grad_norm": 32.08059076203408, "learning_rate": 2.754557466759589e-07, "logits/chosen": -1.296328067779541, "logits/rejected": -1.3102202415466309, "logps/chosen": -209.74093627929688, "logps/rejected": -220.9139862060547, "loss": 0.6487, "rewards/accuracies": 0.5625, "rewards/chosen": 0.611607551574707, "rewards/margins": 0.38875120878219604, "rewards/rejected": 0.22285637259483337, "step": 144 }, { "epoch": 0.948051948051948, "grad_norm": 30.710641933445263, "learning_rate": 2.703773208867658e-07, "logits/chosen": -1.2812106609344482, "logits/rejected": -1.2686681747436523, "logps/chosen": -289.8981018066406, "logps/rejected": -298.4099426269531, "loss": 0.6562, "rewards/accuracies": 0.65625, "rewards/chosen": 0.6197431087493896, "rewards/margins": 0.27826988697052, "rewards/rejected": 0.341473251581192, "step": 146 }, { "epoch": 0.961038961038961, "grad_norm": 32.51693326041137, "learning_rate": 2.652904152054607e-07, "logits/chosen": -1.2882963418960571, "logits/rejected": -1.234848141670227, "logps/chosen": -212.38536071777344, "logps/rejected": -235.0468292236328, "loss": 0.657, "rewards/accuracies": 0.6875, "rewards/chosen": 0.8348413109779358, "rewards/margins": 0.48668912053108215, "rewards/rejected": 0.34815219044685364, "step": 148 }, { "epoch": 0.974025974025974, "grad_norm": 33.13371475089395, "learning_rate": 2.6019714651539645e-07, "logits/chosen": -1.3393000364303589, "logits/rejected": -1.3905009031295776, "logps/chosen": -223.33636474609375, "logps/rejected": -243.6895294189453, "loss": 0.6381, "rewards/accuracies": 0.65625, "rewards/chosen": 0.5773805379867554, "rewards/margins": 0.328535795211792, "rewards/rejected": 0.24884477257728577, "step": 150 }, { "epoch": 0.987012987012987, "grad_norm": 29.25513766134415, "learning_rate": 2.550996343478514e-07, "logits/chosen": -1.210328459739685, "logits/rejected": -1.1693918704986572, "logps/chosen": -266.2218933105469, "logps/rejected": -318.95904541015625, "loss": 0.6367, "rewards/accuracies": 0.625, "rewards/chosen": 0.5575440526008606, "rewards/margins": 0.3903830647468567, "rewards/rejected": 0.1671610325574875, "step": 152 }, { "epoch": 1.0, "grad_norm": 33.4683934031025, "learning_rate": 2.5e-07, "logits/chosen": -1.2191190719604492, "logits/rejected": -1.215990424156189, "logps/chosen": -267.1896667480469, "logps/rejected": -297.6923828125, "loss": 0.6461, "rewards/accuracies": 0.8125, "rewards/chosen": 0.5746477246284485, "rewards/margins": 0.4244813323020935, "rewards/rejected": 0.15016639232635498, "step": 154 }, { "epoch": 1.0129870129870129, "grad_norm": 22.52614785258628, "learning_rate": 2.449003656521487e-07, "logits/chosen": -1.052141785621643, "logits/rejected": -1.2538810968399048, "logps/chosen": -262.6394958496094, "logps/rejected": -265.5911560058594, "loss": 0.5416, "rewards/accuracies": 0.75, "rewards/chosen": 0.68734210729599, "rewards/margins": 0.7715808749198914, "rewards/rejected": -0.08423884212970734, "step": 156 }, { "epoch": 1.025974025974026, "grad_norm": 17.71525419241131, "learning_rate": 2.3980285348460363e-07, "logits/chosen": -1.2752505540847778, "logits/rejected": -1.297239065170288, "logps/chosen": -180.13455200195312, "logps/rejected": -194.86289978027344, "loss": 0.5164, "rewards/accuracies": 0.75, "rewards/chosen": 0.637496292591095, "rewards/margins": 0.591891884803772, "rewards/rejected": 0.045604437589645386, "step": 158 }, { "epoch": 1.0389610389610389, "grad_norm": 17.624016800680035, "learning_rate": 2.3470958479453937e-07, "logits/chosen": -1.3482556343078613, "logits/rejected": -1.2880206108093262, "logps/chosen": -183.36322021484375, "logps/rejected": -208.04705810546875, "loss": 0.524, "rewards/accuracies": 0.75, "rewards/chosen": 0.6870389580726624, "rewards/margins": 0.8608755469322205, "rewards/rejected": -0.17383655905723572, "step": 160 }, { "epoch": 1.051948051948052, "grad_norm": 16.65756304879653, "learning_rate": 2.296226791132342e-07, "logits/chosen": -1.240028977394104, "logits/rejected": -1.2343027591705322, "logps/chosen": -233.16773986816406, "logps/rejected": -250.71063232421875, "loss": 0.5179, "rewards/accuracies": 0.6875, "rewards/chosen": 0.796766996383667, "rewards/margins": 0.684018611907959, "rewards/rejected": 0.11274835467338562, "step": 162 }, { "epoch": 1.0649350649350648, "grad_norm": 18.048904117373002, "learning_rate": 2.245442533240412e-07, "logits/chosen": -1.2781399488449097, "logits/rejected": -1.2129652500152588, "logps/chosen": -244.93289184570312, "logps/rejected": -284.34075927734375, "loss": 0.5295, "rewards/accuracies": 0.84375, "rewards/chosen": 0.6863990426063538, "rewards/margins": 0.871242344379425, "rewards/rejected": -0.1848432570695877, "step": 164 }, { "epoch": 1.077922077922078, "grad_norm": 20.34696734972406, "learning_rate": 2.1947642078146e-07, "logits/chosen": -1.322028398513794, "logits/rejected": -1.3042186498641968, "logps/chosen": -271.44500732421875, "logps/rejected": -270.650146484375, "loss": 0.5102, "rewards/accuracies": 0.8125, "rewards/chosen": 0.853551983833313, "rewards/margins": 0.9862384796142578, "rewards/rejected": -0.1326863318681717, "step": 166 }, { "epoch": 1.0909090909090908, "grad_norm": 17.55986753162257, "learning_rate": 2.1442129043167873e-07, "logits/chosen": -1.328125, "logits/rejected": -1.3073420524597168, "logps/chosen": -206.6400909423828, "logps/rejected": -220.58033752441406, "loss": 0.5149, "rewards/accuracies": 0.78125, "rewards/chosen": 0.7431928515434265, "rewards/margins": 0.8430217504501343, "rewards/rejected": -0.09982895851135254, "step": 168 }, { "epoch": 1.103896103896104, "grad_norm": 17.856674110170477, "learning_rate": 2.0938096593494853e-07, "logits/chosen": -0.9651075005531311, "logits/rejected": -1.0858749151229858, "logps/chosen": -185.41217041015625, "logps/rejected": -206.42196655273438, "loss": 0.5435, "rewards/accuracies": 0.71875, "rewards/chosen": 0.7554150819778442, "rewards/margins": 0.6504155397415161, "rewards/rejected": 0.10499954968690872, "step": 170 }, { "epoch": 1.1168831168831168, "grad_norm": 19.78444270286199, "learning_rate": 2.043575447901596e-07, "logits/chosen": -1.22752845287323, "logits/rejected": -1.306286334991455, "logps/chosen": -281.8180236816406, "logps/rejected": -299.3645935058594, "loss": 0.5278, "rewards/accuracies": 0.71875, "rewards/chosen": 0.7911855578422546, "rewards/margins": 0.6221764087677002, "rewards/rejected": 0.16900911927223206, "step": 172 }, { "epoch": 1.12987012987013, "grad_norm": 20.24750613090253, "learning_rate": 1.9935311746197976e-07, "logits/chosen": -1.0992745161056519, "logits/rejected": -1.0037593841552734, "logps/chosen": -219.87008666992188, "logps/rejected": -246.12489318847656, "loss": 0.5273, "rewards/accuracies": 0.78125, "rewards/chosen": 0.8513480424880981, "rewards/margins": 0.9342952370643616, "rewards/rejected": -0.08294717967510223, "step": 174 }, { "epoch": 1.1428571428571428, "grad_norm": 19.866822049463956, "learning_rate": 1.9436976651092142e-07, "logits/chosen": -1.3867590427398682, "logits/rejected": -1.2858481407165527, "logps/chosen": -258.7950439453125, "logps/rejected": -296.06842041015625, "loss": 0.5338, "rewards/accuracies": 0.90625, "rewards/chosen": 0.7717873454093933, "rewards/margins": 0.8829994201660156, "rewards/rejected": -0.11121205985546112, "step": 176 }, { "epoch": 1.155844155844156, "grad_norm": 18.093341348498807, "learning_rate": 1.8940956572669692e-07, "logits/chosen": -1.0915521383285522, "logits/rejected": -1.1871168613433838, "logps/chosen": -247.3667755126953, "logps/rejected": -254.37728881835938, "loss": 0.4999, "rewards/accuracies": 0.84375, "rewards/chosen": 0.7344910502433777, "rewards/margins": 0.7577481865882874, "rewards/rejected": -0.023257076740264893, "step": 178 }, { "epoch": 1.1688311688311688, "grad_norm": 19.525075850508752, "learning_rate": 1.8447457926522452e-07, "logits/chosen": -0.9950531721115112, "logits/rejected": -0.994107186794281, "logps/chosen": -259.6279296875, "logps/rejected": -308.6029052734375, "loss": 0.5492, "rewards/accuracies": 0.8125, "rewards/chosen": 0.7692892551422119, "rewards/margins": 0.9317154884338379, "rewards/rejected": -0.16242632269859314, "step": 180 }, { "epoch": 1.1818181818181819, "grad_norm": 19.472127981634152, "learning_rate": 1.7956686078964255e-07, "logits/chosen": -1.3088796138763428, "logits/rejected": -1.4440947771072388, "logps/chosen": -304.2322082519531, "logps/rejected": -302.7728271484375, "loss": 0.5274, "rewards/accuracies": 0.84375, "rewards/chosen": 0.815383791923523, "rewards/margins": 0.9992111921310425, "rewards/rejected": -0.18382729589939117, "step": 182 }, { "epoch": 1.1948051948051948, "grad_norm": 17.961643519105607, "learning_rate": 1.7468845261569147e-07, "logits/chosen": -1.3604259490966797, "logits/rejected": -1.1473455429077148, "logps/chosen": -278.00860595703125, "logps/rejected": -333.7747802734375, "loss": 0.5072, "rewards/accuracies": 0.90625, "rewards/chosen": 0.8540161848068237, "rewards/margins": 0.9145561456680298, "rewards/rejected": -0.060539960861206055, "step": 184 }, { "epoch": 1.2077922077922079, "grad_norm": 17.34162227747004, "learning_rate": 1.698413848618161e-07, "logits/chosen": -1.2001099586486816, "logits/rejected": -1.1247773170471191, "logps/chosen": -219.12808227539062, "logps/rejected": -240.87974548339844, "loss": 0.5124, "rewards/accuracies": 0.8125, "rewards/chosen": 0.792206883430481, "rewards/margins": 0.9953896403312683, "rewards/rejected": -0.20318272709846497, "step": 186 }, { "epoch": 1.2207792207792207, "grad_norm": 18.899208870167122, "learning_rate": 1.6502767460434585e-07, "logits/chosen": -1.1938526630401611, "logits/rejected": -1.2088103294372559, "logps/chosen": -343.2425537109375, "logps/rejected": -352.4480895996094, "loss": 0.5276, "rewards/accuracies": 0.875, "rewards/chosen": 0.7310408353805542, "rewards/margins": 0.948591947555542, "rewards/rejected": -0.21755114197731018, "step": 188 }, { "epoch": 1.2337662337662338, "grad_norm": 21.1592051257712, "learning_rate": 1.602493250381003e-07, "logits/chosen": -1.2564481496810913, "logits/rejected": -1.3384077548980713, "logps/chosen": -211.84091186523438, "logps/rejected": -175.0691375732422, "loss": 0.5227, "rewards/accuracies": 0.75, "rewards/chosen": 0.6941075325012207, "rewards/margins": 0.8241455554962158, "rewards/rejected": -0.13003802299499512, "step": 190 }, { "epoch": 1.2467532467532467, "grad_norm": 17.37996372066179, "learning_rate": 1.555083246427734e-07, "logits/chosen": -1.1990139484405518, "logits/rejected": -1.3094348907470703, "logps/chosen": -227.32859802246094, "logps/rejected": -240.32440185546875, "loss": 0.5063, "rewards/accuracies": 0.8125, "rewards/chosen": 0.8420521020889282, "rewards/margins": 0.7185032367706299, "rewards/rejected": 0.12354880571365356, "step": 192 }, { "epoch": 1.2597402597402598, "grad_norm": 19.348154887947697, "learning_rate": 1.5080664635543932e-07, "logits/chosen": -1.2713487148284912, "logits/rejected": -1.1336112022399902, "logps/chosen": -217.08712768554688, "logps/rejected": -286.9374084472656, "loss": 0.5347, "rewards/accuracies": 0.875, "rewards/chosen": 0.9916350245475769, "rewards/margins": 0.9090840220451355, "rewards/rejected": 0.08255089819431305, "step": 194 }, { "epoch": 1.2727272727272727, "grad_norm": 18.399160019731372, "learning_rate": 1.461462467495284e-07, "logits/chosen": -1.1585338115692139, "logits/rejected": -1.1779534816741943, "logps/chosen": -184.70875549316406, "logps/rejected": -199.60733032226562, "loss": 0.5194, "rewards/accuracies": 0.75, "rewards/chosen": 0.6531062722206116, "rewards/margins": 0.5796053409576416, "rewards/rejected": 0.07350096851587296, "step": 196 }, { "epoch": 1.2857142857142856, "grad_norm": 17.142926149258614, "learning_rate": 1.4152906522061047e-07, "logits/chosen": -1.145853877067566, "logits/rejected": -1.1914734840393066, "logps/chosen": -244.29345703125, "logps/rejected": -252.35826110839844, "loss": 0.5101, "rewards/accuracies": 0.9375, "rewards/chosen": 0.8408298492431641, "rewards/margins": 0.9886577725410461, "rewards/rejected": -0.14782798290252686, "step": 198 }, { "epoch": 1.2987012987012987, "grad_norm": 19.542604645193908, "learning_rate": 1.369570231793286e-07, "logits/chosen": -1.2806802988052368, "logits/rejected": -1.2916889190673828, "logps/chosen": -233.01324462890625, "logps/rejected": -256.65887451171875, "loss": 0.515, "rewards/accuracies": 0.75, "rewards/chosen": 0.8423460721969604, "rewards/margins": 0.8902460336685181, "rewards/rejected": -0.04789996147155762, "step": 200 }, { "epoch": 1.3116883116883118, "grad_norm": 18.81859489288187, "learning_rate": 1.3243202325181578e-07, "logits/chosen": -1.0917601585388184, "logits/rejected": -1.1492509841918945, "logps/chosen": -241.70211791992188, "logps/rejected": -247.6362762451172, "loss": 0.5173, "rewards/accuracies": 0.75, "rewards/chosen": 0.8184989094734192, "rewards/margins": 0.8419989347457886, "rewards/rejected": -0.023499924689531326, "step": 202 }, { "epoch": 1.3246753246753247, "grad_norm": 21.98362763457428, "learning_rate": 1.2795594848792974e-07, "logits/chosen": -1.2968580722808838, "logits/rejected": -1.3090989589691162, "logps/chosen": -209.9777374267578, "logps/rejected": -249.19898986816406, "loss": 0.5144, "rewards/accuracies": 0.78125, "rewards/chosen": 0.6729121208190918, "rewards/margins": 0.8331139087677002, "rewards/rejected": -0.16020171344280243, "step": 204 }, { "epoch": 1.3376623376623376, "grad_norm": 19.163927573547067, "learning_rate": 1.2353066157763304e-07, "logits/chosen": -1.165097713470459, "logits/rejected": -1.0794737339019775, "logps/chosen": -243.25994873046875, "logps/rejected": -296.20458984375, "loss": 0.5074, "rewards/accuracies": 0.75, "rewards/chosen": 0.8912444710731506, "rewards/margins": 1.0708703994750977, "rewards/rejected": -0.17962592840194702, "step": 206 }, { "epoch": 1.3506493506493507, "grad_norm": 19.327223964730898, "learning_rate": 1.1915800407584702e-07, "logits/chosen": -1.2650034427642822, "logits/rejected": -1.260176181793213, "logps/chosen": -297.4647216796875, "logps/rejected": -283.3460998535156, "loss": 0.5184, "rewards/accuracies": 0.6875, "rewards/chosen": 0.607912540435791, "rewards/margins": 0.5631698369979858, "rewards/rejected": 0.04474277421832085, "step": 208 }, { "epoch": 1.3636363636363638, "grad_norm": 17.1533818598118, "learning_rate": 1.1483979563610069e-07, "logits/chosen": -1.3990474939346313, "logits/rejected": -1.1992498636245728, "logps/chosen": -179.99732971191406, "logps/rejected": -254.8299560546875, "loss": 0.5, "rewards/accuracies": 0.78125, "rewards/chosen": 0.6216685771942139, "rewards/margins": 0.651849627494812, "rewards/rejected": -0.03018118068575859, "step": 210 }, { "epoch": 1.3766233766233766, "grad_norm": 18.960132927151744, "learning_rate": 1.1057783325329267e-07, "logits/chosen": -1.1669042110443115, "logits/rejected": -1.221852421760559, "logps/chosen": -292.6320495605469, "logps/rejected": -333.9742736816406, "loss": 0.505, "rewards/accuracies": 0.90625, "rewards/chosen": 0.997809648513794, "rewards/margins": 1.133697748184204, "rewards/rejected": -0.13588815927505493, "step": 212 }, { "epoch": 1.3896103896103895, "grad_norm": 16.02074293153386, "learning_rate": 1.0637389051588425e-07, "logits/chosen": -1.1035025119781494, "logits/rejected": -1.113924503326416, "logps/chosen": -217.35040283203125, "logps/rejected": -243.62997436523438, "loss": 0.5012, "rewards/accuracies": 0.84375, "rewards/chosen": 0.7329657673835754, "rewards/margins": 0.827349066734314, "rewards/rejected": -0.09438329190015793, "step": 214 }, { "epoch": 1.4025974025974026, "grad_norm": 16.742187965244536, "learning_rate": 1.0222971686783089e-07, "logits/chosen": -1.392643690109253, "logits/rejected": -1.4035465717315674, "logps/chosen": -252.4545440673828, "logps/rejected": -249.23336791992188, "loss": 0.5069, "rewards/accuracies": 0.78125, "rewards/chosen": 0.7851105332374573, "rewards/margins": 0.8927165865898132, "rewards/rejected": -0.10760608315467834, "step": 216 }, { "epoch": 1.4155844155844157, "grad_norm": 20.851477095001684, "learning_rate": 9.814703688056319e-08, "logits/chosen": -1.1926751136779785, "logits/rejected": -1.2301442623138428, "logps/chosen": -165.44703674316406, "logps/rejected": -194.3583221435547, "loss": 0.5326, "rewards/accuracies": 0.84375, "rewards/chosen": 0.8975037336349487, "rewards/margins": 0.9634904265403748, "rewards/rejected": -0.065986767411232, "step": 218 }, { "epoch": 1.4285714285714286, "grad_norm": 16.315287900746362, "learning_rate": 9.412754953531663e-08, "logits/chosen": -1.3106112480163574, "logits/rejected": -1.3371086120605469, "logps/chosen": -255.00607299804688, "logps/rejected": -258.26361083984375, "loss": 0.5037, "rewards/accuracies": 0.6875, "rewards/chosen": 0.6847482919692993, "rewards/margins": 0.697127103805542, "rewards/rejected": -0.012378760613501072, "step": 220 }, { "epoch": 1.4415584415584415, "grad_norm": 15.292540532901763, "learning_rate": 9.017292751611219e-08, "logits/chosen": -0.9778515100479126, "logits/rejected": -0.9790754318237305, "logps/chosen": -189.63674926757812, "logps/rejected": -222.9846649169922, "loss": 0.5021, "rewards/accuracies": 0.78125, "rewards/chosen": 0.6293007135391235, "rewards/margins": 0.9083631038665771, "rewards/rejected": -0.27906233072280884, "step": 222 }, { "epoch": 1.4545454545454546, "grad_norm": 17.14845228574236, "learning_rate": 8.628481651367875e-08, "logits/chosen": -1.227216362953186, "logits/rejected": -1.306528091430664, "logps/chosen": -212.0201873779297, "logps/rejected": -204.0389404296875, "loss": 0.5137, "rewards/accuracies": 0.90625, "rewards/chosen": 0.9179770946502686, "rewards/margins": 0.7093281745910645, "rewards/rejected": 0.20864906907081604, "step": 224 }, { "epoch": 1.4675324675324675, "grad_norm": 18.007638315726304, "learning_rate": 8.246483454061015e-08, "logits/chosen": -1.3051916360855103, "logits/rejected": -1.121058702468872, "logps/chosen": -251.83587646484375, "logps/rejected": -286.392822265625, "loss": 0.5128, "rewards/accuracies": 0.96875, "rewards/chosen": 0.7694352269172668, "rewards/margins": 1.2012625932693481, "rewards/rejected": -0.43182751536369324, "step": 226 }, { "epoch": 1.4805194805194806, "grad_norm": 16.89635894695866, "learning_rate": 7.871457125803896e-08, "logits/chosen": -1.3657732009887695, "logits/rejected": -1.4764338731765747, "logps/chosen": -242.94061279296875, "logps/rejected": -254.3301239013672, "loss": 0.4956, "rewards/accuracies": 0.9375, "rewards/chosen": 0.9610437750816345, "rewards/margins": 1.0195040702819824, "rewards/rejected": -0.058460310101509094, "step": 228 }, { "epoch": 1.4935064935064934, "grad_norm": 18.122088719985634, "learning_rate": 7.503558731410958e-08, "logits/chosen": -1.1523171663284302, "logits/rejected": -1.183262586593628, "logps/chosen": -255.1221466064453, "logps/rejected": -274.8419494628906, "loss": 0.5169, "rewards/accuracies": 0.6875, "rewards/chosen": 0.6205594539642334, "rewards/margins": 0.730150580406189, "rewards/rejected": -0.10959106683731079, "step": 230 }, { "epoch": 1.5064935064935066, "grad_norm": 17.706521703256886, "learning_rate": 7.14294136945241e-08, "logits/chosen": -1.226336121559143, "logits/rejected": -1.068296194076538, "logps/chosen": -210.2167510986328, "logps/rejected": -264.81085205078125, "loss": 0.5089, "rewards/accuracies": 0.875, "rewards/chosen": 0.7191243171691895, "rewards/margins": 1.1839776039123535, "rewards/rejected": -0.46485334634780884, "step": 232 }, { "epoch": 1.5194805194805194, "grad_norm": 16.66406313160021, "learning_rate": 6.789755108543274e-08, "logits/chosen": -1.228325605392456, "logits/rejected": -1.3681098222732544, "logps/chosen": -172.350830078125, "logps/rejected": -175.53936767578125, "loss": 0.5051, "rewards/accuracies": 0.75, "rewards/chosen": 0.6207537651062012, "rewards/margins": 0.7215192317962646, "rewards/rejected": -0.10076545178890228, "step": 234 }, { "epoch": 1.5324675324675323, "grad_norm": 18.51693060533818, "learning_rate": 6.444146924893251e-08, "logits/chosen": -1.101006269454956, "logits/rejected": -1.0755804777145386, "logps/chosen": -233.41761779785156, "logps/rejected": -264.3063659667969, "loss": 0.5173, "rewards/accuracies": 0.84375, "rewards/chosen": 0.8273410797119141, "rewards/margins": 1.0633716583251953, "rewards/rejected": -0.23603063821792603, "step": 236 }, { "epoch": 1.5454545454545454, "grad_norm": 18.052624843204704, "learning_rate": 6.106260641143546e-08, "logits/chosen": -1.1633551120758057, "logits/rejected": -1.1994410753250122, "logps/chosen": -269.3063659667969, "logps/rejected": -271.7884216308594, "loss": 0.4853, "rewards/accuracies": 0.84375, "rewards/chosen": 0.6259198188781738, "rewards/margins": 0.9741422533988953, "rewards/rejected": -0.34822240471839905, "step": 238 }, { "epoch": 1.5584415584415585, "grad_norm": 24.12818089831554, "learning_rate": 5.776236866515946e-08, "logits/chosen": -1.0952610969543457, "logits/rejected": -1.2616946697235107, "logps/chosen": -209.5765380859375, "logps/rejected": -185.8994140625, "loss": 0.5069, "rewards/accuracies": 0.8125, "rewards/chosen": 0.6148033738136292, "rewards/margins": 0.7846094369888306, "rewards/rejected": -0.16980606317520142, "step": 240 }, { "epoch": 1.5714285714285714, "grad_norm": 17.25585157151766, "learning_rate": 5.454212938299255e-08, "logits/chosen": -1.220323920249939, "logits/rejected": -1.1233211755752563, "logps/chosen": -210.1920623779297, "logps/rejected": -257.0649719238281, "loss": 0.4983, "rewards/accuracies": 0.71875, "rewards/chosen": 0.8331396579742432, "rewards/margins": 0.8172454833984375, "rewards/rejected": 0.01589421182870865, "step": 242 }, { "epoch": 1.5844155844155843, "grad_norm": 18.694264506215937, "learning_rate": 5.140322864697183e-08, "logits/chosen": -1.2713161706924438, "logits/rejected": -1.1681023836135864, "logps/chosen": -250.68112182617188, "logps/rejected": -248.07406616210938, "loss": 0.5152, "rewards/accuracies": 0.90625, "rewards/chosen": 0.8290749192237854, "rewards/margins": 0.7929165363311768, "rewards/rejected": 0.036158446222543716, "step": 244 }, { "epoch": 1.5974025974025974, "grad_norm": 18.951231695832167, "learning_rate": 4.8346972690617494e-08, "logits/chosen": -1.2086520195007324, "logits/rejected": -1.2194980382919312, "logps/chosen": -237.5225067138672, "logps/rejected": -294.5182800292969, "loss": 0.4988, "rewards/accuracies": 0.75, "rewards/chosen": 0.823986291885376, "rewards/margins": 1.3896714448928833, "rewards/rejected": -0.5656850934028625, "step": 246 }, { "epoch": 1.6103896103896105, "grad_norm": 17.838639398070846, "learning_rate": 4.53746333553516e-08, "logits/chosen": -1.3609139919281006, "logits/rejected": -1.3907623291015625, "logps/chosen": -254.64718627929688, "logps/rejected": -267.22265625, "loss": 0.499, "rewards/accuracies": 0.6875, "rewards/chosen": 0.7264419794082642, "rewards/margins": 0.9065679311752319, "rewards/rejected": -0.18012598156929016, "step": 248 }, { "epoch": 1.6233766233766234, "grad_norm": 15.282492364953745, "learning_rate": 4.248744756122985e-08, "logits/chosen": -1.4239810705184937, "logits/rejected": -1.3679332733154297, "logps/chosen": -231.71435546875, "logps/rejected": -287.30572509765625, "loss": 0.5073, "rewards/accuracies": 0.78125, "rewards/chosen": 0.5901214480400085, "rewards/margins": 0.7662074565887451, "rewards/rejected": -0.17608599364757538, "step": 250 }, { "epoch": 1.6363636363636362, "grad_norm": 18.90237932042577, "learning_rate": 3.968661679220467e-08, "logits/chosen": -1.0735975503921509, "logits/rejected": -1.167972207069397, "logps/chosen": -188.3458709716797, "logps/rejected": -179.90293884277344, "loss": 0.5256, "rewards/accuracies": 0.78125, "rewards/chosen": 0.6942843794822693, "rewards/margins": 0.7579864859580994, "rewards/rejected": -0.06370209902524948, "step": 252 }, { "epoch": 1.6493506493506493, "grad_norm": 17.45584095837454, "learning_rate": 3.6973306596135873e-08, "logits/chosen": -1.3392971754074097, "logits/rejected": -1.358009696006775, "logps/chosen": -254.31295776367188, "logps/rejected": -254.20733642578125, "loss": 0.502, "rewards/accuracies": 0.84375, "rewards/chosen": 0.7509973645210266, "rewards/margins": 0.8210946917533875, "rewards/rejected": -0.07009733468294144, "step": 254 }, { "epoch": 1.6623376623376624, "grad_norm": 19.754012599719967, "learning_rate": 3.43486460997548e-08, "logits/chosen": -1.2106963396072388, "logits/rejected": -1.2819626331329346, "logps/chosen": -253.831787109375, "logps/rejected": -303.4576721191406, "loss": 0.5023, "rewards/accuracies": 0.75, "rewards/chosen": 0.732483446598053, "rewards/margins": 1.113276720046997, "rewards/rejected": -0.38079336285591125, "step": 256 }, { "epoch": 1.6753246753246753, "grad_norm": 18.3258569948234, "learning_rate": 3.1813727538785943e-08, "logits/chosen": -1.3981492519378662, "logits/rejected": -1.4056671857833862, "logps/chosen": -356.7002868652344, "logps/rejected": -362.1513366699219, "loss": 0.494, "rewards/accuracies": 0.78125, "rewards/chosen": 0.6453557014465332, "rewards/margins": 1.1145414113998413, "rewards/rejected": -0.46918565034866333, "step": 258 }, { "epoch": 1.6883116883116882, "grad_norm": 19.670963229677074, "learning_rate": 2.936960580341971e-08, "logits/chosen": -1.1725897789001465, "logits/rejected": -1.261063575744629, "logps/chosen": -216.58990478515625, "logps/rejected": -220.63986206054688, "loss": 0.5231, "rewards/accuracies": 0.75, "rewards/chosen": 0.7476328015327454, "rewards/margins": 0.7463090419769287, "rewards/rejected": 0.0013237213715910912, "step": 260 }, { "epoch": 1.7012987012987013, "grad_norm": 17.983233221990766, "learning_rate": 2.701729799932653e-08, "logits/chosen": -1.3052499294281006, "logits/rejected": -1.333134412765503, "logps/chosen": -277.6193542480469, "logps/rejected": -287.4527893066406, "loss": 0.4954, "rewards/accuracies": 0.8125, "rewards/chosen": 0.726482629776001, "rewards/margins": 0.9591711759567261, "rewards/rejected": -0.23268845677375793, "step": 262 }, { "epoch": 1.7142857142857144, "grad_norm": 17.467592052173433, "learning_rate": 2.475778302439524e-08, "logits/chosen": -1.2939479351043701, "logits/rejected": -1.3279248476028442, "logps/chosen": -207.76170349121094, "logps/rejected": -233.1382598876953, "loss": 0.472, "rewards/accuracies": 0.8125, "rewards/chosen": 0.8571805953979492, "rewards/margins": 0.9823651313781738, "rewards/rejected": -0.12518461048603058, "step": 264 }, { "epoch": 1.7272727272727273, "grad_norm": 16.46335190300962, "learning_rate": 2.259200116137039e-08, "logits/chosen": -1.1292874813079834, "logits/rejected": -1.2274911403656006, "logps/chosen": -277.028564453125, "logps/rejected": -308.0982971191406, "loss": 0.4976, "rewards/accuracies": 0.8125, "rewards/chosen": 0.9631680846214294, "rewards/margins": 1.1252551078796387, "rewards/rejected": -0.16208705306053162, "step": 266 }, { "epoch": 1.7402597402597402, "grad_norm": 18.56546944179356, "learning_rate": 2.0520853686560175e-08, "logits/chosen": -1.2864351272583008, "logits/rejected": -1.2748289108276367, "logps/chosen": -203.02589416503906, "logps/rejected": -217.65682983398438, "loss": 0.5072, "rewards/accuracies": 0.8125, "rewards/chosen": 0.5683714747428894, "rewards/margins": 0.9107145667076111, "rewards/rejected": -0.3423430323600769, "step": 268 }, { "epoch": 1.7532467532467533, "grad_norm": 16.59250739344183, "learning_rate": 1.854520249477551e-08, "logits/chosen": -1.3317527770996094, "logits/rejected": -1.3174333572387695, "logps/chosen": -205.539306640625, "logps/rejected": -236.86062622070312, "loss": 0.487, "rewards/accuracies": 0.71875, "rewards/chosen": 0.8909409642219543, "rewards/margins": 1.0206615924835205, "rewards/rejected": -0.1297205686569214, "step": 270 }, { "epoch": 1.7662337662337664, "grad_norm": 18.312480737747663, "learning_rate": 1.666586974065831e-08, "logits/chosen": -1.1755964756011963, "logits/rejected": -1.2511848211288452, "logps/chosen": -244.85443115234375, "logps/rejected": -236.78126525878906, "loss": 0.5334, "rewards/accuracies": 0.71875, "rewards/chosen": 0.5452998876571655, "rewards/margins": 0.6891751885414124, "rewards/rejected": -0.14387527108192444, "step": 272 }, { "epoch": 1.7792207792207793, "grad_norm": 18.481484735169893, "learning_rate": 1.4883637496547141e-08, "logits/chosen": -1.2515636682510376, "logits/rejected": -1.2902557849884033, "logps/chosen": -254.2708740234375, "logps/rejected": -258.9367980957031, "loss": 0.5094, "rewards/accuracies": 0.90625, "rewards/chosen": 0.7975621223449707, "rewards/margins": 1.1254695653915405, "rewards/rejected": -0.327907532453537, "step": 274 }, { "epoch": 1.7922077922077921, "grad_norm": 18.685754884224313, "learning_rate": 1.3199247427022526e-08, "logits/chosen": -1.2295678853988647, "logits/rejected": -1.2680904865264893, "logps/chosen": -196.48304748535156, "logps/rejected": -207.93121337890625, "loss": 0.5151, "rewards/accuracies": 0.71875, "rewards/chosen": 0.8263696432113647, "rewards/margins": 0.9237637519836426, "rewards/rejected": -0.09739402681589127, "step": 276 }, { "epoch": 1.8051948051948052, "grad_norm": 18.498386316775253, "learning_rate": 1.16134004802681e-08, "logits/chosen": -1.0846458673477173, "logits/rejected": -0.9547269940376282, "logps/chosen": -231.32818603515625, "logps/rejected": -263.1034851074219, "loss": 0.5058, "rewards/accuracies": 0.84375, "rewards/chosen": 0.6925534605979919, "rewards/margins": 1.1582170724868774, "rewards/rejected": -0.4656636714935303, "step": 278 }, { "epoch": 1.8181818181818183, "grad_norm": 17.340516668358195, "learning_rate": 1.0126756596375685e-08, "logits/chosen": -1.3053959608078003, "logits/rejected": -1.2687203884124756, "logps/chosen": -206.0025177001953, "logps/rejected": -235.99029541015625, "loss": 0.499, "rewards/accuracies": 0.8125, "rewards/chosen": 0.8264460563659668, "rewards/margins": 0.9402331113815308, "rewards/rejected": -0.11378702521324158, "step": 280 }, { "epoch": 1.8311688311688312, "grad_norm": 20.777883031718197, "learning_rate": 8.739934432715034e-09, "logits/chosen": -1.2384371757507324, "logits/rejected": -1.3831499814987183, "logps/chosen": -226.1611785888672, "logps/rejected": -219.2269287109375, "loss": 0.5138, "rewards/accuracies": 0.78125, "rewards/chosen": 0.514397382736206, "rewards/margins": 0.7629111409187317, "rewards/rejected": -0.24851378798484802, "step": 282 }, { "epoch": 1.844155844155844, "grad_norm": 18.170821830096596, "learning_rate": 7.453511106483901e-09, "logits/chosen": -1.4039347171783447, "logits/rejected": -1.4059028625488281, "logps/chosen": -269.5606689453125, "logps/rejected": -254.4373779296875, "loss": 0.5001, "rewards/accuracies": 0.90625, "rewards/chosen": 0.7399085164070129, "rewards/margins": 1.1216527223587036, "rewards/rejected": -0.3817441165447235, "step": 284 }, { "epoch": 1.8571428571428572, "grad_norm": 21.929296495653098, "learning_rate": 6.268021954544095e-09, "logits/chosen": -1.2451683282852173, "logits/rejected": -1.2341443300247192, "logps/chosen": -249.87173461914062, "logps/rejected": -256.73297119140625, "loss": 0.5052, "rewards/accuracies": 0.75, "rewards/chosen": 0.665867805480957, "rewards/margins": 0.8860350251197815, "rewards/rejected": -0.22016723453998566, "step": 286 }, { "epoch": 1.87012987012987, "grad_norm": 17.66627473591946, "learning_rate": 5.1839603106447475e-09, "logits/chosen": -1.1848328113555908, "logits/rejected": -1.2601234912872314, "logps/chosen": -274.2748718261719, "logps/rejected": -276.2279968261719, "loss": 0.5143, "rewards/accuracies": 0.6875, "rewards/chosen": 0.40582898259162903, "rewards/margins": 0.7143059372901917, "rewards/rejected": -0.3084769546985626, "step": 288 }, { "epoch": 1.883116883116883, "grad_norm": 17.62107555046559, "learning_rate": 4.201777300124249e-09, "logits/chosen": -1.3278290033340454, "logits/rejected": -1.412712574005127, "logps/chosen": -232.41680908203125, "logps/rejected": -261.15911865234375, "loss": 0.4936, "rewards/accuracies": 0.875, "rewards/chosen": 0.6634901165962219, "rewards/margins": 1.093658685684204, "rewards/rejected": -0.43016865849494934, "step": 290 }, { "epoch": 1.896103896103896, "grad_norm": 15.858708451261625, "learning_rate": 3.3218816521777827e-09, "logits/chosen": -1.1245518922805786, "logits/rejected": -1.1566898822784424, "logps/chosen": -204.53695678710938, "logps/rejected": -195.2686309814453, "loss": 0.5017, "rewards/accuracies": 0.8125, "rewards/chosen": 0.6790772080421448, "rewards/margins": 0.7251225709915161, "rewards/rejected": -0.04604536294937134, "step": 292 }, { "epoch": 1.9090909090909092, "grad_norm": 18.1779116105005, "learning_rate": 2.5446395297668287e-09, "logits/chosen": -1.2490640878677368, "logits/rejected": -1.3185060024261475, "logps/chosen": -197.9644775390625, "logps/rejected": -211.1698455810547, "loss": 0.503, "rewards/accuracies": 0.875, "rewards/chosen": 0.6239749789237976, "rewards/margins": 1.010934591293335, "rewards/rejected": -0.38695967197418213, "step": 294 }, { "epoch": 1.922077922077922, "grad_norm": 18.8639806757367, "learning_rate": 1.870374377243078e-09, "logits/chosen": -1.2287623882293701, "logits/rejected": -1.214994192123413, "logps/chosen": -183.43939208984375, "logps/rejected": -200.90785217285156, "loss": 0.5078, "rewards/accuracies": 0.71875, "rewards/chosen": 0.3988313674926758, "rewards/margins": 0.920490026473999, "rewards/rejected": -0.5216587781906128, "step": 296 }, { "epoch": 1.935064935064935, "grad_norm": 19.016439869677555, "learning_rate": 1.2993667857489898e-09, "logits/chosen": -1.365774393081665, "logits/rejected": -1.4144957065582275, "logps/chosen": -299.92120361328125, "logps/rejected": -313.76287841796875, "loss": 0.492, "rewards/accuracies": 0.875, "rewards/chosen": 0.790663480758667, "rewards/margins": 1.3730117082595825, "rewards/rejected": -0.5823482275009155, "step": 298 }, { "epoch": 1.948051948051948, "grad_norm": 14.859752452370639, "learning_rate": 8.318543764516961e-10, "logits/chosen": -1.3680145740509033, "logits/rejected": -1.3740514516830444, "logps/chosen": -264.17486572265625, "logps/rejected": -253.31149291992188, "loss": 0.5, "rewards/accuracies": 0.90625, "rewards/chosen": 0.784439206123352, "rewards/margins": 1.083054542541504, "rewards/rejected": -0.29861530661582947, "step": 300 }, { "epoch": 1.9610389610389611, "grad_norm": 18.397492445789904, "learning_rate": 4.680317016582669e-10, "logits/chosen": -1.3651466369628906, "logits/rejected": -1.2224037647247314, "logps/chosen": -203.9071502685547, "logps/rejected": -270.9750061035156, "loss": 0.4975, "rewards/accuracies": 0.78125, "rewards/chosen": 0.7834948301315308, "rewards/margins": 0.9753363132476807, "rewards/rejected": -0.1918414682149887, "step": 302 }, { "epoch": 1.974025974025974, "grad_norm": 17.010218612769055, "learning_rate": 2.0805016385427865e-10, "logits/chosen": -1.2519904375076294, "logits/rejected": -1.3218713998794556, "logps/chosen": -171.00653076171875, "logps/rejected": -186.12115478515625, "loss": 0.5156, "rewards/accuracies": 0.8125, "rewards/chosen": 0.7086281776428223, "rewards/margins": 0.9142274856567383, "rewards/rejected": -0.20559939742088318, "step": 304 }, { "epoch": 1.987012987012987, "grad_norm": 20.168300406467708, "learning_rate": 5.2017952698379943e-11, "logits/chosen": -1.107275128364563, "logits/rejected": -1.237258791923523, "logps/chosen": -269.6531982421875, "logps/rejected": -244.50665283203125, "loss": 0.5078, "rewards/accuracies": 0.90625, "rewards/chosen": 0.7939411997795105, "rewards/margins": 1.0421514511108398, "rewards/rejected": -0.24821028113365173, "step": 306 }, { "epoch": 2.0, "grad_norm": 19.11260041379628, "learning_rate": 0.0, "logits/chosen": -1.115605115890503, "logits/rejected": -1.0857875347137451, "logps/chosen": -228.82167053222656, "logps/rejected": -256.9181823730469, "loss": 0.5006, "rewards/accuracies": 0.875, "rewards/chosen": 0.7351489067077637, "rewards/margins": 0.7865853905677795, "rewards/rejected": -0.051436468958854675, "step": 308 } ], "logging_steps": 2, "max_steps": 308, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }