| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9998389434691577, |
| "eval_steps": 500, |
| "global_step": 4656, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.010737102056155044, |
| "grad_norm": 59.99636459350586, |
| "learning_rate": 3.5714285714285718e-06, |
| "loss": 12.7755, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.02147420411231009, |
| "grad_norm": 47.23111343383789, |
| "learning_rate": 7.1428571428571436e-06, |
| "loss": 7.8117, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.03221130616846513, |
| "grad_norm": 47.798980712890625, |
| "learning_rate": 9.999879015387978e-06, |
| "loss": 7.1417, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.04294840822462018, |
| "grad_norm": 37.97929763793945, |
| "learning_rate": 9.995645168701038e-06, |
| "loss": 5.285, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.05368551028077522, |
| "grad_norm": 46.82593536376953, |
| "learning_rate": 9.98536794504998e-06, |
| "loss": 5.5196, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.06442261233693027, |
| "grad_norm": 47.67216873168945, |
| "learning_rate": 9.969059777090564e-06, |
| "loss": 5.5595, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.0751597143930853, |
| "grad_norm": 47.013511657714844, |
| "learning_rate": 9.946740393286928e-06, |
| "loss": 5.0251, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.08589681644924035, |
| "grad_norm": 24.63237190246582, |
| "learning_rate": 9.918436794045507e-06, |
| "loss": 4.7625, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.09663391850539539, |
| "grad_norm": 22.05866813659668, |
| "learning_rate": 9.884183219051837e-06, |
| "loss": 5.1414, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.10737102056155044, |
| "grad_norm": 21.014677047729492, |
| "learning_rate": 9.844021105849837e-06, |
| "loss": 3.8645, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.11810812261770548, |
| "grad_norm": 43.38600540161133, |
| "learning_rate": 9.797999039713586e-06, |
| "loss": 4.6035, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.12884522467386053, |
| "grad_norm": 22.228700637817383, |
| "learning_rate": 9.746172694872332e-06, |
| "loss": 4.1814, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.13958232673001558, |
| "grad_norm": 23.762821197509766, |
| "learning_rate": 9.688604767159736e-06, |
| "loss": 4.6889, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.1503194287861706, |
| "grad_norm": 20.935850143432617, |
| "learning_rate": 9.62536489816892e-06, |
| "loss": 4.8945, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.16105653084232566, |
| "grad_norm": 19.82425880432129, |
| "learning_rate": 9.556529591005001e-06, |
| "loss": 4.494, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.1717936328984807, |
| "grad_norm": 22.668556213378906, |
| "learning_rate": 9.482182117737066e-06, |
| "loss": 4.4041, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.18253073495463573, |
| "grad_norm": 22.826379776000977, |
| "learning_rate": 9.402412418661541e-06, |
| "loss": 4.8089, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.19326783701079078, |
| "grad_norm": 19.665138244628906, |
| "learning_rate": 9.317316993498788e-06, |
| "loss": 4.232, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.20400493906694583, |
| "grad_norm": 21.231645584106445, |
| "learning_rate": 9.226998784654606e-06, |
| "loss": 3.7319, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.21474204112310089, |
| "grad_norm": 37.30881118774414, |
| "learning_rate": 9.131567052687811e-06, |
| "loss": 4.4938, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.2254791431792559, |
| "grad_norm": 23.012168884277344, |
| "learning_rate": 9.03113724413456e-06, |
| "loss": 4.9744, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.23621624523541096, |
| "grad_norm": 25.494842529296875, |
| "learning_rate": 8.925830851849338e-06, |
| "loss": 3.9475, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.246953347291566, |
| "grad_norm": 21.22249412536621, |
| "learning_rate": 8.815775268031514e-06, |
| "loss": 4.3448, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.25769044934772106, |
| "grad_norm": 40.145023345947266, |
| "learning_rate": 8.701103630115303e-06, |
| "loss": 4.4259, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.2684275514038761, |
| "grad_norm": 17.972578048706055, |
| "learning_rate": 8.581954659709549e-06, |
| "loss": 4.7292, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.27916465346003116, |
| "grad_norm": 19.54686737060547, |
| "learning_rate": 8.458472494782169e-06, |
| "loss": 4.4949, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.2899017555161862, |
| "grad_norm": 39.886653900146484, |
| "learning_rate": 8.330806515292271e-06, |
| "loss": 4.1289, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.3006388575723412, |
| "grad_norm": 19.24032974243164, |
| "learning_rate": 8.199111162480871e-06, |
| "loss": 4.6771, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.3113759596284963, |
| "grad_norm": 18.134519577026367, |
| "learning_rate": 8.063545752038854e-06, |
| "loss": 3.7991, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.3221130616846513, |
| "grad_norm": 25.192407608032227, |
| "learning_rate": 7.924274281378153e-06, |
| "loss": 3.6166, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.33285016374080634, |
| "grad_norm": 22.86334800720215, |
| "learning_rate": 7.781465231239318e-06, |
| "loss": 5.0183, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.3435872657969614, |
| "grad_norm": 17.691837310791016, |
| "learning_rate": 7.635291361875474e-06, |
| "loss": 4.2544, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.35432436785311644, |
| "grad_norm": 21.726091384887695, |
| "learning_rate": 7.485929504059234e-06, |
| "loss": 4.2973, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.36506146990927146, |
| "grad_norm": 17.235157012939453, |
| "learning_rate": 7.333560345165371e-06, |
| "loss": 3.714, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.37579857196542654, |
| "grad_norm": 16.26618194580078, |
| "learning_rate": 7.178368210588067e-06, |
| "loss": 3.5892, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.38653567402158157, |
| "grad_norm": 14.692360877990723, |
| "learning_rate": 7.020540840757124e-06, |
| "loss": 4.577, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.39727277607773664, |
| "grad_norm": 20.708250045776367, |
| "learning_rate": 6.860269164022921e-06, |
| "loss": 4.4776, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.40800987813389167, |
| "grad_norm": 17.772565841674805, |
| "learning_rate": 6.697747065684851e-06, |
| "loss": 4.6948, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.4187469801900467, |
| "grad_norm": 17.883790969848633, |
| "learning_rate": 6.5331711534426326e-06, |
| "loss": 4.0064, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.42948408224620177, |
| "grad_norm": 20.181188583374023, |
| "learning_rate": 6.366740519554286e-06, |
| "loss": 3.9072, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.4402211843023568, |
| "grad_norm": 37.15888214111328, |
| "learning_rate": 6.198656499988444e-06, |
| "loss": 4.2209, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.4509582863585118, |
| "grad_norm": 17.573535919189453, |
| "learning_rate": 6.029122430862373e-06, |
| "loss": 4.8097, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.4616953884146669, |
| "grad_norm": 38.13616943359375, |
| "learning_rate": 5.858343402460391e-06, |
| "loss": 4.0292, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.4724324904708219, |
| "grad_norm": 17.187131881713867, |
| "learning_rate": 5.68652601113019e-06, |
| "loss": 3.7332, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.48316959252697694, |
| "grad_norm": 35.56867599487305, |
| "learning_rate": 5.513878109357228e-06, |
| "loss": 4.2786, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.493906694583132, |
| "grad_norm": 18.53168487548828, |
| "learning_rate": 5.3406085543195555e-06, |
| "loss": 4.8574, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.504643796639287, |
| "grad_norm": 20.49349594116211, |
| "learning_rate": 5.166926955227224e-06, |
| "loss": 4.9073, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.5153808986954421, |
| "grad_norm": 19.841901779174805, |
| "learning_rate": 4.993043419751933e-06, |
| "loss": 3.827, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.5261180007515971, |
| "grad_norm": 19.35201644897461, |
| "learning_rate": 4.8191682998536905e-06, |
| "loss": 3.4893, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.5368551028077522, |
| "grad_norm": 21.604345321655273, |
| "learning_rate": 4.645511937311934e-06, |
| "loss": 4.367, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.5475922048639073, |
| "grad_norm": 40.630821228027344, |
| "learning_rate": 4.472284409268976e-06, |
| "loss": 5.0816, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.5583293069200623, |
| "grad_norm": 18.870628356933594, |
| "learning_rate": 4.299695274093593e-06, |
| "loss": 4.8803, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.5690664089762173, |
| "grad_norm": 21.05728530883789, |
| "learning_rate": 4.1279533178721755e-06, |
| "loss": 4.6022, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.5798035110323724, |
| "grad_norm": 39.68675231933594, |
| "learning_rate": 3.957266301834145e-06, |
| "loss": 4.212, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.5905406130885275, |
| "grad_norm": 17.26553726196289, |
| "learning_rate": 3.7878407110171646e-06, |
| "loss": 4.0448, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.6012777151446824, |
| "grad_norm": 21.71043586730957, |
| "learning_rate": 3.6198815044761847e-06, |
| "loss": 4.6691, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.6120148172008375, |
| "grad_norm": 21.753311157226562, |
| "learning_rate": 3.4535918673385456e-06, |
| "loss": 4.3453, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.6227519192569926, |
| "grad_norm": 18.6822452545166, |
| "learning_rate": 3.2891729650050096e-06, |
| "loss": 4.2042, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.6334890213131475, |
| "grad_norm": 18.38760757446289, |
| "learning_rate": 3.1268236997941535e-06, |
| "loss": 3.8025, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.6442261233693026, |
| "grad_norm": 39.5139045715332, |
| "learning_rate": 2.966740470324451e-06, |
| "loss": 5.1229, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.6549632254254577, |
| "grad_norm": 20.45186424255371, |
| "learning_rate": 2.8091169339251644e-06, |
| "loss": 4.0329, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.6657003274816127, |
| "grad_norm": 18.005401611328125, |
| "learning_rate": 2.654143772363455e-06, |
| "loss": 4.0975, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.6764374295377678, |
| "grad_norm": 21.24009895324707, |
| "learning_rate": 2.502008461171114e-06, |
| "loss": 3.9577, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.6871745315939228, |
| "grad_norm": 18.686386108398438, |
| "learning_rate": 2.352895042849965e-06, |
| "loss": 3.9965, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.6979116336500778, |
| "grad_norm": 24.032150268554688, |
| "learning_rate": 2.20698390423032e-06, |
| "loss": 4.6225, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.7086487357062329, |
| "grad_norm": 18.239240646362305, |
| "learning_rate": 2.0644515582517803e-06, |
| "loss": 4.2193, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.719385837762388, |
| "grad_norm": 21.52141571044922, |
| "learning_rate": 1.9254704304304174e-06, |
| "loss": 4.5206, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.7301229398185429, |
| "grad_norm": 20.970258712768555, |
| "learning_rate": 1.7902086502706256e-06, |
| "loss": 3.7111, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.740860041874698, |
| "grad_norm": 17.317781448364258, |
| "learning_rate": 1.658829847873965e-06, |
| "loss": 4.1838, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.7515971439308531, |
| "grad_norm": 19.438854217529297, |
| "learning_rate": 1.5314929559910985e-06, |
| "loss": 4.3512, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.762334245987008, |
| "grad_norm": 26.218725204467773, |
| "learning_rate": 1.4083520177562154e-06, |
| "loss": 3.1841, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.7730713480431631, |
| "grad_norm": 26.113292694091797, |
| "learning_rate": 1.2895560003365837e-06, |
| "loss": 4.6759, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.7838084500993182, |
| "grad_norm": 20.18511962890625, |
| "learning_rate": 1.1752486147226505e-06, |
| "loss": 3.9116, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.7945455521554733, |
| "grad_norm": 21.312776565551758, |
| "learning_rate": 1.0655681418766772e-06, |
| "loss": 4.0108, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.8052826542116283, |
| "grad_norm": 23.473669052124023, |
| "learning_rate": 9.60647265450249e-07, |
| "loss": 3.5384, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.8160197562677833, |
| "grad_norm": 34.65909957885742, |
| "learning_rate": 8.60612911273011e-07, |
| "loss": 4.4025, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.8267568583239384, |
| "grad_norm": 18.83558464050293, |
| "learning_rate": 7.655860938068071e-07, |
| "loss": 3.2632, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.8374939603800934, |
| "grad_norm": 21.728172302246094, |
| "learning_rate": 6.756817697509755e-07, |
| "loss": 4.0358, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.8482310624362485, |
| "grad_norm": 18.24883460998535, |
| "learning_rate": 5.910086989758862e-07, |
| "loss": 3.4722, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.8589681644924035, |
| "grad_norm": 16.885324478149414, |
| "learning_rate": 5.11669312952977e-07, |
| "loss": 4.0926, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.8697052665485585, |
| "grad_norm": 16.516597747802734, |
| "learning_rate": 4.377595908404225e-07, |
| "loss": 3.9504, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.8804423686047136, |
| "grad_norm": 17.790292739868164, |
| "learning_rate": 3.693689433743658e-07, |
| "loss": 4.0759, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.8911794706608687, |
| "grad_norm": 18.888835906982422, |
| "learning_rate": 3.065801047061517e-07, |
| "loss": 3.3214, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.9019165727170236, |
| "grad_norm": 36.61859893798828, |
| "learning_rate": 2.4946903231642727e-07, |
| "loss": 4.5779, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.9126536747731787, |
| "grad_norm": 21.771671295166016, |
| "learning_rate": 1.9810481512716638e-07, |
| "loss": 3.8909, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.9233907768293338, |
| "grad_norm": 19.28317642211914, |
| "learning_rate": 1.5254958992280022e-07, |
| "loss": 3.9403, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.9341278788854888, |
| "grad_norm": 16.99443817138672, |
| "learning_rate": 1.128584661815435e-07, |
| "loss": 3.7742, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.9448649809416438, |
| "grad_norm": 20.192611694335938, |
| "learning_rate": 7.907945940786033e-08, |
| "loss": 3.7403, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.9556020829977989, |
| "grad_norm": 18.506389617919922, |
| "learning_rate": 5.125343304671459e-08, |
| "loss": 5.146, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.9663391850539539, |
| "grad_norm": 20.306739807128906, |
| "learning_rate": 2.9414049049872883e-08, |
| "loss": 3.5826, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.977076287110109, |
| "grad_norm": 22.654619216918945, |
| "learning_rate": 1.35877271540652e-08, |
| "loss": 4.2733, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.987813389166264, |
| "grad_norm": 18.207374572753906, |
| "learning_rate": 3.7936129202648106e-09, |
| "loss": 3.4258, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.9985504912224191, |
| "grad_norm": 25.076688766479492, |
| "learning_rate": 4.3554572743409463e-11, |
| "loss": 4.9966, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.9998389434691577, |
| "step": 4656, |
| "total_flos": 3.3921983034765083e+19, |
| "train_loss": 4.446265889606934, |
| "train_runtime": 69353.2916, |
| "train_samples_per_second": 4.297, |
| "train_steps_per_second": 0.067 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 4656, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.3921983034765083e+19, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|