{ "best_global_step": 234, "best_metric": 0.3496658504009247, "best_model_checkpoint": "saves/test/checkpoint-234", "epoch": 1.0, "eval_steps": 13, "global_step": 249, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.020080321285140562, "grad_norm": 30.13371467590332, "learning_rate": 8.000000000000001e-06, "loss": 0.8454, "num_input_tokens_seen": 832, "step": 5 }, { "epoch": 0.040160642570281124, "grad_norm": 20.238059997558594, "learning_rate": 1.8e-05, "loss": 0.7689, "num_input_tokens_seen": 1760, "step": 10 }, { "epoch": 0.05220883534136546, "eval_loss": 0.6837943196296692, "eval_runtime": 0.9982, "eval_samples_per_second": 56.101, "eval_steps_per_second": 28.051, "num_input_tokens_seen": 2288, "step": 13 }, { "epoch": 0.060240963855421686, "grad_norm": 9.776243209838867, "learning_rate": 2.8000000000000003e-05, "loss": 0.6956, "num_input_tokens_seen": 2608, "step": 15 }, { "epoch": 0.08032128514056225, "grad_norm": 17.104515075683594, "learning_rate": 3.8e-05, "loss": 0.5552, "num_input_tokens_seen": 3536, "step": 20 }, { "epoch": 0.10040160642570281, "grad_norm": 11.06977653503418, "learning_rate": 4.8e-05, "loss": 0.6557, "num_input_tokens_seen": 4496, "step": 25 }, { "epoch": 0.10441767068273092, "eval_loss": 0.46041956543922424, "eval_runtime": 0.7224, "eval_samples_per_second": 77.516, "eval_steps_per_second": 38.758, "num_input_tokens_seen": 4656, "step": 26 }, { "epoch": 0.12048192771084337, "grad_norm": 8.536067008972168, "learning_rate": 4.996067037544542e-05, "loss": 0.4952, "num_input_tokens_seen": 5424, "step": 30 }, { "epoch": 0.14056224899598393, "grad_norm": 7.301750659942627, "learning_rate": 4.980110583549062e-05, "loss": 0.3647, "num_input_tokens_seen": 6304, "step": 35 }, { "epoch": 0.1566265060240964, "eval_loss": 0.3834764063358307, "eval_runtime": 0.753, "eval_samples_per_second": 74.371, "eval_steps_per_second": 37.185, "num_input_tokens_seen": 6944, "step": 39 }, { "epoch": 0.1606425702811245, "grad_norm": 2.0650792121887207, "learning_rate": 4.951963201008076e-05, "loss": 0.4023, "num_input_tokens_seen": 7072, "step": 40 }, { "epoch": 0.18072289156626506, "grad_norm": 6.442427158355713, "learning_rate": 4.91176324775594e-05, "loss": 0.3811, "num_input_tokens_seen": 7856, "step": 45 }, { "epoch": 0.20080321285140562, "grad_norm": 9.193267822265625, "learning_rate": 4.8597083257709194e-05, "loss": 0.3506, "num_input_tokens_seen": 8880, "step": 50 }, { "epoch": 0.20883534136546184, "eval_loss": 0.38357535004615784, "eval_runtime": 0.8575, "eval_samples_per_second": 65.305, "eval_steps_per_second": 32.653, "num_input_tokens_seen": 9232, "step": 52 }, { "epoch": 0.22088353413654618, "grad_norm": 2.6214115619659424, "learning_rate": 4.796054309867053e-05, "loss": 0.4133, "num_input_tokens_seen": 9680, "step": 55 }, { "epoch": 0.24096385542168675, "grad_norm": 6.665003776550293, "learning_rate": 4.721114089947181e-05, "loss": 0.3351, "num_input_tokens_seen": 10576, "step": 60 }, { "epoch": 0.26104417670682734, "grad_norm": 2.606006383895874, "learning_rate": 4.6352560329995686e-05, "loss": 0.3084, "num_input_tokens_seen": 11424, "step": 65 }, { "epoch": 0.26104417670682734, "eval_loss": 0.36912769079208374, "eval_runtime": 0.9476, "eval_samples_per_second": 59.094, "eval_steps_per_second": 29.547, "num_input_tokens_seen": 11424, "step": 65 }, { "epoch": 0.28112449799196787, "grad_norm": 2.763232946395874, "learning_rate": 4.538902172398151e-05, "loss": 0.374, "num_input_tokens_seen": 12224, "step": 70 }, { "epoch": 0.30120481927710846, "grad_norm": 1.5414751768112183, "learning_rate": 4.4325261334068426e-05, "loss": 0.3649, "num_input_tokens_seen": 13168, "step": 75 }, { "epoch": 0.3132530120481928, "eval_loss": 0.3668709397315979, "eval_runtime": 0.8423, "eval_samples_per_second": 66.484, "eval_steps_per_second": 33.242, "num_input_tokens_seen": 13760, "step": 78 }, { "epoch": 0.321285140562249, "grad_norm": 7.335175514221191, "learning_rate": 4.316650805085068e-05, "loss": 0.3613, "num_input_tokens_seen": 14080, "step": 80 }, { "epoch": 0.3413654618473896, "grad_norm": 6.502185344696045, "learning_rate": 4.1918457700381855e-05, "loss": 0.3601, "num_input_tokens_seen": 15056, "step": 85 }, { "epoch": 0.3614457831325301, "grad_norm": 6.903041362762451, "learning_rate": 4.058724504646834e-05, "loss": 0.3612, "num_input_tokens_seen": 15904, "step": 90 }, { "epoch": 0.3654618473895582, "eval_loss": 0.36205965280532837, "eval_runtime": 0.7854, "eval_samples_per_second": 71.302, "eval_steps_per_second": 35.651, "num_input_tokens_seen": 16048, "step": 91 }, { "epoch": 0.3815261044176707, "grad_norm": 6.267019748687744, "learning_rate": 3.9179413635373897e-05, "loss": 0.3307, "num_input_tokens_seen": 16688, "step": 95 }, { "epoch": 0.40160642570281124, "grad_norm": 6.203659534454346, "learning_rate": 3.770188363116324e-05, "loss": 0.2896, "num_input_tokens_seen": 17552, "step": 100 }, { "epoch": 0.41767068273092367, "eval_loss": 0.37523433566093445, "eval_runtime": 0.7387, "eval_samples_per_second": 75.808, "eval_steps_per_second": 37.904, "num_input_tokens_seen": 18272, "step": 104 }, { "epoch": 0.42168674698795183, "grad_norm": 2.0911366939544678, "learning_rate": 3.616191779978907e-05, "loss": 0.3154, "num_input_tokens_seen": 18400, "step": 105 }, { "epoch": 0.44176706827309237, "grad_norm": 5.106855392456055, "learning_rate": 3.456708580912725e-05, "loss": 0.3694, "num_input_tokens_seen": 19456, "step": 110 }, { "epoch": 0.46184738955823296, "grad_norm": 8.49660587310791, "learning_rate": 3.292522702044221e-05, "loss": 0.4278, "num_input_tokens_seen": 20288, "step": 115 }, { "epoch": 0.46987951807228917, "eval_loss": 0.3691355884075165, "eval_runtime": 0.7901, "eval_samples_per_second": 70.875, "eval_steps_per_second": 35.437, "num_input_tokens_seen": 20656, "step": 117 }, { "epoch": 0.4819277108433735, "grad_norm": 2.1965019702911377, "learning_rate": 3.1244411954180676e-05, "loss": 0.3821, "num_input_tokens_seen": 21328, "step": 120 }, { "epoch": 0.5020080321285141, "grad_norm": 1.8442065715789795, "learning_rate": 2.9532902619507462e-05, "loss": 0.3707, "num_input_tokens_seen": 22304, "step": 125 }, { "epoch": 0.5220883534136547, "grad_norm": 6.632619857788086, "learning_rate": 2.7799111902582696e-05, "loss": 0.3591, "num_input_tokens_seen": 23056, "step": 130 }, { "epoch": 0.5220883534136547, "eval_loss": 0.3583144545555115, "eval_runtime": 0.9332, "eval_samples_per_second": 60.01, "eval_steps_per_second": 30.005, "num_input_tokens_seen": 23056, "step": 130 }, { "epoch": 0.5421686746987951, "grad_norm": 7.6188459396362305, "learning_rate": 2.6051562213206632e-05, "loss": 0.3438, "num_input_tokens_seen": 23840, "step": 135 }, { "epoch": 0.5622489959839357, "grad_norm": 1.9679665565490723, "learning_rate": 2.429884359310328e-05, "loss": 0.3726, "num_input_tokens_seen": 24832, "step": 140 }, { "epoch": 0.5742971887550201, "eval_loss": 0.3531239330768585, "eval_runtime": 0.8476, "eval_samples_per_second": 66.071, "eval_steps_per_second": 33.036, "num_input_tokens_seen": 25312, "step": 143 }, { "epoch": 0.5823293172690763, "grad_norm": 1.664465069770813, "learning_rate": 2.2549571491760986e-05, "loss": 0.3791, "num_input_tokens_seen": 25648, "step": 145 }, { "epoch": 0.6024096385542169, "grad_norm": 7.226130962371826, "learning_rate": 2.0812344417381595e-05, "loss": 0.3411, "num_input_tokens_seen": 26496, "step": 150 }, { "epoch": 0.6224899598393574, "grad_norm": 7.071829319000244, "learning_rate": 1.909570167110415e-05, "loss": 0.3829, "num_input_tokens_seen": 27392, "step": 155 }, { "epoch": 0.6265060240963856, "eval_loss": 0.35204586386680603, "eval_runtime": 0.805, "eval_samples_per_second": 69.565, "eval_steps_per_second": 34.782, "num_input_tokens_seen": 27552, "step": 156 }, { "epoch": 0.642570281124498, "grad_norm": 1.6194074153900146, "learning_rate": 1.7408081372259632e-05, "loss": 0.3653, "num_input_tokens_seen": 28272, "step": 160 }, { "epoch": 0.6626506024096386, "grad_norm": 5.642275810241699, "learning_rate": 1.5757778980982626e-05, "loss": 0.3318, "num_input_tokens_seen": 29184, "step": 165 }, { "epoch": 0.678714859437751, "eval_loss": 0.35015299916267395, "eval_runtime": 0.7303, "eval_samples_per_second": 76.682, "eval_steps_per_second": 38.341, "num_input_tokens_seen": 29984, "step": 169 }, { "epoch": 0.6827309236947792, "grad_norm": 1.980563759803772, "learning_rate": 1.4152906522061048e-05, "loss": 0.337, "num_input_tokens_seen": 30128, "step": 170 }, { "epoch": 0.7028112449799196, "grad_norm": 2.214860200881958, "learning_rate": 1.2601352710458313e-05, "loss": 0.3303, "num_input_tokens_seen": 30976, "step": 175 }, { "epoch": 0.7228915662650602, "grad_norm": 5.042296886444092, "learning_rate": 1.1110744174509952e-05, "loss": 0.3655, "num_input_tokens_seen": 31776, "step": 180 }, { "epoch": 0.7309236947791165, "eval_loss": 0.3543336093425751, "eval_runtime": 0.8576, "eval_samples_per_second": 65.301, "eval_steps_per_second": 32.65, "num_input_tokens_seen": 32080, "step": 182 }, { "epoch": 0.7429718875502008, "grad_norm": 7.1150078773498535, "learning_rate": 9.688407967401248e-06, "loss": 0.3755, "num_input_tokens_seen": 32608, "step": 185 }, { "epoch": 0.7630522088353414, "grad_norm": 6.07709264755249, "learning_rate": 8.341335551199902e-06, "loss": 0.3967, "num_input_tokens_seen": 33360, "step": 190 }, { "epoch": 0.7831325301204819, "grad_norm": 7.137906074523926, "learning_rate": 7.076148430479321e-06, "loss": 0.3703, "num_input_tokens_seen": 34176, "step": 195 }, { "epoch": 0.7831325301204819, "eval_loss": 0.35257503390312195, "eval_runtime": 0.7915, "eval_samples_per_second": 70.751, "eval_steps_per_second": 35.376, "num_input_tokens_seen": 34176, "step": 195 }, { "epoch": 0.8032128514056225, "grad_norm": 1.6914492845535278, "learning_rate": 5.899065604459814e-06, "loss": 0.355, "num_input_tokens_seen": 34992, "step": 200 }, { "epoch": 0.8232931726907631, "grad_norm": 1.7085988521575928, "learning_rate": 4.81587299765594e-06, "loss": 0.3585, "num_input_tokens_seen": 35888, "step": 205 }, { "epoch": 0.8353413654618473, "eval_loss": 0.3534963130950928, "eval_runtime": 0.8554, "eval_samples_per_second": 65.464, "eval_steps_per_second": 32.732, "num_input_tokens_seen": 36512, "step": 208 }, { "epoch": 0.8433734939759037, "grad_norm": 1.570613145828247, "learning_rate": 3.831895019292897e-06, "loss": 0.3715, "num_input_tokens_seen": 36848, "step": 210 }, { "epoch": 0.8634538152610441, "grad_norm": 2.1737060546875, "learning_rate": 2.9519683912911266e-06, "loss": 0.3719, "num_input_tokens_seen": 37888, "step": 215 }, { "epoch": 0.8835341365461847, "grad_norm": 1.8593710660934448, "learning_rate": 2.1804183734670277e-06, "loss": 0.3626, "num_input_tokens_seen": 38768, "step": 220 }, { "epoch": 0.8875502008032129, "eval_loss": 0.35169586539268494, "eval_runtime": 0.8105, "eval_samples_per_second": 69.096, "eval_steps_per_second": 34.548, "num_input_tokens_seen": 38912, "step": 221 }, { "epoch": 0.9036144578313253, "grad_norm": 1.1120434999465942, "learning_rate": 1.5210375028143097e-06, "loss": 0.3549, "num_input_tokens_seen": 39488, "step": 225 }, { "epoch": 0.9236947791164659, "grad_norm": 1.7197054624557495, "learning_rate": 9.770669513725128e-07, "loss": 0.3419, "num_input_tokens_seen": 40336, "step": 230 }, { "epoch": 0.9397590361445783, "eval_loss": 0.3496658504009247, "eval_runtime": 0.8315, "eval_samples_per_second": 67.352, "eval_steps_per_second": 33.676, "num_input_tokens_seen": 41120, "step": 234 }, { "epoch": 0.9437751004016064, "grad_norm": 1.5268943309783936, "learning_rate": 5.5118059431781e-07, "loss": 0.3635, "num_input_tokens_seen": 41328, "step": 235 }, { "epoch": 0.963855421686747, "grad_norm": 1.8964738845825195, "learning_rate": 2.454718665888589e-07, "loss": 0.3463, "num_input_tokens_seen": 42176, "step": 240 }, { "epoch": 0.9839357429718876, "grad_norm": 1.2437193393707275, "learning_rate": 6.14434726538493e-08, "loss": 0.3311, "num_input_tokens_seen": 43312, "step": 245 }, { "epoch": 0.9919678714859438, "eval_loss": 0.3558541238307953, "eval_runtime": 0.7677, "eval_samples_per_second": 72.948, "eval_steps_per_second": 36.474, "num_input_tokens_seen": 43600, "step": 247 }, { "epoch": 1.0, "num_input_tokens_seen": 43904, "step": 249, "total_flos": 278458437992448.0, "train_loss": 0.3966633373474979, "train_runtime": 65.6097, "train_samples_per_second": 7.59, "train_steps_per_second": 3.795 } ], "logging_steps": 5, "max_steps": 249, "num_input_tokens_seen": 43904, "num_train_epochs": 1, "save_steps": 13, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 278458437992448.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }