| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 8.736842105263158, |
| "eval_steps": 500, |
| "global_step": 81, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.10526315789473684, |
| "grad_norm": 14.31118392944336, |
| "learning_rate": 2e-05, |
| "loss": 2.4434, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.21052631578947367, |
| "grad_norm": 10.385307312011719, |
| "learning_rate": 4e-05, |
| "loss": 2.5855, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.3157894736842105, |
| "grad_norm": 18.257061004638672, |
| "learning_rate": 6e-05, |
| "loss": 2.5327, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.42105263157894735, |
| "grad_norm": 9.803607940673828, |
| "learning_rate": 8e-05, |
| "loss": 2.401, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.5263157894736842, |
| "grad_norm": 10.1410493850708, |
| "learning_rate": 0.0001, |
| "loss": 2.3054, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.631578947368421, |
| "grad_norm": 8.388815879821777, |
| "learning_rate": 0.00012, |
| "loss": 2.3298, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.7368421052631579, |
| "grad_norm": 7.938385963439941, |
| "learning_rate": 0.00014, |
| "loss": 2.0323, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.8421052631578947, |
| "grad_norm": 7.110404968261719, |
| "learning_rate": 0.00016, |
| "loss": 2.0579, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.9473684210526315, |
| "grad_norm": 3.487077474594116, |
| "learning_rate": 0.00018, |
| "loss": 1.6501, |
| "step": 9 |
| }, |
| { |
| "epoch": 1.0789473684210527, |
| "grad_norm": 5.468748092651367, |
| "learning_rate": 0.0002, |
| "loss": 2.8277, |
| "step": 10 |
| }, |
| { |
| "epoch": 1.1842105263157894, |
| "grad_norm": 2.3874928951263428, |
| "learning_rate": 0.0001999229036240723, |
| "loss": 1.2231, |
| "step": 11 |
| }, |
| { |
| "epoch": 1.2894736842105263, |
| "grad_norm": 5.447896480560303, |
| "learning_rate": 0.0001996917333733128, |
| "loss": 1.5471, |
| "step": 12 |
| }, |
| { |
| "epoch": 1.3947368421052633, |
| "grad_norm": 3.8631601333618164, |
| "learning_rate": 0.00019930684569549264, |
| "loss": 1.1826, |
| "step": 13 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 12.651008605957031, |
| "learning_rate": 0.00019876883405951377, |
| "loss": 1.4406, |
| "step": 14 |
| }, |
| { |
| "epoch": 1.6052631578947367, |
| "grad_norm": 2.3405802249908447, |
| "learning_rate": 0.00019807852804032305, |
| "loss": 1.0181, |
| "step": 15 |
| }, |
| { |
| "epoch": 1.7105263157894737, |
| "grad_norm": 6.5888671875, |
| "learning_rate": 0.00019723699203976766, |
| "loss": 1.2595, |
| "step": 16 |
| }, |
| { |
| "epoch": 1.8157894736842106, |
| "grad_norm": 4.1633758544921875, |
| "learning_rate": 0.00019624552364536473, |
| "loss": 1.0053, |
| "step": 17 |
| }, |
| { |
| "epoch": 1.9210526315789473, |
| "grad_norm": 2.3248441219329834, |
| "learning_rate": 0.00019510565162951537, |
| "loss": 0.9642, |
| "step": 18 |
| }, |
| { |
| "epoch": 2.0526315789473686, |
| "grad_norm": 5.530248641967773, |
| "learning_rate": 0.00019381913359224842, |
| "loss": 2.2521, |
| "step": 19 |
| }, |
| { |
| "epoch": 2.1578947368421053, |
| "grad_norm": 1.3485472202301025, |
| "learning_rate": 0.0001923879532511287, |
| "loss": 0.7636, |
| "step": 20 |
| }, |
| { |
| "epoch": 2.263157894736842, |
| "grad_norm": 2.785285472869873, |
| "learning_rate": 0.00019081431738250814, |
| "loss": 1.0499, |
| "step": 21 |
| }, |
| { |
| "epoch": 2.3684210526315788, |
| "grad_norm": 2.2571582794189453, |
| "learning_rate": 0.0001891006524188368, |
| "loss": 0.8719, |
| "step": 22 |
| }, |
| { |
| "epoch": 2.473684210526316, |
| "grad_norm": 1.8401918411254883, |
| "learning_rate": 0.00018724960070727972, |
| "loss": 0.8511, |
| "step": 23 |
| }, |
| { |
| "epoch": 2.5789473684210527, |
| "grad_norm": 2.386539936065674, |
| "learning_rate": 0.00018526401643540922, |
| "loss": 0.8703, |
| "step": 24 |
| }, |
| { |
| "epoch": 2.6842105263157894, |
| "grad_norm": 1.384438157081604, |
| "learning_rate": 0.00018314696123025454, |
| "loss": 0.6987, |
| "step": 25 |
| }, |
| { |
| "epoch": 2.7894736842105265, |
| "grad_norm": 1.705210566520691, |
| "learning_rate": 0.00018090169943749476, |
| "loss": 0.8401, |
| "step": 26 |
| }, |
| { |
| "epoch": 2.8947368421052633, |
| "grad_norm": 1.8555241823196411, |
| "learning_rate": 0.00017853169308807448, |
| "loss": 1.023, |
| "step": 27 |
| }, |
| { |
| "epoch": 3.026315789473684, |
| "grad_norm": 3.781007766723633, |
| "learning_rate": 0.0001760405965600031, |
| "loss": 1.4192, |
| "step": 28 |
| }, |
| { |
| "epoch": 3.1315789473684212, |
| "grad_norm": 1.589735507965088, |
| "learning_rate": 0.00017343225094356855, |
| "loss": 0.7493, |
| "step": 29 |
| }, |
| { |
| "epoch": 3.236842105263158, |
| "grad_norm": 2.163848638534546, |
| "learning_rate": 0.00017071067811865476, |
| "loss": 1.0892, |
| "step": 30 |
| }, |
| { |
| "epoch": 3.3421052631578947, |
| "grad_norm": 1.8226677179336548, |
| "learning_rate": 0.0001678800745532942, |
| "loss": 0.7154, |
| "step": 31 |
| }, |
| { |
| "epoch": 3.4473684210526314, |
| "grad_norm": 1.5850944519042969, |
| "learning_rate": 0.00016494480483301836, |
| "loss": 0.8619, |
| "step": 32 |
| }, |
| { |
| "epoch": 3.5526315789473686, |
| "grad_norm": 1.5230872631072998, |
| "learning_rate": 0.00016190939493098344, |
| "loss": 0.6566, |
| "step": 33 |
| }, |
| { |
| "epoch": 3.6578947368421053, |
| "grad_norm": 1.870915174484253, |
| "learning_rate": 0.00015877852522924732, |
| "loss": 0.8506, |
| "step": 34 |
| }, |
| { |
| "epoch": 3.763157894736842, |
| "grad_norm": 1.698755145072937, |
| "learning_rate": 0.00015555702330196023, |
| "loss": 0.762, |
| "step": 35 |
| }, |
| { |
| "epoch": 3.8684210526315788, |
| "grad_norm": 1.6367840766906738, |
| "learning_rate": 0.0001522498564715949, |
| "loss": 0.7136, |
| "step": 36 |
| }, |
| { |
| "epoch": 3.973684210526316, |
| "grad_norm": 2.9744303226470947, |
| "learning_rate": 0.00014886212414969553, |
| "loss": 0.9568, |
| "step": 37 |
| }, |
| { |
| "epoch": 4.105263157894737, |
| "grad_norm": 1.3073623180389404, |
| "learning_rate": 0.00014539904997395468, |
| "loss": 0.6397, |
| "step": 38 |
| }, |
| { |
| "epoch": 4.2105263157894735, |
| "grad_norm": 1.2846301794052124, |
| "learning_rate": 0.0001418659737537428, |
| "loss": 0.4418, |
| "step": 39 |
| }, |
| { |
| "epoch": 4.315789473684211, |
| "grad_norm": 1.6083595752716064, |
| "learning_rate": 0.000138268343236509, |
| "loss": 0.7064, |
| "step": 40 |
| }, |
| { |
| "epoch": 4.421052631578947, |
| "grad_norm": 1.2033262252807617, |
| "learning_rate": 0.0001346117057077493, |
| "loss": 0.5957, |
| "step": 41 |
| }, |
| { |
| "epoch": 4.526315789473684, |
| "grad_norm": 1.3466635942459106, |
| "learning_rate": 0.00013090169943749476, |
| "loss": 0.7566, |
| "step": 42 |
| }, |
| { |
| "epoch": 4.631578947368421, |
| "grad_norm": 1.616913914680481, |
| "learning_rate": 0.00012714404498650743, |
| "loss": 0.7539, |
| "step": 43 |
| }, |
| { |
| "epoch": 4.7368421052631575, |
| "grad_norm": 1.4130221605300903, |
| "learning_rate": 0.00012334453638559057, |
| "loss": 0.4622, |
| "step": 44 |
| }, |
| { |
| "epoch": 4.842105263157895, |
| "grad_norm": 1.3309547901153564, |
| "learning_rate": 0.00011950903220161285, |
| "loss": 0.7263, |
| "step": 45 |
| }, |
| { |
| "epoch": 4.947368421052632, |
| "grad_norm": 1.354387879371643, |
| "learning_rate": 0.0001156434465040231, |
| "loss": 0.5625, |
| "step": 46 |
| }, |
| { |
| "epoch": 5.078947368421052, |
| "grad_norm": 3.008500337600708, |
| "learning_rate": 0.00011175373974578378, |
| "loss": 0.7619, |
| "step": 47 |
| }, |
| { |
| "epoch": 5.184210526315789, |
| "grad_norm": 1.5682655572891235, |
| "learning_rate": 0.0001078459095727845, |
| "loss": 0.3858, |
| "step": 48 |
| }, |
| { |
| "epoch": 5.2894736842105265, |
| "grad_norm": 1.8495001792907715, |
| "learning_rate": 0.00010392598157590688, |
| "loss": 0.4573, |
| "step": 49 |
| }, |
| { |
| "epoch": 5.394736842105263, |
| "grad_norm": 2.635056972503662, |
| "learning_rate": 0.0001, |
| "loss": 0.7921, |
| "step": 50 |
| }, |
| { |
| "epoch": 5.5, |
| "grad_norm": 1.5783675909042358, |
| "learning_rate": 9.607401842409317e-05, |
| "loss": 0.3204, |
| "step": 51 |
| }, |
| { |
| "epoch": 5.605263157894737, |
| "grad_norm": 2.0661301612854004, |
| "learning_rate": 9.215409042721552e-05, |
| "loss": 0.6708, |
| "step": 52 |
| }, |
| { |
| "epoch": 5.7105263157894735, |
| "grad_norm": 1.4969799518585205, |
| "learning_rate": 8.824626025421626e-05, |
| "loss": 0.4052, |
| "step": 53 |
| }, |
| { |
| "epoch": 5.815789473684211, |
| "grad_norm": 1.6124238967895508, |
| "learning_rate": 8.435655349597689e-05, |
| "loss": 0.4829, |
| "step": 54 |
| }, |
| { |
| "epoch": 5.921052631578947, |
| "grad_norm": 1.431242823600769, |
| "learning_rate": 8.049096779838719e-05, |
| "loss": 0.4273, |
| "step": 55 |
| }, |
| { |
| "epoch": 6.052631578947368, |
| "grad_norm": 3.2095937728881836, |
| "learning_rate": 7.66554636144095e-05, |
| "loss": 0.8437, |
| "step": 56 |
| }, |
| { |
| "epoch": 6.157894736842105, |
| "grad_norm": 0.8892129063606262, |
| "learning_rate": 7.285595501349258e-05, |
| "loss": 0.2127, |
| "step": 57 |
| }, |
| { |
| "epoch": 6.2631578947368425, |
| "grad_norm": 1.6543368101119995, |
| "learning_rate": 6.909830056250527e-05, |
| "loss": 0.4305, |
| "step": 58 |
| }, |
| { |
| "epoch": 6.368421052631579, |
| "grad_norm": 1.4523223638534546, |
| "learning_rate": 6.538829429225069e-05, |
| "loss": 0.3762, |
| "step": 59 |
| }, |
| { |
| "epoch": 6.473684210526316, |
| "grad_norm": 0.9503556489944458, |
| "learning_rate": 6.173165676349103e-05, |
| "loss": 0.3628, |
| "step": 60 |
| }, |
| { |
| "epoch": 6.578947368421053, |
| "grad_norm": 1.4268410205841064, |
| "learning_rate": 5.8134026246257225e-05, |
| "loss": 0.4751, |
| "step": 61 |
| }, |
| { |
| "epoch": 6.684210526315789, |
| "grad_norm": 0.9671533703804016, |
| "learning_rate": 5.4600950026045326e-05, |
| "loss": 0.2934, |
| "step": 62 |
| }, |
| { |
| "epoch": 6.7894736842105265, |
| "grad_norm": 1.1906907558441162, |
| "learning_rate": 5.113787585030454e-05, |
| "loss": 0.2345, |
| "step": 63 |
| }, |
| { |
| "epoch": 6.894736842105263, |
| "grad_norm": 1.5449588298797607, |
| "learning_rate": 4.7750143528405126e-05, |
| "loss": 0.4283, |
| "step": 64 |
| }, |
| { |
| "epoch": 7.026315789473684, |
| "grad_norm": 2.223735809326172, |
| "learning_rate": 4.444297669803981e-05, |
| "loss": 0.4928, |
| "step": 65 |
| }, |
| { |
| "epoch": 7.131578947368421, |
| "grad_norm": 0.9985809326171875, |
| "learning_rate": 4.12214747707527e-05, |
| "loss": 0.3185, |
| "step": 66 |
| }, |
| { |
| "epoch": 7.2368421052631575, |
| "grad_norm": 1.0797033309936523, |
| "learning_rate": 3.8090605069016595e-05, |
| "loss": 0.4603, |
| "step": 67 |
| }, |
| { |
| "epoch": 7.342105263157895, |
| "grad_norm": 1.0436688661575317, |
| "learning_rate": 3.5055195166981645e-05, |
| "loss": 0.4009, |
| "step": 68 |
| }, |
| { |
| "epoch": 7.447368421052632, |
| "grad_norm": 1.1368814706802368, |
| "learning_rate": 3.211992544670582e-05, |
| "loss": 0.2486, |
| "step": 69 |
| }, |
| { |
| "epoch": 7.552631578947368, |
| "grad_norm": 1.0792099237442017, |
| "learning_rate": 2.9289321881345254e-05, |
| "loss": 0.3308, |
| "step": 70 |
| }, |
| { |
| "epoch": 7.657894736842105, |
| "grad_norm": 0.9220963716506958, |
| "learning_rate": 2.6567749056431467e-05, |
| "loss": 0.3044, |
| "step": 71 |
| }, |
| { |
| "epoch": 7.7631578947368425, |
| "grad_norm": 1.1355466842651367, |
| "learning_rate": 2.3959403439996907e-05, |
| "loss": 0.2657, |
| "step": 72 |
| }, |
| { |
| "epoch": 7.868421052631579, |
| "grad_norm": 0.9300779104232788, |
| "learning_rate": 2.146830691192553e-05, |
| "loss": 0.2587, |
| "step": 73 |
| }, |
| { |
| "epoch": 7.973684210526316, |
| "grad_norm": 2.0543935298919678, |
| "learning_rate": 1.9098300562505266e-05, |
| "loss": 0.5869, |
| "step": 74 |
| }, |
| { |
| "epoch": 8.105263157894736, |
| "grad_norm": 1.0328178405761719, |
| "learning_rate": 1.6853038769745467e-05, |
| "loss": 0.3424, |
| "step": 75 |
| }, |
| { |
| "epoch": 8.210526315789474, |
| "grad_norm": 1.0409759283065796, |
| "learning_rate": 1.4735983564590783e-05, |
| "loss": 0.3577, |
| "step": 76 |
| }, |
| { |
| "epoch": 8.31578947368421, |
| "grad_norm": 1.2581232786178589, |
| "learning_rate": 1.2750399292720283e-05, |
| "loss": 0.354, |
| "step": 77 |
| }, |
| { |
| "epoch": 8.421052631578947, |
| "grad_norm": 1.073712706565857, |
| "learning_rate": 1.0899347581163221e-05, |
| "loss": 0.3332, |
| "step": 78 |
| }, |
| { |
| "epoch": 8.526315789473685, |
| "grad_norm": 1.0400298833847046, |
| "learning_rate": 9.185682617491863e-06, |
| "loss": 0.3112, |
| "step": 79 |
| }, |
| { |
| "epoch": 8.631578947368421, |
| "grad_norm": 1.7903478145599365, |
| "learning_rate": 7.612046748871327e-06, |
| "loss": 0.4138, |
| "step": 80 |
| }, |
| { |
| "epoch": 8.736842105263158, |
| "grad_norm": 1.1479917764663696, |
| "learning_rate": 6.180866407751595e-06, |
| "loss": 0.2799, |
| "step": 81 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 90, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 10, |
| "save_steps": 9, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.0747062240097075e+17, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|