| { |
| "best_metric": 1.3649392127990723, |
| "best_model_checkpoint": "./output/checkpoints/2024-06-11_11-02-23/checkpoint-50", |
| "epoch": 1.0, |
| "eval_steps": 1, |
| "global_step": 97, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.010309278350515464, |
| "grad_norm": 2.784351348876953, |
| "learning_rate": 4e-05, |
| "loss": 5.0719, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.010309278350515464, |
| "eval_loss": 4.890735149383545, |
| "eval_runtime": 11.998, |
| "eval_samples_per_second": 11.252, |
| "eval_steps_per_second": 0.75, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.020618556701030927, |
| "grad_norm": 2.8573310375213623, |
| "learning_rate": 8e-05, |
| "loss": 4.9508, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.020618556701030927, |
| "eval_loss": 4.775210857391357, |
| "eval_runtime": 12.0698, |
| "eval_samples_per_second": 11.185, |
| "eval_steps_per_second": 0.746, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.030927835051546393, |
| "grad_norm": 2.920828104019165, |
| "learning_rate": 0.00012, |
| "loss": 4.973, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.030927835051546393, |
| "eval_loss": 4.2439093589782715, |
| "eval_runtime": 12.1674, |
| "eval_samples_per_second": 11.095, |
| "eval_steps_per_second": 0.74, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.041237113402061855, |
| "grad_norm": 3.0682826042175293, |
| "learning_rate": 0.00016, |
| "loss": 4.3398, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.041237113402061855, |
| "eval_loss": 3.316483497619629, |
| "eval_runtime": 12.1421, |
| "eval_samples_per_second": 11.118, |
| "eval_steps_per_second": 0.741, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.05154639175257732, |
| "grad_norm": 3.1937592029571533, |
| "learning_rate": 0.0002, |
| "loss": 3.253, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.05154639175257732, |
| "eval_loss": 2.4086239337921143, |
| "eval_runtime": 12.0796, |
| "eval_samples_per_second": 11.176, |
| "eval_steps_per_second": 0.745, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.061855670103092786, |
| "grad_norm": 2.0619874000549316, |
| "learning_rate": 0.00024, |
| "loss": 2.3725, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.061855670103092786, |
| "eval_loss": 1.9427752494812012, |
| "eval_runtime": 12.1816, |
| "eval_samples_per_second": 11.082, |
| "eval_steps_per_second": 0.739, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.07216494845360824, |
| "grad_norm": 1.4178358316421509, |
| "learning_rate": 0.00028, |
| "loss": 1.7391, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.07216494845360824, |
| "eval_loss": 1.6751385927200317, |
| "eval_runtime": 12.2395, |
| "eval_samples_per_second": 11.03, |
| "eval_steps_per_second": 0.735, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.08247422680412371, |
| "grad_norm": 0.9345605373382568, |
| "learning_rate": 0.00032, |
| "loss": 1.4029, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.08247422680412371, |
| "eval_loss": 1.551221489906311, |
| "eval_runtime": 12.2329, |
| "eval_samples_per_second": 11.036, |
| "eval_steps_per_second": 0.736, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.09278350515463918, |
| "grad_norm": 1.1335052251815796, |
| "learning_rate": 0.00036, |
| "loss": 1.4404, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.09278350515463918, |
| "eval_loss": 1.447738528251648, |
| "eval_runtime": 12.2069, |
| "eval_samples_per_second": 11.059, |
| "eval_steps_per_second": 0.737, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.10309278350515463, |
| "grad_norm": 0.23779241740703583, |
| "learning_rate": 0.0004, |
| "loss": 1.3197, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.10309278350515463, |
| "eval_loss": 1.4138007164001465, |
| "eval_runtime": 12.2609, |
| "eval_samples_per_second": 11.011, |
| "eval_steps_per_second": 0.734, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.1134020618556701, |
| "grad_norm": 0.19672752916812897, |
| "learning_rate": 0.00039540229885057476, |
| "loss": 1.2882, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.1134020618556701, |
| "eval_loss": 1.393249750137329, |
| "eval_runtime": 12.1836, |
| "eval_samples_per_second": 11.08, |
| "eval_steps_per_second": 0.739, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.12371134020618557, |
| "grad_norm": 0.2525981366634369, |
| "learning_rate": 0.00039080459770114945, |
| "loss": 1.2831, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.12371134020618557, |
| "eval_loss": 1.4042422771453857, |
| "eval_runtime": 12.232, |
| "eval_samples_per_second": 11.037, |
| "eval_steps_per_second": 0.736, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.13402061855670103, |
| "grad_norm": 0.23460708558559418, |
| "learning_rate": 0.0003862068965517242, |
| "loss": 1.1615, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.13402061855670103, |
| "eval_loss": 1.4705734252929688, |
| "eval_runtime": 12.2648, |
| "eval_samples_per_second": 11.007, |
| "eval_steps_per_second": 0.734, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.14432989690721648, |
| "grad_norm": 0.21198733150959015, |
| "learning_rate": 0.00038160919540229887, |
| "loss": 1.0625, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.14432989690721648, |
| "eval_loss": 1.5055038928985596, |
| "eval_runtime": 12.2044, |
| "eval_samples_per_second": 11.062, |
| "eval_steps_per_second": 0.737, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.15463917525773196, |
| "grad_norm": 0.3658374845981598, |
| "learning_rate": 0.00037701149425287356, |
| "loss": 1.1191, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.15463917525773196, |
| "eval_loss": 1.4498964548110962, |
| "eval_runtime": 12.2004, |
| "eval_samples_per_second": 11.065, |
| "eval_steps_per_second": 0.738, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.16494845360824742, |
| "grad_norm": 0.222326397895813, |
| "learning_rate": 0.0003724137931034483, |
| "loss": 1.0887, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.16494845360824742, |
| "eval_loss": 1.3819501399993896, |
| "eval_runtime": 12.1444, |
| "eval_samples_per_second": 11.116, |
| "eval_steps_per_second": 0.741, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.17525773195876287, |
| "grad_norm": 0.1682404726743698, |
| "learning_rate": 0.000367816091954023, |
| "loss": 1.0915, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.17525773195876287, |
| "eval_loss": 1.3552738428115845, |
| "eval_runtime": 12.2121, |
| "eval_samples_per_second": 11.055, |
| "eval_steps_per_second": 0.737, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.18556701030927836, |
| "grad_norm": 0.15567483007907867, |
| "learning_rate": 0.0003632183908045977, |
| "loss": 1.0509, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.18556701030927836, |
| "eval_loss": 1.3488575220108032, |
| "eval_runtime": 12.2988, |
| "eval_samples_per_second": 10.977, |
| "eval_steps_per_second": 0.732, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.1958762886597938, |
| "grad_norm": 0.1600300371646881, |
| "learning_rate": 0.0003586206896551724, |
| "loss": 0.9982, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.1958762886597938, |
| "eval_loss": 1.3651177883148193, |
| "eval_runtime": 12.1263, |
| "eval_samples_per_second": 11.133, |
| "eval_steps_per_second": 0.742, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.20618556701030927, |
| "grad_norm": 0.1044178307056427, |
| "learning_rate": 0.00035402298850574715, |
| "loss": 0.9809, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.20618556701030927, |
| "eval_loss": 1.401971697807312, |
| "eval_runtime": 12.2092, |
| "eval_samples_per_second": 11.057, |
| "eval_steps_per_second": 0.737, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.21649484536082475, |
| "grad_norm": 0.12360141426324844, |
| "learning_rate": 0.0003494252873563219, |
| "loss": 1.0549, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.21649484536082475, |
| "eval_loss": 1.426545262336731, |
| "eval_runtime": 12.1581, |
| "eval_samples_per_second": 11.104, |
| "eval_steps_per_second": 0.74, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.2268041237113402, |
| "grad_norm": 0.12509943544864655, |
| "learning_rate": 0.0003448275862068965, |
| "loss": 1.0323, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.2268041237113402, |
| "eval_loss": 1.4258630275726318, |
| "eval_runtime": 12.201, |
| "eval_samples_per_second": 11.065, |
| "eval_steps_per_second": 0.738, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.23711340206185566, |
| "grad_norm": 0.13586747646331787, |
| "learning_rate": 0.00034022988505747127, |
| "loss": 1.0746, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.23711340206185566, |
| "eval_loss": 1.3952091932296753, |
| "eval_runtime": 12.1932, |
| "eval_samples_per_second": 11.072, |
| "eval_steps_per_second": 0.738, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.24742268041237114, |
| "grad_norm": 0.08069202303886414, |
| "learning_rate": 0.000335632183908046, |
| "loss": 0.9645, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.24742268041237114, |
| "eval_loss": 1.376875638961792, |
| "eval_runtime": 12.2172, |
| "eval_samples_per_second": 11.05, |
| "eval_steps_per_second": 0.737, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.25773195876288657, |
| "grad_norm": 0.09109444171190262, |
| "learning_rate": 0.0003310344827586207, |
| "loss": 0.9345, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.25773195876288657, |
| "eval_loss": 1.373625636100769, |
| "eval_runtime": 12.2642, |
| "eval_samples_per_second": 11.008, |
| "eval_steps_per_second": 0.734, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.26804123711340205, |
| "grad_norm": 0.0649966150522232, |
| "learning_rate": 0.00032643678160919543, |
| "loss": 1.0266, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.26804123711340205, |
| "eval_loss": 1.3720897436141968, |
| "eval_runtime": 12.1576, |
| "eval_samples_per_second": 11.104, |
| "eval_steps_per_second": 0.74, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.27835051546391754, |
| "grad_norm": 0.09308775514364243, |
| "learning_rate": 0.0003218390804597701, |
| "loss": 0.9797, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.27835051546391754, |
| "eval_loss": 1.3858685493469238, |
| "eval_runtime": 12.1546, |
| "eval_samples_per_second": 11.107, |
| "eval_steps_per_second": 0.74, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.28865979381443296, |
| "grad_norm": 0.06654678285121918, |
| "learning_rate": 0.00031724137931034486, |
| "loss": 1.0072, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.28865979381443296, |
| "eval_loss": 1.4031970500946045, |
| "eval_runtime": 12.1666, |
| "eval_samples_per_second": 11.096, |
| "eval_steps_per_second": 0.74, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.29896907216494845, |
| "grad_norm": 0.07720344513654709, |
| "learning_rate": 0.0003126436781609196, |
| "loss": 0.923, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.29896907216494845, |
| "eval_loss": 1.4118249416351318, |
| "eval_runtime": 12.1572, |
| "eval_samples_per_second": 11.105, |
| "eval_steps_per_second": 0.74, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.30927835051546393, |
| "grad_norm": 0.10230278223752975, |
| "learning_rate": 0.00030804597701149423, |
| "loss": 0.9821, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.30927835051546393, |
| "eval_loss": 1.4045817852020264, |
| "eval_runtime": 12.1974, |
| "eval_samples_per_second": 11.068, |
| "eval_steps_per_second": 0.738, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.31958762886597936, |
| "grad_norm": 0.07451125234365463, |
| "learning_rate": 0.00030344827586206897, |
| "loss": 1.0021, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.31958762886597936, |
| "eval_loss": 1.391209363937378, |
| "eval_runtime": 12.2389, |
| "eval_samples_per_second": 11.03, |
| "eval_steps_per_second": 0.735, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.32989690721649484, |
| "grad_norm": 0.0714351087808609, |
| "learning_rate": 0.00029885057471264366, |
| "loss": 1.0071, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.32989690721649484, |
| "eval_loss": 1.3730576038360596, |
| "eval_runtime": 12.2102, |
| "eval_samples_per_second": 11.056, |
| "eval_steps_per_second": 0.737, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.3402061855670103, |
| "grad_norm": 0.06839103996753693, |
| "learning_rate": 0.0002942528735632184, |
| "loss": 0.973, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.3402061855670103, |
| "eval_loss": 1.3659778833389282, |
| "eval_runtime": 12.2323, |
| "eval_samples_per_second": 11.036, |
| "eval_steps_per_second": 0.736, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.35051546391752575, |
| "grad_norm": 0.08078178018331528, |
| "learning_rate": 0.00028965517241379314, |
| "loss": 0.964, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.35051546391752575, |
| "eval_loss": 1.3762452602386475, |
| "eval_runtime": 12.2027, |
| "eval_samples_per_second": 11.063, |
| "eval_steps_per_second": 0.738, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.36082474226804123, |
| "grad_norm": 0.06870069354772568, |
| "learning_rate": 0.0002850574712643678, |
| "loss": 0.9148, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.36082474226804123, |
| "eval_loss": 1.3925738334655762, |
| "eval_runtime": 12.2644, |
| "eval_samples_per_second": 11.007, |
| "eval_steps_per_second": 0.734, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.3711340206185567, |
| "grad_norm": 0.06974003463983536, |
| "learning_rate": 0.00028045977011494257, |
| "loss": 1.0128, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.3711340206185567, |
| "eval_loss": 1.4087179899215698, |
| "eval_runtime": 12.2407, |
| "eval_samples_per_second": 11.029, |
| "eval_steps_per_second": 0.735, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.38144329896907214, |
| "grad_norm": 0.08603405207395554, |
| "learning_rate": 0.00027586206896551725, |
| "loss": 0.9776, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.38144329896907214, |
| "eval_loss": 1.4067703485488892, |
| "eval_runtime": 12.173, |
| "eval_samples_per_second": 11.09, |
| "eval_steps_per_second": 0.739, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.3917525773195876, |
| "grad_norm": 0.07761300355195999, |
| "learning_rate": 0.00027126436781609194, |
| "loss": 0.9655, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.3917525773195876, |
| "eval_loss": 1.3932013511657715, |
| "eval_runtime": 12.1941, |
| "eval_samples_per_second": 11.071, |
| "eval_steps_per_second": 0.738, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.4020618556701031, |
| "grad_norm": 0.06392566114664078, |
| "learning_rate": 0.0002666666666666667, |
| "loss": 0.974, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.4020618556701031, |
| "eval_loss": 1.3819767236709595, |
| "eval_runtime": 12.1765, |
| "eval_samples_per_second": 11.087, |
| "eval_steps_per_second": 0.739, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.41237113402061853, |
| "grad_norm": 0.05517549812793732, |
| "learning_rate": 0.00026206896551724137, |
| "loss": 0.9793, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.41237113402061853, |
| "eval_loss": 1.3717000484466553, |
| "eval_runtime": 12.2217, |
| "eval_samples_per_second": 11.046, |
| "eval_steps_per_second": 0.736, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.422680412371134, |
| "grad_norm": 0.0804053246974945, |
| "learning_rate": 0.0002574712643678161, |
| "loss": 0.9585, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.422680412371134, |
| "eval_loss": 1.3579998016357422, |
| "eval_runtime": 12.182, |
| "eval_samples_per_second": 11.082, |
| "eval_steps_per_second": 0.739, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.4329896907216495, |
| "grad_norm": 0.07214821875095367, |
| "learning_rate": 0.0002528735632183908, |
| "loss": 0.9332, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.4329896907216495, |
| "eval_loss": 1.3583327531814575, |
| "eval_runtime": 12.1257, |
| "eval_samples_per_second": 11.133, |
| "eval_steps_per_second": 0.742, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.44329896907216493, |
| "grad_norm": 0.07595060020685196, |
| "learning_rate": 0.00024827586206896553, |
| "loss": 0.8998, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.44329896907216493, |
| "eval_loss": 1.3721704483032227, |
| "eval_runtime": 12.2745, |
| "eval_samples_per_second": 10.998, |
| "eval_steps_per_second": 0.733, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.4536082474226804, |
| "grad_norm": 0.07757716625928879, |
| "learning_rate": 0.00024367816091954025, |
| "loss": 0.9661, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.4536082474226804, |
| "eval_loss": 1.3984168767929077, |
| "eval_runtime": 12.1398, |
| "eval_samples_per_second": 11.12, |
| "eval_steps_per_second": 0.741, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.4639175257731959, |
| "grad_norm": 0.053873661905527115, |
| "learning_rate": 0.00023908045977011496, |
| "loss": 0.9418, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.4639175257731959, |
| "eval_loss": 1.4266961812973022, |
| "eval_runtime": 12.1955, |
| "eval_samples_per_second": 11.07, |
| "eval_steps_per_second": 0.738, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.4742268041237113, |
| "grad_norm": 0.08943776786327362, |
| "learning_rate": 0.00023448275862068965, |
| "loss": 0.9309, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.4742268041237113, |
| "eval_loss": 1.4349451065063477, |
| "eval_runtime": 12.1983, |
| "eval_samples_per_second": 11.067, |
| "eval_steps_per_second": 0.738, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.4845360824742268, |
| "grad_norm": 0.0885058343410492, |
| "learning_rate": 0.00022988505747126436, |
| "loss": 1.0245, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.4845360824742268, |
| "eval_loss": 1.4226434230804443, |
| "eval_runtime": 12.1753, |
| "eval_samples_per_second": 11.088, |
| "eval_steps_per_second": 0.739, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.4948453608247423, |
| "grad_norm": 0.058818139135837555, |
| "learning_rate": 0.00022528735632183907, |
| "loss": 0.9007, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.4948453608247423, |
| "eval_loss": 1.4033018350601196, |
| "eval_runtime": 12.267, |
| "eval_samples_per_second": 11.005, |
| "eval_steps_per_second": 0.734, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.5051546391752577, |
| "grad_norm": 0.07104739546775818, |
| "learning_rate": 0.0002206896551724138, |
| "loss": 0.9469, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.5051546391752577, |
| "eval_loss": 1.3786002397537231, |
| "eval_runtime": 12.2731, |
| "eval_samples_per_second": 11.0, |
| "eval_steps_per_second": 0.733, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.5154639175257731, |
| "grad_norm": 0.05872216075658798, |
| "learning_rate": 0.00021609195402298853, |
| "loss": 0.9671, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.5154639175257731, |
| "eval_loss": 1.3649392127990723, |
| "eval_runtime": 12.1662, |
| "eval_samples_per_second": 11.096, |
| "eval_steps_per_second": 0.74, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.5257731958762887, |
| "grad_norm": 0.07843936234712601, |
| "learning_rate": 0.00021149425287356324, |
| "loss": 0.9052, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.5257731958762887, |
| "eval_loss": 1.3671882152557373, |
| "eval_runtime": 12.1527, |
| "eval_samples_per_second": 11.109, |
| "eval_steps_per_second": 0.741, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.5360824742268041, |
| "grad_norm": 0.07407287508249283, |
| "learning_rate": 0.00020689655172413795, |
| "loss": 0.9221, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.5360824742268041, |
| "eval_loss": 1.3811315298080444, |
| "eval_runtime": 12.1661, |
| "eval_samples_per_second": 11.096, |
| "eval_steps_per_second": 0.74, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.5463917525773195, |
| "grad_norm": 0.06168922409415245, |
| "learning_rate": 0.00020229885057471267, |
| "loss": 0.9809, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.5463917525773195, |
| "eval_loss": 1.3900806903839111, |
| "eval_runtime": 12.1958, |
| "eval_samples_per_second": 11.069, |
| "eval_steps_per_second": 0.738, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.5567010309278351, |
| "grad_norm": 0.05980532988905907, |
| "learning_rate": 0.00019770114942528738, |
| "loss": 0.9492, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.5567010309278351, |
| "eval_loss": 1.4011154174804688, |
| "eval_runtime": 12.1559, |
| "eval_samples_per_second": 11.106, |
| "eval_steps_per_second": 0.74, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.5670103092783505, |
| "grad_norm": 0.05770307034254074, |
| "learning_rate": 0.0001931034482758621, |
| "loss": 0.9377, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.5670103092783505, |
| "eval_loss": 1.4104657173156738, |
| "eval_runtime": 12.2103, |
| "eval_samples_per_second": 11.056, |
| "eval_steps_per_second": 0.737, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.5773195876288659, |
| "grad_norm": 0.05812888965010643, |
| "learning_rate": 0.00018850574712643678, |
| "loss": 0.9139, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.5773195876288659, |
| "eval_loss": 1.4152581691741943, |
| "eval_runtime": 12.1515, |
| "eval_samples_per_second": 11.11, |
| "eval_steps_per_second": 0.741, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.5876288659793815, |
| "grad_norm": 0.0625990554690361, |
| "learning_rate": 0.0001839080459770115, |
| "loss": 0.9111, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.5876288659793815, |
| "eval_loss": 1.4142401218414307, |
| "eval_runtime": 12.208, |
| "eval_samples_per_second": 11.058, |
| "eval_steps_per_second": 0.737, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.5979381443298969, |
| "grad_norm": 0.06904823333024979, |
| "learning_rate": 0.0001793103448275862, |
| "loss": 0.9019, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.5979381443298969, |
| "eval_loss": 1.4036115407943726, |
| "eval_runtime": 12.1626, |
| "eval_samples_per_second": 11.1, |
| "eval_steps_per_second": 0.74, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.6082474226804123, |
| "grad_norm": 0.06511174887418747, |
| "learning_rate": 0.00017471264367816095, |
| "loss": 0.9409, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.6082474226804123, |
| "eval_loss": 1.3875020742416382, |
| "eval_runtime": 12.1594, |
| "eval_samples_per_second": 11.103, |
| "eval_steps_per_second": 0.74, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.6185567010309279, |
| "grad_norm": 0.06755220890045166, |
| "learning_rate": 0.00017011494252873563, |
| "loss": 0.8928, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.6185567010309279, |
| "eval_loss": 1.3814879655838013, |
| "eval_runtime": 12.1832, |
| "eval_samples_per_second": 11.081, |
| "eval_steps_per_second": 0.739, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.6288659793814433, |
| "grad_norm": 0.057419709861278534, |
| "learning_rate": 0.00016551724137931035, |
| "loss": 0.8698, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.6288659793814433, |
| "eval_loss": 1.376993179321289, |
| "eval_runtime": 12.1649, |
| "eval_samples_per_second": 11.097, |
| "eval_steps_per_second": 0.74, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.6391752577319587, |
| "grad_norm": 0.09423535317182541, |
| "learning_rate": 0.00016091954022988506, |
| "loss": 0.9605, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.6391752577319587, |
| "eval_loss": 1.3864612579345703, |
| "eval_runtime": 12.1886, |
| "eval_samples_per_second": 11.076, |
| "eval_steps_per_second": 0.738, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.6494845360824743, |
| "grad_norm": 0.05667712539434433, |
| "learning_rate": 0.0001563218390804598, |
| "loss": 0.9863, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.6494845360824743, |
| "eval_loss": 1.3983639478683472, |
| "eval_runtime": 12.1418, |
| "eval_samples_per_second": 11.119, |
| "eval_steps_per_second": 0.741, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.6597938144329897, |
| "grad_norm": 0.061302803456783295, |
| "learning_rate": 0.00015172413793103449, |
| "loss": 0.9454, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.6597938144329897, |
| "eval_loss": 1.406610369682312, |
| "eval_runtime": 12.2718, |
| "eval_samples_per_second": 11.001, |
| "eval_steps_per_second": 0.733, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.6701030927835051, |
| "grad_norm": 0.06619007140398026, |
| "learning_rate": 0.0001471264367816092, |
| "loss": 0.9302, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.6701030927835051, |
| "eval_loss": 1.408695936203003, |
| "eval_runtime": 12.1873, |
| "eval_samples_per_second": 11.077, |
| "eval_steps_per_second": 0.738, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.6804123711340206, |
| "grad_norm": 0.059212010353803635, |
| "learning_rate": 0.0001425287356321839, |
| "loss": 0.9409, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.6804123711340206, |
| "eval_loss": 1.4100947380065918, |
| "eval_runtime": 12.2058, |
| "eval_samples_per_second": 11.06, |
| "eval_steps_per_second": 0.737, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.6907216494845361, |
| "grad_norm": 0.06854245811700821, |
| "learning_rate": 0.00013793103448275863, |
| "loss": 0.9408, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.6907216494845361, |
| "eval_loss": 1.4050439596176147, |
| "eval_runtime": 12.1478, |
| "eval_samples_per_second": 11.113, |
| "eval_steps_per_second": 0.741, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.7010309278350515, |
| "grad_norm": 0.05722883343696594, |
| "learning_rate": 0.00013333333333333334, |
| "loss": 0.91, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.7010309278350515, |
| "eval_loss": 1.397505283355713, |
| "eval_runtime": 12.1678, |
| "eval_samples_per_second": 11.095, |
| "eval_steps_per_second": 0.74, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.711340206185567, |
| "grad_norm": 0.07448893785476685, |
| "learning_rate": 0.00012873563218390805, |
| "loss": 0.9451, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.711340206185567, |
| "eval_loss": 1.3847121000289917, |
| "eval_runtime": 12.1444, |
| "eval_samples_per_second": 11.116, |
| "eval_steps_per_second": 0.741, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.7216494845360825, |
| "grad_norm": 0.055692195892333984, |
| "learning_rate": 0.00012413793103448277, |
| "loss": 0.8967, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.7216494845360825, |
| "eval_loss": 1.376731038093567, |
| "eval_runtime": 12.1911, |
| "eval_samples_per_second": 11.074, |
| "eval_steps_per_second": 0.738, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.7319587628865979, |
| "grad_norm": 0.06589022278785706, |
| "learning_rate": 0.00011954022988505748, |
| "loss": 0.8795, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.7319587628865979, |
| "eval_loss": 1.3753036260604858, |
| "eval_runtime": 12.1728, |
| "eval_samples_per_second": 11.09, |
| "eval_steps_per_second": 0.739, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.7422680412371134, |
| "grad_norm": 0.12176728248596191, |
| "learning_rate": 0.00011494252873563218, |
| "loss": 0.9092, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.7422680412371134, |
| "eval_loss": 1.3911007642745972, |
| "eval_runtime": 12.1946, |
| "eval_samples_per_second": 11.07, |
| "eval_steps_per_second": 0.738, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.7525773195876289, |
| "grad_norm": 0.05275936424732208, |
| "learning_rate": 0.0001103448275862069, |
| "loss": 0.9621, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.7525773195876289, |
| "eval_loss": 1.407221794128418, |
| "eval_runtime": 12.2388, |
| "eval_samples_per_second": 11.031, |
| "eval_steps_per_second": 0.735, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.7628865979381443, |
| "grad_norm": 0.06748662143945694, |
| "learning_rate": 0.00010574712643678162, |
| "loss": 0.9154, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.7628865979381443, |
| "eval_loss": 1.4170591831207275, |
| "eval_runtime": 12.1869, |
| "eval_samples_per_second": 11.078, |
| "eval_steps_per_second": 0.739, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.7731958762886598, |
| "grad_norm": 0.0736880972981453, |
| "learning_rate": 0.00010114942528735633, |
| "loss": 0.911, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.7731958762886598, |
| "eval_loss": 1.4201780557632446, |
| "eval_runtime": 12.2248, |
| "eval_samples_per_second": 11.043, |
| "eval_steps_per_second": 0.736, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.7835051546391752, |
| "grad_norm": 0.05896177887916565, |
| "learning_rate": 9.655172413793105e-05, |
| "loss": 0.9412, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.7835051546391752, |
| "eval_loss": 1.4200078248977661, |
| "eval_runtime": 12.1599, |
| "eval_samples_per_second": 11.102, |
| "eval_steps_per_second": 0.74, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.7938144329896907, |
| "grad_norm": 0.06385839730501175, |
| "learning_rate": 9.195402298850575e-05, |
| "loss": 0.8999, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.7938144329896907, |
| "eval_loss": 1.4164679050445557, |
| "eval_runtime": 12.1435, |
| "eval_samples_per_second": 11.117, |
| "eval_steps_per_second": 0.741, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.8041237113402062, |
| "grad_norm": 0.0656963661313057, |
| "learning_rate": 8.735632183908047e-05, |
| "loss": 0.8924, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.8041237113402062, |
| "eval_loss": 1.4131464958190918, |
| "eval_runtime": 12.147, |
| "eval_samples_per_second": 11.114, |
| "eval_steps_per_second": 0.741, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.8144329896907216, |
| "grad_norm": 0.07376889884471893, |
| "learning_rate": 8.275862068965517e-05, |
| "loss": 0.9304, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.8144329896907216, |
| "eval_loss": 1.4098708629608154, |
| "eval_runtime": 12.1509, |
| "eval_samples_per_second": 11.11, |
| "eval_steps_per_second": 0.741, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.8247422680412371, |
| "grad_norm": 0.06411939859390259, |
| "learning_rate": 7.81609195402299e-05, |
| "loss": 0.9216, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.8247422680412371, |
| "eval_loss": 1.4088366031646729, |
| "eval_runtime": 12.1696, |
| "eval_samples_per_second": 11.093, |
| "eval_steps_per_second": 0.74, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.8350515463917526, |
| "grad_norm": 0.06034550443291664, |
| "learning_rate": 7.35632183908046e-05, |
| "loss": 0.8914, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.8350515463917526, |
| "eval_loss": 1.4062119722366333, |
| "eval_runtime": 12.1085, |
| "eval_samples_per_second": 11.149, |
| "eval_steps_per_second": 0.743, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.845360824742268, |
| "grad_norm": 0.06504890322685242, |
| "learning_rate": 6.896551724137931e-05, |
| "loss": 0.9608, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.845360824742268, |
| "eval_loss": 1.4031813144683838, |
| "eval_runtime": 12.1849, |
| "eval_samples_per_second": 11.079, |
| "eval_steps_per_second": 0.739, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.8556701030927835, |
| "grad_norm": 0.05961364135146141, |
| "learning_rate": 6.436781609195403e-05, |
| "loss": 0.8992, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.8556701030927835, |
| "eval_loss": 1.4021321535110474, |
| "eval_runtime": 12.2033, |
| "eval_samples_per_second": 11.063, |
| "eval_steps_per_second": 0.738, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.865979381443299, |
| "grad_norm": 0.06472059339284897, |
| "learning_rate": 5.977011494252874e-05, |
| "loss": 0.9458, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.865979381443299, |
| "eval_loss": 1.3988347053527832, |
| "eval_runtime": 12.169, |
| "eval_samples_per_second": 11.094, |
| "eval_steps_per_second": 0.74, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.8762886597938144, |
| "grad_norm": 0.05986656993627548, |
| "learning_rate": 5.517241379310345e-05, |
| "loss": 0.8628, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.8762886597938144, |
| "eval_loss": 1.3969188928604126, |
| "eval_runtime": 12.2292, |
| "eval_samples_per_second": 11.039, |
| "eval_steps_per_second": 0.736, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.8865979381443299, |
| "grad_norm": 0.062148451805114746, |
| "learning_rate": 5.057471264367817e-05, |
| "loss": 0.8841, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.8865979381443299, |
| "eval_loss": 1.3965938091278076, |
| "eval_runtime": 12.1883, |
| "eval_samples_per_second": 11.076, |
| "eval_steps_per_second": 0.738, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.8969072164948454, |
| "grad_norm": 0.05559258908033371, |
| "learning_rate": 4.597701149425287e-05, |
| "loss": 0.8883, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.8969072164948454, |
| "eval_loss": 1.3965078592300415, |
| "eval_runtime": 12.1185, |
| "eval_samples_per_second": 11.14, |
| "eval_steps_per_second": 0.743, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.9072164948453608, |
| "grad_norm": 0.05684094876050949, |
| "learning_rate": 4.1379310344827587e-05, |
| "loss": 0.8765, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.9072164948453608, |
| "eval_loss": 1.3967227935791016, |
| "eval_runtime": 12.1899, |
| "eval_samples_per_second": 11.075, |
| "eval_steps_per_second": 0.738, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.9175257731958762, |
| "grad_norm": 0.05952519550919533, |
| "learning_rate": 3.67816091954023e-05, |
| "loss": 0.8598, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.9175257731958762, |
| "eval_loss": 1.3951915502548218, |
| "eval_runtime": 12.1745, |
| "eval_samples_per_second": 11.089, |
| "eval_steps_per_second": 0.739, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.9278350515463918, |
| "grad_norm": 0.06364478170871735, |
| "learning_rate": 3.218390804597701e-05, |
| "loss": 0.9653, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.9278350515463918, |
| "eval_loss": 1.394257664680481, |
| "eval_runtime": 12.2896, |
| "eval_samples_per_second": 10.985, |
| "eval_steps_per_second": 0.732, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.9381443298969072, |
| "grad_norm": 0.06441052258014679, |
| "learning_rate": 2.7586206896551727e-05, |
| "loss": 0.9397, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.9381443298969072, |
| "eval_loss": 1.392314076423645, |
| "eval_runtime": 12.2278, |
| "eval_samples_per_second": 11.04, |
| "eval_steps_per_second": 0.736, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.9484536082474226, |
| "grad_norm": 0.06320352107286453, |
| "learning_rate": 2.2988505747126437e-05, |
| "loss": 0.8635, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.9484536082474226, |
| "eval_loss": 1.3919230699539185, |
| "eval_runtime": 12.1765, |
| "eval_samples_per_second": 11.087, |
| "eval_steps_per_second": 0.739, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.9587628865979382, |
| "grad_norm": 0.062386397272348404, |
| "learning_rate": 1.839080459770115e-05, |
| "loss": 0.9257, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.9587628865979382, |
| "eval_loss": 1.3923670053482056, |
| "eval_runtime": 12.2566, |
| "eval_samples_per_second": 11.014, |
| "eval_steps_per_second": 0.734, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.9690721649484536, |
| "grad_norm": 0.05672856792807579, |
| "learning_rate": 1.3793103448275863e-05, |
| "loss": 0.8754, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.9690721649484536, |
| "eval_loss": 1.393159031867981, |
| "eval_runtime": 12.1859, |
| "eval_samples_per_second": 11.078, |
| "eval_steps_per_second": 0.739, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.979381443298969, |
| "grad_norm": 0.06344141811132431, |
| "learning_rate": 9.195402298850575e-06, |
| "loss": 0.9454, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.979381443298969, |
| "eval_loss": 1.3938028812408447, |
| "eval_runtime": 12.1382, |
| "eval_samples_per_second": 11.122, |
| "eval_steps_per_second": 0.741, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.9896907216494846, |
| "grad_norm": 0.06258992105722427, |
| "learning_rate": 4.5977011494252875e-06, |
| "loss": 0.9112, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.9896907216494846, |
| "eval_loss": 1.394579529762268, |
| "eval_runtime": 12.1768, |
| "eval_samples_per_second": 11.087, |
| "eval_steps_per_second": 0.739, |
| "step": 96 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.08749664574861526, |
| "learning_rate": 0.0, |
| "loss": 0.8526, |
| "step": 97 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 1.3950037956237793, |
| "eval_runtime": 12.1878, |
| "eval_samples_per_second": 11.077, |
| "eval_steps_per_second": 0.738, |
| "step": 97 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 97, |
| "total_flos": 3.573314566697779e+16, |
| "train_loss": 1.1783372968742527, |
| "train_runtime": 2093.8635, |
| "train_samples_per_second": 1.472, |
| "train_steps_per_second": 0.046 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 97, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 10, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.573314566697779e+16, |
| "train_batch_size": 32, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|