| { |
| "best_global_step": 1400, |
| "best_metric": 0.5561918616294861, |
| "best_model_checkpoint": ".../training_output/checkpoint-1400", |
| "epoch": 10.0, |
| "eval_steps": 50, |
| "global_step": 1420, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.07042253521126761, |
| "grad_norm": 4.834911346435547, |
| "learning_rate": 6.338028169014085e-07, |
| "loss": 3.8016, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.14084507042253522, |
| "grad_norm": 3.7942707538604736, |
| "learning_rate": 1.3380281690140844e-06, |
| "loss": 3.7424, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.2112676056338028, |
| "grad_norm": 4.044854164123535, |
| "learning_rate": 2.0422535211267608e-06, |
| "loss": 3.6583, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.28169014084507044, |
| "grad_norm": 4.419040679931641, |
| "learning_rate": 2.746478873239437e-06, |
| "loss": 3.7144, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.352112676056338, |
| "grad_norm": 4.994226455688477, |
| "learning_rate": 3.4507042253521127e-06, |
| "loss": 3.5497, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.352112676056338, |
| "eval_loss": 1.7768433094024658, |
| "eval_runtime": 10.0717, |
| "eval_samples_per_second": 95.217, |
| "eval_steps_per_second": 5.957, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.4225352112676056, |
| "grad_norm": 4.968951225280762, |
| "learning_rate": 4.154929577464789e-06, |
| "loss": 3.6716, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.49295774647887325, |
| "grad_norm": 4.8160014152526855, |
| "learning_rate": 4.859154929577465e-06, |
| "loss": 3.5206, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.5633802816901409, |
| "grad_norm": 5.351637363433838, |
| "learning_rate": 5.563380281690142e-06, |
| "loss": 3.4553, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.6338028169014085, |
| "grad_norm": 5.762838363647461, |
| "learning_rate": 6.267605633802818e-06, |
| "loss": 3.4572, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.704225352112676, |
| "grad_norm": 5.674283504486084, |
| "learning_rate": 6.9718309859154935e-06, |
| "loss": 3.4715, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.704225352112676, |
| "eval_loss": 1.704325556755066, |
| "eval_runtime": 9.3452, |
| "eval_samples_per_second": 102.619, |
| "eval_steps_per_second": 6.42, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.7746478873239436, |
| "grad_norm": 5.651138782501221, |
| "learning_rate": 7.67605633802817e-06, |
| "loss": 3.4693, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.8450704225352113, |
| "grad_norm": 5.866046905517578, |
| "learning_rate": 8.380281690140846e-06, |
| "loss": 3.482, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.9154929577464789, |
| "grad_norm": 6.652071475982666, |
| "learning_rate": 9.084507042253522e-06, |
| "loss": 3.4296, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.9859154929577465, |
| "grad_norm": 8.077320098876953, |
| "learning_rate": 9.788732394366198e-06, |
| "loss": 3.3627, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.056338028169014, |
| "grad_norm": 6.8231329917907715, |
| "learning_rate": 9.945226917057903e-06, |
| "loss": 3.3338, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.056338028169014, |
| "eval_loss": 1.6177195310592651, |
| "eval_runtime": 9.306, |
| "eval_samples_per_second": 103.052, |
| "eval_steps_per_second": 6.447, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.1267605633802817, |
| "grad_norm": 7.225569725036621, |
| "learning_rate": 9.866979655712052e-06, |
| "loss": 3.2974, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.1971830985915493, |
| "grad_norm": 7.667593479156494, |
| "learning_rate": 9.788732394366198e-06, |
| "loss": 3.3756, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.267605633802817, |
| "grad_norm": 7.092339992523193, |
| "learning_rate": 9.710485133020345e-06, |
| "loss": 3.3302, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.3380281690140845, |
| "grad_norm": 7.544447898864746, |
| "learning_rate": 9.632237871674493e-06, |
| "loss": 3.2841, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.408450704225352, |
| "grad_norm": 8.606858253479004, |
| "learning_rate": 9.55399061032864e-06, |
| "loss": 3.14, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.408450704225352, |
| "eval_loss": 1.5555429458618164, |
| "eval_runtime": 8.9418, |
| "eval_samples_per_second": 107.249, |
| "eval_steps_per_second": 6.71, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.4788732394366197, |
| "grad_norm": 9.349486351013184, |
| "learning_rate": 9.475743348982786e-06, |
| "loss": 3.2447, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.5492957746478875, |
| "grad_norm": 10.120262145996094, |
| "learning_rate": 9.397496087636934e-06, |
| "loss": 3.301, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.619718309859155, |
| "grad_norm": 8.683691024780273, |
| "learning_rate": 9.31924882629108e-06, |
| "loss": 3.2738, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.6901408450704225, |
| "grad_norm": 8.391075134277344, |
| "learning_rate": 9.241001564945227e-06, |
| "loss": 3.1474, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.76056338028169, |
| "grad_norm": 9.799640655517578, |
| "learning_rate": 9.162754303599375e-06, |
| "loss": 3.1417, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.76056338028169, |
| "eval_loss": 1.5090895891189575, |
| "eval_runtime": 9.059, |
| "eval_samples_per_second": 105.862, |
| "eval_steps_per_second": 6.623, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.8309859154929577, |
| "grad_norm": 9.724505424499512, |
| "learning_rate": 9.084507042253522e-06, |
| "loss": 3.1931, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.9014084507042255, |
| "grad_norm": 11.591618537902832, |
| "learning_rate": 9.006259780907669e-06, |
| "loss": 3.1649, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.971830985915493, |
| "grad_norm": 9.90600299835205, |
| "learning_rate": 8.928012519561817e-06, |
| "loss": 3.2231, |
| "step": 280 |
| }, |
| { |
| "epoch": 2.0422535211267605, |
| "grad_norm": 10.291154861450195, |
| "learning_rate": 8.849765258215963e-06, |
| "loss": 3.1172, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.112676056338028, |
| "grad_norm": 11.620966911315918, |
| "learning_rate": 8.77151799687011e-06, |
| "loss": 3.0193, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.112676056338028, |
| "eval_loss": 1.437499761581421, |
| "eval_runtime": 8.9265, |
| "eval_samples_per_second": 107.433, |
| "eval_steps_per_second": 6.722, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.183098591549296, |
| "grad_norm": 12.65719223022461, |
| "learning_rate": 8.693270735524258e-06, |
| "loss": 2.9451, |
| "step": 310 |
| }, |
| { |
| "epoch": 2.2535211267605635, |
| "grad_norm": 12.792698860168457, |
| "learning_rate": 8.615023474178405e-06, |
| "loss": 2.9669, |
| "step": 320 |
| }, |
| { |
| "epoch": 2.323943661971831, |
| "grad_norm": 12.682389259338379, |
| "learning_rate": 8.536776212832551e-06, |
| "loss": 2.9739, |
| "step": 330 |
| }, |
| { |
| "epoch": 2.3943661971830985, |
| "grad_norm": 12.96263313293457, |
| "learning_rate": 8.4585289514867e-06, |
| "loss": 2.8988, |
| "step": 340 |
| }, |
| { |
| "epoch": 2.464788732394366, |
| "grad_norm": 14.160724639892578, |
| "learning_rate": 8.380281690140846e-06, |
| "loss": 2.9009, |
| "step": 350 |
| }, |
| { |
| "epoch": 2.464788732394366, |
| "eval_loss": 1.3772871494293213, |
| "eval_runtime": 9.6484, |
| "eval_samples_per_second": 99.394, |
| "eval_steps_per_second": 6.219, |
| "step": 350 |
| }, |
| { |
| "epoch": 2.535211267605634, |
| "grad_norm": 14.379948616027832, |
| "learning_rate": 8.302034428794992e-06, |
| "loss": 2.8597, |
| "step": 360 |
| }, |
| { |
| "epoch": 2.6056338028169015, |
| "grad_norm": 15.024813652038574, |
| "learning_rate": 8.22378716744914e-06, |
| "loss": 2.8603, |
| "step": 370 |
| }, |
| { |
| "epoch": 2.676056338028169, |
| "grad_norm": 15.947714805603027, |
| "learning_rate": 8.145539906103287e-06, |
| "loss": 2.9262, |
| "step": 380 |
| }, |
| { |
| "epoch": 2.7464788732394365, |
| "grad_norm": 15.812085151672363, |
| "learning_rate": 8.067292644757434e-06, |
| "loss": 2.9301, |
| "step": 390 |
| }, |
| { |
| "epoch": 2.816901408450704, |
| "grad_norm": 15.922052383422852, |
| "learning_rate": 7.989045383411582e-06, |
| "loss": 2.9446, |
| "step": 400 |
| }, |
| { |
| "epoch": 2.816901408450704, |
| "eval_loss": 1.3113040924072266, |
| "eval_runtime": 8.9464, |
| "eval_samples_per_second": 107.194, |
| "eval_steps_per_second": 6.707, |
| "step": 400 |
| }, |
| { |
| "epoch": 2.887323943661972, |
| "grad_norm": 15.517271995544434, |
| "learning_rate": 7.910798122065728e-06, |
| "loss": 2.9103, |
| "step": 410 |
| }, |
| { |
| "epoch": 2.9577464788732395, |
| "grad_norm": 16.0577449798584, |
| "learning_rate": 7.832550860719875e-06, |
| "loss": 2.88, |
| "step": 420 |
| }, |
| { |
| "epoch": 3.028169014084507, |
| "grad_norm": 15.434414863586426, |
| "learning_rate": 7.754303599374023e-06, |
| "loss": 2.7008, |
| "step": 430 |
| }, |
| { |
| "epoch": 3.0985915492957745, |
| "grad_norm": 18.000938415527344, |
| "learning_rate": 7.67605633802817e-06, |
| "loss": 2.669, |
| "step": 440 |
| }, |
| { |
| "epoch": 3.169014084507042, |
| "grad_norm": 17.124902725219727, |
| "learning_rate": 7.597809076682316e-06, |
| "loss": 2.6466, |
| "step": 450 |
| }, |
| { |
| "epoch": 3.169014084507042, |
| "eval_loss": 1.2235872745513916, |
| "eval_runtime": 8.9632, |
| "eval_samples_per_second": 106.993, |
| "eval_steps_per_second": 6.694, |
| "step": 450 |
| }, |
| { |
| "epoch": 3.23943661971831, |
| "grad_norm": 18.728208541870117, |
| "learning_rate": 7.5195618153364636e-06, |
| "loss": 2.6583, |
| "step": 460 |
| }, |
| { |
| "epoch": 3.3098591549295775, |
| "grad_norm": 20.174840927124023, |
| "learning_rate": 7.441314553990611e-06, |
| "loss": 2.5912, |
| "step": 470 |
| }, |
| { |
| "epoch": 3.380281690140845, |
| "grad_norm": 19.468631744384766, |
| "learning_rate": 7.3630672926447575e-06, |
| "loss": 2.5551, |
| "step": 480 |
| }, |
| { |
| "epoch": 3.4507042253521125, |
| "grad_norm": 19.333127975463867, |
| "learning_rate": 7.284820031298905e-06, |
| "loss": 2.5759, |
| "step": 490 |
| }, |
| { |
| "epoch": 3.52112676056338, |
| "grad_norm": 20.050857543945312, |
| "learning_rate": 7.206572769953052e-06, |
| "loss": 2.6059, |
| "step": 500 |
| }, |
| { |
| "epoch": 3.52112676056338, |
| "eval_loss": 1.1633683443069458, |
| "eval_runtime": 8.9644, |
| "eval_samples_per_second": 106.979, |
| "eval_steps_per_second": 6.693, |
| "step": 500 |
| }, |
| { |
| "epoch": 3.591549295774648, |
| "grad_norm": 21.03053855895996, |
| "learning_rate": 7.128325508607199e-06, |
| "loss": 2.5128, |
| "step": 510 |
| }, |
| { |
| "epoch": 3.6619718309859155, |
| "grad_norm": 20.328510284423828, |
| "learning_rate": 7.050078247261346e-06, |
| "loss": 2.5397, |
| "step": 520 |
| }, |
| { |
| "epoch": 3.732394366197183, |
| "grad_norm": 21.760059356689453, |
| "learning_rate": 6.9718309859154935e-06, |
| "loss": 2.5198, |
| "step": 530 |
| }, |
| { |
| "epoch": 3.802816901408451, |
| "grad_norm": 20.767467498779297, |
| "learning_rate": 6.89358372456964e-06, |
| "loss": 2.4605, |
| "step": 540 |
| }, |
| { |
| "epoch": 3.873239436619718, |
| "grad_norm": 24.69249725341797, |
| "learning_rate": 6.815336463223787e-06, |
| "loss": 2.4911, |
| "step": 550 |
| }, |
| { |
| "epoch": 3.873239436619718, |
| "eval_loss": 1.0876926183700562, |
| "eval_runtime": 8.9503, |
| "eval_samples_per_second": 107.147, |
| "eval_steps_per_second": 6.704, |
| "step": 550 |
| }, |
| { |
| "epoch": 3.943661971830986, |
| "grad_norm": 21.451522827148438, |
| "learning_rate": 6.737089201877935e-06, |
| "loss": 2.445, |
| "step": 560 |
| }, |
| { |
| "epoch": 4.014084507042254, |
| "grad_norm": 21.80568504333496, |
| "learning_rate": 6.658841940532081e-06, |
| "loss": 2.4261, |
| "step": 570 |
| }, |
| { |
| "epoch": 4.084507042253521, |
| "grad_norm": 23.770599365234375, |
| "learning_rate": 6.580594679186229e-06, |
| "loss": 2.2965, |
| "step": 580 |
| }, |
| { |
| "epoch": 4.154929577464789, |
| "grad_norm": 22.17987060546875, |
| "learning_rate": 6.502347417840375e-06, |
| "loss": 2.1889, |
| "step": 590 |
| }, |
| { |
| "epoch": 4.225352112676056, |
| "grad_norm": 22.51678466796875, |
| "learning_rate": 6.424100156494523e-06, |
| "loss": 2.2193, |
| "step": 600 |
| }, |
| { |
| "epoch": 4.225352112676056, |
| "eval_loss": 1.049774408340454, |
| "eval_runtime": 8.9455, |
| "eval_samples_per_second": 107.205, |
| "eval_steps_per_second": 6.707, |
| "step": 600 |
| }, |
| { |
| "epoch": 4.295774647887324, |
| "grad_norm": 23.58041000366211, |
| "learning_rate": 6.345852895148671e-06, |
| "loss": 2.2024, |
| "step": 610 |
| }, |
| { |
| "epoch": 4.366197183098592, |
| "grad_norm": 24.69291877746582, |
| "learning_rate": 6.267605633802818e-06, |
| "loss": 2.2096, |
| "step": 620 |
| }, |
| { |
| "epoch": 4.436619718309859, |
| "grad_norm": 24.199907302856445, |
| "learning_rate": 6.189358372456965e-06, |
| "loss": 2.2598, |
| "step": 630 |
| }, |
| { |
| "epoch": 4.507042253521127, |
| "grad_norm": 23.92287826538086, |
| "learning_rate": 6.111111111111112e-06, |
| "loss": 2.2322, |
| "step": 640 |
| }, |
| { |
| "epoch": 4.577464788732394, |
| "grad_norm": 22.647499084472656, |
| "learning_rate": 6.0328638497652595e-06, |
| "loss": 2.1822, |
| "step": 650 |
| }, |
| { |
| "epoch": 4.577464788732394, |
| "eval_loss": 0.9610461592674255, |
| "eval_runtime": 8.9698, |
| "eval_samples_per_second": 106.915, |
| "eval_steps_per_second": 6.689, |
| "step": 650 |
| }, |
| { |
| "epoch": 4.647887323943662, |
| "grad_norm": 23.590543746948242, |
| "learning_rate": 5.954616588419406e-06, |
| "loss": 2.1621, |
| "step": 660 |
| }, |
| { |
| "epoch": 4.71830985915493, |
| "grad_norm": 23.479148864746094, |
| "learning_rate": 5.876369327073553e-06, |
| "loss": 2.1543, |
| "step": 670 |
| }, |
| { |
| "epoch": 4.788732394366197, |
| "grad_norm": 26.514415740966797, |
| "learning_rate": 5.798122065727701e-06, |
| "loss": 2.2376, |
| "step": 680 |
| }, |
| { |
| "epoch": 4.859154929577465, |
| "grad_norm": 25.98239517211914, |
| "learning_rate": 5.719874804381847e-06, |
| "loss": 2.1652, |
| "step": 690 |
| }, |
| { |
| "epoch": 4.929577464788732, |
| "grad_norm": 25.147550582885742, |
| "learning_rate": 5.641627543035995e-06, |
| "loss": 2.1194, |
| "step": 700 |
| }, |
| { |
| "epoch": 4.929577464788732, |
| "eval_loss": 0.8909062743186951, |
| "eval_runtime": 8.9807, |
| "eval_samples_per_second": 106.785, |
| "eval_steps_per_second": 6.681, |
| "step": 700 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 24.74693489074707, |
| "learning_rate": 5.563380281690142e-06, |
| "loss": 2.0919, |
| "step": 710 |
| }, |
| { |
| "epoch": 5.070422535211268, |
| "grad_norm": 27.079837799072266, |
| "learning_rate": 5.4851330203442885e-06, |
| "loss": 1.9779, |
| "step": 720 |
| }, |
| { |
| "epoch": 5.140845070422535, |
| "grad_norm": 25.947982788085938, |
| "learning_rate": 5.406885758998436e-06, |
| "loss": 1.8619, |
| "step": 730 |
| }, |
| { |
| "epoch": 5.211267605633803, |
| "grad_norm": 27.102828979492188, |
| "learning_rate": 5.328638497652583e-06, |
| "loss": 1.9172, |
| "step": 740 |
| }, |
| { |
| "epoch": 5.28169014084507, |
| "grad_norm": 27.407745361328125, |
| "learning_rate": 5.25039123630673e-06, |
| "loss": 1.9355, |
| "step": 750 |
| }, |
| { |
| "epoch": 5.28169014084507, |
| "eval_loss": 0.8482040166854858, |
| "eval_runtime": 9.8318, |
| "eval_samples_per_second": 97.54, |
| "eval_steps_per_second": 6.103, |
| "step": 750 |
| }, |
| { |
| "epoch": 5.352112676056338, |
| "grad_norm": 24.58213233947754, |
| "learning_rate": 5.172143974960877e-06, |
| "loss": 1.9393, |
| "step": 760 |
| }, |
| { |
| "epoch": 5.422535211267606, |
| "grad_norm": 24.623889923095703, |
| "learning_rate": 5.093896713615024e-06, |
| "loss": 1.8727, |
| "step": 770 |
| }, |
| { |
| "epoch": 5.492957746478873, |
| "grad_norm": 26.396087646484375, |
| "learning_rate": 5.015649452269171e-06, |
| "loss": 1.9599, |
| "step": 780 |
| }, |
| { |
| "epoch": 5.563380281690141, |
| "grad_norm": 28.786396026611328, |
| "learning_rate": 4.9374021909233185e-06, |
| "loss": 1.8271, |
| "step": 790 |
| }, |
| { |
| "epoch": 5.633802816901408, |
| "grad_norm": 25.507366180419922, |
| "learning_rate": 4.859154929577465e-06, |
| "loss": 1.9065, |
| "step": 800 |
| }, |
| { |
| "epoch": 5.633802816901408, |
| "eval_loss": 0.8195229768753052, |
| "eval_runtime": 8.968, |
| "eval_samples_per_second": 106.935, |
| "eval_steps_per_second": 6.69, |
| "step": 800 |
| }, |
| { |
| "epoch": 5.704225352112676, |
| "grad_norm": 26.951005935668945, |
| "learning_rate": 4.780907668231612e-06, |
| "loss": 1.8742, |
| "step": 810 |
| }, |
| { |
| "epoch": 5.774647887323944, |
| "grad_norm": 27.19331169128418, |
| "learning_rate": 4.70266040688576e-06, |
| "loss": 1.8754, |
| "step": 820 |
| }, |
| { |
| "epoch": 5.845070422535211, |
| "grad_norm": 26.17990493774414, |
| "learning_rate": 4.624413145539906e-06, |
| "loss": 1.8412, |
| "step": 830 |
| }, |
| { |
| "epoch": 5.915492957746479, |
| "grad_norm": 28.198486328125, |
| "learning_rate": 4.546165884194054e-06, |
| "loss": 1.845, |
| "step": 840 |
| }, |
| { |
| "epoch": 5.985915492957746, |
| "grad_norm": 24.81572914123535, |
| "learning_rate": 4.467918622848201e-06, |
| "loss": 1.8035, |
| "step": 850 |
| }, |
| { |
| "epoch": 5.985915492957746, |
| "eval_loss": 0.7713097929954529, |
| "eval_runtime": 8.914, |
| "eval_samples_per_second": 107.583, |
| "eval_steps_per_second": 6.731, |
| "step": 850 |
| }, |
| { |
| "epoch": 6.056338028169014, |
| "grad_norm": 27.168418884277344, |
| "learning_rate": 4.3896713615023476e-06, |
| "loss": 1.7062, |
| "step": 860 |
| }, |
| { |
| "epoch": 6.126760563380282, |
| "grad_norm": 25.584774017333984, |
| "learning_rate": 4.311424100156495e-06, |
| "loss": 1.6908, |
| "step": 870 |
| }, |
| { |
| "epoch": 6.197183098591549, |
| "grad_norm": 27.289413452148438, |
| "learning_rate": 4.233176838810642e-06, |
| "loss": 1.6868, |
| "step": 880 |
| }, |
| { |
| "epoch": 6.267605633802817, |
| "grad_norm": 26.789995193481445, |
| "learning_rate": 4.154929577464789e-06, |
| "loss": 1.686, |
| "step": 890 |
| }, |
| { |
| "epoch": 6.338028169014084, |
| "grad_norm": 27.878765106201172, |
| "learning_rate": 4.076682316118936e-06, |
| "loss": 1.564, |
| "step": 900 |
| }, |
| { |
| "epoch": 6.338028169014084, |
| "eval_loss": 0.7355391383171082, |
| "eval_runtime": 8.9357, |
| "eval_samples_per_second": 107.322, |
| "eval_steps_per_second": 6.715, |
| "step": 900 |
| }, |
| { |
| "epoch": 6.408450704225352, |
| "grad_norm": 27.236083984375, |
| "learning_rate": 3.998435054773084e-06, |
| "loss": 1.7135, |
| "step": 910 |
| }, |
| { |
| "epoch": 6.47887323943662, |
| "grad_norm": 26.007850646972656, |
| "learning_rate": 3.92018779342723e-06, |
| "loss": 1.6614, |
| "step": 920 |
| }, |
| { |
| "epoch": 6.549295774647887, |
| "grad_norm": 26.08616065979004, |
| "learning_rate": 3.8419405320813775e-06, |
| "loss": 1.6442, |
| "step": 930 |
| }, |
| { |
| "epoch": 6.619718309859155, |
| "grad_norm": 26.021728515625, |
| "learning_rate": 3.7636932707355245e-06, |
| "loss": 1.6117, |
| "step": 940 |
| }, |
| { |
| "epoch": 6.690140845070422, |
| "grad_norm": 26.019577026367188, |
| "learning_rate": 3.6854460093896714e-06, |
| "loss": 1.6317, |
| "step": 950 |
| }, |
| { |
| "epoch": 6.690140845070422, |
| "eval_loss": 0.700791597366333, |
| "eval_runtime": 8.9228, |
| "eval_samples_per_second": 107.477, |
| "eval_steps_per_second": 6.724, |
| "step": 950 |
| }, |
| { |
| "epoch": 6.76056338028169, |
| "grad_norm": 25.988859176635742, |
| "learning_rate": 3.6071987480438188e-06, |
| "loss": 1.6589, |
| "step": 960 |
| }, |
| { |
| "epoch": 6.830985915492958, |
| "grad_norm": 26.973453521728516, |
| "learning_rate": 3.5289514866979657e-06, |
| "loss": 1.673, |
| "step": 970 |
| }, |
| { |
| "epoch": 6.901408450704225, |
| "grad_norm": 27.36475372314453, |
| "learning_rate": 3.4507042253521127e-06, |
| "loss": 1.6018, |
| "step": 980 |
| }, |
| { |
| "epoch": 6.971830985915493, |
| "grad_norm": 28.897504806518555, |
| "learning_rate": 3.37245696400626e-06, |
| "loss": 1.6291, |
| "step": 990 |
| }, |
| { |
| "epoch": 7.042253521126761, |
| "grad_norm": 26.3133544921875, |
| "learning_rate": 3.294209702660407e-06, |
| "loss": 1.5407, |
| "step": 1000 |
| }, |
| { |
| "epoch": 7.042253521126761, |
| "eval_loss": 0.6679695248603821, |
| "eval_runtime": 8.9256, |
| "eval_samples_per_second": 107.444, |
| "eval_steps_per_second": 6.722, |
| "step": 1000 |
| }, |
| { |
| "epoch": 7.112676056338028, |
| "grad_norm": 29.09581184387207, |
| "learning_rate": 3.215962441314554e-06, |
| "loss": 1.4842, |
| "step": 1010 |
| }, |
| { |
| "epoch": 7.183098591549296, |
| "grad_norm": 27.576128005981445, |
| "learning_rate": 3.137715179968701e-06, |
| "loss": 1.5063, |
| "step": 1020 |
| }, |
| { |
| "epoch": 7.253521126760563, |
| "grad_norm": 29.655752182006836, |
| "learning_rate": 3.0594679186228483e-06, |
| "loss": 1.5052, |
| "step": 1030 |
| }, |
| { |
| "epoch": 7.323943661971831, |
| "grad_norm": 25.104883193969727, |
| "learning_rate": 2.9812206572769952e-06, |
| "loss": 1.42, |
| "step": 1040 |
| }, |
| { |
| "epoch": 7.394366197183099, |
| "grad_norm": 27.621112823486328, |
| "learning_rate": 2.902973395931143e-06, |
| "loss": 1.4671, |
| "step": 1050 |
| }, |
| { |
| "epoch": 7.394366197183099, |
| "eval_loss": 0.6450517177581787, |
| "eval_runtime": 8.9546, |
| "eval_samples_per_second": 107.096, |
| "eval_steps_per_second": 6.7, |
| "step": 1050 |
| }, |
| { |
| "epoch": 7.464788732394366, |
| "grad_norm": 27.425331115722656, |
| "learning_rate": 2.82472613458529e-06, |
| "loss": 1.4879, |
| "step": 1060 |
| }, |
| { |
| "epoch": 7.535211267605634, |
| "grad_norm": 27.76177978515625, |
| "learning_rate": 2.746478873239437e-06, |
| "loss": 1.5012, |
| "step": 1070 |
| }, |
| { |
| "epoch": 7.605633802816901, |
| "grad_norm": 24.934165954589844, |
| "learning_rate": 2.6682316118935843e-06, |
| "loss": 1.4386, |
| "step": 1080 |
| }, |
| { |
| "epoch": 7.676056338028169, |
| "grad_norm": 28.298328399658203, |
| "learning_rate": 2.5899843505477313e-06, |
| "loss": 1.4544, |
| "step": 1090 |
| }, |
| { |
| "epoch": 7.746478873239437, |
| "grad_norm": 24.99042320251465, |
| "learning_rate": 2.5117370892018782e-06, |
| "loss": 1.4076, |
| "step": 1100 |
| }, |
| { |
| "epoch": 7.746478873239437, |
| "eval_loss": 0.6194617748260498, |
| "eval_runtime": 8.9294, |
| "eval_samples_per_second": 107.398, |
| "eval_steps_per_second": 6.719, |
| "step": 1100 |
| }, |
| { |
| "epoch": 7.816901408450704, |
| "grad_norm": 27.29831314086914, |
| "learning_rate": 2.433489827856025e-06, |
| "loss": 1.4838, |
| "step": 1110 |
| }, |
| { |
| "epoch": 7.887323943661972, |
| "grad_norm": 27.53099250793457, |
| "learning_rate": 2.3552425665101726e-06, |
| "loss": 1.4745, |
| "step": 1120 |
| }, |
| { |
| "epoch": 7.957746478873239, |
| "grad_norm": 27.09596824645996, |
| "learning_rate": 2.2769953051643195e-06, |
| "loss": 1.3781, |
| "step": 1130 |
| }, |
| { |
| "epoch": 8.028169014084508, |
| "grad_norm": 26.090456008911133, |
| "learning_rate": 2.1987480438184665e-06, |
| "loss": 1.422, |
| "step": 1140 |
| }, |
| { |
| "epoch": 8.098591549295774, |
| "grad_norm": 26.628990173339844, |
| "learning_rate": 2.120500782472614e-06, |
| "loss": 1.345, |
| "step": 1150 |
| }, |
| { |
| "epoch": 8.098591549295774, |
| "eval_loss": 0.603298008441925, |
| "eval_runtime": 8.923, |
| "eval_samples_per_second": 107.475, |
| "eval_steps_per_second": 6.724, |
| "step": 1150 |
| }, |
| { |
| "epoch": 8.169014084507042, |
| "grad_norm": 26.968416213989258, |
| "learning_rate": 2.0422535211267608e-06, |
| "loss": 1.4478, |
| "step": 1160 |
| }, |
| { |
| "epoch": 8.23943661971831, |
| "grad_norm": 28.169918060302734, |
| "learning_rate": 1.9640062597809077e-06, |
| "loss": 1.3111, |
| "step": 1170 |
| }, |
| { |
| "epoch": 8.309859154929578, |
| "grad_norm": 28.084226608276367, |
| "learning_rate": 1.885758998435055e-06, |
| "loss": 1.3754, |
| "step": 1180 |
| }, |
| { |
| "epoch": 8.380281690140846, |
| "grad_norm": 28.451416015625, |
| "learning_rate": 1.807511737089202e-06, |
| "loss": 1.3996, |
| "step": 1190 |
| }, |
| { |
| "epoch": 8.450704225352112, |
| "grad_norm": 27.99500846862793, |
| "learning_rate": 1.729264475743349e-06, |
| "loss": 1.3818, |
| "step": 1200 |
| }, |
| { |
| "epoch": 8.450704225352112, |
| "eval_loss": 0.5862457752227783, |
| "eval_runtime": 8.9231, |
| "eval_samples_per_second": 107.473, |
| "eval_steps_per_second": 6.724, |
| "step": 1200 |
| }, |
| { |
| "epoch": 8.52112676056338, |
| "grad_norm": 26.25287628173828, |
| "learning_rate": 1.6510172143974962e-06, |
| "loss": 1.348, |
| "step": 1210 |
| }, |
| { |
| "epoch": 8.591549295774648, |
| "grad_norm": 27.26688003540039, |
| "learning_rate": 1.5727699530516433e-06, |
| "loss": 1.3567, |
| "step": 1220 |
| }, |
| { |
| "epoch": 8.661971830985916, |
| "grad_norm": 25.22686767578125, |
| "learning_rate": 1.4945226917057903e-06, |
| "loss": 1.331, |
| "step": 1230 |
| }, |
| { |
| "epoch": 8.732394366197184, |
| "grad_norm": 27.34234619140625, |
| "learning_rate": 1.4162754303599375e-06, |
| "loss": 1.2977, |
| "step": 1240 |
| }, |
| { |
| "epoch": 8.80281690140845, |
| "grad_norm": 23.48381233215332, |
| "learning_rate": 1.3380281690140844e-06, |
| "loss": 1.2578, |
| "step": 1250 |
| }, |
| { |
| "epoch": 8.80281690140845, |
| "eval_loss": 0.5742074847221375, |
| "eval_runtime": 8.9308, |
| "eval_samples_per_second": 107.381, |
| "eval_steps_per_second": 6.718, |
| "step": 1250 |
| }, |
| { |
| "epoch": 8.873239436619718, |
| "grad_norm": 25.575672149658203, |
| "learning_rate": 1.2597809076682316e-06, |
| "loss": 1.2807, |
| "step": 1260 |
| }, |
| { |
| "epoch": 8.943661971830986, |
| "grad_norm": 27.154817581176758, |
| "learning_rate": 1.1815336463223787e-06, |
| "loss": 1.3597, |
| "step": 1270 |
| }, |
| { |
| "epoch": 9.014084507042254, |
| "grad_norm": 28.042072296142578, |
| "learning_rate": 1.103286384976526e-06, |
| "loss": 1.3292, |
| "step": 1280 |
| }, |
| { |
| "epoch": 9.084507042253522, |
| "grad_norm": 27.011009216308594, |
| "learning_rate": 1.0250391236306729e-06, |
| "loss": 1.2585, |
| "step": 1290 |
| }, |
| { |
| "epoch": 9.154929577464788, |
| "grad_norm": 27.294761657714844, |
| "learning_rate": 9.467918622848201e-07, |
| "loss": 1.2552, |
| "step": 1300 |
| }, |
| { |
| "epoch": 9.154929577464788, |
| "eval_loss": 0.5647178292274475, |
| "eval_runtime": 8.9528, |
| "eval_samples_per_second": 107.117, |
| "eval_steps_per_second": 6.702, |
| "step": 1300 |
| }, |
| { |
| "epoch": 9.225352112676056, |
| "grad_norm": 25.794416427612305, |
| "learning_rate": 8.685446009389673e-07, |
| "loss": 1.3339, |
| "step": 1310 |
| }, |
| { |
| "epoch": 9.295774647887324, |
| "grad_norm": 26.401464462280273, |
| "learning_rate": 7.902973395931143e-07, |
| "loss": 1.2872, |
| "step": 1320 |
| }, |
| { |
| "epoch": 9.366197183098592, |
| "grad_norm": 28.1634464263916, |
| "learning_rate": 7.120500782472614e-07, |
| "loss": 1.276, |
| "step": 1330 |
| }, |
| { |
| "epoch": 9.43661971830986, |
| "grad_norm": 26.156156539916992, |
| "learning_rate": 6.338028169014085e-07, |
| "loss": 1.2846, |
| "step": 1340 |
| }, |
| { |
| "epoch": 9.507042253521126, |
| "grad_norm": 26.216571807861328, |
| "learning_rate": 5.555555555555555e-07, |
| "loss": 1.2782, |
| "step": 1350 |
| }, |
| { |
| "epoch": 9.507042253521126, |
| "eval_loss": 0.5621166229248047, |
| "eval_runtime": 8.9665, |
| "eval_samples_per_second": 106.953, |
| "eval_steps_per_second": 6.692, |
| "step": 1350 |
| }, |
| { |
| "epoch": 9.577464788732394, |
| "grad_norm": 26.63737678527832, |
| "learning_rate": 4.773082942097027e-07, |
| "loss": 1.2883, |
| "step": 1360 |
| }, |
| { |
| "epoch": 9.647887323943662, |
| "grad_norm": 27.532459259033203, |
| "learning_rate": 3.990610328638498e-07, |
| "loss": 1.2515, |
| "step": 1370 |
| }, |
| { |
| "epoch": 9.71830985915493, |
| "grad_norm": 23.93533706665039, |
| "learning_rate": 3.208137715179969e-07, |
| "loss": 1.3023, |
| "step": 1380 |
| }, |
| { |
| "epoch": 9.788732394366198, |
| "grad_norm": 26.48802375793457, |
| "learning_rate": 2.42566510172144e-07, |
| "loss": 1.3053, |
| "step": 1390 |
| }, |
| { |
| "epoch": 9.859154929577464, |
| "grad_norm": 25.6018123626709, |
| "learning_rate": 1.643192488262911e-07, |
| "loss": 1.2556, |
| "step": 1400 |
| }, |
| { |
| "epoch": 9.859154929577464, |
| "eval_loss": 0.5561918616294861, |
| "eval_runtime": 9.3282, |
| "eval_samples_per_second": 102.806, |
| "eval_steps_per_second": 6.432, |
| "step": 1400 |
| }, |
| { |
| "epoch": 9.929577464788732, |
| "grad_norm": 26.17299461364746, |
| "learning_rate": 8.607198748043818e-08, |
| "loss": 1.2926, |
| "step": 1410 |
| }, |
| { |
| "epoch": 10.0, |
| "grad_norm": 28.140972137451172, |
| "learning_rate": 7.824726134585291e-09, |
| "loss": 1.2199, |
| "step": 1420 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 1420, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 10, |
| "save_steps": 350, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 64, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|