| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 20.217435897435898, |
| "eval_steps": 500, |
| "global_step": 4928, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.020512820512820513, |
| "grad_norm": 0.875, |
| "learning_rate": 2.9999969395283144e-06, |
| "loss": 0.6637, |
| "num_input_tokens_seen": 75456, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.041025641025641026, |
| "grad_norm": 1.4140625, |
| "learning_rate": 2.9999877581257458e-06, |
| "loss": 0.6963, |
| "num_input_tokens_seen": 160992, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.06153846153846154, |
| "grad_norm": 1.59375, |
| "learning_rate": 2.9999724558297605e-06, |
| "loss": 0.6919, |
| "num_input_tokens_seen": 253856, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.08205128205128205, |
| "grad_norm": 2.40625, |
| "learning_rate": 2.999951032702801e-06, |
| "loss": 0.6605, |
| "num_input_tokens_seen": 328640, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.10256410256410256, |
| "grad_norm": 0.75390625, |
| "learning_rate": 2.9999234888322877e-06, |
| "loss": 0.5888, |
| "num_input_tokens_seen": 401504, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.12307692307692308, |
| "grad_norm": 1.578125, |
| "learning_rate": 2.9998898243306162e-06, |
| "loss": 0.6087, |
| "num_input_tokens_seen": 475168, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.14358974358974358, |
| "grad_norm": 0.89453125, |
| "learning_rate": 2.99985003933516e-06, |
| "loss": 0.6405, |
| "num_input_tokens_seen": 547168, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.1641025641025641, |
| "grad_norm": 1.1015625, |
| "learning_rate": 2.999804134008266e-06, |
| "loss": 0.6443, |
| "num_input_tokens_seen": 618176, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.18461538461538463, |
| "grad_norm": 0.69921875, |
| "learning_rate": 2.9997521085372565e-06, |
| "loss": 0.6554, |
| "num_input_tokens_seen": 701184, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.20512820512820512, |
| "grad_norm": 1.0234375, |
| "learning_rate": 2.999693963134429e-06, |
| "loss": 0.6903, |
| "num_input_tokens_seen": 782656, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.22564102564102564, |
| "grad_norm": 0.796875, |
| "learning_rate": 2.9996296980370526e-06, |
| "loss": 0.6915, |
| "num_input_tokens_seen": 869280, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.24615384615384617, |
| "grad_norm": 0.8515625, |
| "learning_rate": 2.99955931350737e-06, |
| "loss": 0.584, |
| "num_input_tokens_seen": 944032, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.26666666666666666, |
| "grad_norm": 0.91796875, |
| "learning_rate": 2.999482809832594e-06, |
| "loss": 0.5908, |
| "num_input_tokens_seen": 1025568, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.28717948717948716, |
| "grad_norm": 0.84375, |
| "learning_rate": 2.9994001873249074e-06, |
| "loss": 0.5811, |
| "num_input_tokens_seen": 1105344, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.3076923076923077, |
| "grad_norm": 0.6953125, |
| "learning_rate": 2.999311446321462e-06, |
| "loss": 0.5792, |
| "num_input_tokens_seen": 1182016, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.3282051282051282, |
| "grad_norm": 0.99609375, |
| "learning_rate": 2.999216587184378e-06, |
| "loss": 0.5923, |
| "num_input_tokens_seen": 1265440, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.3487179487179487, |
| "grad_norm": 0.91015625, |
| "learning_rate": 2.9991156103007394e-06, |
| "loss": 0.5896, |
| "num_input_tokens_seen": 1347296, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.36923076923076925, |
| "grad_norm": 0.90234375, |
| "learning_rate": 2.9990085160825954e-06, |
| "loss": 0.6283, |
| "num_input_tokens_seen": 1424064, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.38974358974358975, |
| "grad_norm": 0.67578125, |
| "learning_rate": 2.9988953049669577e-06, |
| "loss": 0.5329, |
| "num_input_tokens_seen": 1492416, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.41025641025641024, |
| "grad_norm": 0.44921875, |
| "learning_rate": 2.998775977415799e-06, |
| "loss": 0.5582, |
| "num_input_tokens_seen": 1572192, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.4307692307692308, |
| "grad_norm": 0.59375, |
| "learning_rate": 2.998650533916051e-06, |
| "loss": 0.6529, |
| "num_input_tokens_seen": 1650592, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.4512820512820513, |
| "grad_norm": 0.515625, |
| "learning_rate": 2.998518974979602e-06, |
| "loss": 0.5449, |
| "num_input_tokens_seen": 1728128, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.4717948717948718, |
| "grad_norm": 0.494140625, |
| "learning_rate": 2.998381301143295e-06, |
| "loss": 0.573, |
| "num_input_tokens_seen": 1806080, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.49230769230769234, |
| "grad_norm": 0.38671875, |
| "learning_rate": 2.9982375129689253e-06, |
| "loss": 0.6065, |
| "num_input_tokens_seen": 1884544, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.5128205128205128, |
| "grad_norm": 0.396484375, |
| "learning_rate": 2.9980876110432404e-06, |
| "loss": 0.6232, |
| "num_input_tokens_seen": 1961376, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.5333333333333333, |
| "grad_norm": 0.408203125, |
| "learning_rate": 2.9979315959779335e-06, |
| "loss": 0.5061, |
| "num_input_tokens_seen": 2036064, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.5538461538461539, |
| "grad_norm": 0.271484375, |
| "learning_rate": 2.9977694684096447e-06, |
| "loss": 0.5657, |
| "num_input_tokens_seen": 2109376, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.5743589743589743, |
| "grad_norm": 0.27734375, |
| "learning_rate": 2.997601228999956e-06, |
| "loss": 0.5597, |
| "num_input_tokens_seen": 2181344, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.5948717948717949, |
| "grad_norm": 0.2451171875, |
| "learning_rate": 2.99742687843539e-06, |
| "loss": 0.5514, |
| "num_input_tokens_seen": 2256768, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.6153846153846154, |
| "grad_norm": 0.2265625, |
| "learning_rate": 2.997246417427407e-06, |
| "loss": 0.53, |
| "num_input_tokens_seen": 2330144, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.6358974358974359, |
| "grad_norm": 0.19921875, |
| "learning_rate": 2.9970598467124008e-06, |
| "loss": 0.5365, |
| "num_input_tokens_seen": 2402688, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.6564102564102564, |
| "grad_norm": 0.1923828125, |
| "learning_rate": 2.9968671670516983e-06, |
| "loss": 0.5578, |
| "num_input_tokens_seen": 2476800, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.676923076923077, |
| "grad_norm": 0.166015625, |
| "learning_rate": 2.9966683792315528e-06, |
| "loss": 0.5492, |
| "num_input_tokens_seen": 2565792, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.6974358974358974, |
| "grad_norm": 0.1533203125, |
| "learning_rate": 2.9964634840631435e-06, |
| "loss": 0.5144, |
| "num_input_tokens_seen": 2637792, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.717948717948718, |
| "grad_norm": 0.154296875, |
| "learning_rate": 2.9962524823825724e-06, |
| "loss": 0.5741, |
| "num_input_tokens_seen": 2718944, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.7384615384615385, |
| "grad_norm": 0.1513671875, |
| "learning_rate": 2.9960353750508583e-06, |
| "loss": 0.5846, |
| "num_input_tokens_seen": 2802240, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.7589743589743589, |
| "grad_norm": 0.1474609375, |
| "learning_rate": 2.995812162953936e-06, |
| "loss": 0.5834, |
| "num_input_tokens_seen": 2884672, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.7794871794871795, |
| "grad_norm": 0.154296875, |
| "learning_rate": 2.9955828470026515e-06, |
| "loss": 0.5493, |
| "num_input_tokens_seen": 2953536, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.138671875, |
| "learning_rate": 2.9953474281327576e-06, |
| "loss": 0.5503, |
| "num_input_tokens_seen": 3026496, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.8205128205128205, |
| "grad_norm": 0.244140625, |
| "learning_rate": 2.995105907304912e-06, |
| "loss": 0.5835, |
| "num_input_tokens_seen": 3105376, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.841025641025641, |
| "grad_norm": 0.349609375, |
| "learning_rate": 2.9948582855046704e-06, |
| "loss": 0.5583, |
| "num_input_tokens_seen": 3179776, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.8615384615384616, |
| "grad_norm": 0.1826171875, |
| "learning_rate": 2.9946045637424864e-06, |
| "loss": 0.6392, |
| "num_input_tokens_seen": 3262336, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.882051282051282, |
| "grad_norm": 0.1640625, |
| "learning_rate": 2.994344743053704e-06, |
| "loss": 0.5658, |
| "num_input_tokens_seen": 3343904, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.9025641025641026, |
| "grad_norm": 0.50390625, |
| "learning_rate": 2.9940788244985557e-06, |
| "loss": 0.5609, |
| "num_input_tokens_seen": 3419264, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.9230769230769231, |
| "grad_norm": 0.458984375, |
| "learning_rate": 2.9938068091621556e-06, |
| "loss": 0.5305, |
| "num_input_tokens_seen": 3496032, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.9435897435897436, |
| "grad_norm": 0.41015625, |
| "learning_rate": 2.9935286981544975e-06, |
| "loss": 0.5652, |
| "num_input_tokens_seen": 3570880, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.9641025641025641, |
| "grad_norm": 0.7734375, |
| "learning_rate": 2.9932444926104495e-06, |
| "loss": 0.5348, |
| "num_input_tokens_seen": 3644096, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.9846153846153847, |
| "grad_norm": 0.86328125, |
| "learning_rate": 2.992954193689749e-06, |
| "loss": 0.5344, |
| "num_input_tokens_seen": 3717088, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.005128205128205, |
| "grad_norm": 0.435546875, |
| "learning_rate": 2.9926578025769978e-06, |
| "loss": 0.5977, |
| "num_input_tokens_seen": 3783648, |
| "step": 245 |
| }, |
| { |
| "epoch": 1.0256410256410255, |
| "grad_norm": 0.408203125, |
| "learning_rate": 2.992355320481658e-06, |
| "loss": 0.5304, |
| "num_input_tokens_seen": 3852160, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.0461538461538462, |
| "grad_norm": 0.75390625, |
| "learning_rate": 2.9920467486380475e-06, |
| "loss": 0.5617, |
| "num_input_tokens_seen": 3924416, |
| "step": 255 |
| }, |
| { |
| "epoch": 1.0666666666666667, |
| "grad_norm": 0.46875, |
| "learning_rate": 2.991732088305333e-06, |
| "loss": 0.5367, |
| "num_input_tokens_seen": 4006432, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.087179487179487, |
| "grad_norm": 0.5234375, |
| "learning_rate": 2.991411340767526e-06, |
| "loss": 0.5207, |
| "num_input_tokens_seen": 4081888, |
| "step": 265 |
| }, |
| { |
| "epoch": 1.1076923076923078, |
| "grad_norm": 0.41015625, |
| "learning_rate": 2.9910845073334793e-06, |
| "loss": 0.556, |
| "num_input_tokens_seen": 4155968, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.1282051282051282, |
| "grad_norm": 0.421875, |
| "learning_rate": 2.9907515893368784e-06, |
| "loss": 0.537, |
| "num_input_tokens_seen": 4234272, |
| "step": 275 |
| }, |
| { |
| "epoch": 1.1487179487179486, |
| "grad_norm": 0.40234375, |
| "learning_rate": 2.9904125881362378e-06, |
| "loss": 0.5305, |
| "num_input_tokens_seen": 4302368, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.1692307692307693, |
| "grad_norm": 0.5625, |
| "learning_rate": 2.990067505114896e-06, |
| "loss": 0.5337, |
| "num_input_tokens_seen": 4376640, |
| "step": 285 |
| }, |
| { |
| "epoch": 1.1897435897435897, |
| "grad_norm": 0.451171875, |
| "learning_rate": 2.9897163416810084e-06, |
| "loss": 0.592, |
| "num_input_tokens_seen": 4458208, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.2102564102564102, |
| "grad_norm": 0.40625, |
| "learning_rate": 2.9893590992675427e-06, |
| "loss": 0.5808, |
| "num_input_tokens_seen": 4537920, |
| "step": 295 |
| }, |
| { |
| "epoch": 1.2307692307692308, |
| "grad_norm": 0.58984375, |
| "learning_rate": 2.988995779332273e-06, |
| "loss": 0.5569, |
| "num_input_tokens_seen": 4606880, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.2512820512820513, |
| "grad_norm": 0.431640625, |
| "learning_rate": 2.9886263833577725e-06, |
| "loss": 0.5422, |
| "num_input_tokens_seen": 4682816, |
| "step": 305 |
| }, |
| { |
| "epoch": 1.2717948717948717, |
| "grad_norm": 0.5234375, |
| "learning_rate": 2.98825091285141e-06, |
| "loss": 0.5296, |
| "num_input_tokens_seen": 4756032, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.2923076923076924, |
| "grad_norm": 0.4453125, |
| "learning_rate": 2.987869369345341e-06, |
| "loss": 0.5012, |
| "num_input_tokens_seen": 4827232, |
| "step": 315 |
| }, |
| { |
| "epoch": 1.3128205128205128, |
| "grad_norm": 0.392578125, |
| "learning_rate": 2.987481754396502e-06, |
| "loss": 0.5289, |
| "num_input_tokens_seen": 4902368, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.3333333333333333, |
| "grad_norm": 0.326171875, |
| "learning_rate": 2.9870880695866067e-06, |
| "loss": 0.5245, |
| "num_input_tokens_seen": 4978080, |
| "step": 325 |
| }, |
| { |
| "epoch": 1.353846153846154, |
| "grad_norm": 0.62109375, |
| "learning_rate": 2.986688316522136e-06, |
| "loss": 0.5325, |
| "num_input_tokens_seen": 5047456, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.3743589743589744, |
| "grad_norm": 0.390625, |
| "learning_rate": 2.9862824968343352e-06, |
| "loss": 0.5068, |
| "num_input_tokens_seen": 5118720, |
| "step": 335 |
| }, |
| { |
| "epoch": 1.3948717948717948, |
| "grad_norm": 0.345703125, |
| "learning_rate": 2.9858706121792036e-06, |
| "loss": 0.5165, |
| "num_input_tokens_seen": 5196288, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.4153846153846155, |
| "grad_norm": 0.251953125, |
| "learning_rate": 2.985452664237488e-06, |
| "loss": 0.5025, |
| "num_input_tokens_seen": 5272480, |
| "step": 345 |
| }, |
| { |
| "epoch": 1.435897435897436, |
| "grad_norm": 0.302734375, |
| "learning_rate": 2.98502865471468e-06, |
| "loss": 0.5285, |
| "num_input_tokens_seen": 5343296, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.4564102564102563, |
| "grad_norm": 0.314453125, |
| "learning_rate": 2.9845985853410053e-06, |
| "loss": 0.4983, |
| "num_input_tokens_seen": 5415904, |
| "step": 355 |
| }, |
| { |
| "epoch": 1.476923076923077, |
| "grad_norm": 0.236328125, |
| "learning_rate": 2.9841624578714167e-06, |
| "loss": 0.5789, |
| "num_input_tokens_seen": 5502176, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.4974358974358974, |
| "grad_norm": 0.2060546875, |
| "learning_rate": 2.9837202740855897e-06, |
| "loss": 0.5394, |
| "num_input_tokens_seen": 5580352, |
| "step": 365 |
| }, |
| { |
| "epoch": 1.5179487179487179, |
| "grad_norm": 0.267578125, |
| "learning_rate": 2.9832720357879107e-06, |
| "loss": 0.5664, |
| "num_input_tokens_seen": 5662912, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.5384615384615383, |
| "grad_norm": 0.22265625, |
| "learning_rate": 2.9828177448074753e-06, |
| "loss": 0.5546, |
| "num_input_tokens_seen": 5743776, |
| "step": 375 |
| }, |
| { |
| "epoch": 1.558974358974359, |
| "grad_norm": 0.20703125, |
| "learning_rate": 2.9823574029980757e-06, |
| "loss": 0.5412, |
| "num_input_tokens_seen": 5812384, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.5794871794871796, |
| "grad_norm": 0.1630859375, |
| "learning_rate": 2.981891012238196e-06, |
| "loss": 0.5587, |
| "num_input_tokens_seen": 5892768, |
| "step": 385 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.1640625, |
| "learning_rate": 2.9814185744310036e-06, |
| "loss": 0.535, |
| "num_input_tokens_seen": 5974592, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.6205128205128205, |
| "grad_norm": 0.1533203125, |
| "learning_rate": 2.9809400915043424e-06, |
| "loss": 0.512, |
| "num_input_tokens_seen": 6050080, |
| "step": 395 |
| }, |
| { |
| "epoch": 1.641025641025641, |
| "grad_norm": 0.142578125, |
| "learning_rate": 2.9804555654107243e-06, |
| "loss": 0.5392, |
| "num_input_tokens_seen": 6137248, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.6615384615384614, |
| "grad_norm": 0.1376953125, |
| "learning_rate": 2.9799649981273185e-06, |
| "loss": 0.6444, |
| "num_input_tokens_seen": 6232704, |
| "step": 405 |
| }, |
| { |
| "epoch": 1.682051282051282, |
| "grad_norm": 0.1435546875, |
| "learning_rate": 2.9794683916559493e-06, |
| "loss": 0.5202, |
| "num_input_tokens_seen": 6304064, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.7025641025641025, |
| "grad_norm": 0.1455078125, |
| "learning_rate": 2.9789657480230842e-06, |
| "loss": 0.5344, |
| "num_input_tokens_seen": 6392928, |
| "step": 415 |
| }, |
| { |
| "epoch": 1.7230769230769232, |
| "grad_norm": 0.16015625, |
| "learning_rate": 2.9784570692798236e-06, |
| "loss": 0.4614, |
| "num_input_tokens_seen": 6473120, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.7435897435897436, |
| "grad_norm": 0.19921875, |
| "learning_rate": 2.977942357501898e-06, |
| "loss": 0.5036, |
| "num_input_tokens_seen": 6545312, |
| "step": 425 |
| }, |
| { |
| "epoch": 1.764102564102564, |
| "grad_norm": 0.41796875, |
| "learning_rate": 2.977421614789655e-06, |
| "loss": 0.5308, |
| "num_input_tokens_seen": 6629984, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.7846153846153845, |
| "grad_norm": 0.1767578125, |
| "learning_rate": 2.976894843268051e-06, |
| "loss": 0.5475, |
| "num_input_tokens_seen": 6715936, |
| "step": 435 |
| }, |
| { |
| "epoch": 1.8051282051282052, |
| "grad_norm": 0.150390625, |
| "learning_rate": 2.976362045086647e-06, |
| "loss": 0.5704, |
| "num_input_tokens_seen": 6797472, |
| "step": 440 |
| }, |
| { |
| "epoch": 1.8256410256410256, |
| "grad_norm": 0.416015625, |
| "learning_rate": 2.975823222419594e-06, |
| "loss": 0.4781, |
| "num_input_tokens_seen": 6867808, |
| "step": 445 |
| }, |
| { |
| "epoch": 1.8461538461538463, |
| "grad_norm": 0.25, |
| "learning_rate": 2.9752783774656267e-06, |
| "loss": 0.5247, |
| "num_input_tokens_seen": 6944480, |
| "step": 450 |
| }, |
| { |
| "epoch": 1.8666666666666667, |
| "grad_norm": 0.703125, |
| "learning_rate": 2.974727512448056e-06, |
| "loss": 0.5161, |
| "num_input_tokens_seen": 7024064, |
| "step": 455 |
| }, |
| { |
| "epoch": 1.8871794871794871, |
| "grad_norm": 0.65234375, |
| "learning_rate": 2.974170629614757e-06, |
| "loss": 0.5049, |
| "num_input_tokens_seen": 7103488, |
| "step": 460 |
| }, |
| { |
| "epoch": 1.9076923076923076, |
| "grad_norm": 0.734375, |
| "learning_rate": 2.9736077312381624e-06, |
| "loss": 0.5712, |
| "num_input_tokens_seen": 7190304, |
| "step": 465 |
| }, |
| { |
| "epoch": 1.9282051282051282, |
| "grad_norm": 0.40234375, |
| "learning_rate": 2.9730388196152513e-06, |
| "loss": 0.5222, |
| "num_input_tokens_seen": 7265056, |
| "step": 470 |
| }, |
| { |
| "epoch": 1.9487179487179487, |
| "grad_norm": 0.28125, |
| "learning_rate": 2.972463897067541e-06, |
| "loss": 0.4829, |
| "num_input_tokens_seen": 7336992, |
| "step": 475 |
| }, |
| { |
| "epoch": 1.9692307692307693, |
| "grad_norm": 0.56640625, |
| "learning_rate": 2.971882965941077e-06, |
| "loss": 0.5218, |
| "num_input_tokens_seen": 7413984, |
| "step": 480 |
| }, |
| { |
| "epoch": 1.9897435897435898, |
| "grad_norm": 0.357421875, |
| "learning_rate": 2.9712960286064237e-06, |
| "loss": 0.5543, |
| "num_input_tokens_seen": 7492768, |
| "step": 485 |
| }, |
| { |
| "epoch": 2.01025641025641, |
| "grad_norm": 0.26171875, |
| "learning_rate": 2.970703087458655e-06, |
| "loss": 0.5036, |
| "num_input_tokens_seen": 7563904, |
| "step": 490 |
| }, |
| { |
| "epoch": 2.0307692307692307, |
| "grad_norm": 0.38671875, |
| "learning_rate": 2.9701041449173426e-06, |
| "loss": 0.556, |
| "num_input_tokens_seen": 7634464, |
| "step": 495 |
| }, |
| { |
| "epoch": 2.051282051282051, |
| "grad_norm": 0.375, |
| "learning_rate": 2.96949920342655e-06, |
| "loss": 0.5742, |
| "num_input_tokens_seen": 7711168, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.071794871794872, |
| "grad_norm": 0.365234375, |
| "learning_rate": 2.968888265454818e-06, |
| "loss": 0.4905, |
| "num_input_tokens_seen": 7786656, |
| "step": 505 |
| }, |
| { |
| "epoch": 2.0923076923076924, |
| "grad_norm": 0.271484375, |
| "learning_rate": 2.968271333495158e-06, |
| "loss": 0.5134, |
| "num_input_tokens_seen": 7858240, |
| "step": 510 |
| }, |
| { |
| "epoch": 2.112820512820513, |
| "grad_norm": 0.373046875, |
| "learning_rate": 2.967648410065041e-06, |
| "loss": 0.6211, |
| "num_input_tokens_seen": 7952064, |
| "step": 515 |
| }, |
| { |
| "epoch": 2.1333333333333333, |
| "grad_norm": 0.45703125, |
| "learning_rate": 2.9670194977063857e-06, |
| "loss": 0.515, |
| "num_input_tokens_seen": 8026688, |
| "step": 520 |
| }, |
| { |
| "epoch": 2.1538461538461537, |
| "grad_norm": 0.400390625, |
| "learning_rate": 2.96638459898555e-06, |
| "loss": 0.5464, |
| "num_input_tokens_seen": 8104064, |
| "step": 525 |
| }, |
| { |
| "epoch": 2.174358974358974, |
| "grad_norm": 0.421875, |
| "learning_rate": 2.9657437164933205e-06, |
| "loss": 0.5331, |
| "num_input_tokens_seen": 8187968, |
| "step": 530 |
| }, |
| { |
| "epoch": 2.194871794871795, |
| "grad_norm": 0.396484375, |
| "learning_rate": 2.9650968528449e-06, |
| "loss": 0.5236, |
| "num_input_tokens_seen": 8261312, |
| "step": 535 |
| }, |
| { |
| "epoch": 2.2153846153846155, |
| "grad_norm": 0.34765625, |
| "learning_rate": 2.9644440106799e-06, |
| "loss": 0.4345, |
| "num_input_tokens_seen": 8336032, |
| "step": 540 |
| }, |
| { |
| "epoch": 2.235897435897436, |
| "grad_norm": 0.353515625, |
| "learning_rate": 2.963785192662327e-06, |
| "loss": 0.4853, |
| "num_input_tokens_seen": 8405024, |
| "step": 545 |
| }, |
| { |
| "epoch": 2.2564102564102564, |
| "grad_norm": 0.365234375, |
| "learning_rate": 2.9631204014805716e-06, |
| "loss": 0.5, |
| "num_input_tokens_seen": 8483456, |
| "step": 550 |
| }, |
| { |
| "epoch": 2.276923076923077, |
| "grad_norm": 0.38671875, |
| "learning_rate": 2.9624496398474014e-06, |
| "loss": 0.4863, |
| "num_input_tokens_seen": 8560768, |
| "step": 555 |
| }, |
| { |
| "epoch": 2.2974358974358973, |
| "grad_norm": 0.32421875, |
| "learning_rate": 2.961772910499945e-06, |
| "loss": 0.48, |
| "num_input_tokens_seen": 8637888, |
| "step": 560 |
| }, |
| { |
| "epoch": 2.3179487179487177, |
| "grad_norm": 0.376953125, |
| "learning_rate": 2.9610902161996838e-06, |
| "loss": 0.5768, |
| "num_input_tokens_seen": 8715360, |
| "step": 565 |
| }, |
| { |
| "epoch": 2.3384615384615386, |
| "grad_norm": 0.265625, |
| "learning_rate": 2.96040155973244e-06, |
| "loss": 0.4751, |
| "num_input_tokens_seen": 8795200, |
| "step": 570 |
| }, |
| { |
| "epoch": 2.358974358974359, |
| "grad_norm": 0.283203125, |
| "learning_rate": 2.959706943908365e-06, |
| "loss": 0.5161, |
| "num_input_tokens_seen": 8883136, |
| "step": 575 |
| }, |
| { |
| "epoch": 2.3794871794871795, |
| "grad_norm": 0.271484375, |
| "learning_rate": 2.9590063715619287e-06, |
| "loss": 0.5588, |
| "num_input_tokens_seen": 8961568, |
| "step": 580 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.2373046875, |
| "learning_rate": 2.9582998455519062e-06, |
| "loss": 0.5527, |
| "num_input_tokens_seen": 9043360, |
| "step": 585 |
| }, |
| { |
| "epoch": 2.4205128205128204, |
| "grad_norm": 0.21875, |
| "learning_rate": 2.9575873687613676e-06, |
| "loss": 0.4897, |
| "num_input_tokens_seen": 9116448, |
| "step": 590 |
| }, |
| { |
| "epoch": 2.4410256410256412, |
| "grad_norm": 0.21484375, |
| "learning_rate": 2.9568689440976676e-06, |
| "loss": 0.5359, |
| "num_input_tokens_seen": 9193120, |
| "step": 595 |
| }, |
| { |
| "epoch": 2.4615384615384617, |
| "grad_norm": 0.205078125, |
| "learning_rate": 2.95614457449243e-06, |
| "loss": 0.5763, |
| "num_input_tokens_seen": 9269920, |
| "step": 600 |
| }, |
| { |
| "epoch": 2.482051282051282, |
| "grad_norm": 0.1845703125, |
| "learning_rate": 2.9554142629015382e-06, |
| "loss": 0.4631, |
| "num_input_tokens_seen": 9339968, |
| "step": 605 |
| }, |
| { |
| "epoch": 2.5025641025641026, |
| "grad_norm": 0.1650390625, |
| "learning_rate": 2.954678012305123e-06, |
| "loss": 0.5349, |
| "num_input_tokens_seen": 9426976, |
| "step": 610 |
| }, |
| { |
| "epoch": 2.523076923076923, |
| "grad_norm": 0.1708984375, |
| "learning_rate": 2.9539358257075495e-06, |
| "loss": 0.5532, |
| "num_input_tokens_seen": 9509056, |
| "step": 615 |
| }, |
| { |
| "epoch": 2.5435897435897434, |
| "grad_norm": 0.1611328125, |
| "learning_rate": 2.9531877061374066e-06, |
| "loss": 0.4748, |
| "num_input_tokens_seen": 9590720, |
| "step": 620 |
| }, |
| { |
| "epoch": 2.564102564102564, |
| "grad_norm": 0.15234375, |
| "learning_rate": 2.9524336566474915e-06, |
| "loss": 0.5022, |
| "num_input_tokens_seen": 9667648, |
| "step": 625 |
| }, |
| { |
| "epoch": 2.5846153846153848, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 2.9516736803148014e-06, |
| "loss": 0.5005, |
| "num_input_tokens_seen": 9738016, |
| "step": 630 |
| }, |
| { |
| "epoch": 2.605128205128205, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 2.9509077802405174e-06, |
| "loss": 0.5297, |
| "num_input_tokens_seen": 9816224, |
| "step": 635 |
| }, |
| { |
| "epoch": 2.6256410256410256, |
| "grad_norm": 0.13671875, |
| "learning_rate": 2.9501359595499933e-06, |
| "loss": 0.5399, |
| "num_input_tokens_seen": 9891104, |
| "step": 640 |
| }, |
| { |
| "epoch": 2.646153846153846, |
| "grad_norm": 0.177734375, |
| "learning_rate": 2.9493582213927425e-06, |
| "loss": 0.4901, |
| "num_input_tokens_seen": 9969792, |
| "step": 645 |
| }, |
| { |
| "epoch": 2.6666666666666665, |
| "grad_norm": 0.142578125, |
| "learning_rate": 2.9485745689424267e-06, |
| "loss": 0.4591, |
| "num_input_tokens_seen": 10044608, |
| "step": 650 |
| }, |
| { |
| "epoch": 2.6871794871794874, |
| "grad_norm": 0.2451171875, |
| "learning_rate": 2.9477850053968405e-06, |
| "loss": 0.5729, |
| "num_input_tokens_seen": 10132640, |
| "step": 655 |
| }, |
| { |
| "epoch": 2.707692307692308, |
| "grad_norm": 0.150390625, |
| "learning_rate": 2.9469895339778995e-06, |
| "loss": 0.5405, |
| "num_input_tokens_seen": 10207968, |
| "step": 660 |
| }, |
| { |
| "epoch": 2.7282051282051283, |
| "grad_norm": 0.30859375, |
| "learning_rate": 2.946188157931627e-06, |
| "loss": 0.4786, |
| "num_input_tokens_seen": 10277408, |
| "step": 665 |
| }, |
| { |
| "epoch": 2.7487179487179487, |
| "grad_norm": 0.314453125, |
| "learning_rate": 2.9453808805281423e-06, |
| "loss": 0.5035, |
| "num_input_tokens_seen": 10349184, |
| "step": 670 |
| }, |
| { |
| "epoch": 2.769230769230769, |
| "grad_norm": 0.2490234375, |
| "learning_rate": 2.944567705061644e-06, |
| "loss": 0.4719, |
| "num_input_tokens_seen": 10434112, |
| "step": 675 |
| }, |
| { |
| "epoch": 2.7897435897435896, |
| "grad_norm": 0.65234375, |
| "learning_rate": 2.9437486348504e-06, |
| "loss": 0.5118, |
| "num_input_tokens_seen": 10506208, |
| "step": 680 |
| }, |
| { |
| "epoch": 2.81025641025641, |
| "grad_norm": 0.68359375, |
| "learning_rate": 2.9429236732367318e-06, |
| "loss": 0.5014, |
| "num_input_tokens_seen": 10577696, |
| "step": 685 |
| }, |
| { |
| "epoch": 2.830769230769231, |
| "grad_norm": 0.71484375, |
| "learning_rate": 2.942092823587001e-06, |
| "loss": 0.4827, |
| "num_input_tokens_seen": 10657984, |
| "step": 690 |
| }, |
| { |
| "epoch": 2.8512820512820514, |
| "grad_norm": 0.71484375, |
| "learning_rate": 2.941256089291597e-06, |
| "loss": 0.5177, |
| "num_input_tokens_seen": 10734688, |
| "step": 695 |
| }, |
| { |
| "epoch": 2.871794871794872, |
| "grad_norm": 0.6328125, |
| "learning_rate": 2.940413473764923e-06, |
| "loss": 0.4517, |
| "num_input_tokens_seen": 10812640, |
| "step": 700 |
| }, |
| { |
| "epoch": 2.8923076923076922, |
| "grad_norm": 0.70703125, |
| "learning_rate": 2.9395649804453786e-06, |
| "loss": 0.4574, |
| "num_input_tokens_seen": 10884800, |
| "step": 705 |
| }, |
| { |
| "epoch": 2.9128205128205127, |
| "grad_norm": 0.27734375, |
| "learning_rate": 2.9387106127953515e-06, |
| "loss": 0.5092, |
| "num_input_tokens_seen": 10962016, |
| "step": 710 |
| }, |
| { |
| "epoch": 2.9333333333333336, |
| "grad_norm": 0.34375, |
| "learning_rate": 2.937850374301198e-06, |
| "loss": 0.4888, |
| "num_input_tokens_seen": 11033280, |
| "step": 715 |
| }, |
| { |
| "epoch": 2.953846153846154, |
| "grad_norm": 0.375, |
| "learning_rate": 2.9369842684732336e-06, |
| "loss": 0.5447, |
| "num_input_tokens_seen": 11113696, |
| "step": 720 |
| }, |
| { |
| "epoch": 2.9743589743589745, |
| "grad_norm": 0.396484375, |
| "learning_rate": 2.936112298845713e-06, |
| "loss": 0.5438, |
| "num_input_tokens_seen": 11195104, |
| "step": 725 |
| }, |
| { |
| "epoch": 2.994871794871795, |
| "grad_norm": 0.2470703125, |
| "learning_rate": 2.935234468976822e-06, |
| "loss": 0.46, |
| "num_input_tokens_seen": 11270304, |
| "step": 730 |
| }, |
| { |
| "epoch": 3.0153846153846153, |
| "grad_norm": 0.392578125, |
| "learning_rate": 2.934350782448658e-06, |
| "loss": 0.569, |
| "num_input_tokens_seen": 11350784, |
| "step": 735 |
| }, |
| { |
| "epoch": 3.0358974358974358, |
| "grad_norm": 0.318359375, |
| "learning_rate": 2.9334612428672175e-06, |
| "loss": 0.5246, |
| "num_input_tokens_seen": 11429568, |
| "step": 740 |
| }, |
| { |
| "epoch": 3.056410256410256, |
| "grad_norm": 0.35546875, |
| "learning_rate": 2.9325658538623822e-06, |
| "loss": 0.4587, |
| "num_input_tokens_seen": 11502784, |
| "step": 745 |
| }, |
| { |
| "epoch": 3.076923076923077, |
| "grad_norm": 0.3671875, |
| "learning_rate": 2.931664619087902e-06, |
| "loss": 0.5095, |
| "num_input_tokens_seen": 11575680, |
| "step": 750 |
| }, |
| { |
| "epoch": 3.0974358974358975, |
| "grad_norm": 0.35546875, |
| "learning_rate": 2.9307575422213813e-06, |
| "loss": 0.4916, |
| "num_input_tokens_seen": 11649856, |
| "step": 755 |
| }, |
| { |
| "epoch": 3.117948717948718, |
| "grad_norm": 0.404296875, |
| "learning_rate": 2.929844626964265e-06, |
| "loss": 0.5647, |
| "num_input_tokens_seen": 11727616, |
| "step": 760 |
| }, |
| { |
| "epoch": 3.1384615384615384, |
| "grad_norm": 0.421875, |
| "learning_rate": 2.9289258770418208e-06, |
| "loss": 0.4448, |
| "num_input_tokens_seen": 11806208, |
| "step": 765 |
| }, |
| { |
| "epoch": 3.158974358974359, |
| "grad_norm": 0.39453125, |
| "learning_rate": 2.9280012962031263e-06, |
| "loss": 0.5086, |
| "num_input_tokens_seen": 11884096, |
| "step": 770 |
| }, |
| { |
| "epoch": 3.1794871794871793, |
| "grad_norm": 0.29296875, |
| "learning_rate": 2.9270708882210525e-06, |
| "loss": 0.4796, |
| "num_input_tokens_seen": 11956416, |
| "step": 775 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 0.33203125, |
| "learning_rate": 2.92613465689225e-06, |
| "loss": 0.4797, |
| "num_input_tokens_seen": 12032384, |
| "step": 780 |
| }, |
| { |
| "epoch": 3.2205128205128206, |
| "grad_norm": 0.287109375, |
| "learning_rate": 2.92519260603713e-06, |
| "loss": 0.4523, |
| "num_input_tokens_seen": 12107328, |
| "step": 785 |
| }, |
| { |
| "epoch": 3.241025641025641, |
| "grad_norm": 0.326171875, |
| "learning_rate": 2.9242447394998545e-06, |
| "loss": 0.4795, |
| "num_input_tokens_seen": 12178848, |
| "step": 790 |
| }, |
| { |
| "epoch": 3.2615384615384615, |
| "grad_norm": 0.2734375, |
| "learning_rate": 2.923291061148314e-06, |
| "loss": 0.5164, |
| "num_input_tokens_seen": 12252160, |
| "step": 795 |
| }, |
| { |
| "epoch": 3.282051282051282, |
| "grad_norm": 0.271484375, |
| "learning_rate": 2.9223315748741146e-06, |
| "loss": 0.4949, |
| "num_input_tokens_seen": 12325120, |
| "step": 800 |
| }, |
| { |
| "epoch": 3.3025641025641024, |
| "grad_norm": 0.244140625, |
| "learning_rate": 2.9213662845925662e-06, |
| "loss": 0.4848, |
| "num_input_tokens_seen": 12398144, |
| "step": 805 |
| }, |
| { |
| "epoch": 3.3230769230769233, |
| "grad_norm": 0.2265625, |
| "learning_rate": 2.9203951942426586e-06, |
| "loss": 0.5114, |
| "num_input_tokens_seen": 12475008, |
| "step": 810 |
| }, |
| { |
| "epoch": 3.3435897435897437, |
| "grad_norm": 0.2109375, |
| "learning_rate": 2.9194183077870516e-06, |
| "loss": 0.6022, |
| "num_input_tokens_seen": 12562336, |
| "step": 815 |
| }, |
| { |
| "epoch": 3.364102564102564, |
| "grad_norm": 0.1923828125, |
| "learning_rate": 2.9184356292120562e-06, |
| "loss": 0.4922, |
| "num_input_tokens_seen": 12646560, |
| "step": 820 |
| }, |
| { |
| "epoch": 3.3846153846153846, |
| "grad_norm": 0.1962890625, |
| "learning_rate": 2.9174471625276198e-06, |
| "loss": 0.5707, |
| "num_input_tokens_seen": 12718848, |
| "step": 825 |
| }, |
| { |
| "epoch": 3.405128205128205, |
| "grad_norm": 0.1650390625, |
| "learning_rate": 2.916452911767307e-06, |
| "loss": 0.4784, |
| "num_input_tokens_seen": 12798240, |
| "step": 830 |
| }, |
| { |
| "epoch": 3.4256410256410255, |
| "grad_norm": 0.1552734375, |
| "learning_rate": 2.915452880988287e-06, |
| "loss": 0.4423, |
| "num_input_tokens_seen": 12872608, |
| "step": 835 |
| }, |
| { |
| "epoch": 3.4461538461538463, |
| "grad_norm": 0.1396484375, |
| "learning_rate": 2.914447074271314e-06, |
| "loss": 0.4809, |
| "num_input_tokens_seen": 12952896, |
| "step": 840 |
| }, |
| { |
| "epoch": 3.466666666666667, |
| "grad_norm": 0.1650390625, |
| "learning_rate": 2.913435495720712e-06, |
| "loss": 0.5316, |
| "num_input_tokens_seen": 13036768, |
| "step": 845 |
| }, |
| { |
| "epoch": 3.4871794871794872, |
| "grad_norm": 0.140625, |
| "learning_rate": 2.9124181494643574e-06, |
| "loss": 0.4592, |
| "num_input_tokens_seen": 13114784, |
| "step": 850 |
| }, |
| { |
| "epoch": 3.5076923076923077, |
| "grad_norm": 0.138671875, |
| "learning_rate": 2.911395039653663e-06, |
| "loss": 0.4878, |
| "num_input_tokens_seen": 13188448, |
| "step": 855 |
| }, |
| { |
| "epoch": 3.528205128205128, |
| "grad_norm": 0.1298828125, |
| "learning_rate": 2.9103661704635604e-06, |
| "loss": 0.5066, |
| "num_input_tokens_seen": 13262592, |
| "step": 860 |
| }, |
| { |
| "epoch": 3.5487179487179485, |
| "grad_norm": 0.11962890625, |
| "learning_rate": 2.909331546092483e-06, |
| "loss": 0.4649, |
| "num_input_tokens_seen": 13339936, |
| "step": 865 |
| }, |
| { |
| "epoch": 3.569230769230769, |
| "grad_norm": 0.14453125, |
| "learning_rate": 2.908291170762349e-06, |
| "loss": 0.5233, |
| "num_input_tokens_seen": 13416256, |
| "step": 870 |
| }, |
| { |
| "epoch": 3.58974358974359, |
| "grad_norm": 0.306640625, |
| "learning_rate": 2.9072450487185434e-06, |
| "loss": 0.5018, |
| "num_input_tokens_seen": 13487392, |
| "step": 875 |
| }, |
| { |
| "epoch": 3.6102564102564103, |
| "grad_norm": 0.337890625, |
| "learning_rate": 2.9061931842299026e-06, |
| "loss": 0.4602, |
| "num_input_tokens_seen": 13569984, |
| "step": 880 |
| }, |
| { |
| "epoch": 3.6307692307692307, |
| "grad_norm": 0.1640625, |
| "learning_rate": 2.9051355815886952e-06, |
| "loss": 0.5309, |
| "num_input_tokens_seen": 13650944, |
| "step": 885 |
| }, |
| { |
| "epoch": 3.651282051282051, |
| "grad_norm": 0.30859375, |
| "learning_rate": 2.904072245110605e-06, |
| "loss": 0.5186, |
| "num_input_tokens_seen": 13720736, |
| "step": 890 |
| }, |
| { |
| "epoch": 3.6717948717948716, |
| "grad_norm": 0.333984375, |
| "learning_rate": 2.9030031791347136e-06, |
| "loss": 0.4839, |
| "num_input_tokens_seen": 13791616, |
| "step": 895 |
| }, |
| { |
| "epoch": 3.6923076923076925, |
| "grad_norm": 0.25, |
| "learning_rate": 2.901928388023483e-06, |
| "loss": 0.5199, |
| "num_input_tokens_seen": 13867488, |
| "step": 900 |
| }, |
| { |
| "epoch": 3.712820512820513, |
| "grad_norm": 0.69921875, |
| "learning_rate": 2.900847876162736e-06, |
| "loss": 0.5414, |
| "num_input_tokens_seen": 13954848, |
| "step": 905 |
| }, |
| { |
| "epoch": 3.7333333333333334, |
| "grad_norm": 0.80859375, |
| "learning_rate": 2.899761647961641e-06, |
| "loss": 0.5451, |
| "num_input_tokens_seen": 14037792, |
| "step": 910 |
| }, |
| { |
| "epoch": 3.753846153846154, |
| "grad_norm": 1.0, |
| "learning_rate": 2.898669707852692e-06, |
| "loss": 0.4448, |
| "num_input_tokens_seen": 14107520, |
| "step": 915 |
| }, |
| { |
| "epoch": 3.7743589743589743, |
| "grad_norm": 0.3125, |
| "learning_rate": 2.897572060291692e-06, |
| "loss": 0.5213, |
| "num_input_tokens_seen": 14193888, |
| "step": 920 |
| }, |
| { |
| "epoch": 3.7948717948717947, |
| "grad_norm": 0.5, |
| "learning_rate": 2.896468709757733e-06, |
| "loss": 0.4968, |
| "num_input_tokens_seen": 14270976, |
| "step": 925 |
| }, |
| { |
| "epoch": 3.815384615384615, |
| "grad_norm": 0.58203125, |
| "learning_rate": 2.8953596607531788e-06, |
| "loss": 0.5769, |
| "num_input_tokens_seen": 14351232, |
| "step": 930 |
| }, |
| { |
| "epoch": 3.835897435897436, |
| "grad_norm": 0.26171875, |
| "learning_rate": 2.894244917803647e-06, |
| "loss": 0.4925, |
| "num_input_tokens_seen": 14426912, |
| "step": 935 |
| }, |
| { |
| "epoch": 3.8564102564102565, |
| "grad_norm": 0.333984375, |
| "learning_rate": 2.8931244854579904e-06, |
| "loss": 0.481, |
| "num_input_tokens_seen": 14515776, |
| "step": 940 |
| }, |
| { |
| "epoch": 3.876923076923077, |
| "grad_norm": 0.322265625, |
| "learning_rate": 2.891998368288277e-06, |
| "loss": 0.4699, |
| "num_input_tokens_seen": 14587104, |
| "step": 945 |
| }, |
| { |
| "epoch": 3.8974358974358974, |
| "grad_norm": 0.341796875, |
| "learning_rate": 2.890866570889773e-06, |
| "loss": 0.5206, |
| "num_input_tokens_seen": 14663680, |
| "step": 950 |
| }, |
| { |
| "epoch": 3.917948717948718, |
| "grad_norm": 0.35546875, |
| "learning_rate": 2.8897290978809245e-06, |
| "loss": 0.5117, |
| "num_input_tokens_seen": 14747360, |
| "step": 955 |
| }, |
| { |
| "epoch": 3.9384615384615387, |
| "grad_norm": 0.3671875, |
| "learning_rate": 2.888585953903336e-06, |
| "loss": 0.4891, |
| "num_input_tokens_seen": 14826464, |
| "step": 960 |
| }, |
| { |
| "epoch": 3.958974358974359, |
| "grad_norm": 0.376953125, |
| "learning_rate": 2.8874371436217534e-06, |
| "loss": 0.4943, |
| "num_input_tokens_seen": 14916416, |
| "step": 965 |
| }, |
| { |
| "epoch": 3.9794871794871796, |
| "grad_norm": 0.32421875, |
| "learning_rate": 2.8862826717240464e-06, |
| "loss": 0.5222, |
| "num_input_tokens_seen": 14995072, |
| "step": 970 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 0.337890625, |
| "learning_rate": 2.8851225429211855e-06, |
| "loss": 0.5197, |
| "num_input_tokens_seen": 15070304, |
| "step": 975 |
| }, |
| { |
| "epoch": 4.02051282051282, |
| "grad_norm": 0.380859375, |
| "learning_rate": 2.883956761947226e-06, |
| "loss": 0.501, |
| "num_input_tokens_seen": 15152480, |
| "step": 980 |
| }, |
| { |
| "epoch": 4.041025641025641, |
| "grad_norm": 0.478515625, |
| "learning_rate": 2.8827853335592876e-06, |
| "loss": 0.5142, |
| "num_input_tokens_seen": 15229184, |
| "step": 985 |
| }, |
| { |
| "epoch": 4.061538461538461, |
| "grad_norm": 0.376953125, |
| "learning_rate": 2.8816082625375353e-06, |
| "loss": 0.5239, |
| "num_input_tokens_seen": 15311072, |
| "step": 990 |
| }, |
| { |
| "epoch": 4.082051282051282, |
| "grad_norm": 0.380859375, |
| "learning_rate": 2.8804255536851584e-06, |
| "loss": 0.4823, |
| "num_input_tokens_seen": 15383232, |
| "step": 995 |
| }, |
| { |
| "epoch": 4.102564102564102, |
| "grad_norm": 0.34375, |
| "learning_rate": 2.8792372118283528e-06, |
| "loss": 0.5416, |
| "num_input_tokens_seen": 15464064, |
| "step": 1000 |
| }, |
| { |
| "epoch": 4.123076923076923, |
| "grad_norm": 0.416015625, |
| "learning_rate": 2.878043241816301e-06, |
| "loss": 0.4889, |
| "num_input_tokens_seen": 15536480, |
| "step": 1005 |
| }, |
| { |
| "epoch": 4.143589743589744, |
| "grad_norm": 0.3828125, |
| "learning_rate": 2.876843648521152e-06, |
| "loss": 0.5338, |
| "num_input_tokens_seen": 15618816, |
| "step": 1010 |
| }, |
| { |
| "epoch": 4.164102564102564, |
| "grad_norm": 0.2578125, |
| "learning_rate": 2.8756384368380003e-06, |
| "loss": 0.5101, |
| "num_input_tokens_seen": 15694304, |
| "step": 1015 |
| }, |
| { |
| "epoch": 4.184615384615385, |
| "grad_norm": 0.25390625, |
| "learning_rate": 2.874427611684867e-06, |
| "loss": 0.4792, |
| "num_input_tokens_seen": 15770784, |
| "step": 1020 |
| }, |
| { |
| "epoch": 4.205128205128205, |
| "grad_norm": 0.27734375, |
| "learning_rate": 2.8732111780026813e-06, |
| "loss": 0.4959, |
| "num_input_tokens_seen": 15837312, |
| "step": 1025 |
| }, |
| { |
| "epoch": 4.225641025641026, |
| "grad_norm": 0.22265625, |
| "learning_rate": 2.871989140755257e-06, |
| "loss": 0.5227, |
| "num_input_tokens_seen": 15908096, |
| "step": 1030 |
| }, |
| { |
| "epoch": 4.246153846153846, |
| "grad_norm": 0.2392578125, |
| "learning_rate": 2.870761504929275e-06, |
| "loss": 0.473, |
| "num_input_tokens_seen": 15982720, |
| "step": 1035 |
| }, |
| { |
| "epoch": 4.266666666666667, |
| "grad_norm": 0.1943359375, |
| "learning_rate": 2.869528275534261e-06, |
| "loss": 0.4911, |
| "num_input_tokens_seen": 16056256, |
| "step": 1040 |
| }, |
| { |
| "epoch": 4.287179487179487, |
| "grad_norm": 0.1787109375, |
| "learning_rate": 2.8682894576025677e-06, |
| "loss": 0.4678, |
| "num_input_tokens_seen": 16128256, |
| "step": 1045 |
| }, |
| { |
| "epoch": 4.3076923076923075, |
| "grad_norm": 0.1650390625, |
| "learning_rate": 2.8670450561893498e-06, |
| "loss": 0.4534, |
| "num_input_tokens_seen": 16203808, |
| "step": 1050 |
| }, |
| { |
| "epoch": 4.328205128205128, |
| "grad_norm": 0.1796875, |
| "learning_rate": 2.865795076372549e-06, |
| "loss": 0.5788, |
| "num_input_tokens_seen": 16290464, |
| "step": 1055 |
| }, |
| { |
| "epoch": 4.348717948717948, |
| "grad_norm": 0.1845703125, |
| "learning_rate": 2.8645395232528683e-06, |
| "loss": 0.4744, |
| "num_input_tokens_seen": 16362688, |
| "step": 1060 |
| }, |
| { |
| "epoch": 4.36923076923077, |
| "grad_norm": 0.146484375, |
| "learning_rate": 2.863278401953754e-06, |
| "loss": 0.4801, |
| "num_input_tokens_seen": 16434976, |
| "step": 1065 |
| }, |
| { |
| "epoch": 4.38974358974359, |
| "grad_norm": 0.134765625, |
| "learning_rate": 2.862011717621375e-06, |
| "loss": 0.5035, |
| "num_input_tokens_seen": 16514880, |
| "step": 1070 |
| }, |
| { |
| "epoch": 4.410256410256411, |
| "grad_norm": 0.15625, |
| "learning_rate": 2.860739475424599e-06, |
| "loss": 0.5456, |
| "num_input_tokens_seen": 16592544, |
| "step": 1075 |
| }, |
| { |
| "epoch": 4.430769230769231, |
| "grad_norm": 0.125, |
| "learning_rate": 2.859461680554975e-06, |
| "loss": 0.4773, |
| "num_input_tokens_seen": 16675360, |
| "step": 1080 |
| }, |
| { |
| "epoch": 4.4512820512820515, |
| "grad_norm": 0.1103515625, |
| "learning_rate": 2.858178338226709e-06, |
| "loss": 0.4793, |
| "num_input_tokens_seen": 16753728, |
| "step": 1085 |
| }, |
| { |
| "epoch": 4.471794871794872, |
| "grad_norm": 0.11572265625, |
| "learning_rate": 2.8568894536766462e-06, |
| "loss": 0.4698, |
| "num_input_tokens_seen": 16829696, |
| "step": 1090 |
| }, |
| { |
| "epoch": 4.492307692307692, |
| "grad_norm": 0.19140625, |
| "learning_rate": 2.8555950321642444e-06, |
| "loss": 0.4648, |
| "num_input_tokens_seen": 16908128, |
| "step": 1095 |
| }, |
| { |
| "epoch": 4.512820512820513, |
| "grad_norm": 0.2333984375, |
| "learning_rate": 2.8542950789715587e-06, |
| "loss": 0.4473, |
| "num_input_tokens_seen": 16979136, |
| "step": 1100 |
| }, |
| { |
| "epoch": 4.533333333333333, |
| "grad_norm": 0.216796875, |
| "learning_rate": 2.8529895994032153e-06, |
| "loss": 0.5128, |
| "num_input_tokens_seen": 17063296, |
| "step": 1105 |
| }, |
| { |
| "epoch": 4.553846153846154, |
| "grad_norm": 0.15234375, |
| "learning_rate": 2.851678598786392e-06, |
| "loss": 0.4834, |
| "num_input_tokens_seen": 17145984, |
| "step": 1110 |
| }, |
| { |
| "epoch": 4.574358974358974, |
| "grad_norm": 0.296875, |
| "learning_rate": 2.8503620824707946e-06, |
| "loss": 0.4581, |
| "num_input_tokens_seen": 17221152, |
| "step": 1115 |
| }, |
| { |
| "epoch": 4.5948717948717945, |
| "grad_norm": 0.267578125, |
| "learning_rate": 2.8490400558286395e-06, |
| "loss": 0.5189, |
| "num_input_tokens_seen": 17298688, |
| "step": 1120 |
| }, |
| { |
| "epoch": 4.615384615384615, |
| "grad_norm": 0.302734375, |
| "learning_rate": 2.847712524254626e-06, |
| "loss": 0.4739, |
| "num_input_tokens_seen": 17374240, |
| "step": 1125 |
| }, |
| { |
| "epoch": 4.635897435897435, |
| "grad_norm": 0.6640625, |
| "learning_rate": 2.846379493165918e-06, |
| "loss": 0.5376, |
| "num_input_tokens_seen": 17451360, |
| "step": 1130 |
| }, |
| { |
| "epoch": 4.656410256410257, |
| "grad_norm": 0.82421875, |
| "learning_rate": 2.8450409680021204e-06, |
| "loss": 0.5152, |
| "num_input_tokens_seen": 17535776, |
| "step": 1135 |
| }, |
| { |
| "epoch": 4.676923076923077, |
| "grad_norm": 0.90234375, |
| "learning_rate": 2.8436969542252576e-06, |
| "loss": 0.4957, |
| "num_input_tokens_seen": 17616256, |
| "step": 1140 |
| }, |
| { |
| "epoch": 4.697435897435898, |
| "grad_norm": 0.31640625, |
| "learning_rate": 2.842347457319752e-06, |
| "loss": 0.4963, |
| "num_input_tokens_seen": 17691616, |
| "step": 1145 |
| }, |
| { |
| "epoch": 4.717948717948718, |
| "grad_norm": 0.5078125, |
| "learning_rate": 2.8409924827923985e-06, |
| "loss": 0.4868, |
| "num_input_tokens_seen": 17767264, |
| "step": 1150 |
| }, |
| { |
| "epoch": 4.7384615384615385, |
| "grad_norm": 0.36328125, |
| "learning_rate": 2.839632036172346e-06, |
| "loss": 0.5595, |
| "num_input_tokens_seen": 17847648, |
| "step": 1155 |
| }, |
| { |
| "epoch": 4.758974358974359, |
| "grad_norm": 0.296875, |
| "learning_rate": 2.8382661230110716e-06, |
| "loss": 0.6248, |
| "num_input_tokens_seen": 17932992, |
| "step": 1160 |
| }, |
| { |
| "epoch": 4.779487179487179, |
| "grad_norm": 0.326171875, |
| "learning_rate": 2.8368947488823613e-06, |
| "loss": 0.4935, |
| "num_input_tokens_seen": 18008544, |
| "step": 1165 |
| }, |
| { |
| "epoch": 4.8, |
| "grad_norm": 0.3203125, |
| "learning_rate": 2.8355179193822834e-06, |
| "loss": 0.472, |
| "num_input_tokens_seen": 18091904, |
| "step": 1170 |
| }, |
| { |
| "epoch": 4.82051282051282, |
| "grad_norm": 0.396484375, |
| "learning_rate": 2.834135640129168e-06, |
| "loss": 0.4618, |
| "num_input_tokens_seen": 18169760, |
| "step": 1175 |
| }, |
| { |
| "epoch": 4.841025641025641, |
| "grad_norm": 0.2890625, |
| "learning_rate": 2.8327479167635834e-06, |
| "loss": 0.4375, |
| "num_input_tokens_seen": 18242016, |
| "step": 1180 |
| }, |
| { |
| "epoch": 4.861538461538462, |
| "grad_norm": 0.314453125, |
| "learning_rate": 2.831354754948315e-06, |
| "loss": 0.4954, |
| "num_input_tokens_seen": 18316192, |
| "step": 1185 |
| }, |
| { |
| "epoch": 4.8820512820512825, |
| "grad_norm": 0.330078125, |
| "learning_rate": 2.829956160368338e-06, |
| "loss": 0.4885, |
| "num_input_tokens_seen": 18402720, |
| "step": 1190 |
| }, |
| { |
| "epoch": 4.902564102564103, |
| "grad_norm": 0.482421875, |
| "learning_rate": 2.828552138730798e-06, |
| "loss": 0.4452, |
| "num_input_tokens_seen": 18472768, |
| "step": 1195 |
| }, |
| { |
| "epoch": 4.923076923076923, |
| "grad_norm": 0.33984375, |
| "learning_rate": 2.8271426957649868e-06, |
| "loss": 0.4602, |
| "num_input_tokens_seen": 18548128, |
| "step": 1200 |
| }, |
| { |
| "epoch": 4.943589743589744, |
| "grad_norm": 0.412109375, |
| "learning_rate": 2.8257278372223177e-06, |
| "loss": 0.4391, |
| "num_input_tokens_seen": 18622112, |
| "step": 1205 |
| }, |
| { |
| "epoch": 4.964102564102564, |
| "grad_norm": 0.484375, |
| "learning_rate": 2.824307568876304e-06, |
| "loss": 0.4614, |
| "num_input_tokens_seen": 18704288, |
| "step": 1210 |
| }, |
| { |
| "epoch": 4.984615384615385, |
| "grad_norm": 0.35546875, |
| "learning_rate": 2.8228818965225326e-06, |
| "loss": 0.5284, |
| "num_input_tokens_seen": 18780128, |
| "step": 1215 |
| }, |
| { |
| "epoch": 5.005128205128205, |
| "grad_norm": 0.326171875, |
| "learning_rate": 2.8214508259786443e-06, |
| "loss": 0.5213, |
| "num_input_tokens_seen": 18850496, |
| "step": 1220 |
| }, |
| { |
| "epoch": 5.0256410256410255, |
| "grad_norm": 0.333984375, |
| "learning_rate": 2.820014363084307e-06, |
| "loss": 0.5071, |
| "num_input_tokens_seen": 18926816, |
| "step": 1225 |
| }, |
| { |
| "epoch": 5.046153846153846, |
| "grad_norm": 0.341796875, |
| "learning_rate": 2.8185725137011922e-06, |
| "loss": 0.4964, |
| "num_input_tokens_seen": 19002624, |
| "step": 1230 |
| }, |
| { |
| "epoch": 5.066666666666666, |
| "grad_norm": 0.3359375, |
| "learning_rate": 2.8171252837129523e-06, |
| "loss": 0.5196, |
| "num_input_tokens_seen": 19083296, |
| "step": 1235 |
| }, |
| { |
| "epoch": 5.087179487179487, |
| "grad_norm": 0.28125, |
| "learning_rate": 2.815672679025196e-06, |
| "loss": 0.5272, |
| "num_input_tokens_seen": 19158048, |
| "step": 1240 |
| }, |
| { |
| "epoch": 5.107692307692307, |
| "grad_norm": 0.26171875, |
| "learning_rate": 2.814214705565464e-06, |
| "loss": 0.5034, |
| "num_input_tokens_seen": 19233888, |
| "step": 1245 |
| }, |
| { |
| "epoch": 5.128205128205128, |
| "grad_norm": 0.2890625, |
| "learning_rate": 2.8127513692832047e-06, |
| "loss": 0.5069, |
| "num_input_tokens_seen": 19317472, |
| "step": 1250 |
| }, |
| { |
| "epoch": 5.148717948717949, |
| "grad_norm": 0.2265625, |
| "learning_rate": 2.8112826761497507e-06, |
| "loss": 0.5116, |
| "num_input_tokens_seen": 19398496, |
| "step": 1255 |
| }, |
| { |
| "epoch": 5.1692307692307695, |
| "grad_norm": 0.21484375, |
| "learning_rate": 2.8098086321582937e-06, |
| "loss": 0.4286, |
| "num_input_tokens_seen": 19466400, |
| "step": 1260 |
| }, |
| { |
| "epoch": 5.18974358974359, |
| "grad_norm": 0.2119140625, |
| "learning_rate": 2.8083292433238602e-06, |
| "loss": 0.5058, |
| "num_input_tokens_seen": 19550336, |
| "step": 1265 |
| }, |
| { |
| "epoch": 5.21025641025641, |
| "grad_norm": 0.2177734375, |
| "learning_rate": 2.8068445156832864e-06, |
| "loss": 0.4587, |
| "num_input_tokens_seen": 19625792, |
| "step": 1270 |
| }, |
| { |
| "epoch": 5.230769230769231, |
| "grad_norm": 0.1943359375, |
| "learning_rate": 2.805354455295196e-06, |
| "loss": 0.4901, |
| "num_input_tokens_seen": 19711776, |
| "step": 1275 |
| }, |
| { |
| "epoch": 5.251282051282051, |
| "grad_norm": 0.173828125, |
| "learning_rate": 2.8038590682399718e-06, |
| "loss": 0.5474, |
| "num_input_tokens_seen": 19794208, |
| "step": 1280 |
| }, |
| { |
| "epoch": 5.271794871794872, |
| "grad_norm": 0.1611328125, |
| "learning_rate": 2.8023583606197336e-06, |
| "loss": 0.4452, |
| "num_input_tokens_seen": 19872128, |
| "step": 1285 |
| }, |
| { |
| "epoch": 5.292307692307692, |
| "grad_norm": 0.1357421875, |
| "learning_rate": 2.800852338558312e-06, |
| "loss": 0.5081, |
| "num_input_tokens_seen": 19953856, |
| "step": 1290 |
| }, |
| { |
| "epoch": 5.312820512820513, |
| "grad_norm": 0.1376953125, |
| "learning_rate": 2.7993410082012247e-06, |
| "loss": 0.4863, |
| "num_input_tokens_seen": 20026848, |
| "step": 1295 |
| }, |
| { |
| "epoch": 5.333333333333333, |
| "grad_norm": 0.138671875, |
| "learning_rate": 2.79782437571565e-06, |
| "loss": 0.4979, |
| "num_input_tokens_seen": 20100928, |
| "step": 1300 |
| }, |
| { |
| "epoch": 5.3538461538461535, |
| "grad_norm": 0.1318359375, |
| "learning_rate": 2.7963024472904013e-06, |
| "loss": 0.4676, |
| "num_input_tokens_seen": 20173504, |
| "step": 1305 |
| }, |
| { |
| "epoch": 5.374358974358975, |
| "grad_norm": 0.12451171875, |
| "learning_rate": 2.7947752291359053e-06, |
| "loss": 0.4256, |
| "num_input_tokens_seen": 20244608, |
| "step": 1310 |
| }, |
| { |
| "epoch": 5.394871794871795, |
| "grad_norm": 0.130859375, |
| "learning_rate": 2.7932427274841715e-06, |
| "loss": 0.4576, |
| "num_input_tokens_seen": 20324992, |
| "step": 1315 |
| }, |
| { |
| "epoch": 5.415384615384616, |
| "grad_norm": 0.1376953125, |
| "learning_rate": 2.7917049485887705e-06, |
| "loss": 0.5155, |
| "num_input_tokens_seen": 20402304, |
| "step": 1320 |
| }, |
| { |
| "epoch": 5.435897435897436, |
| "grad_norm": 0.20703125, |
| "learning_rate": 2.790161898724808e-06, |
| "loss": 0.4304, |
| "num_input_tokens_seen": 20480800, |
| "step": 1325 |
| }, |
| { |
| "epoch": 5.456410256410257, |
| "grad_norm": 0.138671875, |
| "learning_rate": 2.7886135841888973e-06, |
| "loss": 0.4759, |
| "num_input_tokens_seen": 20560096, |
| "step": 1330 |
| }, |
| { |
| "epoch": 5.476923076923077, |
| "grad_norm": 0.15234375, |
| "learning_rate": 2.787060011299135e-06, |
| "loss": 0.4974, |
| "num_input_tokens_seen": 20645216, |
| "step": 1335 |
| }, |
| { |
| "epoch": 5.4974358974358974, |
| "grad_norm": 0.341796875, |
| "learning_rate": 2.785501186395077e-06, |
| "loss": 0.5174, |
| "num_input_tokens_seen": 20738688, |
| "step": 1340 |
| }, |
| { |
| "epoch": 5.517948717948718, |
| "grad_norm": 0.271484375, |
| "learning_rate": 2.7839371158377077e-06, |
| "loss": 0.5272, |
| "num_input_tokens_seen": 20812928, |
| "step": 1345 |
| }, |
| { |
| "epoch": 5.538461538461538, |
| "grad_norm": 0.296875, |
| "learning_rate": 2.78236780600942e-06, |
| "loss": 0.5129, |
| "num_input_tokens_seen": 20890592, |
| "step": 1350 |
| }, |
| { |
| "epoch": 5.558974358974359, |
| "grad_norm": 0.57421875, |
| "learning_rate": 2.780793263313984e-06, |
| "loss": 0.484, |
| "num_input_tokens_seen": 20961984, |
| "step": 1355 |
| }, |
| { |
| "epoch": 5.579487179487179, |
| "grad_norm": 0.8671875, |
| "learning_rate": 2.7792134941765247e-06, |
| "loss": 0.4793, |
| "num_input_tokens_seen": 21030784, |
| "step": 1360 |
| }, |
| { |
| "epoch": 5.6, |
| "grad_norm": 0.267578125, |
| "learning_rate": 2.7776285050434937e-06, |
| "loss": 0.4521, |
| "num_input_tokens_seen": 21108960, |
| "step": 1365 |
| }, |
| { |
| "epoch": 5.62051282051282, |
| "grad_norm": 0.267578125, |
| "learning_rate": 2.7760383023826425e-06, |
| "loss": 0.4192, |
| "num_input_tokens_seen": 21181728, |
| "step": 1370 |
| }, |
| { |
| "epoch": 5.641025641025641, |
| "grad_norm": 0.64453125, |
| "learning_rate": 2.7744428926829993e-06, |
| "loss": 0.5131, |
| "num_input_tokens_seen": 21255328, |
| "step": 1375 |
| }, |
| { |
| "epoch": 5.661538461538462, |
| "grad_norm": 0.271484375, |
| "learning_rate": 2.7728422824548387e-06, |
| "loss": 0.483, |
| "num_input_tokens_seen": 21324064, |
| "step": 1380 |
| }, |
| { |
| "epoch": 5.682051282051282, |
| "grad_norm": 0.341796875, |
| "learning_rate": 2.7712364782296567e-06, |
| "loss": 0.4858, |
| "num_input_tokens_seen": 21399040, |
| "step": 1385 |
| }, |
| { |
| "epoch": 5.702564102564103, |
| "grad_norm": 0.318359375, |
| "learning_rate": 2.769625486560145e-06, |
| "loss": 0.4629, |
| "num_input_tokens_seen": 21472640, |
| "step": 1390 |
| }, |
| { |
| "epoch": 5.723076923076923, |
| "grad_norm": 0.39453125, |
| "learning_rate": 2.7680093140201625e-06, |
| "loss": 0.5023, |
| "num_input_tokens_seen": 21544448, |
| "step": 1395 |
| }, |
| { |
| "epoch": 5.743589743589744, |
| "grad_norm": 0.28515625, |
| "learning_rate": 2.766387967204709e-06, |
| "loss": 0.4903, |
| "num_input_tokens_seen": 21611136, |
| "step": 1400 |
| }, |
| { |
| "epoch": 5.764102564102564, |
| "grad_norm": 0.2578125, |
| "learning_rate": 2.7647614527299007e-06, |
| "loss": 0.5558, |
| "num_input_tokens_seen": 21703040, |
| "step": 1405 |
| }, |
| { |
| "epoch": 5.7846153846153845, |
| "grad_norm": 0.45703125, |
| "learning_rate": 2.763129777232938e-06, |
| "loss": 0.5612, |
| "num_input_tokens_seen": 21784096, |
| "step": 1410 |
| }, |
| { |
| "epoch": 5.805128205128205, |
| "grad_norm": 0.333984375, |
| "learning_rate": 2.7614929473720847e-06, |
| "loss": 0.4683, |
| "num_input_tokens_seen": 21855072, |
| "step": 1415 |
| }, |
| { |
| "epoch": 5.825641025641025, |
| "grad_norm": 0.38671875, |
| "learning_rate": 2.7598509698266346e-06, |
| "loss": 0.5171, |
| "num_input_tokens_seen": 21933312, |
| "step": 1420 |
| }, |
| { |
| "epoch": 5.846153846153846, |
| "grad_norm": 0.359375, |
| "learning_rate": 2.758203851296889e-06, |
| "loss": 0.5214, |
| "num_input_tokens_seen": 22019008, |
| "step": 1425 |
| }, |
| { |
| "epoch": 5.866666666666667, |
| "grad_norm": 0.326171875, |
| "learning_rate": 2.756551598504128e-06, |
| "loss": 0.4975, |
| "num_input_tokens_seen": 22092864, |
| "step": 1430 |
| }, |
| { |
| "epoch": 5.887179487179488, |
| "grad_norm": 0.392578125, |
| "learning_rate": 2.7548942181905816e-06, |
| "loss": 0.4853, |
| "num_input_tokens_seen": 22171584, |
| "step": 1435 |
| }, |
| { |
| "epoch": 5.907692307692308, |
| "grad_norm": 0.375, |
| "learning_rate": 2.753231717119405e-06, |
| "loss": 0.483, |
| "num_input_tokens_seen": 22241376, |
| "step": 1440 |
| }, |
| { |
| "epoch": 5.9282051282051285, |
| "grad_norm": 0.373046875, |
| "learning_rate": 2.751564102074646e-06, |
| "loss": 0.4965, |
| "num_input_tokens_seen": 22313664, |
| "step": 1445 |
| }, |
| { |
| "epoch": 5.948717948717949, |
| "grad_norm": 0.34765625, |
| "learning_rate": 2.749891379861225e-06, |
| "loss": 0.5342, |
| "num_input_tokens_seen": 22397408, |
| "step": 1450 |
| }, |
| { |
| "epoch": 5.969230769230769, |
| "grad_norm": 0.3359375, |
| "learning_rate": 2.748213557304899e-06, |
| "loss": 0.4233, |
| "num_input_tokens_seen": 22473664, |
| "step": 1455 |
| }, |
| { |
| "epoch": 5.98974358974359, |
| "grad_norm": 0.349609375, |
| "learning_rate": 2.74653064125224e-06, |
| "loss": 0.5244, |
| "num_input_tokens_seen": 22553760, |
| "step": 1460 |
| }, |
| { |
| "epoch": 6.01025641025641, |
| "grad_norm": 0.298828125, |
| "learning_rate": 2.7448426385706036e-06, |
| "loss": 0.5211, |
| "num_input_tokens_seen": 22624608, |
| "step": 1465 |
| }, |
| { |
| "epoch": 6.030769230769231, |
| "grad_norm": 0.271484375, |
| "learning_rate": 2.7431495561481027e-06, |
| "loss": 0.5618, |
| "num_input_tokens_seen": 22710048, |
| "step": 1470 |
| }, |
| { |
| "epoch": 6.051282051282051, |
| "grad_norm": 0.2734375, |
| "learning_rate": 2.741451400893578e-06, |
| "loss": 0.5172, |
| "num_input_tokens_seen": 22787392, |
| "step": 1475 |
| }, |
| { |
| "epoch": 6.0717948717948715, |
| "grad_norm": 0.22265625, |
| "learning_rate": 2.739748179736571e-06, |
| "loss": 0.5035, |
| "num_input_tokens_seen": 22865120, |
| "step": 1480 |
| }, |
| { |
| "epoch": 6.092307692307692, |
| "grad_norm": 0.205078125, |
| "learning_rate": 2.7380398996272955e-06, |
| "loss": 0.519, |
| "num_input_tokens_seen": 22952832, |
| "step": 1485 |
| }, |
| { |
| "epoch": 6.112820512820512, |
| "grad_norm": 0.2138671875, |
| "learning_rate": 2.736326567536609e-06, |
| "loss": 0.4438, |
| "num_input_tokens_seen": 23028544, |
| "step": 1490 |
| }, |
| { |
| "epoch": 6.133333333333334, |
| "grad_norm": 0.2021484375, |
| "learning_rate": 2.7346081904559827e-06, |
| "loss": 0.4669, |
| "num_input_tokens_seen": 23100096, |
| "step": 1495 |
| }, |
| { |
| "epoch": 6.153846153846154, |
| "grad_norm": 0.1533203125, |
| "learning_rate": 2.732884775397477e-06, |
| "loss": 0.4702, |
| "num_input_tokens_seen": 23183392, |
| "step": 1500 |
| }, |
| { |
| "epoch": 6.174358974358975, |
| "grad_norm": 0.1533203125, |
| "learning_rate": 2.731156329393709e-06, |
| "loss": 0.5031, |
| "num_input_tokens_seen": 23266208, |
| "step": 1505 |
| }, |
| { |
| "epoch": 6.194871794871795, |
| "grad_norm": 0.1484375, |
| "learning_rate": 2.729422859497825e-06, |
| "loss": 0.5005, |
| "num_input_tokens_seen": 23348064, |
| "step": 1510 |
| }, |
| { |
| "epoch": 6.2153846153846155, |
| "grad_norm": 0.1552734375, |
| "learning_rate": 2.7276843727834727e-06, |
| "loss": 0.4798, |
| "num_input_tokens_seen": 23420128, |
| "step": 1515 |
| }, |
| { |
| "epoch": 6.235897435897436, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 2.725940876344771e-06, |
| "loss": 0.5059, |
| "num_input_tokens_seen": 23497056, |
| "step": 1520 |
| }, |
| { |
| "epoch": 6.256410256410256, |
| "grad_norm": 0.1376953125, |
| "learning_rate": 2.7241923772962823e-06, |
| "loss": 0.4582, |
| "num_input_tokens_seen": 23564928, |
| "step": 1525 |
| }, |
| { |
| "epoch": 6.276923076923077, |
| "grad_norm": 0.1220703125, |
| "learning_rate": 2.722438882772982e-06, |
| "loss": 0.4295, |
| "num_input_tokens_seen": 23646624, |
| "step": 1530 |
| }, |
| { |
| "epoch": 6.297435897435897, |
| "grad_norm": 0.1484375, |
| "learning_rate": 2.720680399930231e-06, |
| "loss": 0.4682, |
| "num_input_tokens_seen": 23716960, |
| "step": 1535 |
| }, |
| { |
| "epoch": 6.317948717948718, |
| "grad_norm": 0.1611328125, |
| "learning_rate": 2.7189169359437443e-06, |
| "loss": 0.4944, |
| "num_input_tokens_seen": 23796032, |
| "step": 1540 |
| }, |
| { |
| "epoch": 6.338461538461538, |
| "grad_norm": 0.1318359375, |
| "learning_rate": 2.7171484980095653e-06, |
| "loss": 0.4405, |
| "num_input_tokens_seen": 23868768, |
| "step": 1545 |
| }, |
| { |
| "epoch": 6.358974358974359, |
| "grad_norm": 0.1953125, |
| "learning_rate": 2.715375093344032e-06, |
| "loss": 0.4742, |
| "num_input_tokens_seen": 23937824, |
| "step": 1550 |
| }, |
| { |
| "epoch": 6.37948717948718, |
| "grad_norm": 0.146484375, |
| "learning_rate": 2.713596729183751e-06, |
| "loss": 0.4654, |
| "num_input_tokens_seen": 24009472, |
| "step": 1555 |
| }, |
| { |
| "epoch": 6.4, |
| "grad_norm": 0.146484375, |
| "learning_rate": 2.7118134127855667e-06, |
| "loss": 0.4686, |
| "num_input_tokens_seen": 24096256, |
| "step": 1560 |
| }, |
| { |
| "epoch": 6.420512820512821, |
| "grad_norm": 0.3203125, |
| "learning_rate": 2.7100251514265317e-06, |
| "loss": 0.5152, |
| "num_input_tokens_seen": 24180640, |
| "step": 1565 |
| }, |
| { |
| "epoch": 6.441025641025641, |
| "grad_norm": 0.392578125, |
| "learning_rate": 2.7082319524038764e-06, |
| "loss": 0.4762, |
| "num_input_tokens_seen": 24251296, |
| "step": 1570 |
| }, |
| { |
| "epoch": 6.461538461538462, |
| "grad_norm": 0.74609375, |
| "learning_rate": 2.706433823034981e-06, |
| "loss": 0.5113, |
| "num_input_tokens_seen": 24329760, |
| "step": 1575 |
| }, |
| { |
| "epoch": 6.482051282051282, |
| "grad_norm": 0.55078125, |
| "learning_rate": 2.7046307706573445e-06, |
| "loss": 0.4942, |
| "num_input_tokens_seen": 24420896, |
| "step": 1580 |
| }, |
| { |
| "epoch": 6.5025641025641026, |
| "grad_norm": 0.703125, |
| "learning_rate": 2.702822802628554e-06, |
| "loss": 0.475, |
| "num_input_tokens_seen": 24495360, |
| "step": 1585 |
| }, |
| { |
| "epoch": 6.523076923076923, |
| "grad_norm": 0.2216796875, |
| "learning_rate": 2.701009926326256e-06, |
| "loss": 0.4801, |
| "num_input_tokens_seen": 24572192, |
| "step": 1590 |
| }, |
| { |
| "epoch": 6.543589743589743, |
| "grad_norm": 0.267578125, |
| "learning_rate": 2.6991921491481267e-06, |
| "loss": 0.4776, |
| "num_input_tokens_seen": 24647552, |
| "step": 1595 |
| }, |
| { |
| "epoch": 6.564102564102564, |
| "grad_norm": 0.578125, |
| "learning_rate": 2.6973694785118394e-06, |
| "loss": 0.4878, |
| "num_input_tokens_seen": 24719136, |
| "step": 1600 |
| }, |
| { |
| "epoch": 6.584615384615384, |
| "grad_norm": 0.37109375, |
| "learning_rate": 2.695541921855037e-06, |
| "loss": 0.5138, |
| "num_input_tokens_seen": 24800320, |
| "step": 1605 |
| }, |
| { |
| "epoch": 6.605128205128205, |
| "grad_norm": 0.310546875, |
| "learning_rate": 2.6937094866353006e-06, |
| "loss": 0.4782, |
| "num_input_tokens_seen": 24877088, |
| "step": 1610 |
| }, |
| { |
| "epoch": 6.625641025641025, |
| "grad_norm": 0.498046875, |
| "learning_rate": 2.6918721803301174e-06, |
| "loss": 0.5043, |
| "num_input_tokens_seen": 24954272, |
| "step": 1615 |
| }, |
| { |
| "epoch": 6.6461538461538465, |
| "grad_norm": 0.294921875, |
| "learning_rate": 2.690030010436853e-06, |
| "loss": 0.4237, |
| "num_input_tokens_seen": 25023744, |
| "step": 1620 |
| }, |
| { |
| "epoch": 6.666666666666667, |
| "grad_norm": 0.64453125, |
| "learning_rate": 2.688182984472719e-06, |
| "loss": 0.5302, |
| "num_input_tokens_seen": 25105664, |
| "step": 1625 |
| }, |
| { |
| "epoch": 6.687179487179487, |
| "grad_norm": 0.3515625, |
| "learning_rate": 2.686331109974743e-06, |
| "loss": 0.4991, |
| "num_input_tokens_seen": 25183680, |
| "step": 1630 |
| }, |
| { |
| "epoch": 6.707692307692308, |
| "grad_norm": 0.443359375, |
| "learning_rate": 2.684474394499738e-06, |
| "loss": 0.5142, |
| "num_input_tokens_seen": 25265920, |
| "step": 1635 |
| }, |
| { |
| "epoch": 6.728205128205128, |
| "grad_norm": 0.4609375, |
| "learning_rate": 2.6826128456242708e-06, |
| "loss": 0.4651, |
| "num_input_tokens_seen": 25343648, |
| "step": 1640 |
| }, |
| { |
| "epoch": 6.748717948717949, |
| "grad_norm": 0.361328125, |
| "learning_rate": 2.680746470944631e-06, |
| "loss": 0.5633, |
| "num_input_tokens_seen": 25418176, |
| "step": 1645 |
| }, |
| { |
| "epoch": 6.769230769230769, |
| "grad_norm": 0.375, |
| "learning_rate": 2.6788752780768007e-06, |
| "loss": 0.5124, |
| "num_input_tokens_seen": 25504832, |
| "step": 1650 |
| }, |
| { |
| "epoch": 6.78974358974359, |
| "grad_norm": 0.62109375, |
| "learning_rate": 2.6769992746564256e-06, |
| "loss": 0.5046, |
| "num_input_tokens_seen": 25582112, |
| "step": 1655 |
| }, |
| { |
| "epoch": 6.81025641025641, |
| "grad_norm": 0.35546875, |
| "learning_rate": 2.6751184683387777e-06, |
| "loss": 0.484, |
| "num_input_tokens_seen": 25656992, |
| "step": 1660 |
| }, |
| { |
| "epoch": 6.8307692307692305, |
| "grad_norm": 0.3203125, |
| "learning_rate": 2.67323286679873e-06, |
| "loss": 0.4526, |
| "num_input_tokens_seen": 25729600, |
| "step": 1665 |
| }, |
| { |
| "epoch": 6.851282051282051, |
| "grad_norm": 0.365234375, |
| "learning_rate": 2.671342477730723e-06, |
| "loss": 0.4563, |
| "num_input_tokens_seen": 25801536, |
| "step": 1670 |
| }, |
| { |
| "epoch": 6.871794871794872, |
| "grad_norm": 0.380859375, |
| "learning_rate": 2.6694473088487324e-06, |
| "loss": 0.4951, |
| "num_input_tokens_seen": 25882912, |
| "step": 1675 |
| }, |
| { |
| "epoch": 6.892307692307693, |
| "grad_norm": 0.32421875, |
| "learning_rate": 2.6675473678862403e-06, |
| "loss": 0.5223, |
| "num_input_tokens_seen": 25957952, |
| "step": 1680 |
| }, |
| { |
| "epoch": 6.912820512820513, |
| "grad_norm": 0.353515625, |
| "learning_rate": 2.6656426625961993e-06, |
| "loss": 0.5471, |
| "num_input_tokens_seen": 26034432, |
| "step": 1685 |
| }, |
| { |
| "epoch": 6.933333333333334, |
| "grad_norm": 0.251953125, |
| "learning_rate": 2.6637332007510063e-06, |
| "loss": 0.4252, |
| "num_input_tokens_seen": 26106656, |
| "step": 1690 |
| }, |
| { |
| "epoch": 6.953846153846154, |
| "grad_norm": 0.31640625, |
| "learning_rate": 2.661818990142465e-06, |
| "loss": 0.5269, |
| "num_input_tokens_seen": 26186976, |
| "step": 1695 |
| }, |
| { |
| "epoch": 6.9743589743589745, |
| "grad_norm": 0.25, |
| "learning_rate": 2.65990003858176e-06, |
| "loss": 0.4487, |
| "num_input_tokens_seen": 26259264, |
| "step": 1700 |
| }, |
| { |
| "epoch": 6.994871794871795, |
| "grad_norm": 0.2353515625, |
| "learning_rate": 2.6579763538994197e-06, |
| "loss": 0.4705, |
| "num_input_tokens_seen": 26333248, |
| "step": 1705 |
| }, |
| { |
| "epoch": 7.015384615384615, |
| "grad_norm": 0.205078125, |
| "learning_rate": 2.656047943945287e-06, |
| "loss": 0.4443, |
| "num_input_tokens_seen": 26404832, |
| "step": 1710 |
| }, |
| { |
| "epoch": 7.035897435897436, |
| "grad_norm": 0.1865234375, |
| "learning_rate": 2.6541148165884885e-06, |
| "loss": 0.4615, |
| "num_input_tokens_seen": 26484608, |
| "step": 1715 |
| }, |
| { |
| "epoch": 7.056410256410256, |
| "grad_norm": 0.1943359375, |
| "learning_rate": 2.652176979717399e-06, |
| "loss": 0.5042, |
| "num_input_tokens_seen": 26556224, |
| "step": 1720 |
| }, |
| { |
| "epoch": 7.076923076923077, |
| "grad_norm": 0.16796875, |
| "learning_rate": 2.6502344412396116e-06, |
| "loss": 0.4601, |
| "num_input_tokens_seen": 26629632, |
| "step": 1725 |
| }, |
| { |
| "epoch": 7.097435897435897, |
| "grad_norm": 0.146484375, |
| "learning_rate": 2.6482872090819053e-06, |
| "loss": 0.534, |
| "num_input_tokens_seen": 26709568, |
| "step": 1730 |
| }, |
| { |
| "epoch": 7.1179487179487175, |
| "grad_norm": 0.201171875, |
| "learning_rate": 2.646335291190211e-06, |
| "loss": 0.4875, |
| "num_input_tokens_seen": 26785728, |
| "step": 1735 |
| }, |
| { |
| "epoch": 7.138461538461539, |
| "grad_norm": 0.150390625, |
| "learning_rate": 2.6443786955295827e-06, |
| "loss": 0.5223, |
| "num_input_tokens_seen": 26865024, |
| "step": 1740 |
| }, |
| { |
| "epoch": 7.158974358974359, |
| "grad_norm": 0.140625, |
| "learning_rate": 2.6424174300841606e-06, |
| "loss": 0.4365, |
| "num_input_tokens_seen": 26934720, |
| "step": 1745 |
| }, |
| { |
| "epoch": 7.17948717948718, |
| "grad_norm": 0.138671875, |
| "learning_rate": 2.6404515028571406e-06, |
| "loss": 0.4951, |
| "num_input_tokens_seen": 27008192, |
| "step": 1750 |
| }, |
| { |
| "epoch": 7.2, |
| "grad_norm": 0.11962890625, |
| "learning_rate": 2.638480921870743e-06, |
| "loss": 0.5132, |
| "num_input_tokens_seen": 27092000, |
| "step": 1755 |
| }, |
| { |
| "epoch": 7.220512820512821, |
| "grad_norm": 0.1298828125, |
| "learning_rate": 2.636505695166177e-06, |
| "loss": 0.4713, |
| "num_input_tokens_seen": 27172160, |
| "step": 1760 |
| }, |
| { |
| "epoch": 7.241025641025641, |
| "grad_norm": 0.16015625, |
| "learning_rate": 2.63452583080361e-06, |
| "loss": 0.479, |
| "num_input_tokens_seen": 27255712, |
| "step": 1765 |
| }, |
| { |
| "epoch": 7.2615384615384615, |
| "grad_norm": 0.1396484375, |
| "learning_rate": 2.6325413368621337e-06, |
| "loss": 0.4967, |
| "num_input_tokens_seen": 27343136, |
| "step": 1770 |
| }, |
| { |
| "epoch": 7.282051282051282, |
| "grad_norm": 0.2470703125, |
| "learning_rate": 2.630552221439732e-06, |
| "loss": 0.4843, |
| "num_input_tokens_seen": 27417312, |
| "step": 1775 |
| }, |
| { |
| "epoch": 7.302564102564102, |
| "grad_norm": 0.1396484375, |
| "learning_rate": 2.6285584926532465e-06, |
| "loss": 0.4738, |
| "num_input_tokens_seen": 27505824, |
| "step": 1780 |
| }, |
| { |
| "epoch": 7.323076923076923, |
| "grad_norm": 0.2890625, |
| "learning_rate": 2.626560158638344e-06, |
| "loss": 0.5716, |
| "num_input_tokens_seen": 27583776, |
| "step": 1785 |
| }, |
| { |
| "epoch": 7.343589743589743, |
| "grad_norm": 0.357421875, |
| "learning_rate": 2.6245572275494845e-06, |
| "loss": 0.515, |
| "num_input_tokens_seen": 27658912, |
| "step": 1790 |
| }, |
| { |
| "epoch": 7.364102564102564, |
| "grad_norm": 0.2314453125, |
| "learning_rate": 2.6225497075598865e-06, |
| "loss": 0.47, |
| "num_input_tokens_seen": 27733472, |
| "step": 1795 |
| }, |
| { |
| "epoch": 7.384615384615385, |
| "grad_norm": 0.51953125, |
| "learning_rate": 2.6205376068614943e-06, |
| "loss": 0.4749, |
| "num_input_tokens_seen": 27812160, |
| "step": 1800 |
| }, |
| { |
| "epoch": 7.4051282051282055, |
| "grad_norm": 0.55078125, |
| "learning_rate": 2.6185209336649438e-06, |
| "loss": 0.4727, |
| "num_input_tokens_seen": 27885024, |
| "step": 1805 |
| }, |
| { |
| "epoch": 7.425641025641026, |
| "grad_norm": 0.75, |
| "learning_rate": 2.61649969619953e-06, |
| "loss": 0.478, |
| "num_input_tokens_seen": 27956480, |
| "step": 1810 |
| }, |
| { |
| "epoch": 7.446153846153846, |
| "grad_norm": 0.28125, |
| "learning_rate": 2.614473902713173e-06, |
| "loss": 0.4778, |
| "num_input_tokens_seen": 28028032, |
| "step": 1815 |
| }, |
| { |
| "epoch": 7.466666666666667, |
| "grad_norm": 0.49609375, |
| "learning_rate": 2.612443561472385e-06, |
| "loss": 0.4443, |
| "num_input_tokens_seen": 28112992, |
| "step": 1820 |
| }, |
| { |
| "epoch": 7.487179487179487, |
| "grad_norm": 0.5, |
| "learning_rate": 2.610408680762234e-06, |
| "loss": 0.5186, |
| "num_input_tokens_seen": 28191520, |
| "step": 1825 |
| }, |
| { |
| "epoch": 7.507692307692308, |
| "grad_norm": 0.314453125, |
| "learning_rate": 2.6083692688863135e-06, |
| "loss": 0.5152, |
| "num_input_tokens_seen": 28277440, |
| "step": 1830 |
| }, |
| { |
| "epoch": 7.528205128205128, |
| "grad_norm": 0.29296875, |
| "learning_rate": 2.6063253341667064e-06, |
| "loss": 0.5173, |
| "num_input_tokens_seen": 28357440, |
| "step": 1835 |
| }, |
| { |
| "epoch": 7.5487179487179485, |
| "grad_norm": 0.326171875, |
| "learning_rate": 2.604276884943953e-06, |
| "loss": 0.4585, |
| "num_input_tokens_seen": 28426656, |
| "step": 1840 |
| }, |
| { |
| "epoch": 7.569230769230769, |
| "grad_norm": 0.279296875, |
| "learning_rate": 2.602223929577013e-06, |
| "loss": 0.4611, |
| "num_input_tokens_seen": 28499968, |
| "step": 1845 |
| }, |
| { |
| "epoch": 7.589743589743589, |
| "grad_norm": 0.310546875, |
| "learning_rate": 2.6001664764432363e-06, |
| "loss": 0.4929, |
| "num_input_tokens_seen": 28573664, |
| "step": 1850 |
| }, |
| { |
| "epoch": 7.61025641025641, |
| "grad_norm": 0.318359375, |
| "learning_rate": 2.5981045339383244e-06, |
| "loss": 0.5018, |
| "num_input_tokens_seen": 28658144, |
| "step": 1855 |
| }, |
| { |
| "epoch": 7.63076923076923, |
| "grad_norm": 0.328125, |
| "learning_rate": 2.596038110476301e-06, |
| "loss": 0.483, |
| "num_input_tokens_seen": 28730944, |
| "step": 1860 |
| }, |
| { |
| "epoch": 7.651282051282052, |
| "grad_norm": 0.52734375, |
| "learning_rate": 2.593967214489473e-06, |
| "loss": 0.5111, |
| "num_input_tokens_seen": 28816384, |
| "step": 1865 |
| }, |
| { |
| "epoch": 7.671794871794872, |
| "grad_norm": 0.37109375, |
| "learning_rate": 2.591891854428398e-06, |
| "loss": 0.4689, |
| "num_input_tokens_seen": 28891616, |
| "step": 1870 |
| }, |
| { |
| "epoch": 7.6923076923076925, |
| "grad_norm": 0.33203125, |
| "learning_rate": 2.5898120387618507e-06, |
| "loss": 0.4917, |
| "num_input_tokens_seen": 28970400, |
| "step": 1875 |
| }, |
| { |
| "epoch": 7.712820512820513, |
| "grad_norm": 0.33203125, |
| "learning_rate": 2.587727775976787e-06, |
| "loss": 0.4956, |
| "num_input_tokens_seen": 29051520, |
| "step": 1880 |
| }, |
| { |
| "epoch": 7.733333333333333, |
| "grad_norm": 0.3046875, |
| "learning_rate": 2.585639074578309e-06, |
| "loss": 0.438, |
| "num_input_tokens_seen": 29128544, |
| "step": 1885 |
| }, |
| { |
| "epoch": 7.753846153846154, |
| "grad_norm": 0.306640625, |
| "learning_rate": 2.5835459430896333e-06, |
| "loss": 0.4644, |
| "num_input_tokens_seen": 29210496, |
| "step": 1890 |
| }, |
| { |
| "epoch": 7.774358974358974, |
| "grad_norm": 0.365234375, |
| "learning_rate": 2.5814483900520522e-06, |
| "loss": 0.4901, |
| "num_input_tokens_seen": 29282400, |
| "step": 1895 |
| }, |
| { |
| "epoch": 7.794871794871795, |
| "grad_norm": 0.4140625, |
| "learning_rate": 2.5793464240249014e-06, |
| "loss": 0.4879, |
| "num_input_tokens_seen": 29352256, |
| "step": 1900 |
| }, |
| { |
| "epoch": 7.815384615384615, |
| "grad_norm": 0.330078125, |
| "learning_rate": 2.5772400535855242e-06, |
| "loss": 0.4552, |
| "num_input_tokens_seen": 29426336, |
| "step": 1905 |
| }, |
| { |
| "epoch": 7.835897435897436, |
| "grad_norm": 0.353515625, |
| "learning_rate": 2.575129287329237e-06, |
| "loss": 0.5417, |
| "num_input_tokens_seen": 29512224, |
| "step": 1910 |
| }, |
| { |
| "epoch": 7.856410256410256, |
| "grad_norm": 0.27734375, |
| "learning_rate": 2.5730141338692926e-06, |
| "loss": 0.4637, |
| "num_input_tokens_seen": 29590112, |
| "step": 1915 |
| }, |
| { |
| "epoch": 7.876923076923077, |
| "grad_norm": 0.31640625, |
| "learning_rate": 2.5708946018368487e-06, |
| "loss": 0.4486, |
| "num_input_tokens_seen": 29672608, |
| "step": 1920 |
| }, |
| { |
| "epoch": 7.897435897435898, |
| "grad_norm": 0.228515625, |
| "learning_rate": 2.568770699880928e-06, |
| "loss": 0.5094, |
| "num_input_tokens_seen": 29755520, |
| "step": 1925 |
| }, |
| { |
| "epoch": 7.917948717948718, |
| "grad_norm": 0.2197265625, |
| "learning_rate": 2.566642436668387e-06, |
| "loss": 0.5111, |
| "num_input_tokens_seen": 29833344, |
| "step": 1930 |
| }, |
| { |
| "epoch": 7.938461538461539, |
| "grad_norm": 0.1923828125, |
| "learning_rate": 2.5645098208838774e-06, |
| "loss": 0.4737, |
| "num_input_tokens_seen": 29904800, |
| "step": 1935 |
| }, |
| { |
| "epoch": 7.958974358974359, |
| "grad_norm": 0.2080078125, |
| "learning_rate": 2.562372861229813e-06, |
| "loss": 0.4384, |
| "num_input_tokens_seen": 29975488, |
| "step": 1940 |
| }, |
| { |
| "epoch": 7.97948717948718, |
| "grad_norm": 0.197265625, |
| "learning_rate": 2.5602315664263337e-06, |
| "loss": 0.4383, |
| "num_input_tokens_seen": 30046496, |
| "step": 1945 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 0.1611328125, |
| "learning_rate": 2.5580859452112685e-06, |
| "loss": 0.4782, |
| "num_input_tokens_seen": 30119840, |
| "step": 1950 |
| }, |
| { |
| "epoch": 8.02051282051282, |
| "grad_norm": 0.1484375, |
| "learning_rate": 2.555936006340101e-06, |
| "loss": 0.5371, |
| "num_input_tokens_seen": 30207040, |
| "step": 1955 |
| }, |
| { |
| "epoch": 8.04102564102564, |
| "grad_norm": 0.1376953125, |
| "learning_rate": 2.553781758585935e-06, |
| "loss": 0.4867, |
| "num_input_tokens_seen": 30283968, |
| "step": 1960 |
| }, |
| { |
| "epoch": 8.061538461538461, |
| "grad_norm": 0.1376953125, |
| "learning_rate": 2.551623210739455e-06, |
| "loss": 0.4309, |
| "num_input_tokens_seen": 30355552, |
| "step": 1965 |
| }, |
| { |
| "epoch": 8.082051282051282, |
| "grad_norm": 0.12109375, |
| "learning_rate": 2.549460371608895e-06, |
| "loss": 0.5087, |
| "num_input_tokens_seen": 30435776, |
| "step": 1970 |
| }, |
| { |
| "epoch": 8.102564102564102, |
| "grad_norm": 0.1328125, |
| "learning_rate": 2.5472932500199976e-06, |
| "loss": 0.4746, |
| "num_input_tokens_seen": 30507616, |
| "step": 1975 |
| }, |
| { |
| "epoch": 8.123076923076923, |
| "grad_norm": 0.1396484375, |
| "learning_rate": 2.5451218548159823e-06, |
| "loss": 0.4833, |
| "num_input_tokens_seen": 30583456, |
| "step": 1980 |
| }, |
| { |
| "epoch": 8.143589743589743, |
| "grad_norm": 0.12890625, |
| "learning_rate": 2.5429461948575077e-06, |
| "loss": 0.4849, |
| "num_input_tokens_seen": 30654176, |
| "step": 1985 |
| }, |
| { |
| "epoch": 8.164102564102564, |
| "grad_norm": 0.1416015625, |
| "learning_rate": 2.540766279022634e-06, |
| "loss": 0.4812, |
| "num_input_tokens_seen": 30721920, |
| "step": 1990 |
| }, |
| { |
| "epoch": 8.184615384615384, |
| "grad_norm": 0.216796875, |
| "learning_rate": 2.53858211620679e-06, |
| "loss": 0.4976, |
| "num_input_tokens_seen": 30791264, |
| "step": 1995 |
| }, |
| { |
| "epoch": 8.205128205128204, |
| "grad_norm": 0.20703125, |
| "learning_rate": 2.536393715322732e-06, |
| "loss": 0.4556, |
| "num_input_tokens_seen": 30862336, |
| "step": 2000 |
| }, |
| { |
| "epoch": 8.225641025641025, |
| "grad_norm": 0.150390625, |
| "learning_rate": 2.5342010853005127e-06, |
| "loss": 0.4496, |
| "num_input_tokens_seen": 30940064, |
| "step": 2005 |
| }, |
| { |
| "epoch": 8.246153846153845, |
| "grad_norm": 0.318359375, |
| "learning_rate": 2.532004235087441e-06, |
| "loss": 0.4722, |
| "num_input_tokens_seen": 31013248, |
| "step": 2010 |
| }, |
| { |
| "epoch": 8.266666666666667, |
| "grad_norm": 0.26171875, |
| "learning_rate": 2.529803173648049e-06, |
| "loss": 0.4875, |
| "num_input_tokens_seen": 31094496, |
| "step": 2015 |
| }, |
| { |
| "epoch": 8.287179487179488, |
| "grad_norm": 0.2578125, |
| "learning_rate": 2.52759790996405e-06, |
| "loss": 0.4598, |
| "num_input_tokens_seen": 31171680, |
| "step": 2020 |
| }, |
| { |
| "epoch": 8.307692307692308, |
| "grad_norm": 0.77734375, |
| "learning_rate": 2.525388453034307e-06, |
| "loss": 0.5069, |
| "num_input_tokens_seen": 31252064, |
| "step": 2025 |
| }, |
| { |
| "epoch": 8.328205128205129, |
| "grad_norm": 0.6875, |
| "learning_rate": 2.5231748118747945e-06, |
| "loss": 0.5155, |
| "num_input_tokens_seen": 31329696, |
| "step": 2030 |
| }, |
| { |
| "epoch": 8.34871794871795, |
| "grad_norm": 0.96484375, |
| "learning_rate": 2.5209569955185604e-06, |
| "loss": 0.5436, |
| "num_input_tokens_seen": 31407648, |
| "step": 2035 |
| }, |
| { |
| "epoch": 8.36923076923077, |
| "grad_norm": 0.30078125, |
| "learning_rate": 2.51873501301569e-06, |
| "loss": 0.4953, |
| "num_input_tokens_seen": 31475200, |
| "step": 2040 |
| }, |
| { |
| "epoch": 8.38974358974359, |
| "grad_norm": 0.6328125, |
| "learning_rate": 2.5165088734332695e-06, |
| "loss": 0.4804, |
| "num_input_tokens_seen": 31547104, |
| "step": 2045 |
| }, |
| { |
| "epoch": 8.41025641025641, |
| "grad_norm": 0.52734375, |
| "learning_rate": 2.5142785858553486e-06, |
| "loss": 0.5533, |
| "num_input_tokens_seen": 31629440, |
| "step": 2050 |
| }, |
| { |
| "epoch": 8.430769230769231, |
| "grad_norm": 0.310546875, |
| "learning_rate": 2.512044159382903e-06, |
| "loss": 0.541, |
| "num_input_tokens_seen": 31713024, |
| "step": 2055 |
| }, |
| { |
| "epoch": 8.451282051282051, |
| "grad_norm": 0.71484375, |
| "learning_rate": 2.5098056031337975e-06, |
| "loss": 0.4444, |
| "num_input_tokens_seen": 31790432, |
| "step": 2060 |
| }, |
| { |
| "epoch": 8.471794871794872, |
| "grad_norm": 0.3046875, |
| "learning_rate": 2.5075629262427507e-06, |
| "loss": 0.4869, |
| "num_input_tokens_seen": 31870592, |
| "step": 2065 |
| }, |
| { |
| "epoch": 8.492307692307692, |
| "grad_norm": 0.6015625, |
| "learning_rate": 2.505316137861294e-06, |
| "loss": 0.4855, |
| "num_input_tokens_seen": 31945344, |
| "step": 2070 |
| }, |
| { |
| "epoch": 8.512820512820513, |
| "grad_norm": 0.283203125, |
| "learning_rate": 2.503065247157737e-06, |
| "loss": 0.5027, |
| "num_input_tokens_seen": 32030016, |
| "step": 2075 |
| }, |
| { |
| "epoch": 8.533333333333333, |
| "grad_norm": 0.3515625, |
| "learning_rate": 2.500810263317129e-06, |
| "loss": 0.4885, |
| "num_input_tokens_seen": 32108160, |
| "step": 2080 |
| }, |
| { |
| "epoch": 8.553846153846154, |
| "grad_norm": 0.33203125, |
| "learning_rate": 2.4985511955412238e-06, |
| "loss": 0.4451, |
| "num_input_tokens_seen": 32188288, |
| "step": 2085 |
| }, |
| { |
| "epoch": 8.574358974358974, |
| "grad_norm": 0.35546875, |
| "learning_rate": 2.4962880530484375e-06, |
| "loss": 0.4899, |
| "num_input_tokens_seen": 32266656, |
| "step": 2090 |
| }, |
| { |
| "epoch": 8.594871794871795, |
| "grad_norm": 0.3515625, |
| "learning_rate": 2.4940208450738146e-06, |
| "loss": 0.5083, |
| "num_input_tokens_seen": 32356544, |
| "step": 2095 |
| }, |
| { |
| "epoch": 8.615384615384615, |
| "grad_norm": 0.37890625, |
| "learning_rate": 2.49174958086899e-06, |
| "loss": 0.4839, |
| "num_input_tokens_seen": 32434720, |
| "step": 2100 |
| }, |
| { |
| "epoch": 8.635897435897435, |
| "grad_norm": 0.373046875, |
| "learning_rate": 2.48947426970215e-06, |
| "loss": 0.437, |
| "num_input_tokens_seen": 32507712, |
| "step": 2105 |
| }, |
| { |
| "epoch": 8.656410256410256, |
| "grad_norm": 0.361328125, |
| "learning_rate": 2.487194920857995e-06, |
| "loss": 0.435, |
| "num_input_tokens_seen": 32577216, |
| "step": 2110 |
| }, |
| { |
| "epoch": 8.676923076923076, |
| "grad_norm": 0.3359375, |
| "learning_rate": 2.484911543637702e-06, |
| "loss": 0.4768, |
| "num_input_tokens_seen": 32647552, |
| "step": 2115 |
| }, |
| { |
| "epoch": 8.697435897435897, |
| "grad_norm": 0.34765625, |
| "learning_rate": 2.4826241473588855e-06, |
| "loss": 0.4578, |
| "num_input_tokens_seen": 32727520, |
| "step": 2120 |
| }, |
| { |
| "epoch": 8.717948717948717, |
| "grad_norm": 0.365234375, |
| "learning_rate": 2.4803327413555623e-06, |
| "loss": 0.5142, |
| "num_input_tokens_seen": 32805440, |
| "step": 2125 |
| }, |
| { |
| "epoch": 8.73846153846154, |
| "grad_norm": 0.35546875, |
| "learning_rate": 2.4780373349781083e-06, |
| "loss": 0.4013, |
| "num_input_tokens_seen": 32880480, |
| "step": 2130 |
| }, |
| { |
| "epoch": 8.75897435897436, |
| "grad_norm": 0.30078125, |
| "learning_rate": 2.4757379375932265e-06, |
| "loss": 0.4616, |
| "num_input_tokens_seen": 32951936, |
| "step": 2135 |
| }, |
| { |
| "epoch": 8.77948717948718, |
| "grad_norm": 0.287109375, |
| "learning_rate": 2.473434558583903e-06, |
| "loss": 0.4791, |
| "num_input_tokens_seen": 33027104, |
| "step": 2140 |
| }, |
| { |
| "epoch": 8.8, |
| "grad_norm": 0.2451171875, |
| "learning_rate": 2.4711272073493745e-06, |
| "loss": 0.5163, |
| "num_input_tokens_seen": 33111040, |
| "step": 2145 |
| }, |
| { |
| "epoch": 8.820512820512821, |
| "grad_norm": 0.2470703125, |
| "learning_rate": 2.468815893305084e-06, |
| "loss": 0.4761, |
| "num_input_tokens_seen": 33202304, |
| "step": 2150 |
| }, |
| { |
| "epoch": 8.841025641025642, |
| "grad_norm": 0.2275390625, |
| "learning_rate": 2.466500625882646e-06, |
| "loss": 0.4405, |
| "num_input_tokens_seen": 33281376, |
| "step": 2155 |
| }, |
| { |
| "epoch": 8.861538461538462, |
| "grad_norm": 0.21484375, |
| "learning_rate": 2.464181414529809e-06, |
| "loss": 0.4538, |
| "num_input_tokens_seen": 33352640, |
| "step": 2160 |
| }, |
| { |
| "epoch": 8.882051282051282, |
| "grad_norm": 0.19140625, |
| "learning_rate": 2.4618582687104132e-06, |
| "loss": 0.4598, |
| "num_input_tokens_seen": 33423232, |
| "step": 2165 |
| }, |
| { |
| "epoch": 8.902564102564103, |
| "grad_norm": 0.169921875, |
| "learning_rate": 2.4595311979043545e-06, |
| "loss": 0.4556, |
| "num_input_tokens_seen": 33503744, |
| "step": 2170 |
| }, |
| { |
| "epoch": 8.923076923076923, |
| "grad_norm": 0.16796875, |
| "learning_rate": 2.4572002116075454e-06, |
| "loss": 0.4665, |
| "num_input_tokens_seen": 33576800, |
| "step": 2175 |
| }, |
| { |
| "epoch": 8.943589743589744, |
| "grad_norm": 0.1748046875, |
| "learning_rate": 2.454865319331876e-06, |
| "loss": 0.4683, |
| "num_input_tokens_seen": 33661120, |
| "step": 2180 |
| }, |
| { |
| "epoch": 8.964102564102564, |
| "grad_norm": 0.1484375, |
| "learning_rate": 2.4525265306051755e-06, |
| "loss": 0.5183, |
| "num_input_tokens_seen": 33733568, |
| "step": 2185 |
| }, |
| { |
| "epoch": 8.984615384615385, |
| "grad_norm": 0.15234375, |
| "learning_rate": 2.4501838549711723e-06, |
| "loss": 0.4671, |
| "num_input_tokens_seen": 33802592, |
| "step": 2190 |
| }, |
| { |
| "epoch": 9.005128205128205, |
| "grad_norm": 0.126953125, |
| "learning_rate": 2.447837301989457e-06, |
| "loss": 0.4858, |
| "num_input_tokens_seen": 33882272, |
| "step": 2195 |
| }, |
| { |
| "epoch": 9.025641025641026, |
| "grad_norm": 0.138671875, |
| "learning_rate": 2.4454868812354403e-06, |
| "loss": 0.4574, |
| "num_input_tokens_seen": 33953920, |
| "step": 2200 |
| }, |
| { |
| "epoch": 9.046153846153846, |
| "grad_norm": 0.11669921875, |
| "learning_rate": 2.4431326023003188e-06, |
| "loss": 0.4419, |
| "num_input_tokens_seen": 34027552, |
| "step": 2205 |
| }, |
| { |
| "epoch": 9.066666666666666, |
| "grad_norm": 0.11669921875, |
| "learning_rate": 2.44077447479103e-06, |
| "loss": 0.5126, |
| "num_input_tokens_seen": 34112480, |
| "step": 2210 |
| }, |
| { |
| "epoch": 9.087179487179487, |
| "grad_norm": 0.1357421875, |
| "learning_rate": 2.4384125083302178e-06, |
| "loss": 0.4517, |
| "num_input_tokens_seen": 34183840, |
| "step": 2215 |
| }, |
| { |
| "epoch": 9.107692307692307, |
| "grad_norm": 0.251953125, |
| "learning_rate": 2.4360467125561907e-06, |
| "loss": 0.5161, |
| "num_input_tokens_seen": 34258912, |
| "step": 2220 |
| }, |
| { |
| "epoch": 9.128205128205128, |
| "grad_norm": 0.240234375, |
| "learning_rate": 2.433677097122883e-06, |
| "loss": 0.4486, |
| "num_input_tokens_seen": 34329248, |
| "step": 2225 |
| }, |
| { |
| "epoch": 9.148717948717948, |
| "grad_norm": 0.166015625, |
| "learning_rate": 2.4313036716998154e-06, |
| "loss": 0.5191, |
| "num_input_tokens_seen": 34411232, |
| "step": 2230 |
| }, |
| { |
| "epoch": 9.169230769230769, |
| "grad_norm": 0.3125, |
| "learning_rate": 2.428926445972058e-06, |
| "loss": 0.5117, |
| "num_input_tokens_seen": 34486144, |
| "step": 2235 |
| }, |
| { |
| "epoch": 9.189743589743589, |
| "grad_norm": 0.2734375, |
| "learning_rate": 2.4265454296401857e-06, |
| "loss": 0.4739, |
| "num_input_tokens_seen": 34564864, |
| "step": 2240 |
| }, |
| { |
| "epoch": 9.21025641025641, |
| "grad_norm": 0.404296875, |
| "learning_rate": 2.4241606324202426e-06, |
| "loss": 0.4468, |
| "num_input_tokens_seen": 34640704, |
| "step": 2245 |
| }, |
| { |
| "epoch": 9.23076923076923, |
| "grad_norm": 0.53125, |
| "learning_rate": 2.4217720640437015e-06, |
| "loss": 0.457, |
| "num_input_tokens_seen": 34715744, |
| "step": 2250 |
| }, |
| { |
| "epoch": 9.25128205128205, |
| "grad_norm": 0.64453125, |
| "learning_rate": 2.4193797342574235e-06, |
| "loss": 0.4915, |
| "num_input_tokens_seen": 34798144, |
| "step": 2255 |
| }, |
| { |
| "epoch": 9.271794871794873, |
| "grad_norm": 0.71875, |
| "learning_rate": 2.4169836528236187e-06, |
| "loss": 0.4417, |
| "num_input_tokens_seen": 34873440, |
| "step": 2260 |
| }, |
| { |
| "epoch": 9.292307692307693, |
| "grad_norm": 0.26953125, |
| "learning_rate": 2.4145838295198066e-06, |
| "loss": 0.4999, |
| "num_input_tokens_seen": 34951552, |
| "step": 2265 |
| }, |
| { |
| "epoch": 9.312820512820513, |
| "grad_norm": 1.203125, |
| "learning_rate": 2.4121802741387743e-06, |
| "loss": 0.453, |
| "num_input_tokens_seen": 35021184, |
| "step": 2270 |
| }, |
| { |
| "epoch": 9.333333333333334, |
| "grad_norm": 0.298828125, |
| "learning_rate": 2.4097729964885407e-06, |
| "loss": 0.4473, |
| "num_input_tokens_seen": 35098080, |
| "step": 2275 |
| }, |
| { |
| "epoch": 9.353846153846154, |
| "grad_norm": 0.53515625, |
| "learning_rate": 2.4073620063923123e-06, |
| "loss": 0.4749, |
| "num_input_tokens_seen": 35170336, |
| "step": 2280 |
| }, |
| { |
| "epoch": 9.374358974358975, |
| "grad_norm": 0.314453125, |
| "learning_rate": 2.404947313688445e-06, |
| "loss": 0.5229, |
| "num_input_tokens_seen": 35255200, |
| "step": 2285 |
| }, |
| { |
| "epoch": 9.394871794871795, |
| "grad_norm": 0.4140625, |
| "learning_rate": 2.4025289282304037e-06, |
| "loss": 0.5158, |
| "num_input_tokens_seen": 35328928, |
| "step": 2290 |
| }, |
| { |
| "epoch": 9.415384615384616, |
| "grad_norm": 0.34375, |
| "learning_rate": 2.4001068598867216e-06, |
| "loss": 0.4548, |
| "num_input_tokens_seen": 35402976, |
| "step": 2295 |
| }, |
| { |
| "epoch": 9.435897435897436, |
| "grad_norm": 0.2373046875, |
| "learning_rate": 2.397681118540961e-06, |
| "loss": 0.4313, |
| "num_input_tokens_seen": 35481344, |
| "step": 2300 |
| }, |
| { |
| "epoch": 9.456410256410257, |
| "grad_norm": 0.369140625, |
| "learning_rate": 2.3952517140916724e-06, |
| "loss": 0.4664, |
| "num_input_tokens_seen": 35553664, |
| "step": 2305 |
| }, |
| { |
| "epoch": 9.476923076923077, |
| "grad_norm": 0.37890625, |
| "learning_rate": 2.392818656452354e-06, |
| "loss": 0.4948, |
| "num_input_tokens_seen": 35630592, |
| "step": 2310 |
| }, |
| { |
| "epoch": 9.497435897435897, |
| "grad_norm": 0.3046875, |
| "learning_rate": 2.39038195555141e-06, |
| "loss": 0.5035, |
| "num_input_tokens_seen": 35712864, |
| "step": 2315 |
| }, |
| { |
| "epoch": 9.517948717948718, |
| "grad_norm": 0.34765625, |
| "learning_rate": 2.387941621332114e-06, |
| "loss": 0.4964, |
| "num_input_tokens_seen": 35790784, |
| "step": 2320 |
| }, |
| { |
| "epoch": 9.538461538461538, |
| "grad_norm": 0.33203125, |
| "learning_rate": 2.3854976637525637e-06, |
| "loss": 0.4684, |
| "num_input_tokens_seen": 35868960, |
| "step": 2325 |
| }, |
| { |
| "epoch": 9.558974358974359, |
| "grad_norm": 0.345703125, |
| "learning_rate": 2.3830500927856433e-06, |
| "loss": 0.5117, |
| "num_input_tokens_seen": 35956832, |
| "step": 2330 |
| }, |
| { |
| "epoch": 9.57948717948718, |
| "grad_norm": 0.345703125, |
| "learning_rate": 2.3805989184189813e-06, |
| "loss": 0.4574, |
| "num_input_tokens_seen": 36027520, |
| "step": 2335 |
| }, |
| { |
| "epoch": 9.6, |
| "grad_norm": 0.30859375, |
| "learning_rate": 2.378144150654911e-06, |
| "loss": 0.5291, |
| "num_input_tokens_seen": 36109248, |
| "step": 2340 |
| }, |
| { |
| "epoch": 9.62051282051282, |
| "grad_norm": 0.357421875, |
| "learning_rate": 2.3756857995104286e-06, |
| "loss": 0.4528, |
| "num_input_tokens_seen": 36179584, |
| "step": 2345 |
| }, |
| { |
| "epoch": 9.64102564102564, |
| "grad_norm": 0.3671875, |
| "learning_rate": 2.3732238750171527e-06, |
| "loss": 0.4976, |
| "num_input_tokens_seen": 36257216, |
| "step": 2350 |
| }, |
| { |
| "epoch": 9.661538461538461, |
| "grad_norm": 0.30859375, |
| "learning_rate": 2.3707583872212837e-06, |
| "loss": 0.4576, |
| "num_input_tokens_seen": 36326272, |
| "step": 2355 |
| }, |
| { |
| "epoch": 9.682051282051281, |
| "grad_norm": 0.2294921875, |
| "learning_rate": 2.3682893461835626e-06, |
| "loss": 0.5315, |
| "num_input_tokens_seen": 36405504, |
| "step": 2360 |
| }, |
| { |
| "epoch": 9.702564102564102, |
| "grad_norm": 0.28515625, |
| "learning_rate": 2.3658167619792294e-06, |
| "loss": 0.5407, |
| "num_input_tokens_seen": 36481536, |
| "step": 2365 |
| }, |
| { |
| "epoch": 9.723076923076922, |
| "grad_norm": 0.267578125, |
| "learning_rate": 2.363340644697983e-06, |
| "loss": 0.4525, |
| "num_input_tokens_seen": 36563200, |
| "step": 2370 |
| }, |
| { |
| "epoch": 9.743589743589745, |
| "grad_norm": 0.205078125, |
| "learning_rate": 2.360861004443939e-06, |
| "loss": 0.5628, |
| "num_input_tokens_seen": 36661120, |
| "step": 2375 |
| }, |
| { |
| "epoch": 9.764102564102565, |
| "grad_norm": 0.2119140625, |
| "learning_rate": 2.358377851335589e-06, |
| "loss": 0.4633, |
| "num_input_tokens_seen": 36738880, |
| "step": 2380 |
| }, |
| { |
| "epoch": 9.784615384615385, |
| "grad_norm": 0.21484375, |
| "learning_rate": 2.3558911955057592e-06, |
| "loss": 0.4438, |
| "num_input_tokens_seen": 36820128, |
| "step": 2385 |
| }, |
| { |
| "epoch": 9.805128205128206, |
| "grad_norm": 0.1806640625, |
| "learning_rate": 2.35340104710157e-06, |
| "loss": 0.4609, |
| "num_input_tokens_seen": 36900128, |
| "step": 2390 |
| }, |
| { |
| "epoch": 9.825641025641026, |
| "grad_norm": 0.1708984375, |
| "learning_rate": 2.350907416284392e-06, |
| "loss": 0.4843, |
| "num_input_tokens_seen": 36969024, |
| "step": 2395 |
| }, |
| { |
| "epoch": 9.846153846153847, |
| "grad_norm": 0.162109375, |
| "learning_rate": 2.348410313229808e-06, |
| "loss": 0.4607, |
| "num_input_tokens_seen": 37053440, |
| "step": 2400 |
| }, |
| { |
| "epoch": 9.866666666666667, |
| "grad_norm": 0.146484375, |
| "learning_rate": 2.3459097481275687e-06, |
| "loss": 0.5134, |
| "num_input_tokens_seen": 37132128, |
| "step": 2405 |
| }, |
| { |
| "epoch": 9.887179487179488, |
| "grad_norm": 0.1396484375, |
| "learning_rate": 2.343405731181552e-06, |
| "loss": 0.478, |
| "num_input_tokens_seen": 37209664, |
| "step": 2410 |
| }, |
| { |
| "epoch": 9.907692307692308, |
| "grad_norm": 0.146484375, |
| "learning_rate": 2.3408982726097227e-06, |
| "loss": 0.4864, |
| "num_input_tokens_seen": 37283936, |
| "step": 2415 |
| }, |
| { |
| "epoch": 9.928205128205128, |
| "grad_norm": 0.1455078125, |
| "learning_rate": 2.3383873826440878e-06, |
| "loss": 0.4876, |
| "num_input_tokens_seen": 37359552, |
| "step": 2420 |
| }, |
| { |
| "epoch": 9.948717948717949, |
| "grad_norm": 0.13671875, |
| "learning_rate": 2.3358730715306574e-06, |
| "loss": 0.5265, |
| "num_input_tokens_seen": 37434112, |
| "step": 2425 |
| }, |
| { |
| "epoch": 9.96923076923077, |
| "grad_norm": 0.12353515625, |
| "learning_rate": 2.3333553495294033e-06, |
| "loss": 0.4759, |
| "num_input_tokens_seen": 37511456, |
| "step": 2430 |
| }, |
| { |
| "epoch": 9.98974358974359, |
| "grad_norm": 0.11865234375, |
| "learning_rate": 2.330834226914214e-06, |
| "loss": 0.4336, |
| "num_input_tokens_seen": 37584320, |
| "step": 2435 |
| }, |
| { |
| "epoch": 10.01025641025641, |
| "grad_norm": 0.123046875, |
| "learning_rate": 2.3283097139728557e-06, |
| "loss": 0.5338, |
| "num_input_tokens_seen": 37672864, |
| "step": 2440 |
| }, |
| { |
| "epoch": 10.03076923076923, |
| "grad_norm": 0.220703125, |
| "learning_rate": 2.3257818210069277e-06, |
| "loss": 0.4542, |
| "num_input_tokens_seen": 37743488, |
| "step": 2445 |
| }, |
| { |
| "epoch": 10.051282051282051, |
| "grad_norm": 0.14453125, |
| "learning_rate": 2.3232505583318246e-06, |
| "loss": 0.5379, |
| "num_input_tokens_seen": 37832256, |
| "step": 2450 |
| }, |
| { |
| "epoch": 10.071794871794872, |
| "grad_norm": 0.1533203125, |
| "learning_rate": 2.320715936276689e-06, |
| "loss": 0.4418, |
| "num_input_tokens_seen": 37908832, |
| "step": 2455 |
| }, |
| { |
| "epoch": 10.092307692307692, |
| "grad_norm": 0.3984375, |
| "learning_rate": 2.3181779651843738e-06, |
| "loss": 0.5294, |
| "num_input_tokens_seen": 37989088, |
| "step": 2460 |
| }, |
| { |
| "epoch": 10.112820512820512, |
| "grad_norm": 0.26953125, |
| "learning_rate": 2.3156366554113967e-06, |
| "loss": 0.4438, |
| "num_input_tokens_seen": 38061024, |
| "step": 2465 |
| }, |
| { |
| "epoch": 10.133333333333333, |
| "grad_norm": 0.318359375, |
| "learning_rate": 2.3130920173278997e-06, |
| "loss": 0.4544, |
| "num_input_tokens_seen": 38136480, |
| "step": 2470 |
| }, |
| { |
| "epoch": 10.153846153846153, |
| "grad_norm": 0.55078125, |
| "learning_rate": 2.310544061317607e-06, |
| "loss": 0.4918, |
| "num_input_tokens_seen": 38213280, |
| "step": 2475 |
| }, |
| { |
| "epoch": 10.174358974358974, |
| "grad_norm": 0.8984375, |
| "learning_rate": 2.307992797777782e-06, |
| "loss": 0.4821, |
| "num_input_tokens_seen": 38283296, |
| "step": 2480 |
| }, |
| { |
| "epoch": 10.194871794871794, |
| "grad_norm": 0.27734375, |
| "learning_rate": 2.3054382371191836e-06, |
| "loss": 0.464, |
| "num_input_tokens_seen": 38360320, |
| "step": 2485 |
| }, |
| { |
| "epoch": 10.215384615384615, |
| "grad_norm": 0.251953125, |
| "learning_rate": 2.3028803897660256e-06, |
| "loss": 0.4829, |
| "num_input_tokens_seen": 38440000, |
| "step": 2490 |
| }, |
| { |
| "epoch": 10.235897435897435, |
| "grad_norm": 0.53125, |
| "learning_rate": 2.3003192661559346e-06, |
| "loss": 0.4495, |
| "num_input_tokens_seen": 38510208, |
| "step": 2495 |
| }, |
| { |
| "epoch": 10.256410256410255, |
| "grad_norm": 0.30859375, |
| "learning_rate": 2.297754876739905e-06, |
| "loss": 0.5503, |
| "num_input_tokens_seen": 38600256, |
| "step": 2500 |
| }, |
| { |
| "epoch": 10.276923076923078, |
| "grad_norm": 0.384765625, |
| "learning_rate": 2.2951872319822597e-06, |
| "loss": 0.5341, |
| "num_input_tokens_seen": 38679168, |
| "step": 2505 |
| }, |
| { |
| "epoch": 10.297435897435898, |
| "grad_norm": 0.3046875, |
| "learning_rate": 2.2926163423606027e-06, |
| "loss": 0.4903, |
| "num_input_tokens_seen": 38752064, |
| "step": 2510 |
| }, |
| { |
| "epoch": 10.317948717948719, |
| "grad_norm": 0.353515625, |
| "learning_rate": 2.2900422183657816e-06, |
| "loss": 0.4543, |
| "num_input_tokens_seen": 38829504, |
| "step": 2515 |
| }, |
| { |
| "epoch": 10.338461538461539, |
| "grad_norm": 0.330078125, |
| "learning_rate": 2.2874648705018403e-06, |
| "loss": 0.5428, |
| "num_input_tokens_seen": 38915904, |
| "step": 2520 |
| }, |
| { |
| "epoch": 10.35897435897436, |
| "grad_norm": 0.349609375, |
| "learning_rate": 2.28488430928598e-06, |
| "loss": 0.4588, |
| "num_input_tokens_seen": 38997760, |
| "step": 2525 |
| }, |
| { |
| "epoch": 10.37948717948718, |
| "grad_norm": 0.37109375, |
| "learning_rate": 2.282300545248512e-06, |
| "loss": 0.4441, |
| "num_input_tokens_seen": 39068000, |
| "step": 2530 |
| }, |
| { |
| "epoch": 10.4, |
| "grad_norm": 0.375, |
| "learning_rate": 2.27971358893282e-06, |
| "loss": 0.4441, |
| "num_input_tokens_seen": 39136480, |
| "step": 2535 |
| }, |
| { |
| "epoch": 10.42051282051282, |
| "grad_norm": 0.421875, |
| "learning_rate": 2.2771234508953116e-06, |
| "loss": 0.442, |
| "num_input_tokens_seen": 39219488, |
| "step": 2540 |
| }, |
| { |
| "epoch": 10.441025641025641, |
| "grad_norm": 0.625, |
| "learning_rate": 2.27453014170538e-06, |
| "loss": 0.5199, |
| "num_input_tokens_seen": 39295712, |
| "step": 2545 |
| }, |
| { |
| "epoch": 10.461538461538462, |
| "grad_norm": 0.345703125, |
| "learning_rate": 2.271933671945357e-06, |
| "loss": 0.5147, |
| "num_input_tokens_seen": 39371008, |
| "step": 2550 |
| }, |
| { |
| "epoch": 10.482051282051282, |
| "grad_norm": 0.359375, |
| "learning_rate": 2.2693340522104727e-06, |
| "loss": 0.5296, |
| "num_input_tokens_seen": 39450944, |
| "step": 2555 |
| }, |
| { |
| "epoch": 10.502564102564103, |
| "grad_norm": 0.310546875, |
| "learning_rate": 2.2667312931088096e-06, |
| "loss": 0.4707, |
| "num_input_tokens_seen": 39530912, |
| "step": 2560 |
| }, |
| { |
| "epoch": 10.523076923076923, |
| "grad_norm": 0.34375, |
| "learning_rate": 2.264125405261263e-06, |
| "loss": 0.4726, |
| "num_input_tokens_seen": 39602400, |
| "step": 2565 |
| }, |
| { |
| "epoch": 10.543589743589743, |
| "grad_norm": 0.37890625, |
| "learning_rate": 2.261516399301493e-06, |
| "loss": 0.4344, |
| "num_input_tokens_seen": 39679232, |
| "step": 2570 |
| }, |
| { |
| "epoch": 10.564102564102564, |
| "grad_norm": 0.302734375, |
| "learning_rate": 2.2589042858758853e-06, |
| "loss": 0.4427, |
| "num_input_tokens_seen": 39755904, |
| "step": 2575 |
| }, |
| { |
| "epoch": 10.584615384615384, |
| "grad_norm": 0.326171875, |
| "learning_rate": 2.256289075643506e-06, |
| "loss": 0.4975, |
| "num_input_tokens_seen": 39826368, |
| "step": 2580 |
| }, |
| { |
| "epoch": 10.605128205128205, |
| "grad_norm": 0.287109375, |
| "learning_rate": 2.2536707792760566e-06, |
| "loss": 0.5045, |
| "num_input_tokens_seen": 39906464, |
| "step": 2585 |
| }, |
| { |
| "epoch": 10.625641025641025, |
| "grad_norm": 0.27734375, |
| "learning_rate": 2.251049407457833e-06, |
| "loss": 0.4833, |
| "num_input_tokens_seen": 39975712, |
| "step": 2590 |
| }, |
| { |
| "epoch": 10.646153846153846, |
| "grad_norm": 0.263671875, |
| "learning_rate": 2.2484249708856823e-06, |
| "loss": 0.4984, |
| "num_input_tokens_seen": 40060032, |
| "step": 2595 |
| }, |
| { |
| "epoch": 10.666666666666666, |
| "grad_norm": 0.2333984375, |
| "learning_rate": 2.2457974802689545e-06, |
| "loss": 0.4186, |
| "num_input_tokens_seen": 40131520, |
| "step": 2600 |
| }, |
| { |
| "epoch": 10.687179487179487, |
| "grad_norm": 0.1962890625, |
| "learning_rate": 2.2431669463294646e-06, |
| "loss": 0.4441, |
| "num_input_tokens_seen": 40205760, |
| "step": 2605 |
| }, |
| { |
| "epoch": 10.707692307692307, |
| "grad_norm": 0.2119140625, |
| "learning_rate": 2.2405333798014453e-06, |
| "loss": 0.5337, |
| "num_input_tokens_seen": 40288992, |
| "step": 2610 |
| }, |
| { |
| "epoch": 10.728205128205127, |
| "grad_norm": 0.208984375, |
| "learning_rate": 2.237896791431505e-06, |
| "loss": 0.4703, |
| "num_input_tokens_seen": 40369440, |
| "step": 2615 |
| }, |
| { |
| "epoch": 10.74871794871795, |
| "grad_norm": 0.1630859375, |
| "learning_rate": 2.2352571919785812e-06, |
| "loss": 0.5217, |
| "num_input_tokens_seen": 40452288, |
| "step": 2620 |
| }, |
| { |
| "epoch": 10.76923076923077, |
| "grad_norm": 0.1572265625, |
| "learning_rate": 2.2326145922139004e-06, |
| "loss": 0.4475, |
| "num_input_tokens_seen": 40523808, |
| "step": 2625 |
| }, |
| { |
| "epoch": 10.78974358974359, |
| "grad_norm": 0.1884765625, |
| "learning_rate": 2.2299690029209313e-06, |
| "loss": 0.4734, |
| "num_input_tokens_seen": 40606496, |
| "step": 2630 |
| }, |
| { |
| "epoch": 10.810256410256411, |
| "grad_norm": 0.1259765625, |
| "learning_rate": 2.227320434895343e-06, |
| "loss": 0.4686, |
| "num_input_tokens_seen": 40684672, |
| "step": 2635 |
| }, |
| { |
| "epoch": 10.830769230769231, |
| "grad_norm": 0.1279296875, |
| "learning_rate": 2.2246688989449577e-06, |
| "loss": 0.5027, |
| "num_input_tokens_seen": 40762752, |
| "step": 2640 |
| }, |
| { |
| "epoch": 10.851282051282052, |
| "grad_norm": 0.12109375, |
| "learning_rate": 2.2220144058897104e-06, |
| "loss": 0.4582, |
| "num_input_tokens_seen": 40851776, |
| "step": 2645 |
| }, |
| { |
| "epoch": 10.871794871794872, |
| "grad_norm": 0.1318359375, |
| "learning_rate": 2.2193569665616017e-06, |
| "loss": 0.4516, |
| "num_input_tokens_seen": 40922304, |
| "step": 2650 |
| }, |
| { |
| "epoch": 10.892307692307693, |
| "grad_norm": 0.12109375, |
| "learning_rate": 2.2166965918046554e-06, |
| "loss": 0.4346, |
| "num_input_tokens_seen": 40998784, |
| "step": 2655 |
| }, |
| { |
| "epoch": 10.912820512820513, |
| "grad_norm": 0.16015625, |
| "learning_rate": 2.214033292474874e-06, |
| "loss": 0.5093, |
| "num_input_tokens_seen": 41075872, |
| "step": 2660 |
| }, |
| { |
| "epoch": 10.933333333333334, |
| "grad_norm": 0.154296875, |
| "learning_rate": 2.2113670794401935e-06, |
| "loss": 0.4853, |
| "num_input_tokens_seen": 41149728, |
| "step": 2665 |
| }, |
| { |
| "epoch": 10.953846153846154, |
| "grad_norm": 0.2333984375, |
| "learning_rate": 2.20869796358044e-06, |
| "loss": 0.5042, |
| "num_input_tokens_seen": 41220160, |
| "step": 2670 |
| }, |
| { |
| "epoch": 10.974358974358974, |
| "grad_norm": 0.1630859375, |
| "learning_rate": 2.2060259557872845e-06, |
| "loss": 0.4601, |
| "num_input_tokens_seen": 41287712, |
| "step": 2675 |
| }, |
| { |
| "epoch": 10.994871794871795, |
| "grad_norm": 0.1533203125, |
| "learning_rate": 2.2033510669641997e-06, |
| "loss": 0.4138, |
| "num_input_tokens_seen": 41363264, |
| "step": 2680 |
| }, |
| { |
| "epoch": 11.015384615384615, |
| "grad_norm": 0.306640625, |
| "learning_rate": 2.2006733080264144e-06, |
| "loss": 0.4724, |
| "num_input_tokens_seen": 41437152, |
| "step": 2685 |
| }, |
| { |
| "epoch": 11.035897435897436, |
| "grad_norm": 0.26953125, |
| "learning_rate": 2.197992689900869e-06, |
| "loss": 0.4932, |
| "num_input_tokens_seen": 41515520, |
| "step": 2690 |
| }, |
| { |
| "epoch": 11.056410256410256, |
| "grad_norm": 0.5859375, |
| "learning_rate": 2.195309223526171e-06, |
| "loss": 0.4299, |
| "num_input_tokens_seen": 41587200, |
| "step": 2695 |
| }, |
| { |
| "epoch": 11.076923076923077, |
| "grad_norm": 0.59375, |
| "learning_rate": 2.192622919852551e-06, |
| "loss": 0.4774, |
| "num_input_tokens_seen": 41663008, |
| "step": 2700 |
| }, |
| { |
| "epoch": 11.097435897435897, |
| "grad_norm": 0.6640625, |
| "learning_rate": 2.1899337898418174e-06, |
| "loss": 0.5241, |
| "num_input_tokens_seen": 41743264, |
| "step": 2705 |
| }, |
| { |
| "epoch": 11.117948717948718, |
| "grad_norm": 0.29296875, |
| "learning_rate": 2.187241844467313e-06, |
| "loss": 0.498, |
| "num_input_tokens_seen": 41823264, |
| "step": 2710 |
| }, |
| { |
| "epoch": 11.138461538461538, |
| "grad_norm": 0.2890625, |
| "learning_rate": 2.1845470947138658e-06, |
| "loss": 0.5311, |
| "num_input_tokens_seen": 41914560, |
| "step": 2715 |
| }, |
| { |
| "epoch": 11.158974358974358, |
| "grad_norm": 0.58984375, |
| "learning_rate": 2.181849551577751e-06, |
| "loss": 0.4464, |
| "num_input_tokens_seen": 41988288, |
| "step": 2720 |
| }, |
| { |
| "epoch": 11.179487179487179, |
| "grad_norm": 0.34375, |
| "learning_rate": 2.179149226066641e-06, |
| "loss": 0.4905, |
| "num_input_tokens_seen": 42068416, |
| "step": 2725 |
| }, |
| { |
| "epoch": 11.2, |
| "grad_norm": 0.275390625, |
| "learning_rate": 2.1764461291995618e-06, |
| "loss": 0.4629, |
| "num_input_tokens_seen": 42139744, |
| "step": 2730 |
| }, |
| { |
| "epoch": 11.22051282051282, |
| "grad_norm": 0.2578125, |
| "learning_rate": 2.173740272006849e-06, |
| "loss": 0.5278, |
| "num_input_tokens_seen": 42228672, |
| "step": 2735 |
| }, |
| { |
| "epoch": 11.24102564102564, |
| "grad_norm": 0.6640625, |
| "learning_rate": 2.1710316655301016e-06, |
| "loss": 0.4513, |
| "num_input_tokens_seen": 42306528, |
| "step": 2740 |
| }, |
| { |
| "epoch": 11.261538461538462, |
| "grad_norm": 0.2470703125, |
| "learning_rate": 2.1683203208221375e-06, |
| "loss": 0.4917, |
| "num_input_tokens_seen": 42389024, |
| "step": 2745 |
| }, |
| { |
| "epoch": 11.282051282051283, |
| "grad_norm": 0.333984375, |
| "learning_rate": 2.165606248946948e-06, |
| "loss": 0.4159, |
| "num_input_tokens_seen": 42473088, |
| "step": 2750 |
| }, |
| { |
| "epoch": 11.302564102564103, |
| "grad_norm": 0.291015625, |
| "learning_rate": 2.1628894609796533e-06, |
| "loss": 0.4917, |
| "num_input_tokens_seen": 42551360, |
| "step": 2755 |
| }, |
| { |
| "epoch": 11.323076923076924, |
| "grad_norm": 0.390625, |
| "learning_rate": 2.1601699680064573e-06, |
| "loss": 0.5037, |
| "num_input_tokens_seen": 42626688, |
| "step": 2760 |
| }, |
| { |
| "epoch": 11.343589743589744, |
| "grad_norm": 0.369140625, |
| "learning_rate": 2.1574477811246014e-06, |
| "loss": 0.4756, |
| "num_input_tokens_seen": 42705056, |
| "step": 2765 |
| }, |
| { |
| "epoch": 11.364102564102565, |
| "grad_norm": 0.421875, |
| "learning_rate": 2.1547229114423207e-06, |
| "loss": 0.4985, |
| "num_input_tokens_seen": 42777632, |
| "step": 2770 |
| }, |
| { |
| "epoch": 11.384615384615385, |
| "grad_norm": 0.326171875, |
| "learning_rate": 2.1519953700787963e-06, |
| "loss": 0.4561, |
| "num_input_tokens_seen": 42845888, |
| "step": 2775 |
| }, |
| { |
| "epoch": 11.405128205128205, |
| "grad_norm": 0.36328125, |
| "learning_rate": 2.149265168164113e-06, |
| "loss": 0.5091, |
| "num_input_tokens_seen": 42922976, |
| "step": 2780 |
| }, |
| { |
| "epoch": 11.425641025641026, |
| "grad_norm": 0.353515625, |
| "learning_rate": 2.146532316839211e-06, |
| "loss": 0.4711, |
| "num_input_tokens_seen": 42996000, |
| "step": 2785 |
| }, |
| { |
| "epoch": 11.446153846153846, |
| "grad_norm": 0.38671875, |
| "learning_rate": 2.1437968272558435e-06, |
| "loss": 0.457, |
| "num_input_tokens_seen": 43074688, |
| "step": 2790 |
| }, |
| { |
| "epoch": 11.466666666666667, |
| "grad_norm": 0.31640625, |
| "learning_rate": 2.1410587105765275e-06, |
| "loss": 0.541, |
| "num_input_tokens_seen": 43157280, |
| "step": 2795 |
| }, |
| { |
| "epoch": 11.487179487179487, |
| "grad_norm": 0.361328125, |
| "learning_rate": 2.138317977974501e-06, |
| "loss": 0.4279, |
| "num_input_tokens_seen": 43234016, |
| "step": 2800 |
| }, |
| { |
| "epoch": 11.507692307692308, |
| "grad_norm": 0.306640625, |
| "learning_rate": 2.135574640633678e-06, |
| "loss": 0.5213, |
| "num_input_tokens_seen": 43310816, |
| "step": 2805 |
| }, |
| { |
| "epoch": 11.528205128205128, |
| "grad_norm": 0.275390625, |
| "learning_rate": 2.132828709748598e-06, |
| "loss": 0.4444, |
| "num_input_tokens_seen": 43382976, |
| "step": 2810 |
| }, |
| { |
| "epoch": 11.548717948717949, |
| "grad_norm": 0.279296875, |
| "learning_rate": 2.130080196524388e-06, |
| "loss": 0.4768, |
| "num_input_tokens_seen": 43462944, |
| "step": 2815 |
| }, |
| { |
| "epoch": 11.569230769230769, |
| "grad_norm": 0.283203125, |
| "learning_rate": 2.1273291121767094e-06, |
| "loss": 0.4376, |
| "num_input_tokens_seen": 43535232, |
| "step": 2820 |
| }, |
| { |
| "epoch": 11.58974358974359, |
| "grad_norm": 0.23828125, |
| "learning_rate": 2.124575467931717e-06, |
| "loss": 0.4342, |
| "num_input_tokens_seen": 43618528, |
| "step": 2825 |
| }, |
| { |
| "epoch": 11.61025641025641, |
| "grad_norm": 0.1982421875, |
| "learning_rate": 2.1218192750260114e-06, |
| "loss": 0.4596, |
| "num_input_tokens_seen": 43691904, |
| "step": 2830 |
| }, |
| { |
| "epoch": 11.63076923076923, |
| "grad_norm": 0.1904296875, |
| "learning_rate": 2.119060544706592e-06, |
| "loss": 0.4811, |
| "num_input_tokens_seen": 43760480, |
| "step": 2835 |
| }, |
| { |
| "epoch": 11.65128205128205, |
| "grad_norm": 0.1923828125, |
| "learning_rate": 2.1162992882308147e-06, |
| "loss": 0.4864, |
| "num_input_tokens_seen": 43829984, |
| "step": 2840 |
| }, |
| { |
| "epoch": 11.671794871794871, |
| "grad_norm": 0.1875, |
| "learning_rate": 2.1135355168663417e-06, |
| "loss": 0.4678, |
| "num_input_tokens_seen": 43906816, |
| "step": 2845 |
| }, |
| { |
| "epoch": 11.692307692307692, |
| "grad_norm": 0.1484375, |
| "learning_rate": 2.1107692418910985e-06, |
| "loss": 0.5001, |
| "num_input_tokens_seen": 43984960, |
| "step": 2850 |
| }, |
| { |
| "epoch": 11.712820512820512, |
| "grad_norm": 0.154296875, |
| "learning_rate": 2.1080004745932274e-06, |
| "loss": 0.4662, |
| "num_input_tokens_seen": 44061440, |
| "step": 2855 |
| }, |
| { |
| "epoch": 11.733333333333333, |
| "grad_norm": 0.1611328125, |
| "learning_rate": 2.1052292262710392e-06, |
| "loss": 0.526, |
| "num_input_tokens_seen": 44147008, |
| "step": 2860 |
| }, |
| { |
| "epoch": 11.753846153846155, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 2.102455508232971e-06, |
| "loss": 0.4821, |
| "num_input_tokens_seen": 44224224, |
| "step": 2865 |
| }, |
| { |
| "epoch": 11.774358974358975, |
| "grad_norm": 0.146484375, |
| "learning_rate": 2.099679331797536e-06, |
| "loss": 0.4923, |
| "num_input_tokens_seen": 44302816, |
| "step": 2870 |
| }, |
| { |
| "epoch": 11.794871794871796, |
| "grad_norm": 0.1650390625, |
| "learning_rate": 2.0969007082932803e-06, |
| "loss": 0.521, |
| "num_input_tokens_seen": 44376160, |
| "step": 2875 |
| }, |
| { |
| "epoch": 11.815384615384616, |
| "grad_norm": 0.125, |
| "learning_rate": 2.0941196490587354e-06, |
| "loss": 0.4932, |
| "num_input_tokens_seen": 44459200, |
| "step": 2880 |
| }, |
| { |
| "epoch": 11.835897435897436, |
| "grad_norm": 0.140625, |
| "learning_rate": 2.0913361654423723e-06, |
| "loss": 0.4866, |
| "num_input_tokens_seen": 44536128, |
| "step": 2885 |
| }, |
| { |
| "epoch": 11.856410256410257, |
| "grad_norm": 0.1357421875, |
| "learning_rate": 2.0885502688025538e-06, |
| "loss": 0.4826, |
| "num_input_tokens_seen": 44605088, |
| "step": 2890 |
| }, |
| { |
| "epoch": 11.876923076923077, |
| "grad_norm": 0.2177734375, |
| "learning_rate": 2.0857619705074912e-06, |
| "loss": 0.4433, |
| "num_input_tokens_seen": 44677984, |
| "step": 2895 |
| }, |
| { |
| "epoch": 11.897435897435898, |
| "grad_norm": 0.166015625, |
| "learning_rate": 2.082971281935195e-06, |
| "loss": 0.4122, |
| "num_input_tokens_seen": 44751200, |
| "step": 2900 |
| }, |
| { |
| "epoch": 11.917948717948718, |
| "grad_norm": 0.28515625, |
| "learning_rate": 2.0801782144734295e-06, |
| "loss": 0.4266, |
| "num_input_tokens_seen": 44824672, |
| "step": 2905 |
| }, |
| { |
| "epoch": 11.938461538461539, |
| "grad_norm": 0.341796875, |
| "learning_rate": 2.0773827795196667e-06, |
| "loss": 0.4574, |
| "num_input_tokens_seen": 44912768, |
| "step": 2910 |
| }, |
| { |
| "epoch": 11.95897435897436, |
| "grad_norm": 0.2578125, |
| "learning_rate": 2.074584988481039e-06, |
| "loss": 0.5026, |
| "num_input_tokens_seen": 44993632, |
| "step": 2915 |
| }, |
| { |
| "epoch": 11.97948717948718, |
| "grad_norm": 0.79296875, |
| "learning_rate": 2.0717848527742935e-06, |
| "loss": 0.5444, |
| "num_input_tokens_seen": 45070848, |
| "step": 2920 |
| }, |
| { |
| "epoch": 12.0, |
| "grad_norm": 0.546875, |
| "learning_rate": 2.0689823838257455e-06, |
| "loss": 0.4509, |
| "num_input_tokens_seen": 45150496, |
| "step": 2925 |
| }, |
| { |
| "epoch": 12.02051282051282, |
| "grad_norm": 0.7421875, |
| "learning_rate": 2.0661775930712297e-06, |
| "loss": 0.4534, |
| "num_input_tokens_seen": 45225440, |
| "step": 2930 |
| }, |
| { |
| "epoch": 12.04102564102564, |
| "grad_norm": 0.3984375, |
| "learning_rate": 2.0633704919560573e-06, |
| "loss": 0.5264, |
| "num_input_tokens_seen": 45300992, |
| "step": 2935 |
| }, |
| { |
| "epoch": 12.061538461538461, |
| "grad_norm": 0.84765625, |
| "learning_rate": 2.0605610919349658e-06, |
| "loss": 0.4373, |
| "num_input_tokens_seen": 45378944, |
| "step": 2940 |
| }, |
| { |
| "epoch": 12.082051282051282, |
| "grad_norm": 0.53515625, |
| "learning_rate": 2.0577494044720746e-06, |
| "loss": 0.4779, |
| "num_input_tokens_seen": 45453504, |
| "step": 2945 |
| }, |
| { |
| "epoch": 12.102564102564102, |
| "grad_norm": 0.265625, |
| "learning_rate": 2.0549354410408364e-06, |
| "loss": 0.556, |
| "num_input_tokens_seen": 45532992, |
| "step": 2950 |
| }, |
| { |
| "epoch": 12.123076923076923, |
| "grad_norm": 0.2890625, |
| "learning_rate": 2.052119213123992e-06, |
| "loss": 0.5152, |
| "num_input_tokens_seen": 45609120, |
| "step": 2955 |
| }, |
| { |
| "epoch": 12.143589743589743, |
| "grad_norm": 0.318359375, |
| "learning_rate": 2.049300732213522e-06, |
| "loss": 0.4412, |
| "num_input_tokens_seen": 45690624, |
| "step": 2960 |
| }, |
| { |
| "epoch": 12.164102564102564, |
| "grad_norm": 0.46875, |
| "learning_rate": 2.046480009810602e-06, |
| "loss": 0.4553, |
| "num_input_tokens_seen": 45763264, |
| "step": 2965 |
| }, |
| { |
| "epoch": 12.184615384615384, |
| "grad_norm": 0.275390625, |
| "learning_rate": 2.0436570574255523e-06, |
| "loss": 0.4913, |
| "num_input_tokens_seen": 45849472, |
| "step": 2970 |
| }, |
| { |
| "epoch": 12.205128205128204, |
| "grad_norm": 0.41015625, |
| "learning_rate": 2.0408318865777953e-06, |
| "loss": 0.5487, |
| "num_input_tokens_seen": 45927552, |
| "step": 2975 |
| }, |
| { |
| "epoch": 12.225641025641025, |
| "grad_norm": 0.291015625, |
| "learning_rate": 2.0380045087958036e-06, |
| "loss": 0.465, |
| "num_input_tokens_seen": 46002656, |
| "step": 2980 |
| }, |
| { |
| "epoch": 12.246153846153845, |
| "grad_norm": 0.404296875, |
| "learning_rate": 2.0351749356170574e-06, |
| "loss": 0.4854, |
| "num_input_tokens_seen": 46087904, |
| "step": 2985 |
| }, |
| { |
| "epoch": 12.266666666666667, |
| "grad_norm": 0.37109375, |
| "learning_rate": 2.032343178587995e-06, |
| "loss": 0.4568, |
| "num_input_tokens_seen": 46165408, |
| "step": 2990 |
| }, |
| { |
| "epoch": 12.287179487179488, |
| "grad_norm": 0.3515625, |
| "learning_rate": 2.0295092492639657e-06, |
| "loss": 0.4926, |
| "num_input_tokens_seen": 46237344, |
| "step": 2995 |
| }, |
| { |
| "epoch": 12.307692307692308, |
| "grad_norm": 0.5, |
| "learning_rate": 2.0266731592091834e-06, |
| "loss": 0.5093, |
| "num_input_tokens_seen": 46307456, |
| "step": 3000 |
| }, |
| { |
| "epoch": 12.328205128205129, |
| "grad_norm": 0.33984375, |
| "learning_rate": 2.0238349199966793e-06, |
| "loss": 0.5077, |
| "num_input_tokens_seen": 46390688, |
| "step": 3005 |
| }, |
| { |
| "epoch": 12.34871794871795, |
| "grad_norm": 0.32421875, |
| "learning_rate": 2.020994543208254e-06, |
| "loss": 0.4541, |
| "num_input_tokens_seen": 46467904, |
| "step": 3010 |
| }, |
| { |
| "epoch": 12.36923076923077, |
| "grad_norm": 0.33203125, |
| "learning_rate": 2.018152040434432e-06, |
| "loss": 0.4975, |
| "num_input_tokens_seen": 46545632, |
| "step": 3015 |
| }, |
| { |
| "epoch": 12.38974358974359, |
| "grad_norm": 0.40234375, |
| "learning_rate": 2.015307423274411e-06, |
| "loss": 0.4988, |
| "num_input_tokens_seen": 46624768, |
| "step": 3020 |
| }, |
| { |
| "epoch": 12.41025641025641, |
| "grad_norm": 0.298828125, |
| "learning_rate": 2.0124607033360193e-06, |
| "loss": 0.4877, |
| "num_input_tokens_seen": 46694528, |
| "step": 3025 |
| }, |
| { |
| "epoch": 12.430769230769231, |
| "grad_norm": 0.3046875, |
| "learning_rate": 2.0096118922356646e-06, |
| "loss": 0.4621, |
| "num_input_tokens_seen": 46767520, |
| "step": 3030 |
| }, |
| { |
| "epoch": 12.451282051282051, |
| "grad_norm": 0.255859375, |
| "learning_rate": 2.0067610015982868e-06, |
| "loss": 0.4742, |
| "num_input_tokens_seen": 46843616, |
| "step": 3035 |
| }, |
| { |
| "epoch": 12.471794871794872, |
| "grad_norm": 0.27734375, |
| "learning_rate": 2.0039080430573133e-06, |
| "loss": 0.3993, |
| "num_input_tokens_seen": 46915136, |
| "step": 3040 |
| }, |
| { |
| "epoch": 12.492307692307692, |
| "grad_norm": 0.232421875, |
| "learning_rate": 2.0010530282546093e-06, |
| "loss": 0.464, |
| "num_input_tokens_seen": 46992000, |
| "step": 3045 |
| }, |
| { |
| "epoch": 12.512820512820513, |
| "grad_norm": 0.2060546875, |
| "learning_rate": 1.9981959688404303e-06, |
| "loss": 0.4744, |
| "num_input_tokens_seen": 47075360, |
| "step": 3050 |
| }, |
| { |
| "epoch": 12.533333333333333, |
| "grad_norm": 0.1953125, |
| "learning_rate": 1.9953368764733763e-06, |
| "loss": 0.4788, |
| "num_input_tokens_seen": 47152704, |
| "step": 3055 |
| }, |
| { |
| "epoch": 12.553846153846154, |
| "grad_norm": 0.1923828125, |
| "learning_rate": 1.992475762820342e-06, |
| "loss": 0.4704, |
| "num_input_tokens_seen": 47229696, |
| "step": 3060 |
| }, |
| { |
| "epoch": 12.574358974358974, |
| "grad_norm": 0.193359375, |
| "learning_rate": 1.9896126395564695e-06, |
| "loss": 0.4645, |
| "num_input_tokens_seen": 47307360, |
| "step": 3065 |
| }, |
| { |
| "epoch": 12.594871794871795, |
| "grad_norm": 0.1728515625, |
| "learning_rate": 1.986747518365104e-06, |
| "loss": 0.5485, |
| "num_input_tokens_seen": 47393824, |
| "step": 3070 |
| }, |
| { |
| "epoch": 12.615384615384615, |
| "grad_norm": 0.154296875, |
| "learning_rate": 1.9838804109377405e-06, |
| "loss": 0.511, |
| "num_input_tokens_seen": 47475104, |
| "step": 3075 |
| }, |
| { |
| "epoch": 12.635897435897435, |
| "grad_norm": 0.1630859375, |
| "learning_rate": 1.9810113289739818e-06, |
| "loss": 0.4624, |
| "num_input_tokens_seen": 47548704, |
| "step": 3080 |
| }, |
| { |
| "epoch": 12.656410256410256, |
| "grad_norm": 0.1435546875, |
| "learning_rate": 1.9781402841814855e-06, |
| "loss": 0.5197, |
| "num_input_tokens_seen": 47638880, |
| "step": 3085 |
| }, |
| { |
| "epoch": 12.676923076923076, |
| "grad_norm": 0.134765625, |
| "learning_rate": 1.9752672882759204e-06, |
| "loss": 0.4277, |
| "num_input_tokens_seen": 47711456, |
| "step": 3090 |
| }, |
| { |
| "epoch": 12.697435897435897, |
| "grad_norm": 0.138671875, |
| "learning_rate": 1.972392352980917e-06, |
| "loss": 0.5166, |
| "num_input_tokens_seen": 47787776, |
| "step": 3095 |
| }, |
| { |
| "epoch": 12.717948717948717, |
| "grad_norm": 0.1357421875, |
| "learning_rate": 1.969515490028019e-06, |
| "loss": 0.4201, |
| "num_input_tokens_seen": 47860480, |
| "step": 3100 |
| }, |
| { |
| "epoch": 12.73846153846154, |
| "grad_norm": 0.1064453125, |
| "learning_rate": 1.966636711156636e-06, |
| "loss": 0.4425, |
| "num_input_tokens_seen": 47936704, |
| "step": 3105 |
| }, |
| { |
| "epoch": 12.75897435897436, |
| "grad_norm": 0.154296875, |
| "learning_rate": 1.9637560281139982e-06, |
| "loss": 0.5056, |
| "num_input_tokens_seen": 48021792, |
| "step": 3110 |
| }, |
| { |
| "epoch": 12.77948717948718, |
| "grad_norm": 0.1953125, |
| "learning_rate": 1.960873452655102e-06, |
| "loss": 0.4929, |
| "num_input_tokens_seen": 48101984, |
| "step": 3115 |
| }, |
| { |
| "epoch": 12.8, |
| "grad_norm": 0.2392578125, |
| "learning_rate": 1.95798899654267e-06, |
| "loss": 0.4868, |
| "num_input_tokens_seen": 48186688, |
| "step": 3120 |
| }, |
| { |
| "epoch": 12.820512820512821, |
| "grad_norm": 0.134765625, |
| "learning_rate": 1.9551026715470954e-06, |
| "loss": 0.4668, |
| "num_input_tokens_seen": 48259872, |
| "step": 3125 |
| }, |
| { |
| "epoch": 12.841025641025642, |
| "grad_norm": 0.6953125, |
| "learning_rate": 1.952214489446401e-06, |
| "loss": 0.53, |
| "num_input_tokens_seen": 48347232, |
| "step": 3130 |
| }, |
| { |
| "epoch": 12.861538461538462, |
| "grad_norm": 0.296875, |
| "learning_rate": 1.949324462026185e-06, |
| "loss": 0.4388, |
| "num_input_tokens_seen": 48422624, |
| "step": 3135 |
| }, |
| { |
| "epoch": 12.882051282051282, |
| "grad_norm": 0.28125, |
| "learning_rate": 1.9464326010795776e-06, |
| "loss": 0.4246, |
| "num_input_tokens_seen": 48492288, |
| "step": 3140 |
| }, |
| { |
| "epoch": 12.902564102564103, |
| "grad_norm": 0.72265625, |
| "learning_rate": 1.9435389184071895e-06, |
| "loss": 0.5186, |
| "num_input_tokens_seen": 48568224, |
| "step": 3145 |
| }, |
| { |
| "epoch": 12.923076923076923, |
| "grad_norm": 0.7734375, |
| "learning_rate": 1.9406434258170666e-06, |
| "loss": 0.4351, |
| "num_input_tokens_seen": 48637280, |
| "step": 3150 |
| }, |
| { |
| "epoch": 12.943589743589744, |
| "grad_norm": 2.03125, |
| "learning_rate": 1.9377461351246395e-06, |
| "loss": 0.5281, |
| "num_input_tokens_seen": 48717088, |
| "step": 3155 |
| }, |
| { |
| "epoch": 12.964102564102564, |
| "grad_norm": 0.236328125, |
| "learning_rate": 1.9348470581526763e-06, |
| "loss": 0.4308, |
| "num_input_tokens_seen": 48787584, |
| "step": 3160 |
| }, |
| { |
| "epoch": 12.984615384615385, |
| "grad_norm": 0.52734375, |
| "learning_rate": 1.9319462067312344e-06, |
| "loss": 0.4133, |
| "num_input_tokens_seen": 48864640, |
| "step": 3165 |
| }, |
| { |
| "epoch": 13.005128205128205, |
| "grad_norm": 0.55859375, |
| "learning_rate": 1.929043592697612e-06, |
| "loss": 0.4802, |
| "num_input_tokens_seen": 48938176, |
| "step": 3170 |
| }, |
| { |
| "epoch": 13.025641025641026, |
| "grad_norm": 0.306640625, |
| "learning_rate": 1.926139227896299e-06, |
| "loss": 0.4504, |
| "num_input_tokens_seen": 49015040, |
| "step": 3175 |
| }, |
| { |
| "epoch": 13.046153846153846, |
| "grad_norm": 0.4140625, |
| "learning_rate": 1.923233124178932e-06, |
| "loss": 0.494, |
| "num_input_tokens_seen": 49091680, |
| "step": 3180 |
| }, |
| { |
| "epoch": 13.066666666666666, |
| "grad_norm": 0.3671875, |
| "learning_rate": 1.9203252934042403e-06, |
| "loss": 0.4628, |
| "num_input_tokens_seen": 49165600, |
| "step": 3185 |
| }, |
| { |
| "epoch": 13.087179487179487, |
| "grad_norm": 0.318359375, |
| "learning_rate": 1.9174157474380034e-06, |
| "loss": 0.4431, |
| "num_input_tokens_seen": 49247136, |
| "step": 3190 |
| }, |
| { |
| "epoch": 13.107692307692307, |
| "grad_norm": 0.27734375, |
| "learning_rate": 1.914504498152998e-06, |
| "loss": 0.4479, |
| "num_input_tokens_seen": 49319648, |
| "step": 3195 |
| }, |
| { |
| "epoch": 13.128205128205128, |
| "grad_norm": 0.40625, |
| "learning_rate": 1.9115915574289525e-06, |
| "loss": 0.4278, |
| "num_input_tokens_seen": 49393216, |
| "step": 3200 |
| }, |
| { |
| "epoch": 13.148717948717948, |
| "grad_norm": 0.302734375, |
| "learning_rate": 1.9086769371524966e-06, |
| "loss": 0.5124, |
| "num_input_tokens_seen": 49480864, |
| "step": 3205 |
| }, |
| { |
| "epoch": 13.169230769230769, |
| "grad_norm": 0.3515625, |
| "learning_rate": 1.9057606492171144e-06, |
| "loss": 0.4438, |
| "num_input_tokens_seen": 49559904, |
| "step": 3210 |
| }, |
| { |
| "epoch": 13.189743589743589, |
| "grad_norm": 0.412109375, |
| "learning_rate": 1.9028427055230948e-06, |
| "loss": 0.4312, |
| "num_input_tokens_seen": 49630240, |
| "step": 3215 |
| }, |
| { |
| "epoch": 13.21025641025641, |
| "grad_norm": 0.443359375, |
| "learning_rate": 1.8999231179774833e-06, |
| "loss": 0.5431, |
| "num_input_tokens_seen": 49709888, |
| "step": 3220 |
| }, |
| { |
| "epoch": 13.23076923076923, |
| "grad_norm": 0.447265625, |
| "learning_rate": 1.897001898494033e-06, |
| "loss": 0.4567, |
| "num_input_tokens_seen": 49778848, |
| "step": 3225 |
| }, |
| { |
| "epoch": 13.25128205128205, |
| "grad_norm": 0.396484375, |
| "learning_rate": 1.8940790589931568e-06, |
| "loss": 0.4227, |
| "num_input_tokens_seen": 49849024, |
| "step": 3230 |
| }, |
| { |
| "epoch": 13.271794871794873, |
| "grad_norm": 0.306640625, |
| "learning_rate": 1.8911546114018775e-06, |
| "loss": 0.4738, |
| "num_input_tokens_seen": 49917952, |
| "step": 3235 |
| }, |
| { |
| "epoch": 13.292307692307693, |
| "grad_norm": 0.3203125, |
| "learning_rate": 1.888228567653781e-06, |
| "loss": 0.4661, |
| "num_input_tokens_seen": 50003392, |
| "step": 3240 |
| }, |
| { |
| "epoch": 13.312820512820513, |
| "grad_norm": 0.330078125, |
| "learning_rate": 1.8853009396889665e-06, |
| "loss": 0.5039, |
| "num_input_tokens_seen": 50085280, |
| "step": 3245 |
| }, |
| { |
| "epoch": 13.333333333333334, |
| "grad_norm": 0.33984375, |
| "learning_rate": 1.8823717394539966e-06, |
| "loss": 0.4442, |
| "num_input_tokens_seen": 50166304, |
| "step": 3250 |
| }, |
| { |
| "epoch": 13.353846153846154, |
| "grad_norm": 0.271484375, |
| "learning_rate": 1.8794409789018507e-06, |
| "loss": 0.5048, |
| "num_input_tokens_seen": 50247200, |
| "step": 3255 |
| }, |
| { |
| "epoch": 13.374358974358975, |
| "grad_norm": 0.275390625, |
| "learning_rate": 1.8765086699918747e-06, |
| "loss": 0.4615, |
| "num_input_tokens_seen": 50321120, |
| "step": 3260 |
| }, |
| { |
| "epoch": 13.394871794871795, |
| "grad_norm": 0.275390625, |
| "learning_rate": 1.8735748246897337e-06, |
| "loss": 0.4838, |
| "num_input_tokens_seen": 50403680, |
| "step": 3265 |
| }, |
| { |
| "epoch": 13.415384615384616, |
| "grad_norm": 0.203125, |
| "learning_rate": 1.8706394549673615e-06, |
| "loss": 0.4454, |
| "num_input_tokens_seen": 50481504, |
| "step": 3270 |
| }, |
| { |
| "epoch": 13.435897435897436, |
| "grad_norm": 0.220703125, |
| "learning_rate": 1.8677025728029122e-06, |
| "loss": 0.4806, |
| "num_input_tokens_seen": 50556448, |
| "step": 3275 |
| }, |
| { |
| "epoch": 13.456410256410257, |
| "grad_norm": 0.2060546875, |
| "learning_rate": 1.8647641901807126e-06, |
| "loss": 0.4601, |
| "num_input_tokens_seen": 50630304, |
| "step": 3280 |
| }, |
| { |
| "epoch": 13.476923076923077, |
| "grad_norm": 0.2236328125, |
| "learning_rate": 1.861824319091212e-06, |
| "loss": 0.4513, |
| "num_input_tokens_seen": 50704704, |
| "step": 3285 |
| }, |
| { |
| "epoch": 13.497435897435897, |
| "grad_norm": 0.1572265625, |
| "learning_rate": 1.8588829715309324e-06, |
| "loss": 0.4704, |
| "num_input_tokens_seen": 50782464, |
| "step": 3290 |
| }, |
| { |
| "epoch": 13.517948717948718, |
| "grad_norm": 0.166015625, |
| "learning_rate": 1.855940159502423e-06, |
| "loss": 0.4959, |
| "num_input_tokens_seen": 50861696, |
| "step": 3295 |
| }, |
| { |
| "epoch": 13.538461538461538, |
| "grad_norm": 0.1591796875, |
| "learning_rate": 1.8529958950142066e-06, |
| "loss": 0.5215, |
| "num_input_tokens_seen": 50949568, |
| "step": 3300 |
| }, |
| { |
| "epoch": 13.558974358974359, |
| "grad_norm": 0.14453125, |
| "learning_rate": 1.8500501900807345e-06, |
| "loss": 0.4692, |
| "num_input_tokens_seen": 51018432, |
| "step": 3305 |
| }, |
| { |
| "epoch": 13.57948717948718, |
| "grad_norm": 0.1435546875, |
| "learning_rate": 1.847103056722335e-06, |
| "loss": 0.5015, |
| "num_input_tokens_seen": 51100352, |
| "step": 3310 |
| }, |
| { |
| "epoch": 13.6, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 1.8441545069651665e-06, |
| "loss": 0.4779, |
| "num_input_tokens_seen": 51178144, |
| "step": 3315 |
| }, |
| { |
| "epoch": 13.62051282051282, |
| "grad_norm": 0.146484375, |
| "learning_rate": 1.8412045528411652e-06, |
| "loss": 0.4971, |
| "num_input_tokens_seen": 51262432, |
| "step": 3320 |
| }, |
| { |
| "epoch": 13.64102564102564, |
| "grad_norm": 0.11279296875, |
| "learning_rate": 1.8382532063880005e-06, |
| "loss": 0.5063, |
| "num_input_tokens_seen": 51342624, |
| "step": 3325 |
| }, |
| { |
| "epoch": 13.661538461538461, |
| "grad_norm": 0.1328125, |
| "learning_rate": 1.8353004796490212e-06, |
| "loss": 0.4273, |
| "num_input_tokens_seen": 51413920, |
| "step": 3330 |
| }, |
| { |
| "epoch": 13.682051282051281, |
| "grad_norm": 0.14453125, |
| "learning_rate": 1.8323463846732099e-06, |
| "loss": 0.5491, |
| "num_input_tokens_seen": 51501184, |
| "step": 3335 |
| }, |
| { |
| "epoch": 13.702564102564102, |
| "grad_norm": 0.2109375, |
| "learning_rate": 1.8293909335151316e-06, |
| "loss": 0.4199, |
| "num_input_tokens_seen": 51570944, |
| "step": 3340 |
| }, |
| { |
| "epoch": 13.723076923076922, |
| "grad_norm": 0.236328125, |
| "learning_rate": 1.826434138234886e-06, |
| "loss": 0.4468, |
| "num_input_tokens_seen": 51650784, |
| "step": 3345 |
| }, |
| { |
| "epoch": 13.743589743589745, |
| "grad_norm": 0.150390625, |
| "learning_rate": 1.8234760108980572e-06, |
| "loss": 0.5225, |
| "num_input_tokens_seen": 51731008, |
| "step": 3350 |
| }, |
| { |
| "epoch": 13.764102564102565, |
| "grad_norm": 0.33203125, |
| "learning_rate": 1.820516563575665e-06, |
| "loss": 0.4634, |
| "num_input_tokens_seen": 51809056, |
| "step": 3355 |
| }, |
| { |
| "epoch": 13.784615384615385, |
| "grad_norm": 0.298828125, |
| "learning_rate": 1.8175558083441164e-06, |
| "loss": 0.5321, |
| "num_input_tokens_seen": 51886688, |
| "step": 3360 |
| }, |
| { |
| "epoch": 13.805128205128206, |
| "grad_norm": 0.2119140625, |
| "learning_rate": 1.814593757285154e-06, |
| "loss": 0.4633, |
| "num_input_tokens_seen": 51959488, |
| "step": 3365 |
| }, |
| { |
| "epoch": 13.825641025641026, |
| "grad_norm": 0.59765625, |
| "learning_rate": 1.8116304224858092e-06, |
| "loss": 0.4812, |
| "num_input_tokens_seen": 52035424, |
| "step": 3370 |
| }, |
| { |
| "epoch": 13.846153846153847, |
| "grad_norm": 0.859375, |
| "learning_rate": 1.8086658160383524e-06, |
| "loss": 0.506, |
| "num_input_tokens_seen": 52112576, |
| "step": 3375 |
| }, |
| { |
| "epoch": 13.866666666666667, |
| "grad_norm": 0.83984375, |
| "learning_rate": 1.8056999500402414e-06, |
| "loss": 0.5091, |
| "num_input_tokens_seen": 52200480, |
| "step": 3380 |
| }, |
| { |
| "epoch": 13.887179487179488, |
| "grad_norm": 0.296875, |
| "learning_rate": 1.8027328365940755e-06, |
| "loss": 0.4861, |
| "num_input_tokens_seen": 52274368, |
| "step": 3385 |
| }, |
| { |
| "epoch": 13.907692307692308, |
| "grad_norm": 0.462890625, |
| "learning_rate": 1.799764487807543e-06, |
| "loss": 0.5051, |
| "num_input_tokens_seen": 52350656, |
| "step": 3390 |
| }, |
| { |
| "epoch": 13.928205128205128, |
| "grad_norm": 0.279296875, |
| "learning_rate": 1.7967949157933742e-06, |
| "loss": 0.4608, |
| "num_input_tokens_seen": 52427328, |
| "step": 3395 |
| }, |
| { |
| "epoch": 13.948717948717949, |
| "grad_norm": 0.3203125, |
| "learning_rate": 1.7938241326692907e-06, |
| "loss": 0.4719, |
| "num_input_tokens_seen": 52503840, |
| "step": 3400 |
| }, |
| { |
| "epoch": 13.96923076923077, |
| "grad_norm": 0.291015625, |
| "learning_rate": 1.7908521505579554e-06, |
| "loss": 0.5062, |
| "num_input_tokens_seen": 52577056, |
| "step": 3405 |
| }, |
| { |
| "epoch": 13.98974358974359, |
| "grad_norm": 0.310546875, |
| "learning_rate": 1.7878789815869247e-06, |
| "loss": 0.5096, |
| "num_input_tokens_seen": 52656000, |
| "step": 3410 |
| }, |
| { |
| "epoch": 14.01025641025641, |
| "grad_norm": 0.291015625, |
| "learning_rate": 1.7849046378885977e-06, |
| "loss": 0.4476, |
| "num_input_tokens_seen": 52725312, |
| "step": 3415 |
| }, |
| { |
| "epoch": 14.03076923076923, |
| "grad_norm": 0.361328125, |
| "learning_rate": 1.7819291316001679e-06, |
| "loss": 0.436, |
| "num_input_tokens_seen": 52804736, |
| "step": 3420 |
| }, |
| { |
| "epoch": 14.051282051282051, |
| "grad_norm": 0.353515625, |
| "learning_rate": 1.7789524748635717e-06, |
| "loss": 0.4761, |
| "num_input_tokens_seen": 52885056, |
| "step": 3425 |
| }, |
| { |
| "epoch": 14.071794871794872, |
| "grad_norm": 0.36328125, |
| "learning_rate": 1.775974679825441e-06, |
| "loss": 0.5039, |
| "num_input_tokens_seen": 52963136, |
| "step": 3430 |
| }, |
| { |
| "epoch": 14.092307692307692, |
| "grad_norm": 0.470703125, |
| "learning_rate": 1.7729957586370525e-06, |
| "loss": 0.5039, |
| "num_input_tokens_seen": 53049856, |
| "step": 3435 |
| }, |
| { |
| "epoch": 14.112820512820512, |
| "grad_norm": 0.45703125, |
| "learning_rate": 1.7700157234542773e-06, |
| "loss": 0.5397, |
| "num_input_tokens_seen": 53127616, |
| "step": 3440 |
| }, |
| { |
| "epoch": 14.133333333333333, |
| "grad_norm": 0.34765625, |
| "learning_rate": 1.7670345864375339e-06, |
| "loss": 0.4747, |
| "num_input_tokens_seen": 53201600, |
| "step": 3445 |
| }, |
| { |
| "epoch": 14.153846153846153, |
| "grad_norm": 0.37890625, |
| "learning_rate": 1.7640523597517357e-06, |
| "loss": 0.4582, |
| "num_input_tokens_seen": 53275520, |
| "step": 3450 |
| }, |
| { |
| "epoch": 14.174358974358974, |
| "grad_norm": 0.328125, |
| "learning_rate": 1.7610690555662435e-06, |
| "loss": 0.4747, |
| "num_input_tokens_seen": 53346720, |
| "step": 3455 |
| }, |
| { |
| "epoch": 14.194871794871794, |
| "grad_norm": 0.337890625, |
| "learning_rate": 1.7580846860548146e-06, |
| "loss": 0.4758, |
| "num_input_tokens_seen": 53420864, |
| "step": 3460 |
| }, |
| { |
| "epoch": 14.215384615384615, |
| "grad_norm": 0.380859375, |
| "learning_rate": 1.7550992633955535e-06, |
| "loss": 0.4305, |
| "num_input_tokens_seen": 53490976, |
| "step": 3465 |
| }, |
| { |
| "epoch": 14.235897435897435, |
| "grad_norm": 0.357421875, |
| "learning_rate": 1.7521127997708621e-06, |
| "loss": 0.4533, |
| "num_input_tokens_seen": 53570816, |
| "step": 3470 |
| }, |
| { |
| "epoch": 14.256410256410255, |
| "grad_norm": 0.30078125, |
| "learning_rate": 1.7491253073673903e-06, |
| "loss": 0.5019, |
| "num_input_tokens_seen": 53647072, |
| "step": 3475 |
| }, |
| { |
| "epoch": 14.276923076923078, |
| "grad_norm": 0.28515625, |
| "learning_rate": 1.7461367983759862e-06, |
| "loss": 0.4575, |
| "num_input_tokens_seen": 53727648, |
| "step": 3480 |
| }, |
| { |
| "epoch": 14.297435897435898, |
| "grad_norm": 0.28515625, |
| "learning_rate": 1.7431472849916455e-06, |
| "loss": 0.4728, |
| "num_input_tokens_seen": 53800544, |
| "step": 3485 |
| }, |
| { |
| "epoch": 14.317948717948719, |
| "grad_norm": 0.298828125, |
| "learning_rate": 1.7401567794134636e-06, |
| "loss": 0.4677, |
| "num_input_tokens_seen": 53875552, |
| "step": 3490 |
| }, |
| { |
| "epoch": 14.338461538461539, |
| "grad_norm": 0.216796875, |
| "learning_rate": 1.737165293844583e-06, |
| "loss": 0.456, |
| "num_input_tokens_seen": 53954208, |
| "step": 3495 |
| }, |
| { |
| "epoch": 14.35897435897436, |
| "grad_norm": 0.205078125, |
| "learning_rate": 1.7341728404921471e-06, |
| "loss": 0.4798, |
| "num_input_tokens_seen": 54033248, |
| "step": 3500 |
| }, |
| { |
| "epoch": 14.37948717948718, |
| "grad_norm": 0.208984375, |
| "learning_rate": 1.7311794315672477e-06, |
| "loss": 0.4166, |
| "num_input_tokens_seen": 54109440, |
| "step": 3505 |
| }, |
| { |
| "epoch": 14.4, |
| "grad_norm": 0.1953125, |
| "learning_rate": 1.7281850792848752e-06, |
| "loss": 0.4663, |
| "num_input_tokens_seen": 54186368, |
| "step": 3510 |
| }, |
| { |
| "epoch": 14.42051282051282, |
| "grad_norm": 0.1591796875, |
| "learning_rate": 1.7251897958638704e-06, |
| "loss": 0.4053, |
| "num_input_tokens_seen": 54267648, |
| "step": 3515 |
| }, |
| { |
| "epoch": 14.441025641025641, |
| "grad_norm": 0.193359375, |
| "learning_rate": 1.7221935935268735e-06, |
| "loss": 0.5342, |
| "num_input_tokens_seen": 54342304, |
| "step": 3520 |
| }, |
| { |
| "epoch": 14.461538461538462, |
| "grad_norm": 0.1455078125, |
| "learning_rate": 1.719196484500274e-06, |
| "loss": 0.4992, |
| "num_input_tokens_seen": 54423808, |
| "step": 3525 |
| }, |
| { |
| "epoch": 14.482051282051282, |
| "grad_norm": 0.1494140625, |
| "learning_rate": 1.7161984810141625e-06, |
| "loss": 0.5067, |
| "num_input_tokens_seen": 54501760, |
| "step": 3530 |
| }, |
| { |
| "epoch": 14.502564102564103, |
| "grad_norm": 0.126953125, |
| "learning_rate": 1.7131995953022776e-06, |
| "loss": 0.4286, |
| "num_input_tokens_seen": 54576512, |
| "step": 3535 |
| }, |
| { |
| "epoch": 14.523076923076923, |
| "grad_norm": 0.1611328125, |
| "learning_rate": 1.7101998396019593e-06, |
| "loss": 0.436, |
| "num_input_tokens_seen": 54649696, |
| "step": 3540 |
| }, |
| { |
| "epoch": 14.543589743589743, |
| "grad_norm": 0.138671875, |
| "learning_rate": 1.7071992261540983e-06, |
| "loss": 0.4889, |
| "num_input_tokens_seen": 54728864, |
| "step": 3545 |
| }, |
| { |
| "epoch": 14.564102564102564, |
| "grad_norm": 0.1328125, |
| "learning_rate": 1.7041977672030842e-06, |
| "loss": 0.4585, |
| "num_input_tokens_seen": 54797856, |
| "step": 3550 |
| }, |
| { |
| "epoch": 14.584615384615384, |
| "grad_norm": 0.1064453125, |
| "learning_rate": 1.7011954749967564e-06, |
| "loss": 0.4646, |
| "num_input_tokens_seen": 54875168, |
| "step": 3555 |
| }, |
| { |
| "epoch": 14.605128205128205, |
| "grad_norm": 0.1474609375, |
| "learning_rate": 1.6981923617863566e-06, |
| "loss": 0.4894, |
| "num_input_tokens_seen": 54949312, |
| "step": 3560 |
| }, |
| { |
| "epoch": 14.625641025641025, |
| "grad_norm": 0.2255859375, |
| "learning_rate": 1.6951884398264742e-06, |
| "loss": 0.5159, |
| "num_input_tokens_seen": 55024768, |
| "step": 3565 |
| }, |
| { |
| "epoch": 14.646153846153846, |
| "grad_norm": 0.150390625, |
| "learning_rate": 1.692183721375001e-06, |
| "loss": 0.4221, |
| "num_input_tokens_seen": 55100128, |
| "step": 3570 |
| }, |
| { |
| "epoch": 14.666666666666666, |
| "grad_norm": 0.1435546875, |
| "learning_rate": 1.6891782186930767e-06, |
| "loss": 0.5417, |
| "num_input_tokens_seen": 55182080, |
| "step": 3575 |
| }, |
| { |
| "epoch": 14.687179487179487, |
| "grad_norm": 0.28515625, |
| "learning_rate": 1.6861719440450437e-06, |
| "loss": 0.5072, |
| "num_input_tokens_seen": 55274048, |
| "step": 3580 |
| }, |
| { |
| "epoch": 14.707692307692307, |
| "grad_norm": 0.267578125, |
| "learning_rate": 1.6831649096983923e-06, |
| "loss": 0.5008, |
| "num_input_tokens_seen": 55347648, |
| "step": 3585 |
| }, |
| { |
| "epoch": 14.728205128205127, |
| "grad_norm": 0.33203125, |
| "learning_rate": 1.680157127923715e-06, |
| "loss": 0.4411, |
| "num_input_tokens_seen": 55418368, |
| "step": 3590 |
| }, |
| { |
| "epoch": 14.74871794871795, |
| "grad_norm": 0.74609375, |
| "learning_rate": 1.677148610994652e-06, |
| "loss": 0.4654, |
| "num_input_tokens_seen": 55490176, |
| "step": 3595 |
| }, |
| { |
| "epoch": 14.76923076923077, |
| "grad_norm": 0.70703125, |
| "learning_rate": 1.6741393711878454e-06, |
| "loss": 0.4576, |
| "num_input_tokens_seen": 55570784, |
| "step": 3600 |
| }, |
| { |
| "epoch": 14.78974358974359, |
| "grad_norm": 0.2353515625, |
| "learning_rate": 1.6711294207828852e-06, |
| "loss": 0.4488, |
| "num_input_tokens_seen": 55643936, |
| "step": 3605 |
| }, |
| { |
| "epoch": 14.810256410256411, |
| "grad_norm": 0.349609375, |
| "learning_rate": 1.6681187720622627e-06, |
| "loss": 0.5175, |
| "num_input_tokens_seen": 55727936, |
| "step": 3610 |
| }, |
| { |
| "epoch": 14.830769230769231, |
| "grad_norm": 0.484375, |
| "learning_rate": 1.6651074373113176e-06, |
| "loss": 0.4271, |
| "num_input_tokens_seen": 55804992, |
| "step": 3615 |
| }, |
| { |
| "epoch": 14.851282051282052, |
| "grad_norm": 0.3203125, |
| "learning_rate": 1.66209542881819e-06, |
| "loss": 0.4631, |
| "num_input_tokens_seen": 55878976, |
| "step": 3620 |
| }, |
| { |
| "epoch": 14.871794871794872, |
| "grad_norm": 0.26171875, |
| "learning_rate": 1.6590827588737685e-06, |
| "loss": 0.5453, |
| "num_input_tokens_seen": 55969376, |
| "step": 3625 |
| }, |
| { |
| "epoch": 14.892307692307693, |
| "grad_norm": 0.34765625, |
| "learning_rate": 1.6560694397716412e-06, |
| "loss": 0.5086, |
| "num_input_tokens_seen": 56041024, |
| "step": 3630 |
| }, |
| { |
| "epoch": 14.912820512820513, |
| "grad_norm": 0.28515625, |
| "learning_rate": 1.6530554838080458e-06, |
| "loss": 0.4602, |
| "num_input_tokens_seen": 56117152, |
| "step": 3635 |
| }, |
| { |
| "epoch": 14.933333333333334, |
| "grad_norm": 0.251953125, |
| "learning_rate": 1.6500409032818175e-06, |
| "loss": 0.5046, |
| "num_input_tokens_seen": 56191264, |
| "step": 3640 |
| }, |
| { |
| "epoch": 14.953846153846154, |
| "grad_norm": 0.271484375, |
| "learning_rate": 1.6470257104943413e-06, |
| "loss": 0.53, |
| "num_input_tokens_seen": 56279264, |
| "step": 3645 |
| }, |
| { |
| "epoch": 14.974358974358974, |
| "grad_norm": 0.341796875, |
| "learning_rate": 1.6440099177494991e-06, |
| "loss": 0.4639, |
| "num_input_tokens_seen": 56350784, |
| "step": 3650 |
| }, |
| { |
| "epoch": 14.994871794871795, |
| "grad_norm": 0.32421875, |
| "learning_rate": 1.6409935373536227e-06, |
| "loss": 0.5015, |
| "num_input_tokens_seen": 56430432, |
| "step": 3655 |
| }, |
| { |
| "epoch": 15.015384615384615, |
| "grad_norm": 0.33203125, |
| "learning_rate": 1.6379765816154413e-06, |
| "loss": 0.4503, |
| "num_input_tokens_seen": 56509056, |
| "step": 3660 |
| }, |
| { |
| "epoch": 15.035897435897436, |
| "grad_norm": 0.322265625, |
| "learning_rate": 1.634959062846031e-06, |
| "loss": 0.4688, |
| "num_input_tokens_seen": 56578016, |
| "step": 3665 |
| }, |
| { |
| "epoch": 15.056410256410256, |
| "grad_norm": 0.34375, |
| "learning_rate": 1.631940993358767e-06, |
| "loss": 0.4761, |
| "num_input_tokens_seen": 56653536, |
| "step": 3670 |
| }, |
| { |
| "epoch": 15.076923076923077, |
| "grad_norm": 0.328125, |
| "learning_rate": 1.6289223854692708e-06, |
| "loss": 0.4433, |
| "num_input_tokens_seen": 56730560, |
| "step": 3675 |
| }, |
| { |
| "epoch": 15.097435897435897, |
| "grad_norm": 0.3046875, |
| "learning_rate": 1.6259032514953601e-06, |
| "loss": 0.4315, |
| "num_input_tokens_seen": 56813312, |
| "step": 3680 |
| }, |
| { |
| "epoch": 15.117948717948718, |
| "grad_norm": 0.283203125, |
| "learning_rate": 1.6228836037570015e-06, |
| "loss": 0.4885, |
| "num_input_tokens_seen": 56891264, |
| "step": 3685 |
| }, |
| { |
| "epoch": 15.138461538461538, |
| "grad_norm": 0.400390625, |
| "learning_rate": 1.619863454576256e-06, |
| "loss": 0.4703, |
| "num_input_tokens_seen": 56964864, |
| "step": 3690 |
| }, |
| { |
| "epoch": 15.158974358974358, |
| "grad_norm": 0.322265625, |
| "learning_rate": 1.6168428162772322e-06, |
| "loss": 0.4518, |
| "num_input_tokens_seen": 57050144, |
| "step": 3695 |
| }, |
| { |
| "epoch": 15.179487179487179, |
| "grad_norm": 0.328125, |
| "learning_rate": 1.6138217011860336e-06, |
| "loss": 0.4858, |
| "num_input_tokens_seen": 57125632, |
| "step": 3700 |
| }, |
| { |
| "epoch": 15.2, |
| "grad_norm": 0.296875, |
| "learning_rate": 1.6108001216307107e-06, |
| "loss": 0.459, |
| "num_input_tokens_seen": 57204128, |
| "step": 3705 |
| }, |
| { |
| "epoch": 15.22051282051282, |
| "grad_norm": 0.33984375, |
| "learning_rate": 1.6077780899412068e-06, |
| "loss": 0.4639, |
| "num_input_tokens_seen": 57281536, |
| "step": 3710 |
| }, |
| { |
| "epoch": 15.24102564102564, |
| "grad_norm": 0.27734375, |
| "learning_rate": 1.6047556184493133e-06, |
| "loss": 0.4496, |
| "num_input_tokens_seen": 57354816, |
| "step": 3715 |
| }, |
| { |
| "epoch": 15.261538461538462, |
| "grad_norm": 0.22265625, |
| "learning_rate": 1.601732719488614e-06, |
| "loss": 0.5043, |
| "num_input_tokens_seen": 57435328, |
| "step": 3720 |
| }, |
| { |
| "epoch": 15.282051282051283, |
| "grad_norm": 0.2109375, |
| "learning_rate": 1.5987094053944384e-06, |
| "loss": 0.4357, |
| "num_input_tokens_seen": 57507424, |
| "step": 3725 |
| }, |
| { |
| "epoch": 15.302564102564103, |
| "grad_norm": 0.228515625, |
| "learning_rate": 1.5956856885038086e-06, |
| "loss": 0.5211, |
| "num_input_tokens_seen": 57586880, |
| "step": 3730 |
| }, |
| { |
| "epoch": 15.323076923076924, |
| "grad_norm": 0.1982421875, |
| "learning_rate": 1.592661581155392e-06, |
| "loss": 0.5036, |
| "num_input_tokens_seen": 57669600, |
| "step": 3735 |
| }, |
| { |
| "epoch": 15.343589743589744, |
| "grad_norm": 0.1708984375, |
| "learning_rate": 1.5896370956894477e-06, |
| "loss": 0.4751, |
| "num_input_tokens_seen": 57744608, |
| "step": 3740 |
| }, |
| { |
| "epoch": 15.364102564102565, |
| "grad_norm": 0.1533203125, |
| "learning_rate": 1.5866122444477794e-06, |
| "loss": 0.4764, |
| "num_input_tokens_seen": 57818496, |
| "step": 3745 |
| }, |
| { |
| "epoch": 15.384615384615385, |
| "grad_norm": 0.1376953125, |
| "learning_rate": 1.5835870397736817e-06, |
| "loss": 0.4551, |
| "num_input_tokens_seen": 57898432, |
| "step": 3750 |
| }, |
| { |
| "epoch": 15.405128205128205, |
| "grad_norm": 0.166015625, |
| "learning_rate": 1.5805614940118928e-06, |
| "loss": 0.5198, |
| "num_input_tokens_seen": 57974624, |
| "step": 3755 |
| }, |
| { |
| "epoch": 15.425641025641026, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 1.577535619508542e-06, |
| "loss": 0.5164, |
| "num_input_tokens_seen": 58052576, |
| "step": 3760 |
| }, |
| { |
| "epoch": 15.446153846153846, |
| "grad_norm": 0.130859375, |
| "learning_rate": 1.5745094286111004e-06, |
| "loss": 0.4699, |
| "num_input_tokens_seen": 58126592, |
| "step": 3765 |
| }, |
| { |
| "epoch": 15.466666666666667, |
| "grad_norm": 0.125, |
| "learning_rate": 1.5714829336683297e-06, |
| "loss": 0.4804, |
| "num_input_tokens_seen": 58214784, |
| "step": 3770 |
| }, |
| { |
| "epoch": 15.487179487179487, |
| "grad_norm": 0.1259765625, |
| "learning_rate": 1.5684561470302337e-06, |
| "loss": 0.4731, |
| "num_input_tokens_seen": 58284480, |
| "step": 3775 |
| }, |
| { |
| "epoch": 15.507692307692308, |
| "grad_norm": 0.1279296875, |
| "learning_rate": 1.5654290810480041e-06, |
| "loss": 0.5115, |
| "num_input_tokens_seen": 58359488, |
| "step": 3780 |
| }, |
| { |
| "epoch": 15.528205128205128, |
| "grad_norm": 0.1201171875, |
| "learning_rate": 1.562401748073975e-06, |
| "loss": 0.4713, |
| "num_input_tokens_seen": 58433632, |
| "step": 3785 |
| }, |
| { |
| "epoch": 15.548717948717949, |
| "grad_norm": 0.275390625, |
| "learning_rate": 1.5593741604615679e-06, |
| "loss": 0.5218, |
| "num_input_tokens_seen": 58526048, |
| "step": 3790 |
| }, |
| { |
| "epoch": 15.569230769230769, |
| "grad_norm": 0.15234375, |
| "learning_rate": 1.5563463305652454e-06, |
| "loss": 0.4432, |
| "num_input_tokens_seen": 58593440, |
| "step": 3795 |
| }, |
| { |
| "epoch": 15.58974358974359, |
| "grad_norm": 0.1552734375, |
| "learning_rate": 1.5533182707404563e-06, |
| "loss": 0.4793, |
| "num_input_tokens_seen": 58664832, |
| "step": 3800 |
| }, |
| { |
| "epoch": 15.61025641025641, |
| "grad_norm": 0.29296875, |
| "learning_rate": 1.550289993343591e-06, |
| "loss": 0.4287, |
| "num_input_tokens_seen": 58734976, |
| "step": 3805 |
| }, |
| { |
| "epoch": 15.63076923076923, |
| "grad_norm": 0.37109375, |
| "learning_rate": 1.547261510731924e-06, |
| "loss": 0.5701, |
| "num_input_tokens_seen": 58826304, |
| "step": 3810 |
| }, |
| { |
| "epoch": 15.65128205128205, |
| "grad_norm": 0.515625, |
| "learning_rate": 1.5442328352635706e-06, |
| "loss": 0.4858, |
| "num_input_tokens_seen": 58897216, |
| "step": 3815 |
| }, |
| { |
| "epoch": 15.671794871794871, |
| "grad_norm": 0.63671875, |
| "learning_rate": 1.54120397929743e-06, |
| "loss": 0.4541, |
| "num_input_tokens_seen": 58970880, |
| "step": 3820 |
| }, |
| { |
| "epoch": 15.692307692307692, |
| "grad_norm": 0.91015625, |
| "learning_rate": 1.5381749551931405e-06, |
| "loss": 0.5178, |
| "num_input_tokens_seen": 59049408, |
| "step": 3825 |
| }, |
| { |
| "epoch": 15.712820512820512, |
| "grad_norm": 0.298828125, |
| "learning_rate": 1.5351457753110244e-06, |
| "loss": 0.4799, |
| "num_input_tokens_seen": 59122336, |
| "step": 3830 |
| }, |
| { |
| "epoch": 15.733333333333333, |
| "grad_norm": 0.2333984375, |
| "learning_rate": 1.5321164520120421e-06, |
| "loss": 0.4284, |
| "num_input_tokens_seen": 59197600, |
| "step": 3835 |
| }, |
| { |
| "epoch": 15.753846153846155, |
| "grad_norm": 0.640625, |
| "learning_rate": 1.5290869976577366e-06, |
| "loss": 0.4343, |
| "num_input_tokens_seen": 59268224, |
| "step": 3840 |
| }, |
| { |
| "epoch": 15.774358974358975, |
| "grad_norm": 0.279296875, |
| "learning_rate": 1.5260574246101875e-06, |
| "loss": 0.4371, |
| "num_input_tokens_seen": 59341408, |
| "step": 3845 |
| }, |
| { |
| "epoch": 15.794871794871796, |
| "grad_norm": 0.27734375, |
| "learning_rate": 1.5230277452319585e-06, |
| "loss": 0.5017, |
| "num_input_tokens_seen": 59423648, |
| "step": 3850 |
| }, |
| { |
| "epoch": 15.815384615384616, |
| "grad_norm": 0.35546875, |
| "learning_rate": 1.519997971886046e-06, |
| "loss": 0.4473, |
| "num_input_tokens_seen": 59495872, |
| "step": 3855 |
| }, |
| { |
| "epoch": 15.835897435897436, |
| "grad_norm": 0.31640625, |
| "learning_rate": 1.5169681169358314e-06, |
| "loss": 0.4693, |
| "num_input_tokens_seen": 59574880, |
| "step": 3860 |
| }, |
| { |
| "epoch": 15.856410256410257, |
| "grad_norm": 0.3046875, |
| "learning_rate": 1.513938192745028e-06, |
| "loss": 0.4955, |
| "num_input_tokens_seen": 59657472, |
| "step": 3865 |
| }, |
| { |
| "epoch": 15.876923076923077, |
| "grad_norm": 0.337890625, |
| "learning_rate": 1.5109082116776328e-06, |
| "loss": 0.4219, |
| "num_input_tokens_seen": 59733888, |
| "step": 3870 |
| }, |
| { |
| "epoch": 15.897435897435898, |
| "grad_norm": 0.423828125, |
| "learning_rate": 1.5078781860978732e-06, |
| "loss": 0.4462, |
| "num_input_tokens_seen": 59811392, |
| "step": 3875 |
| }, |
| { |
| "epoch": 15.917948717948718, |
| "grad_norm": 0.359375, |
| "learning_rate": 1.5048481283701594e-06, |
| "loss": 0.4938, |
| "num_input_tokens_seen": 59886784, |
| "step": 3880 |
| }, |
| { |
| "epoch": 15.938461538461539, |
| "grad_norm": 0.390625, |
| "learning_rate": 1.501818050859033e-06, |
| "loss": 0.4465, |
| "num_input_tokens_seen": 59959072, |
| "step": 3885 |
| }, |
| { |
| "epoch": 15.95897435897436, |
| "grad_norm": 0.408203125, |
| "learning_rate": 1.498787965929116e-06, |
| "loss": 0.462, |
| "num_input_tokens_seen": 60032832, |
| "step": 3890 |
| }, |
| { |
| "epoch": 15.97948717948718, |
| "grad_norm": 0.28515625, |
| "learning_rate": 1.49575788594506e-06, |
| "loss": 0.5279, |
| "num_input_tokens_seen": 60119456, |
| "step": 3895 |
| }, |
| { |
| "epoch": 16.0, |
| "grad_norm": 0.30859375, |
| "learning_rate": 1.4927278232714975e-06, |
| "loss": 0.4908, |
| "num_input_tokens_seen": 60204448, |
| "step": 3900 |
| }, |
| { |
| "epoch": 16.02051282051282, |
| "grad_norm": 0.37109375, |
| "learning_rate": 1.4896977902729893e-06, |
| "loss": 0.4607, |
| "num_input_tokens_seen": 60279104, |
| "step": 3905 |
| }, |
| { |
| "epoch": 16.04102564102564, |
| "grad_norm": 0.30859375, |
| "learning_rate": 1.4866677993139761e-06, |
| "loss": 0.4405, |
| "num_input_tokens_seen": 60349504, |
| "step": 3910 |
| }, |
| { |
| "epoch": 16.06153846153846, |
| "grad_norm": 0.34375, |
| "learning_rate": 1.4836378627587266e-06, |
| "loss": 0.4744, |
| "num_input_tokens_seen": 60420640, |
| "step": 3915 |
| }, |
| { |
| "epoch": 16.08205128205128, |
| "grad_norm": 0.32421875, |
| "learning_rate": 1.4806079929712874e-06, |
| "loss": 0.4868, |
| "num_input_tokens_seen": 60497728, |
| "step": 3920 |
| }, |
| { |
| "epoch": 16.102564102564102, |
| "grad_norm": 0.341796875, |
| "learning_rate": 1.477578202315433e-06, |
| "loss": 0.5095, |
| "num_input_tokens_seen": 60579008, |
| "step": 3925 |
| }, |
| { |
| "epoch": 16.123076923076923, |
| "grad_norm": 0.271484375, |
| "learning_rate": 1.4745485031546143e-06, |
| "loss": 0.5101, |
| "num_input_tokens_seen": 60657088, |
| "step": 3930 |
| }, |
| { |
| "epoch": 16.143589743589743, |
| "grad_norm": 0.28515625, |
| "learning_rate": 1.4715189078519094e-06, |
| "loss": 0.4938, |
| "num_input_tokens_seen": 60734560, |
| "step": 3935 |
| }, |
| { |
| "epoch": 16.164102564102564, |
| "grad_norm": 0.265625, |
| "learning_rate": 1.468489428769973e-06, |
| "loss": 0.4785, |
| "num_input_tokens_seen": 60812960, |
| "step": 3940 |
| }, |
| { |
| "epoch": 16.184615384615384, |
| "grad_norm": 0.2138671875, |
| "learning_rate": 1.4654600782709843e-06, |
| "loss": 0.4953, |
| "num_input_tokens_seen": 60892768, |
| "step": 3945 |
| }, |
| { |
| "epoch": 16.205128205128204, |
| "grad_norm": 0.208984375, |
| "learning_rate": 1.4624308687165985e-06, |
| "loss": 0.4472, |
| "num_input_tokens_seen": 60964096, |
| "step": 3950 |
| }, |
| { |
| "epoch": 16.225641025641025, |
| "grad_norm": 0.1943359375, |
| "learning_rate": 1.4594018124678965e-06, |
| "loss": 0.4191, |
| "num_input_tokens_seen": 61034304, |
| "step": 3955 |
| }, |
| { |
| "epoch": 16.246153846153845, |
| "grad_norm": 0.181640625, |
| "learning_rate": 1.4563729218853329e-06, |
| "loss": 0.465, |
| "num_input_tokens_seen": 61113024, |
| "step": 3960 |
| }, |
| { |
| "epoch": 16.266666666666666, |
| "grad_norm": 0.1787109375, |
| "learning_rate": 1.4533442093286853e-06, |
| "loss": 0.4703, |
| "num_input_tokens_seen": 61184256, |
| "step": 3965 |
| }, |
| { |
| "epoch": 16.287179487179486, |
| "grad_norm": 0.1416015625, |
| "learning_rate": 1.4503156871570062e-06, |
| "loss": 0.4773, |
| "num_input_tokens_seen": 61262208, |
| "step": 3970 |
| }, |
| { |
| "epoch": 16.307692307692307, |
| "grad_norm": 0.1630859375, |
| "learning_rate": 1.4472873677285706e-06, |
| "loss": 0.4172, |
| "num_input_tokens_seen": 61334560, |
| "step": 3975 |
| }, |
| { |
| "epoch": 16.328205128205127, |
| "grad_norm": 0.1484375, |
| "learning_rate": 1.4442592634008266e-06, |
| "loss": 0.4326, |
| "num_input_tokens_seen": 61405888, |
| "step": 3980 |
| }, |
| { |
| "epoch": 16.348717948717947, |
| "grad_norm": 0.1240234375, |
| "learning_rate": 1.4412313865303438e-06, |
| "loss": 0.5303, |
| "num_input_tokens_seen": 61489280, |
| "step": 3985 |
| }, |
| { |
| "epoch": 16.369230769230768, |
| "grad_norm": 0.16015625, |
| "learning_rate": 1.4382037494727649e-06, |
| "loss": 0.4726, |
| "num_input_tokens_seen": 61562912, |
| "step": 3990 |
| }, |
| { |
| "epoch": 16.38974358974359, |
| "grad_norm": 0.1328125, |
| "learning_rate": 1.435176364582752e-06, |
| "loss": 0.458, |
| "num_input_tokens_seen": 61652064, |
| "step": 3995 |
| }, |
| { |
| "epoch": 16.41025641025641, |
| "grad_norm": 0.1171875, |
| "learning_rate": 1.4321492442139405e-06, |
| "loss": 0.5266, |
| "num_input_tokens_seen": 61737536, |
| "step": 4000 |
| }, |
| { |
| "epoch": 16.43076923076923, |
| "grad_norm": 0.158203125, |
| "learning_rate": 1.4291224007188849e-06, |
| "loss": 0.4394, |
| "num_input_tokens_seen": 61811840, |
| "step": 4005 |
| }, |
| { |
| "epoch": 16.45128205128205, |
| "grad_norm": 0.142578125, |
| "learning_rate": 1.4260958464490104e-06, |
| "loss": 0.4474, |
| "num_input_tokens_seen": 61884288, |
| "step": 4010 |
| }, |
| { |
| "epoch": 16.47179487179487, |
| "grad_norm": 0.22265625, |
| "learning_rate": 1.4230695937545616e-06, |
| "loss": 0.5, |
| "num_input_tokens_seen": 61958944, |
| "step": 4015 |
| }, |
| { |
| "epoch": 16.49230769230769, |
| "grad_norm": 0.1455078125, |
| "learning_rate": 1.420043654984553e-06, |
| "loss": 0.4766, |
| "num_input_tokens_seen": 62038016, |
| "step": 4020 |
| }, |
| { |
| "epoch": 16.51282051282051, |
| "grad_norm": 0.2890625, |
| "learning_rate": 1.4170180424867176e-06, |
| "loss": 0.4593, |
| "num_input_tokens_seen": 62107616, |
| "step": 4025 |
| }, |
| { |
| "epoch": 16.533333333333335, |
| "grad_norm": 0.359375, |
| "learning_rate": 1.4139927686074577e-06, |
| "loss": 0.4446, |
| "num_input_tokens_seen": 62188352, |
| "step": 4030 |
| }, |
| { |
| "epoch": 16.553846153846155, |
| "grad_norm": 0.259765625, |
| "learning_rate": 1.4109678456917926e-06, |
| "loss": 0.4658, |
| "num_input_tokens_seen": 62259456, |
| "step": 4035 |
| }, |
| { |
| "epoch": 16.574358974358976, |
| "grad_norm": 0.50390625, |
| "learning_rate": 1.4079432860833104e-06, |
| "loss": 0.5363, |
| "num_input_tokens_seen": 62356448, |
| "step": 4040 |
| }, |
| { |
| "epoch": 16.594871794871796, |
| "grad_norm": 0.78515625, |
| "learning_rate": 1.4049191021241176e-06, |
| "loss": 0.5006, |
| "num_input_tokens_seen": 62429824, |
| "step": 4045 |
| }, |
| { |
| "epoch": 16.615384615384617, |
| "grad_norm": 0.66796875, |
| "learning_rate": 1.4018953061547853e-06, |
| "loss": 0.4638, |
| "num_input_tokens_seen": 62514528, |
| "step": 4050 |
| }, |
| { |
| "epoch": 16.635897435897437, |
| "grad_norm": 0.267578125, |
| "learning_rate": 1.3988719105143038e-06, |
| "loss": 0.4413, |
| "num_input_tokens_seen": 62593472, |
| "step": 4055 |
| }, |
| { |
| "epoch": 16.656410256410258, |
| "grad_norm": 0.48046875, |
| "learning_rate": 1.395848927540028e-06, |
| "loss": 0.4506, |
| "num_input_tokens_seen": 62671520, |
| "step": 4060 |
| }, |
| { |
| "epoch": 16.676923076923078, |
| "grad_norm": 0.5078125, |
| "learning_rate": 1.39282636956763e-06, |
| "loss": 0.4616, |
| "num_input_tokens_seen": 62749856, |
| "step": 4065 |
| }, |
| { |
| "epoch": 16.6974358974359, |
| "grad_norm": 0.32421875, |
| "learning_rate": 1.3898042489310471e-06, |
| "loss": 0.456, |
| "num_input_tokens_seen": 62823616, |
| "step": 4070 |
| }, |
| { |
| "epoch": 16.71794871794872, |
| "grad_norm": 0.357421875, |
| "learning_rate": 1.3867825779624325e-06, |
| "loss": 0.45, |
| "num_input_tokens_seen": 62908416, |
| "step": 4075 |
| }, |
| { |
| "epoch": 16.73846153846154, |
| "grad_norm": 0.41015625, |
| "learning_rate": 1.3837613689921037e-06, |
| "loss": 0.4563, |
| "num_input_tokens_seen": 62978016, |
| "step": 4080 |
| }, |
| { |
| "epoch": 16.75897435897436, |
| "grad_norm": 0.259765625, |
| "learning_rate": 1.380740634348494e-06, |
| "loss": 0.4952, |
| "num_input_tokens_seen": 63065952, |
| "step": 4085 |
| }, |
| { |
| "epoch": 16.77948717948718, |
| "grad_norm": 0.3359375, |
| "learning_rate": 1.3777203863580993e-06, |
| "loss": 0.4327, |
| "num_input_tokens_seen": 63141792, |
| "step": 4090 |
| }, |
| { |
| "epoch": 16.8, |
| "grad_norm": 0.361328125, |
| "learning_rate": 1.3747006373454321e-06, |
| "loss": 0.5152, |
| "num_input_tokens_seen": 63223520, |
| "step": 4095 |
| }, |
| { |
| "epoch": 16.82051282051282, |
| "grad_norm": 0.326171875, |
| "learning_rate": 1.3716813996329669e-06, |
| "loss": 0.4952, |
| "num_input_tokens_seen": 63297696, |
| "step": 4100 |
| }, |
| { |
| "epoch": 16.84102564102564, |
| "grad_norm": 0.4609375, |
| "learning_rate": 1.3686626855410929e-06, |
| "loss": 0.4873, |
| "num_input_tokens_seen": 63378304, |
| "step": 4105 |
| }, |
| { |
| "epoch": 16.861538461538462, |
| "grad_norm": 0.388671875, |
| "learning_rate": 1.3656445073880615e-06, |
| "loss": 0.4461, |
| "num_input_tokens_seen": 63454400, |
| "step": 4110 |
| }, |
| { |
| "epoch": 16.882051282051282, |
| "grad_norm": 0.3671875, |
| "learning_rate": 1.3626268774899381e-06, |
| "loss": 0.4612, |
| "num_input_tokens_seen": 63526400, |
| "step": 4115 |
| }, |
| { |
| "epoch": 16.902564102564103, |
| "grad_norm": 0.33984375, |
| "learning_rate": 1.3596098081605505e-06, |
| "loss": 0.5061, |
| "num_input_tokens_seen": 63612448, |
| "step": 4120 |
| }, |
| { |
| "epoch": 16.923076923076923, |
| "grad_norm": 0.34765625, |
| "learning_rate": 1.3565933117114385e-06, |
| "loss": 0.486, |
| "num_input_tokens_seen": 63693248, |
| "step": 4125 |
| }, |
| { |
| "epoch": 16.943589743589744, |
| "grad_norm": 0.35546875, |
| "learning_rate": 1.3535774004518057e-06, |
| "loss": 0.4772, |
| "num_input_tokens_seen": 63765504, |
| "step": 4130 |
| }, |
| { |
| "epoch": 16.964102564102564, |
| "grad_norm": 0.330078125, |
| "learning_rate": 1.3505620866884666e-06, |
| "loss": 0.457, |
| "num_input_tokens_seen": 63839616, |
| "step": 4135 |
| }, |
| { |
| "epoch": 16.984615384615385, |
| "grad_norm": 0.341796875, |
| "learning_rate": 1.347547382725797e-06, |
| "loss": 0.4961, |
| "num_input_tokens_seen": 63924736, |
| "step": 4140 |
| }, |
| { |
| "epoch": 17.005128205128205, |
| "grad_norm": 0.349609375, |
| "learning_rate": 1.344533300865686e-06, |
| "loss": 0.5218, |
| "num_input_tokens_seen": 64000256, |
| "step": 4145 |
| }, |
| { |
| "epoch": 17.025641025641026, |
| "grad_norm": 0.44921875, |
| "learning_rate": 1.3415198534074823e-06, |
| "loss": 0.4501, |
| "num_input_tokens_seen": 64077472, |
| "step": 4150 |
| }, |
| { |
| "epoch": 17.046153846153846, |
| "grad_norm": 0.2578125, |
| "learning_rate": 1.3385070526479475e-06, |
| "loss": 0.4704, |
| "num_input_tokens_seen": 64157024, |
| "step": 4155 |
| }, |
| { |
| "epoch": 17.066666666666666, |
| "grad_norm": 0.2734375, |
| "learning_rate": 1.3354949108812026e-06, |
| "loss": 0.485, |
| "num_input_tokens_seen": 64226656, |
| "step": 4160 |
| }, |
| { |
| "epoch": 17.087179487179487, |
| "grad_norm": 0.2236328125, |
| "learning_rate": 1.3324834403986815e-06, |
| "loss": 0.4316, |
| "num_input_tokens_seen": 64302816, |
| "step": 4165 |
| }, |
| { |
| "epoch": 17.107692307692307, |
| "grad_norm": 0.2373046875, |
| "learning_rate": 1.3294726534890766e-06, |
| "loss": 0.4781, |
| "num_input_tokens_seen": 64371840, |
| "step": 4170 |
| }, |
| { |
| "epoch": 17.128205128205128, |
| "grad_norm": 0.189453125, |
| "learning_rate": 1.326462562438293e-06, |
| "loss": 0.4331, |
| "num_input_tokens_seen": 64451776, |
| "step": 4175 |
| }, |
| { |
| "epoch": 17.148717948717948, |
| "grad_norm": 0.2080078125, |
| "learning_rate": 1.3234531795293945e-06, |
| "loss": 0.4754, |
| "num_input_tokens_seen": 64523712, |
| "step": 4180 |
| }, |
| { |
| "epoch": 17.16923076923077, |
| "grad_norm": 0.1904296875, |
| "learning_rate": 1.3204445170425565e-06, |
| "loss": 0.4983, |
| "num_input_tokens_seen": 64594272, |
| "step": 4185 |
| }, |
| { |
| "epoch": 17.18974358974359, |
| "grad_norm": 0.1650390625, |
| "learning_rate": 1.3174365872550138e-06, |
| "loss": 0.4852, |
| "num_input_tokens_seen": 64677280, |
| "step": 4190 |
| }, |
| { |
| "epoch": 17.21025641025641, |
| "grad_norm": 0.1591796875, |
| "learning_rate": 1.3144294024410122e-06, |
| "loss": 0.4843, |
| "num_input_tokens_seen": 64758080, |
| "step": 4195 |
| }, |
| { |
| "epoch": 17.23076923076923, |
| "grad_norm": 0.1435546875, |
| "learning_rate": 1.3114229748717563e-06, |
| "loss": 0.4564, |
| "num_input_tokens_seen": 64830848, |
| "step": 4200 |
| }, |
| { |
| "epoch": 17.25128205128205, |
| "grad_norm": 0.1474609375, |
| "learning_rate": 1.308417316815362e-06, |
| "loss": 0.4432, |
| "num_input_tokens_seen": 64907296, |
| "step": 4205 |
| }, |
| { |
| "epoch": 17.27179487179487, |
| "grad_norm": 0.134765625, |
| "learning_rate": 1.3054124405368036e-06, |
| "loss": 0.4901, |
| "num_input_tokens_seen": 64987392, |
| "step": 4210 |
| }, |
| { |
| "epoch": 17.29230769230769, |
| "grad_norm": 0.1318359375, |
| "learning_rate": 1.3024083582978668e-06, |
| "loss": 0.5076, |
| "num_input_tokens_seen": 65077408, |
| "step": 4215 |
| }, |
| { |
| "epoch": 17.31282051282051, |
| "grad_norm": 0.12158203125, |
| "learning_rate": 1.2994050823570968e-06, |
| "loss": 0.4818, |
| "num_input_tokens_seen": 65159104, |
| "step": 4220 |
| }, |
| { |
| "epoch": 17.333333333333332, |
| "grad_norm": 0.099609375, |
| "learning_rate": 1.2964026249697475e-06, |
| "loss": 0.4525, |
| "num_input_tokens_seen": 65242816, |
| "step": 4225 |
| }, |
| { |
| "epoch": 17.353846153846153, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 1.293400998387734e-06, |
| "loss": 0.4444, |
| "num_input_tokens_seen": 65317376, |
| "step": 4230 |
| }, |
| { |
| "epoch": 17.374358974358973, |
| "grad_norm": 0.2314453125, |
| "learning_rate": 1.2904002148595797e-06, |
| "loss": 0.4897, |
| "num_input_tokens_seen": 65391264, |
| "step": 4235 |
| }, |
| { |
| "epoch": 17.394871794871793, |
| "grad_norm": 0.208984375, |
| "learning_rate": 1.2874002866303695e-06, |
| "loss": 0.5116, |
| "num_input_tokens_seen": 65467744, |
| "step": 4240 |
| }, |
| { |
| "epoch": 17.415384615384614, |
| "grad_norm": 0.1298828125, |
| "learning_rate": 1.2844012259416965e-06, |
| "loss": 0.4498, |
| "num_input_tokens_seen": 65535680, |
| "step": 4245 |
| }, |
| { |
| "epoch": 17.435897435897434, |
| "grad_norm": 0.2890625, |
| "learning_rate": 1.2814030450316151e-06, |
| "loss": 0.4984, |
| "num_input_tokens_seen": 65611680, |
| "step": 4250 |
| }, |
| { |
| "epoch": 17.456410256410255, |
| "grad_norm": 0.318359375, |
| "learning_rate": 1.2784057561345885e-06, |
| "loss": 0.5276, |
| "num_input_tokens_seen": 65684160, |
| "step": 4255 |
| }, |
| { |
| "epoch": 17.476923076923075, |
| "grad_norm": 0.291015625, |
| "learning_rate": 1.2754093714814407e-06, |
| "loss": 0.4434, |
| "num_input_tokens_seen": 65755104, |
| "step": 4260 |
| }, |
| { |
| "epoch": 17.4974358974359, |
| "grad_norm": 0.59765625, |
| "learning_rate": 1.2724139032993057e-06, |
| "loss": 0.4743, |
| "num_input_tokens_seen": 65835072, |
| "step": 4265 |
| }, |
| { |
| "epoch": 17.51794871794872, |
| "grad_norm": 0.89453125, |
| "learning_rate": 1.269419363811577e-06, |
| "loss": 0.4691, |
| "num_input_tokens_seen": 65916704, |
| "step": 4270 |
| }, |
| { |
| "epoch": 17.53846153846154, |
| "grad_norm": 0.6484375, |
| "learning_rate": 1.2664257652378598e-06, |
| "loss": 0.5129, |
| "num_input_tokens_seen": 65989696, |
| "step": 4275 |
| }, |
| { |
| "epoch": 17.55897435897436, |
| "grad_norm": 0.259765625, |
| "learning_rate": 1.2634331197939183e-06, |
| "loss": 0.4868, |
| "num_input_tokens_seen": 66071744, |
| "step": 4280 |
| }, |
| { |
| "epoch": 17.57948717948718, |
| "grad_norm": 0.54296875, |
| "learning_rate": 1.2604414396916286e-06, |
| "loss": 0.489, |
| "num_input_tokens_seen": 66153024, |
| "step": 4285 |
| }, |
| { |
| "epoch": 17.6, |
| "grad_norm": 0.55078125, |
| "learning_rate": 1.2574507371389267e-06, |
| "loss": 0.4855, |
| "num_input_tokens_seen": 66220384, |
| "step": 4290 |
| }, |
| { |
| "epoch": 17.620512820512822, |
| "grad_norm": 0.298828125, |
| "learning_rate": 1.25446102433976e-06, |
| "loss": 0.4791, |
| "num_input_tokens_seen": 66298912, |
| "step": 4295 |
| }, |
| { |
| "epoch": 17.641025641025642, |
| "grad_norm": 0.263671875, |
| "learning_rate": 1.2514723134940365e-06, |
| "loss": 0.4563, |
| "num_input_tokens_seen": 66373120, |
| "step": 4300 |
| }, |
| { |
| "epoch": 17.661538461538463, |
| "grad_norm": 0.26171875, |
| "learning_rate": 1.2484846167975767e-06, |
| "loss": 0.4577, |
| "num_input_tokens_seen": 66452096, |
| "step": 4305 |
| }, |
| { |
| "epoch": 17.682051282051283, |
| "grad_norm": 0.302734375, |
| "learning_rate": 1.2454979464420624e-06, |
| "loss": 0.4875, |
| "num_input_tokens_seen": 66530784, |
| "step": 4310 |
| }, |
| { |
| "epoch": 17.702564102564104, |
| "grad_norm": 0.34375, |
| "learning_rate": 1.2425123146149863e-06, |
| "loss": 0.4551, |
| "num_input_tokens_seen": 66607264, |
| "step": 4315 |
| }, |
| { |
| "epoch": 17.723076923076924, |
| "grad_norm": 0.328125, |
| "learning_rate": 1.2395277334996047e-06, |
| "loss": 0.5378, |
| "num_input_tokens_seen": 66692608, |
| "step": 4320 |
| }, |
| { |
| "epoch": 17.743589743589745, |
| "grad_norm": 0.341796875, |
| "learning_rate": 1.2365442152748846e-06, |
| "loss": 0.4518, |
| "num_input_tokens_seen": 66773504, |
| "step": 4325 |
| }, |
| { |
| "epoch": 17.764102564102565, |
| "grad_norm": 0.625, |
| "learning_rate": 1.2335617721154577e-06, |
| "loss": 0.4412, |
| "num_input_tokens_seen": 66844928, |
| "step": 4330 |
| }, |
| { |
| "epoch": 17.784615384615385, |
| "grad_norm": 0.443359375, |
| "learning_rate": 1.2305804161915671e-06, |
| "loss": 0.5214, |
| "num_input_tokens_seen": 66918080, |
| "step": 4335 |
| }, |
| { |
| "epoch": 17.805128205128206, |
| "grad_norm": 0.32421875, |
| "learning_rate": 1.2276001596690205e-06, |
| "loss": 0.47, |
| "num_input_tokens_seen": 66988992, |
| "step": 4340 |
| }, |
| { |
| "epoch": 17.825641025641026, |
| "grad_norm": 0.419921875, |
| "learning_rate": 1.2246210147091382e-06, |
| "loss": 0.4545, |
| "num_input_tokens_seen": 67068288, |
| "step": 4345 |
| }, |
| { |
| "epoch": 17.846153846153847, |
| "grad_norm": 0.3125, |
| "learning_rate": 1.2216429934687062e-06, |
| "loss": 0.4574, |
| "num_input_tokens_seen": 67142560, |
| "step": 4350 |
| }, |
| { |
| "epoch": 17.866666666666667, |
| "grad_norm": 0.40234375, |
| "learning_rate": 1.2186661080999234e-06, |
| "loss": 0.4717, |
| "num_input_tokens_seen": 67222560, |
| "step": 4355 |
| }, |
| { |
| "epoch": 17.887179487179488, |
| "grad_norm": 0.318359375, |
| "learning_rate": 1.2156903707503544e-06, |
| "loss": 0.4662, |
| "num_input_tokens_seen": 67301760, |
| "step": 4360 |
| }, |
| { |
| "epoch": 17.907692307692308, |
| "grad_norm": 0.345703125, |
| "learning_rate": 1.2127157935628789e-06, |
| "loss": 0.4958, |
| "num_input_tokens_seen": 67381856, |
| "step": 4365 |
| }, |
| { |
| "epoch": 17.92820512820513, |
| "grad_norm": 0.318359375, |
| "learning_rate": 1.2097423886756433e-06, |
| "loss": 0.5327, |
| "num_input_tokens_seen": 67467968, |
| "step": 4370 |
| }, |
| { |
| "epoch": 17.94871794871795, |
| "grad_norm": 0.248046875, |
| "learning_rate": 1.2067701682220084e-06, |
| "loss": 0.4606, |
| "num_input_tokens_seen": 67548672, |
| "step": 4375 |
| }, |
| { |
| "epoch": 17.96923076923077, |
| "grad_norm": 0.255859375, |
| "learning_rate": 1.2037991443305043e-06, |
| "loss": 0.4451, |
| "num_input_tokens_seen": 67625248, |
| "step": 4380 |
| }, |
| { |
| "epoch": 17.98974358974359, |
| "grad_norm": 0.28125, |
| "learning_rate": 1.2008293291247754e-06, |
| "loss": 0.4416, |
| "num_input_tokens_seen": 67695808, |
| "step": 4385 |
| }, |
| { |
| "epoch": 18.01025641025641, |
| "grad_norm": 0.23046875, |
| "learning_rate": 1.1978607347235367e-06, |
| "loss": 0.4741, |
| "num_input_tokens_seen": 67774400, |
| "step": 4390 |
| }, |
| { |
| "epoch": 18.03076923076923, |
| "grad_norm": 0.236328125, |
| "learning_rate": 1.1948933732405205e-06, |
| "loss": 0.446, |
| "num_input_tokens_seen": 67846848, |
| "step": 4395 |
| }, |
| { |
| "epoch": 18.05128205128205, |
| "grad_norm": 0.2060546875, |
| "learning_rate": 1.191927256784427e-06, |
| "loss": 0.4315, |
| "num_input_tokens_seen": 67917440, |
| "step": 4400 |
| }, |
| { |
| "epoch": 18.07179487179487, |
| "grad_norm": 0.181640625, |
| "learning_rate": 1.1889623974588772e-06, |
| "loss": 0.4055, |
| "num_input_tokens_seen": 67990656, |
| "step": 4405 |
| }, |
| { |
| "epoch": 18.092307692307692, |
| "grad_norm": 0.1591796875, |
| "learning_rate": 1.185998807362362e-06, |
| "loss": 0.499, |
| "num_input_tokens_seen": 68067872, |
| "step": 4410 |
| }, |
| { |
| "epoch": 18.112820512820512, |
| "grad_norm": 0.162109375, |
| "learning_rate": 1.1830364985881924e-06, |
| "loss": 0.481, |
| "num_input_tokens_seen": 68145632, |
| "step": 4415 |
| }, |
| { |
| "epoch": 18.133333333333333, |
| "grad_norm": 0.1650390625, |
| "learning_rate": 1.1800754832244515e-06, |
| "loss": 0.4931, |
| "num_input_tokens_seen": 68221472, |
| "step": 4420 |
| }, |
| { |
| "epoch": 18.153846153846153, |
| "grad_norm": 0.1513671875, |
| "learning_rate": 1.1771157733539442e-06, |
| "loss": 0.4377, |
| "num_input_tokens_seen": 68292352, |
| "step": 4425 |
| }, |
| { |
| "epoch": 18.174358974358974, |
| "grad_norm": 0.150390625, |
| "learning_rate": 1.174157381054148e-06, |
| "loss": 0.5265, |
| "num_input_tokens_seen": 68374880, |
| "step": 4430 |
| }, |
| { |
| "epoch": 18.194871794871794, |
| "grad_norm": 0.1416015625, |
| "learning_rate": 1.1712003183971644e-06, |
| "loss": 0.5103, |
| "num_input_tokens_seen": 68457280, |
| "step": 4435 |
| }, |
| { |
| "epoch": 18.215384615384615, |
| "grad_norm": 0.119140625, |
| "learning_rate": 1.1682445974496686e-06, |
| "loss": 0.5133, |
| "num_input_tokens_seen": 68540224, |
| "step": 4440 |
| }, |
| { |
| "epoch": 18.235897435897435, |
| "grad_norm": 0.11865234375, |
| "learning_rate": 1.1652902302728607e-06, |
| "loss": 0.4805, |
| "num_input_tokens_seen": 68621056, |
| "step": 4445 |
| }, |
| { |
| "epoch": 18.256410256410255, |
| "grad_norm": 0.12353515625, |
| "learning_rate": 1.1623372289224172e-06, |
| "loss": 0.4863, |
| "num_input_tokens_seen": 68697440, |
| "step": 4450 |
| }, |
| { |
| "epoch": 18.276923076923076, |
| "grad_norm": 0.125, |
| "learning_rate": 1.1593856054484403e-06, |
| "loss": 0.4685, |
| "num_input_tokens_seen": 68778944, |
| "step": 4455 |
| }, |
| { |
| "epoch": 18.297435897435896, |
| "grad_norm": 0.1982421875, |
| "learning_rate": 1.156435371895411e-06, |
| "loss": 0.4934, |
| "num_input_tokens_seen": 68859712, |
| "step": 4460 |
| }, |
| { |
| "epoch": 18.317948717948717, |
| "grad_norm": 0.1962890625, |
| "learning_rate": 1.1534865403021366e-06, |
| "loss": 0.4439, |
| "num_input_tokens_seen": 68932512, |
| "step": 4465 |
| }, |
| { |
| "epoch": 18.338461538461537, |
| "grad_norm": 0.15625, |
| "learning_rate": 1.1505391227017046e-06, |
| "loss": 0.4295, |
| "num_input_tokens_seen": 69006080, |
| "step": 4470 |
| }, |
| { |
| "epoch": 18.358974358974358, |
| "grad_norm": 0.29296875, |
| "learning_rate": 1.1475931311214338e-06, |
| "loss": 0.4644, |
| "num_input_tokens_seen": 69081984, |
| "step": 4475 |
| }, |
| { |
| "epoch": 18.379487179487178, |
| "grad_norm": 0.2421875, |
| "learning_rate": 1.144648577582821e-06, |
| "loss": 0.4899, |
| "num_input_tokens_seen": 69158240, |
| "step": 4480 |
| }, |
| { |
| "epoch": 18.4, |
| "grad_norm": 0.3671875, |
| "learning_rate": 1.141705474101498e-06, |
| "loss": 0.4763, |
| "num_input_tokens_seen": 69241184, |
| "step": 4485 |
| }, |
| { |
| "epoch": 18.42051282051282, |
| "grad_norm": 0.5859375, |
| "learning_rate": 1.138763832687177e-06, |
| "loss": 0.4784, |
| "num_input_tokens_seen": 69326112, |
| "step": 4490 |
| }, |
| { |
| "epoch": 18.44102564102564, |
| "grad_norm": 0.765625, |
| "learning_rate": 1.1358236653436052e-06, |
| "loss": 0.4624, |
| "num_input_tokens_seen": 69400832, |
| "step": 4495 |
| }, |
| { |
| "epoch": 18.46153846153846, |
| "grad_norm": 0.8984375, |
| "learning_rate": 1.1328849840685143e-06, |
| "loss": 0.4523, |
| "num_input_tokens_seen": 69478464, |
| "step": 4500 |
| }, |
| { |
| "epoch": 18.48205128205128, |
| "grad_norm": 0.43359375, |
| "learning_rate": 1.1299478008535726e-06, |
| "loss": 0.5372, |
| "num_input_tokens_seen": 69559712, |
| "step": 4505 |
| }, |
| { |
| "epoch": 18.5025641025641, |
| "grad_norm": 0.458984375, |
| "learning_rate": 1.1270121276843342e-06, |
| "loss": 0.4224, |
| "num_input_tokens_seen": 69638528, |
| "step": 4510 |
| }, |
| { |
| "epoch": 18.523076923076925, |
| "grad_norm": 0.23828125, |
| "learning_rate": 1.1240779765401926e-06, |
| "loss": 0.4726, |
| "num_input_tokens_seen": 69719264, |
| "step": 4515 |
| }, |
| { |
| "epoch": 18.543589743589745, |
| "grad_norm": 0.30078125, |
| "learning_rate": 1.1211453593943293e-06, |
| "loss": 0.4407, |
| "num_input_tokens_seen": 69798432, |
| "step": 4520 |
| }, |
| { |
| "epoch": 18.564102564102566, |
| "grad_norm": 0.310546875, |
| "learning_rate": 1.118214288213667e-06, |
| "loss": 0.4594, |
| "num_input_tokens_seen": 69869984, |
| "step": 4525 |
| }, |
| { |
| "epoch": 18.584615384615386, |
| "grad_norm": 0.318359375, |
| "learning_rate": 1.1152847749588186e-06, |
| "loss": 0.5029, |
| "num_input_tokens_seen": 69949088, |
| "step": 4530 |
| }, |
| { |
| "epoch": 18.605128205128207, |
| "grad_norm": 0.6796875, |
| "learning_rate": 1.1123568315840419e-06, |
| "loss": 0.4386, |
| "num_input_tokens_seen": 70023904, |
| "step": 4535 |
| }, |
| { |
| "epoch": 18.625641025641027, |
| "grad_norm": 0.5625, |
| "learning_rate": 1.1094304700371863e-06, |
| "loss": 0.5118, |
| "num_input_tokens_seen": 70098400, |
| "step": 4540 |
| }, |
| { |
| "epoch": 18.646153846153847, |
| "grad_norm": 0.31640625, |
| "learning_rate": 1.1065057022596483e-06, |
| "loss": 0.4559, |
| "num_input_tokens_seen": 70176576, |
| "step": 4545 |
| }, |
| { |
| "epoch": 18.666666666666668, |
| "grad_norm": 0.345703125, |
| "learning_rate": 1.1035825401863185e-06, |
| "loss": 0.4418, |
| "num_input_tokens_seen": 70255264, |
| "step": 4550 |
| }, |
| { |
| "epoch": 18.68717948717949, |
| "grad_norm": 0.345703125, |
| "learning_rate": 1.100660995745538e-06, |
| "loss": 0.435, |
| "num_input_tokens_seen": 70324160, |
| "step": 4555 |
| }, |
| { |
| "epoch": 18.70769230769231, |
| "grad_norm": 0.365234375, |
| "learning_rate": 1.0977410808590437e-06, |
| "loss": 0.5499, |
| "num_input_tokens_seen": 70411616, |
| "step": 4560 |
| }, |
| { |
| "epoch": 18.72820512820513, |
| "grad_norm": 0.470703125, |
| "learning_rate": 1.0948228074419269e-06, |
| "loss": 0.4705, |
| "num_input_tokens_seen": 70482592, |
| "step": 4565 |
| }, |
| { |
| "epoch": 18.74871794871795, |
| "grad_norm": 0.44921875, |
| "learning_rate": 1.0919061874025774e-06, |
| "loss": 0.4466, |
| "num_input_tokens_seen": 70565120, |
| "step": 4570 |
| }, |
| { |
| "epoch": 18.76923076923077, |
| "grad_norm": 0.3984375, |
| "learning_rate": 1.0889912326426393e-06, |
| "loss": 0.4323, |
| "num_input_tokens_seen": 70635168, |
| "step": 4575 |
| }, |
| { |
| "epoch": 18.78974358974359, |
| "grad_norm": 0.310546875, |
| "learning_rate": 1.0860779550569609e-06, |
| "loss": 0.5235, |
| "num_input_tokens_seen": 70720096, |
| "step": 4580 |
| }, |
| { |
| "epoch": 18.81025641025641, |
| "grad_norm": 0.322265625, |
| "learning_rate": 1.083166366533548e-06, |
| "loss": 0.4659, |
| "num_input_tokens_seen": 70807648, |
| "step": 4585 |
| }, |
| { |
| "epoch": 18.83076923076923, |
| "grad_norm": 0.314453125, |
| "learning_rate": 1.080256478953512e-06, |
| "loss": 0.4968, |
| "num_input_tokens_seen": 70888832, |
| "step": 4590 |
| }, |
| { |
| "epoch": 18.851282051282052, |
| "grad_norm": 0.33203125, |
| "learning_rate": 1.0773483041910247e-06, |
| "loss": 0.5112, |
| "num_input_tokens_seen": 70971136, |
| "step": 4595 |
| }, |
| { |
| "epoch": 18.871794871794872, |
| "grad_norm": 0.29296875, |
| "learning_rate": 1.0744418541132676e-06, |
| "loss": 0.4414, |
| "num_input_tokens_seen": 71038816, |
| "step": 4600 |
| }, |
| { |
| "epoch": 18.892307692307693, |
| "grad_norm": 0.314453125, |
| "learning_rate": 1.0715371405803858e-06, |
| "loss": 0.4874, |
| "num_input_tokens_seen": 71114720, |
| "step": 4605 |
| }, |
| { |
| "epoch": 18.912820512820513, |
| "grad_norm": 0.294921875, |
| "learning_rate": 1.0686341754454364e-06, |
| "loss": 0.4571, |
| "num_input_tokens_seen": 71189856, |
| "step": 4610 |
| }, |
| { |
| "epoch": 18.933333333333334, |
| "grad_norm": 0.23046875, |
| "learning_rate": 1.0657329705543439e-06, |
| "loss": 0.4661, |
| "num_input_tokens_seen": 71271136, |
| "step": 4615 |
| }, |
| { |
| "epoch": 18.953846153846154, |
| "grad_norm": 0.2109375, |
| "learning_rate": 1.0628335377458477e-06, |
| "loss": 0.435, |
| "num_input_tokens_seen": 71343488, |
| "step": 4620 |
| }, |
| { |
| "epoch": 18.974358974358974, |
| "grad_norm": 0.21484375, |
| "learning_rate": 1.0599358888514582e-06, |
| "loss": 0.4845, |
| "num_input_tokens_seen": 71419104, |
| "step": 4625 |
| }, |
| { |
| "epoch": 18.994871794871795, |
| "grad_norm": 0.2021484375, |
| "learning_rate": 1.0570400356954044e-06, |
| "loss": 0.5111, |
| "num_input_tokens_seen": 71491776, |
| "step": 4630 |
| }, |
| { |
| "epoch": 19.015384615384615, |
| "grad_norm": 0.18359375, |
| "learning_rate": 1.0541459900945892e-06, |
| "loss": 0.5006, |
| "num_input_tokens_seen": 71573184, |
| "step": 4635 |
| }, |
| { |
| "epoch": 19.035897435897436, |
| "grad_norm": 0.1630859375, |
| "learning_rate": 1.0512537638585379e-06, |
| "loss": 0.4828, |
| "num_input_tokens_seen": 71648768, |
| "step": 4640 |
| }, |
| { |
| "epoch": 19.056410256410256, |
| "grad_norm": 0.1572265625, |
| "learning_rate": 1.0483633687893526e-06, |
| "loss": 0.4607, |
| "num_input_tokens_seen": 71721824, |
| "step": 4645 |
| }, |
| { |
| "epoch": 19.076923076923077, |
| "grad_norm": 0.1376953125, |
| "learning_rate": 1.0454748166816645e-06, |
| "loss": 0.497, |
| "num_input_tokens_seen": 71797568, |
| "step": 4650 |
| }, |
| { |
| "epoch": 19.097435897435897, |
| "grad_norm": 0.13671875, |
| "learning_rate": 1.0425881193225808e-06, |
| "loss": 0.4607, |
| "num_input_tokens_seen": 71871136, |
| "step": 4655 |
| }, |
| { |
| "epoch": 19.117948717948718, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 1.0397032884916438e-06, |
| "loss": 0.4232, |
| "num_input_tokens_seen": 71941984, |
| "step": 4660 |
| }, |
| { |
| "epoch": 19.138461538461538, |
| "grad_norm": 0.123046875, |
| "learning_rate": 1.0368203359607767e-06, |
| "loss": 0.5216, |
| "num_input_tokens_seen": 72024736, |
| "step": 4665 |
| }, |
| { |
| "epoch": 19.15897435897436, |
| "grad_norm": 0.12109375, |
| "learning_rate": 1.0339392734942393e-06, |
| "loss": 0.4565, |
| "num_input_tokens_seen": 72096064, |
| "step": 4670 |
| }, |
| { |
| "epoch": 19.17948717948718, |
| "grad_norm": 0.1337890625, |
| "learning_rate": 1.031060112848578e-06, |
| "loss": 0.4488, |
| "num_input_tokens_seen": 72171168, |
| "step": 4675 |
| }, |
| { |
| "epoch": 19.2, |
| "grad_norm": 0.126953125, |
| "learning_rate": 1.0281828657725798e-06, |
| "loss": 0.4567, |
| "num_input_tokens_seen": 72249056, |
| "step": 4680 |
| }, |
| { |
| "epoch": 19.22051282051282, |
| "grad_norm": 0.2080078125, |
| "learning_rate": 1.0253075440072212e-06, |
| "loss": 0.4556, |
| "num_input_tokens_seen": 72338944, |
| "step": 4685 |
| }, |
| { |
| "epoch": 19.24102564102564, |
| "grad_norm": 0.1484375, |
| "learning_rate": 1.0224341592856245e-06, |
| "loss": 0.4686, |
| "num_input_tokens_seen": 72412032, |
| "step": 4690 |
| }, |
| { |
| "epoch": 19.26153846153846, |
| "grad_norm": 0.1435546875, |
| "learning_rate": 1.0195627233330052e-06, |
| "loss": 0.4087, |
| "num_input_tokens_seen": 72485760, |
| "step": 4695 |
| }, |
| { |
| "epoch": 19.28205128205128, |
| "grad_norm": 0.3046875, |
| "learning_rate": 1.0166932478666292e-06, |
| "loss": 0.4634, |
| "num_input_tokens_seen": 72565344, |
| "step": 4700 |
| }, |
| { |
| "epoch": 19.3025641025641, |
| "grad_norm": 0.275390625, |
| "learning_rate": 1.0138257445957601e-06, |
| "loss": 0.5316, |
| "num_input_tokens_seen": 72644800, |
| "step": 4705 |
| }, |
| { |
| "epoch": 19.323076923076922, |
| "grad_norm": 0.31640625, |
| "learning_rate": 1.0109602252216153e-06, |
| "loss": 0.4929, |
| "num_input_tokens_seen": 72725440, |
| "step": 4710 |
| }, |
| { |
| "epoch": 19.343589743589742, |
| "grad_norm": 0.640625, |
| "learning_rate": 1.0080967014373152e-06, |
| "loss": 0.442, |
| "num_input_tokens_seen": 72796064, |
| "step": 4715 |
| }, |
| { |
| "epoch": 19.364102564102563, |
| "grad_norm": 0.91015625, |
| "learning_rate": 1.0052351849278385e-06, |
| "loss": 0.4442, |
| "num_input_tokens_seen": 72870080, |
| "step": 4720 |
| }, |
| { |
| "epoch": 19.384615384615383, |
| "grad_norm": 0.2490234375, |
| "learning_rate": 1.0023756873699723e-06, |
| "loss": 0.4751, |
| "num_input_tokens_seen": 72944288, |
| "step": 4725 |
| }, |
| { |
| "epoch": 19.405128205128204, |
| "grad_norm": 0.333984375, |
| "learning_rate": 9.995182204322637e-07, |
| "loss": 0.5129, |
| "num_input_tokens_seen": 73021664, |
| "step": 4730 |
| }, |
| { |
| "epoch": 19.425641025641024, |
| "grad_norm": 0.81640625, |
| "learning_rate": 9.966627957749767e-07, |
| "loss": 0.4796, |
| "num_input_tokens_seen": 73097920, |
| "step": 4735 |
| }, |
| { |
| "epoch": 19.446153846153845, |
| "grad_norm": 0.333984375, |
| "learning_rate": 9.93809425050039e-07, |
| "loss": 0.4785, |
| "num_input_tokens_seen": 73170112, |
| "step": 4740 |
| }, |
| { |
| "epoch": 19.466666666666665, |
| "grad_norm": 0.2255859375, |
| "learning_rate": 9.909581199009971e-07, |
| "loss": 0.4379, |
| "num_input_tokens_seen": 73241312, |
| "step": 4745 |
| }, |
| { |
| "epoch": 19.487179487179485, |
| "grad_norm": 0.267578125, |
| "learning_rate": 9.8810889196297e-07, |
| "loss": 0.5408, |
| "num_input_tokens_seen": 73319936, |
| "step": 4750 |
| }, |
| { |
| "epoch": 19.50769230769231, |
| "grad_norm": 0.3046875, |
| "learning_rate": 9.852617528625992e-07, |
| "loss": 0.4878, |
| "num_input_tokens_seen": 73397824, |
| "step": 4755 |
| }, |
| { |
| "epoch": 19.52820512820513, |
| "grad_norm": 0.29296875, |
| "learning_rate": 9.824167142180032e-07, |
| "loss": 0.4627, |
| "num_input_tokens_seen": 73470688, |
| "step": 4760 |
| }, |
| { |
| "epoch": 19.54871794871795, |
| "grad_norm": 0.283203125, |
| "learning_rate": 9.795737876387285e-07, |
| "loss": 0.4799, |
| "num_input_tokens_seen": 73545792, |
| "step": 4765 |
| }, |
| { |
| "epoch": 19.56923076923077, |
| "grad_norm": 0.34765625, |
| "learning_rate": 9.76732984725704e-07, |
| "loss": 0.428, |
| "num_input_tokens_seen": 73617760, |
| "step": 4770 |
| }, |
| { |
| "epoch": 19.58974358974359, |
| "grad_norm": 0.3515625, |
| "learning_rate": 9.738943170711916e-07, |
| "loss": 0.4608, |
| "num_input_tokens_seen": 73707328, |
| "step": 4775 |
| }, |
| { |
| "epoch": 19.61025641025641, |
| "grad_norm": 0.515625, |
| "learning_rate": 9.710577962587412e-07, |
| "loss": 0.4951, |
| "num_input_tokens_seen": 73782976, |
| "step": 4780 |
| }, |
| { |
| "epoch": 19.630769230769232, |
| "grad_norm": 0.3828125, |
| "learning_rate": 9.68223433863141e-07, |
| "loss": 0.53, |
| "num_input_tokens_seen": 73872256, |
| "step": 4785 |
| }, |
| { |
| "epoch": 19.651282051282053, |
| "grad_norm": 0.341796875, |
| "learning_rate": 9.653912414503731e-07, |
| "loss": 0.45, |
| "num_input_tokens_seen": 73947520, |
| "step": 4790 |
| }, |
| { |
| "epoch": 19.671794871794873, |
| "grad_norm": 0.337890625, |
| "learning_rate": 9.625612305775626e-07, |
| "loss": 0.4218, |
| "num_input_tokens_seen": 74023104, |
| "step": 4795 |
| }, |
| { |
| "epoch": 19.692307692307693, |
| "grad_norm": 0.333984375, |
| "learning_rate": 9.597334127929346e-07, |
| "loss": 0.4531, |
| "num_input_tokens_seen": 74105920, |
| "step": 4800 |
| }, |
| { |
| "epoch": 19.712820512820514, |
| "grad_norm": 0.306640625, |
| "learning_rate": 9.569077996357638e-07, |
| "loss": 0.4142, |
| "num_input_tokens_seen": 74182752, |
| "step": 4805 |
| }, |
| { |
| "epoch": 19.733333333333334, |
| "grad_norm": 0.37109375, |
| "learning_rate": 9.54084402636329e-07, |
| "loss": 0.5121, |
| "num_input_tokens_seen": 74265248, |
| "step": 4810 |
| }, |
| { |
| "epoch": 19.753846153846155, |
| "grad_norm": 0.328125, |
| "learning_rate": 9.512632333158653e-07, |
| "loss": 0.523, |
| "num_input_tokens_seen": 74363232, |
| "step": 4815 |
| }, |
| { |
| "epoch": 19.774358974358975, |
| "grad_norm": 0.353515625, |
| "learning_rate": 9.484443031865176e-07, |
| "loss": 0.5103, |
| "num_input_tokens_seen": 74442848, |
| "step": 4820 |
| }, |
| { |
| "epoch": 19.794871794871796, |
| "grad_norm": 0.27734375, |
| "learning_rate": 9.456276237512949e-07, |
| "loss": 0.4725, |
| "num_input_tokens_seen": 74513376, |
| "step": 4825 |
| }, |
| { |
| "epoch": 19.815384615384616, |
| "grad_norm": 0.271484375, |
| "learning_rate": 9.428132065040198e-07, |
| "loss": 0.4921, |
| "num_input_tokens_seen": 74588928, |
| "step": 4830 |
| }, |
| { |
| "epoch": 19.835897435897436, |
| "grad_norm": 0.291015625, |
| "learning_rate": 9.40001062929285e-07, |
| "loss": 0.4331, |
| "num_input_tokens_seen": 74659040, |
| "step": 4835 |
| }, |
| { |
| "epoch": 19.856410256410257, |
| "grad_norm": 0.216796875, |
| "learning_rate": 9.371912045024046e-07, |
| "loss": 0.44, |
| "num_input_tokens_seen": 74737472, |
| "step": 4840 |
| }, |
| { |
| "epoch": 19.876923076923077, |
| "grad_norm": 0.2158203125, |
| "learning_rate": 9.343836426893687e-07, |
| "loss": 0.4831, |
| "num_input_tokens_seen": 74810592, |
| "step": 4845 |
| }, |
| { |
| "epoch": 19.897435897435898, |
| "grad_norm": 0.1943359375, |
| "learning_rate": 9.315783889467943e-07, |
| "loss": 0.4494, |
| "num_input_tokens_seen": 74880096, |
| "step": 4850 |
| }, |
| { |
| "epoch": 19.91794871794872, |
| "grad_norm": 0.1953125, |
| "learning_rate": 9.287754547218821e-07, |
| "loss": 0.5063, |
| "num_input_tokens_seen": 74951168, |
| "step": 4855 |
| }, |
| { |
| "epoch": 19.93846153846154, |
| "grad_norm": 0.1611328125, |
| "learning_rate": 9.259748514523654e-07, |
| "loss": 0.4975, |
| "num_input_tokens_seen": 75034592, |
| "step": 4860 |
| }, |
| { |
| "epoch": 19.95897435897436, |
| "grad_norm": 0.1728515625, |
| "learning_rate": 9.231765905664677e-07, |
| "loss": 0.4802, |
| "num_input_tokens_seen": 75108992, |
| "step": 4865 |
| }, |
| { |
| "epoch": 19.97948717948718, |
| "grad_norm": 0.166015625, |
| "learning_rate": 9.20380683482853e-07, |
| "loss": 0.4473, |
| "num_input_tokens_seen": 75185824, |
| "step": 4870 |
| }, |
| { |
| "epoch": 20.0, |
| "grad_norm": 0.140625, |
| "learning_rate": 9.175871416105802e-07, |
| "loss": 0.4971, |
| "num_input_tokens_seen": 75263936, |
| "step": 4875 |
| }, |
| { |
| "epoch": 20.02051282051282, |
| "grad_norm": 0.134765625, |
| "learning_rate": 9.147959763490578e-07, |
| "loss": 0.493, |
| "num_input_tokens_seen": 75342752, |
| "step": 4880 |
| }, |
| { |
| "epoch": 20.04102564102564, |
| "grad_norm": 0.11669921875, |
| "learning_rate": 9.120071990879949e-07, |
| "loss": 0.4709, |
| "num_input_tokens_seen": 75420704, |
| "step": 4885 |
| }, |
| { |
| "epoch": 20.06153846153846, |
| "grad_norm": 0.13671875, |
| "learning_rate": 9.092208212073569e-07, |
| "loss": 0.5248, |
| "num_input_tokens_seen": 75496128, |
| "step": 4890 |
| }, |
| { |
| "epoch": 20.08205128205128, |
| "grad_norm": 0.1220703125, |
| "learning_rate": 9.064368540773177e-07, |
| "loss": 0.525, |
| "num_input_tokens_seen": 75585984, |
| "step": 4895 |
| }, |
| { |
| "epoch": 20.102564102564102, |
| "grad_norm": 0.1357421875, |
| "learning_rate": 9.036553090582145e-07, |
| "loss": 0.5186, |
| "num_input_tokens_seen": 75668416, |
| "step": 4900 |
| }, |
| { |
| "epoch": 20.123076923076923, |
| "grad_norm": 0.15234375, |
| "learning_rate": 9.008761975004994e-07, |
| "loss": 0.456, |
| "num_input_tokens_seen": 75744960, |
| "step": 4905 |
| }, |
| { |
| "epoch": 20.143589743589743, |
| "grad_norm": 0.203125, |
| "learning_rate": 8.98099530744696e-07, |
| "loss": 0.4507, |
| "num_input_tokens_seen": 75818688, |
| "step": 4910 |
| }, |
| { |
| "epoch": 20.164102564102564, |
| "grad_norm": 0.150390625, |
| "learning_rate": 8.953253201213517e-07, |
| "loss": 0.4552, |
| "num_input_tokens_seen": 75901824, |
| "step": 4915 |
| }, |
| { |
| "epoch": 20.184615384615384, |
| "grad_norm": 0.146484375, |
| "learning_rate": 8.925535769509895e-07, |
| "loss": 0.4153, |
| "num_input_tokens_seen": 75973248, |
| "step": 4920 |
| }, |
| { |
| "epoch": 20.205128205128204, |
| "grad_norm": 0.310546875, |
| "learning_rate": 8.897843125440653e-07, |
| "loss": 0.4452, |
| "num_input_tokens_seen": 76043840, |
| "step": 4925 |
| }, |
| { |
| "epoch": 20.217435897435898, |
| "num_input_tokens_seen": 76105440, |
| "step": 4928, |
| "total_flos": 1.6998510508071322e+18, |
| "train_loss": 0.4908539823365289, |
| "train_runtime": 36017.1103, |
| "train_samples_per_second": 3.465, |
| "train_steps_per_second": 0.216 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 7776, |
| "num_input_tokens_seen": 76105440, |
| "num_train_epochs": 32, |
| "save_steps": 2000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.6998510508071322e+18, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|