{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 279, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.010752688172043012, "grad_norm": 19.659445278083677, "learning_rate": 7.142857142857143e-07, "loss": 0.8801538944244385, "step": 1 }, { "epoch": 0.021505376344086023, "grad_norm": 21.183622990197563, "learning_rate": 1.4285714285714286e-06, "loss": 0.9106118679046631, "step": 2 }, { "epoch": 0.03225806451612903, "grad_norm": 9.680562864297194, "learning_rate": 2.1428571428571427e-06, "loss": 0.89518141746521, "step": 3 }, { "epoch": 0.043010752688172046, "grad_norm": 8.463557758001873, "learning_rate": 2.8571428571428573e-06, "loss": 0.8813375234603882, "step": 4 }, { "epoch": 0.053763440860215055, "grad_norm": 8.000585049846563, "learning_rate": 3.5714285714285718e-06, "loss": 0.8478757739067078, "step": 5 }, { "epoch": 0.06451612903225806, "grad_norm": 6.392760393586216, "learning_rate": 4.2857142857142855e-06, "loss": 0.7342901825904846, "step": 6 }, { "epoch": 0.07526881720430108, "grad_norm": 4.291231505071156, "learning_rate": 5e-06, "loss": 0.6894474029541016, "step": 7 }, { "epoch": 0.08602150537634409, "grad_norm": 3.3545639671049603, "learning_rate": 5.7142857142857145e-06, "loss": 0.6144750118255615, "step": 8 }, { "epoch": 0.0967741935483871, "grad_norm": 1.3577798099653526, "learning_rate": 6.4285714285714295e-06, "loss": 0.5370485186576843, "step": 9 }, { "epoch": 0.10752688172043011, "grad_norm": 1.2774532761940443, "learning_rate": 7.1428571428571436e-06, "loss": 0.5807953476905823, "step": 10 }, { "epoch": 0.11827956989247312, "grad_norm": 1.0447472651159857, "learning_rate": 7.857142857142858e-06, "loss": 0.5280188918113708, "step": 11 }, { "epoch": 0.12903225806451613, "grad_norm": 1.3301070676426008, "learning_rate": 8.571428571428571e-06, "loss": 0.5441749691963196, "step": 12 }, { "epoch": 0.13978494623655913, "grad_norm": 1.0565188353675214, "learning_rate": 9.285714285714288e-06, "loss": 0.5112073421478271, "step": 13 }, { "epoch": 0.15053763440860216, "grad_norm": 1.0753192556611026, "learning_rate": 1e-05, "loss": 0.5184136629104614, "step": 14 }, { "epoch": 0.16129032258064516, "grad_norm": 0.8209457367101312, "learning_rate": 9.999648647603774e-06, "loss": 0.5023149251937866, "step": 15 }, { "epoch": 0.17204301075268819, "grad_norm": 0.7282985426170734, "learning_rate": 9.998594639794502e-06, "loss": 0.48974722623825073, "step": 16 }, { "epoch": 0.1827956989247312, "grad_norm": 0.7177574512039909, "learning_rate": 9.996838124703448e-06, "loss": 0.47443556785583496, "step": 17 }, { "epoch": 0.1935483870967742, "grad_norm": 0.7867041042254088, "learning_rate": 9.994379349192927e-06, "loss": 0.48248669505119324, "step": 18 }, { "epoch": 0.20430107526881722, "grad_norm": 0.7192865433337082, "learning_rate": 9.991218658821609e-06, "loss": 0.46316730976104736, "step": 19 }, { "epoch": 0.21505376344086022, "grad_norm": 0.8279557032400371, "learning_rate": 9.987356497795944e-06, "loss": 0.5312884449958801, "step": 20 }, { "epoch": 0.22580645161290322, "grad_norm": 0.7907337063911983, "learning_rate": 9.982793408907747e-06, "loss": 0.48629915714263916, "step": 21 }, { "epoch": 0.23655913978494625, "grad_norm": 0.634234358721607, "learning_rate": 9.977530033457906e-06, "loss": 0.47711604833602905, "step": 22 }, { "epoch": 0.24731182795698925, "grad_norm": 0.7748372497671804, "learning_rate": 9.971567111166246e-06, "loss": 0.4741020202636719, "step": 23 }, { "epoch": 0.25806451612903225, "grad_norm": 0.6968307669517513, "learning_rate": 9.964905480067585e-06, "loss": 0.48086652159690857, "step": 24 }, { "epoch": 0.26881720430107525, "grad_norm": 0.6205197336090342, "learning_rate": 9.957546076393944e-06, "loss": 0.45759013295173645, "step": 25 }, { "epoch": 0.27956989247311825, "grad_norm": 0.5942106287222008, "learning_rate": 9.949489934442966e-06, "loss": 0.454744428396225, "step": 26 }, { "epoch": 0.2903225806451613, "grad_norm": 0.6374339549026193, "learning_rate": 9.940738186432565e-06, "loss": 0.4751560091972351, "step": 27 }, { "epoch": 0.3010752688172043, "grad_norm": 0.5187971760627564, "learning_rate": 9.931292062341793e-06, "loss": 0.39658498764038086, "step": 28 }, { "epoch": 0.3118279569892473, "grad_norm": 0.5360240389517962, "learning_rate": 9.921152889737985e-06, "loss": 0.46163371205329895, "step": 29 }, { "epoch": 0.3225806451612903, "grad_norm": 0.5585203893563241, "learning_rate": 9.910322093590177e-06, "loss": 0.46062612533569336, "step": 30 }, { "epoch": 0.3333333333333333, "grad_norm": 0.5610561072744311, "learning_rate": 9.898801196068839e-06, "loss": 0.4484734833240509, "step": 31 }, { "epoch": 0.34408602150537637, "grad_norm": 0.5533641378752504, "learning_rate": 9.886591816331953e-06, "loss": 0.41486942768096924, "step": 32 }, { "epoch": 0.3548387096774194, "grad_norm": 0.5365723882895206, "learning_rate": 9.87369567029745e-06, "loss": 0.43502286076545715, "step": 33 }, { "epoch": 0.3655913978494624, "grad_norm": 0.5536007189636761, "learning_rate": 9.860114570402055e-06, "loss": 0.4295029640197754, "step": 34 }, { "epoch": 0.3763440860215054, "grad_norm": 0.537839879636324, "learning_rate": 9.845850425346563e-06, "loss": 0.45029592514038086, "step": 35 }, { "epoch": 0.3870967741935484, "grad_norm": 0.5299693550104672, "learning_rate": 9.830905239827592e-06, "loss": 0.38297146558761597, "step": 36 }, { "epoch": 0.3978494623655914, "grad_norm": 0.5451092272235792, "learning_rate": 9.815281114255841e-06, "loss": 0.4300140142440796, "step": 37 }, { "epoch": 0.40860215053763443, "grad_norm": 0.6192689983212899, "learning_rate": 9.798980244460892e-06, "loss": 0.47214269638061523, "step": 38 }, { "epoch": 0.41935483870967744, "grad_norm": 0.5870329353069306, "learning_rate": 9.782004921382612e-06, "loss": 0.43783068656921387, "step": 39 }, { "epoch": 0.43010752688172044, "grad_norm": 0.5903637860392952, "learning_rate": 9.764357530749178e-06, "loss": 0.4084509611129761, "step": 40 }, { "epoch": 0.44086021505376344, "grad_norm": 0.61705449873411, "learning_rate": 9.74604055274178e-06, "loss": 0.42812418937683105, "step": 41 }, { "epoch": 0.45161290322580644, "grad_norm": 1.156325694795339, "learning_rate": 9.727056561646067e-06, "loss": 0.4196915626525879, "step": 42 }, { "epoch": 0.46236559139784944, "grad_norm": 0.580641483078281, "learning_rate": 9.707408225490343e-06, "loss": 0.4251587986946106, "step": 43 }, { "epoch": 0.4731182795698925, "grad_norm": 0.5510691280763641, "learning_rate": 9.687098305670606e-06, "loss": 0.4205142855644226, "step": 44 }, { "epoch": 0.4838709677419355, "grad_norm": 0.5225683335691166, "learning_rate": 9.66612965656245e-06, "loss": 0.37856176495552063, "step": 45 }, { "epoch": 0.4946236559139785, "grad_norm": 0.5740926054795668, "learning_rate": 9.644505225119922e-06, "loss": 0.4447917342185974, "step": 46 }, { "epoch": 0.5053763440860215, "grad_norm": 0.5873250280785997, "learning_rate": 9.622228050461345e-06, "loss": 0.3922080397605896, "step": 47 }, { "epoch": 0.5161290322580645, "grad_norm": 0.6219913849412524, "learning_rate": 9.599301263442194e-06, "loss": 0.43804931640625, "step": 48 }, { "epoch": 0.5268817204301075, "grad_norm": 0.5613998719760911, "learning_rate": 9.575728086215093e-06, "loss": 0.39639532566070557, "step": 49 }, { "epoch": 0.5376344086021505, "grad_norm": 0.5892845271021349, "learning_rate": 9.551511831776966e-06, "loss": 0.4323637783527374, "step": 50 }, { "epoch": 0.5483870967741935, "grad_norm": 0.6452327069861976, "learning_rate": 9.526655903503423e-06, "loss": 0.4111916422843933, "step": 51 }, { "epoch": 0.5591397849462365, "grad_norm": 0.6139840452181067, "learning_rate": 9.501163794670445e-06, "loss": 0.41024255752563477, "step": 52 }, { "epoch": 0.5698924731182796, "grad_norm": 0.5311229042653062, "learning_rate": 9.475039087963443e-06, "loss": 0.39047467708587646, "step": 53 }, { "epoch": 0.5806451612903226, "grad_norm": 0.5686779350086011, "learning_rate": 9.448285454973739e-06, "loss": 0.4083505868911743, "step": 54 }, { "epoch": 0.5913978494623656, "grad_norm": 0.5795028580373828, "learning_rate": 9.420906655682553e-06, "loss": 0.41430288553237915, "step": 55 }, { "epoch": 0.6021505376344086, "grad_norm": 0.5426200748349074, "learning_rate": 9.392906537932582e-06, "loss": 0.39692986011505127, "step": 56 }, { "epoch": 0.6129032258064516, "grad_norm": 0.5096666206067523, "learning_rate": 9.364289036887214e-06, "loss": 0.39456355571746826, "step": 57 }, { "epoch": 0.6236559139784946, "grad_norm": 0.6214693543974321, "learning_rate": 9.335058174477472e-06, "loss": 0.43065640330314636, "step": 58 }, { "epoch": 0.6344086021505376, "grad_norm": 0.6080775052566749, "learning_rate": 9.305218058836778e-06, "loss": 0.4233320653438568, "step": 59 }, { "epoch": 0.6451612903225806, "grad_norm": 0.5891380183146641, "learning_rate": 9.274772883723587e-06, "loss": 0.44017481803894043, "step": 60 }, { "epoch": 0.6559139784946236, "grad_norm": 0.5317932432917034, "learning_rate": 9.24372692793199e-06, "loss": 0.39426931738853455, "step": 61 }, { "epoch": 0.6666666666666666, "grad_norm": 0.6464375491620328, "learning_rate": 9.21208455469037e-06, "loss": 0.4302670657634735, "step": 62 }, { "epoch": 0.6774193548387096, "grad_norm": 0.6191245668003084, "learning_rate": 9.179850211048193e-06, "loss": 0.3996105492115021, "step": 63 }, { "epoch": 0.6881720430107527, "grad_norm": 0.5209850545783036, "learning_rate": 9.14702842725101e-06, "loss": 0.3926236629486084, "step": 64 }, { "epoch": 0.6989247311827957, "grad_norm": 0.5469602060829455, "learning_rate": 9.113623816103775e-06, "loss": 0.4129854142665863, "step": 65 }, { "epoch": 0.7096774193548387, "grad_norm": 0.6436176685575749, "learning_rate": 9.079641072322555e-06, "loss": 0.4262722432613373, "step": 66 }, { "epoch": 0.7204301075268817, "grad_norm": 0.6241121753870925, "learning_rate": 9.045084971874738e-06, "loss": 0.4259442687034607, "step": 67 }, { "epoch": 0.7311827956989247, "grad_norm": 0.520706057326722, "learning_rate": 9.009960371307798e-06, "loss": 0.40881264209747314, "step": 68 }, { "epoch": 0.7419354838709677, "grad_norm": 0.5406305590251429, "learning_rate": 8.974272207066767e-06, "loss": 0.4136165976524353, "step": 69 }, { "epoch": 0.7526881720430108, "grad_norm": 0.5589669382622552, "learning_rate": 8.938025494800454e-06, "loss": 0.3692484498023987, "step": 70 }, { "epoch": 0.7634408602150538, "grad_norm": 0.5484586018961858, "learning_rate": 8.901225328656543e-06, "loss": 0.392721563577652, "step": 71 }, { "epoch": 0.7741935483870968, "grad_norm": 0.5720189126246105, "learning_rate": 8.863876880565656e-06, "loss": 0.4131432771682739, "step": 72 }, { "epoch": 0.7849462365591398, "grad_norm": 0.5475650690180592, "learning_rate": 8.825985399514488e-06, "loss": 0.39399850368499756, "step": 73 }, { "epoch": 0.7956989247311828, "grad_norm": 0.571446157132029, "learning_rate": 8.787556210808101e-06, "loss": 0.3765556216239929, "step": 74 }, { "epoch": 0.8064516129032258, "grad_norm": 0.5096610649520538, "learning_rate": 8.748594715321512e-06, "loss": 0.38086819648742676, "step": 75 }, { "epoch": 0.8172043010752689, "grad_norm": 0.5696584948279285, "learning_rate": 8.70910638874064e-06, "loss": 0.40993741154670715, "step": 76 }, { "epoch": 0.8279569892473119, "grad_norm": 0.5448260281274895, "learning_rate": 8.669096780792754e-06, "loss": 0.3912901282310486, "step": 77 }, { "epoch": 0.8387096774193549, "grad_norm": 0.584810442023009, "learning_rate": 8.628571514466502e-06, "loss": 0.39722931385040283, "step": 78 }, { "epoch": 0.8494623655913979, "grad_norm": 0.5590387314291252, "learning_rate": 8.587536285221656e-06, "loss": 0.38008394837379456, "step": 79 }, { "epoch": 0.8602150537634409, "grad_norm": 0.5199976478034525, "learning_rate": 8.545996860188668e-06, "loss": 0.3779188394546509, "step": 80 }, { "epoch": 0.8709677419354839, "grad_norm": 0.6237910452530793, "learning_rate": 8.503959077358143e-06, "loss": 0.40831005573272705, "step": 81 }, { "epoch": 0.8817204301075269, "grad_norm": 0.5103230716146216, "learning_rate": 8.46142884476038e-06, "loss": 0.3830574154853821, "step": 82 }, { "epoch": 0.8924731182795699, "grad_norm": 0.6022470754794768, "learning_rate": 8.418412139635026e-06, "loss": 0.3760561943054199, "step": 83 }, { "epoch": 0.9032258064516129, "grad_norm": 0.5676501864622416, "learning_rate": 8.374915007591053e-06, "loss": 0.3981160521507263, "step": 84 }, { "epoch": 0.9139784946236559, "grad_norm": 0.62733814659435, "learning_rate": 8.330943561757092e-06, "loss": 0.38736072182655334, "step": 85 }, { "epoch": 0.9247311827956989, "grad_norm": 0.581917413328386, "learning_rate": 8.286503981922284e-06, "loss": 0.40054696798324585, "step": 86 }, { "epoch": 0.9354838709677419, "grad_norm": 0.5069721523225358, "learning_rate": 8.241602513667775e-06, "loss": 0.36416178941726685, "step": 87 }, { "epoch": 0.946236559139785, "grad_norm": 0.6484709424945745, "learning_rate": 8.19624546748895e-06, "loss": 0.3703634738922119, "step": 88 }, { "epoch": 0.956989247311828, "grad_norm": 0.5102251649420729, "learning_rate": 8.150439217908557e-06, "loss": 0.39143842458724976, "step": 89 }, { "epoch": 0.967741935483871, "grad_norm": 0.5485517368687335, "learning_rate": 8.104190202580811e-06, "loss": 0.3922935128211975, "step": 90 }, { "epoch": 0.978494623655914, "grad_norm": 0.5680193851368315, "learning_rate": 8.057504921386661e-06, "loss": 0.3872210383415222, "step": 91 }, { "epoch": 0.989247311827957, "grad_norm": 0.5405961708214808, "learning_rate": 8.010389935520269e-06, "loss": 0.4088854193687439, "step": 92 }, { "epoch": 1.0, "grad_norm": 0.5234958056904553, "learning_rate": 7.962851866566912e-06, "loss": 0.3855600953102112, "step": 93 }, { "epoch": 1.010752688172043, "grad_norm": 0.5751268630788401, "learning_rate": 7.914897395572362e-06, "loss": 0.32572415471076965, "step": 94 }, { "epoch": 1.021505376344086, "grad_norm": 0.5181295042077261, "learning_rate": 7.866533262103937e-06, "loss": 0.3046826720237732, "step": 95 }, { "epoch": 1.032258064516129, "grad_norm": 0.5508917328926465, "learning_rate": 7.817766263303312e-06, "loss": 0.29789605736732483, "step": 96 }, { "epoch": 1.043010752688172, "grad_norm": 0.5870437887252276, "learning_rate": 7.768603252931243e-06, "loss": 0.3265905976295471, "step": 97 }, { "epoch": 1.053763440860215, "grad_norm": 0.5133318653975599, "learning_rate": 7.719051140404327e-06, "loss": 0.31667008996009827, "step": 98 }, { "epoch": 1.064516129032258, "grad_norm": 0.5679147172610466, "learning_rate": 7.669116889823955e-06, "loss": 0.2714020311832428, "step": 99 }, { "epoch": 1.075268817204301, "grad_norm": 0.6225396396665326, "learning_rate": 7.6188075189975644e-06, "loss": 0.3139406740665436, "step": 100 }, { "epoch": 1.086021505376344, "grad_norm": 0.714553736071244, "learning_rate": 7.568130098452352e-06, "loss": 0.2961091995239258, "step": 101 }, { "epoch": 1.096774193548387, "grad_norm": 0.523889168005371, "learning_rate": 7.517091750441576e-06, "loss": 0.27245283126831055, "step": 102 }, { "epoch": 1.10752688172043, "grad_norm": 0.639295771779283, "learning_rate": 7.465699647943586e-06, "loss": 0.29422637820243835, "step": 103 }, { "epoch": 1.118279569892473, "grad_norm": 0.6150692376811179, "learning_rate": 7.413961013653725e-06, "loss": 0.29594993591308594, "step": 104 }, { "epoch": 1.129032258064516, "grad_norm": 0.6901790920161586, "learning_rate": 7.361883118969248e-06, "loss": 0.3163909912109375, "step": 105 }, { "epoch": 1.139784946236559, "grad_norm": 0.5947317629259291, "learning_rate": 7.309473282967387e-06, "loss": 0.2941242456436157, "step": 106 }, { "epoch": 1.1505376344086022, "grad_norm": 0.5873657820382887, "learning_rate": 7.256738871376733e-06, "loss": 0.29748812317848206, "step": 107 }, { "epoch": 1.1612903225806452, "grad_norm": 0.5914901548215277, "learning_rate": 7.203687295542032e-06, "loss": 0.3174619972705841, "step": 108 }, { "epoch": 1.1720430107526882, "grad_norm": 0.5757880846821776, "learning_rate": 7.1503260113826035e-06, "loss": 0.31319767236709595, "step": 109 }, { "epoch": 1.1827956989247312, "grad_norm": 0.7019483863871168, "learning_rate": 7.09666251834447e-06, "loss": 0.3302695155143738, "step": 110 }, { "epoch": 1.1935483870967742, "grad_norm": 0.5613831627968954, "learning_rate": 7.042704358346375e-06, "loss": 0.30153942108154297, "step": 111 }, { "epoch": 1.2043010752688172, "grad_norm": 0.5906296533769152, "learning_rate": 6.988459114719849e-06, "loss": 0.3321286141872406, "step": 112 }, { "epoch": 1.2150537634408602, "grad_norm": 0.6185930436417313, "learning_rate": 6.933934411143419e-06, "loss": 0.3183019161224365, "step": 113 }, { "epoch": 1.2258064516129032, "grad_norm": 0.5223193968800454, "learning_rate": 6.879137910571191e-06, "loss": 0.28522804379463196, "step": 114 }, { "epoch": 1.2365591397849462, "grad_norm": 0.5306097120719284, "learning_rate": 6.824077314155877e-06, "loss": 0.27248305082321167, "step": 115 }, { "epoch": 1.2473118279569892, "grad_norm": 0.5410094711890291, "learning_rate": 6.768760360166471e-06, "loss": 0.29798364639282227, "step": 116 }, { "epoch": 1.2580645161290323, "grad_norm": 0.594316663926518, "learning_rate": 6.713194822900707e-06, "loss": 0.27839338779449463, "step": 117 }, { "epoch": 1.2688172043010753, "grad_norm": 0.6110653082771621, "learning_rate": 6.657388511592453e-06, "loss": 0.30802613496780396, "step": 118 }, { "epoch": 1.2795698924731183, "grad_norm": 0.5260300979877027, "learning_rate": 6.601349269314188e-06, "loss": 0.29552769660949707, "step": 119 }, { "epoch": 1.2903225806451613, "grad_norm": 0.541255561766811, "learning_rate": 6.545084971874738e-06, "loss": 0.28333544731140137, "step": 120 }, { "epoch": 1.3010752688172043, "grad_norm": 0.5631658544435217, "learning_rate": 6.488603526712391e-06, "loss": 0.3072783946990967, "step": 121 }, { "epoch": 1.3118279569892473, "grad_norm": 0.5999697642319023, "learning_rate": 6.431912871783587e-06, "loss": 0.2956047058105469, "step": 122 }, { "epoch": 1.3225806451612903, "grad_norm": 0.5550402984265157, "learning_rate": 6.3750209744473105e-06, "loss": 0.3064526319503784, "step": 123 }, { "epoch": 1.3333333333333333, "grad_norm": 0.6418624402271341, "learning_rate": 6.3179358303453386e-06, "loss": 0.30682218074798584, "step": 124 }, { "epoch": 1.3440860215053765, "grad_norm": 0.5305043074037712, "learning_rate": 6.260665462278544e-06, "loss": 0.3036462068557739, "step": 125 }, { "epoch": 1.3548387096774195, "grad_norm": 0.5732839993822031, "learning_rate": 6.203217919079343e-06, "loss": 0.32605141401290894, "step": 126 }, { "epoch": 1.3655913978494625, "grad_norm": 0.5272514261375651, "learning_rate": 6.145601274480521e-06, "loss": 0.2800780236721039, "step": 127 }, { "epoch": 1.3763440860215055, "grad_norm": 0.5383888711112677, "learning_rate": 6.08782362598054e-06, "loss": 0.29685452580451965, "step": 128 }, { "epoch": 1.3870967741935485, "grad_norm": 0.5260069329849035, "learning_rate": 6.029893093705492e-06, "loss": 0.30100512504577637, "step": 129 }, { "epoch": 1.3978494623655915, "grad_norm": 0.5194176965790988, "learning_rate": 5.971817819267914e-06, "loss": 0.28969407081604004, "step": 130 }, { "epoch": 1.4086021505376345, "grad_norm": 0.5795379145956892, "learning_rate": 5.9136059646225375e-06, "loss": 0.3088012635707855, "step": 131 }, { "epoch": 1.4193548387096775, "grad_norm": 0.5605938731509191, "learning_rate": 5.855265710919211e-06, "loss": 0.2954477369785309, "step": 132 }, { "epoch": 1.4301075268817205, "grad_norm": 0.5513949309182947, "learning_rate": 5.796805257353109e-06, "loss": 0.31166109442710876, "step": 133 }, { "epoch": 1.4408602150537635, "grad_norm": 0.6271394444282732, "learning_rate": 5.738232820012407e-06, "loss": 0.33265459537506104, "step": 134 }, { "epoch": 1.4516129032258065, "grad_norm": 0.5513926189063966, "learning_rate": 5.679556630723592e-06, "loss": 0.271618515253067, "step": 135 }, { "epoch": 1.4623655913978495, "grad_norm": 0.5799088021416959, "learning_rate": 5.620784935894548e-06, "loss": 0.2902123034000397, "step": 136 }, { "epoch": 1.4731182795698925, "grad_norm": 0.5662089313253346, "learning_rate": 5.561925995355595e-06, "loss": 0.2844197154045105, "step": 137 }, { "epoch": 1.4838709677419355, "grad_norm": 0.9937555515678825, "learning_rate": 5.5029880811986546e-06, "loss": 0.2922922372817993, "step": 138 }, { "epoch": 1.4946236559139785, "grad_norm": 0.591368004900944, "learning_rate": 5.443979476614674e-06, "loss": 0.28497251868247986, "step": 139 }, { "epoch": 1.5053763440860215, "grad_norm": 0.6076112891374724, "learning_rate": 5.384908474729501e-06, "loss": 0.31142377853393555, "step": 140 }, { "epoch": 1.5161290322580645, "grad_norm": 0.5765965412966652, "learning_rate": 5.325783377438357e-06, "loss": 0.3015650510787964, "step": 141 }, { "epoch": 1.5268817204301075, "grad_norm": 0.5232918774810716, "learning_rate": 5.266612494239088e-06, "loss": 0.275716096162796, "step": 142 }, { "epoch": 1.5376344086021505, "grad_norm": 0.5680197113944581, "learning_rate": 5.207404141064334e-06, "loss": 0.2873622179031372, "step": 143 }, { "epoch": 1.5483870967741935, "grad_norm": 0.49399262225350316, "learning_rate": 5.148166639112799e-06, "loss": 0.26688140630722046, "step": 144 }, { "epoch": 1.5591397849462365, "grad_norm": 0.5949244178581083, "learning_rate": 5.088908313679788e-06, "loss": 0.29102981090545654, "step": 145 }, { "epoch": 1.5698924731182795, "grad_norm": 0.6289302836737687, "learning_rate": 5.029637492987153e-06, "loss": 0.2931191027164459, "step": 146 }, { "epoch": 1.5806451612903225, "grad_norm": 0.6088833365757095, "learning_rate": 4.970362507012848e-06, "loss": 0.2844638228416443, "step": 147 }, { "epoch": 1.5913978494623655, "grad_norm": 0.6971118150263104, "learning_rate": 4.911091686320213e-06, "loss": 0.3226025104522705, "step": 148 }, { "epoch": 1.6021505376344085, "grad_norm": 0.5612811561354285, "learning_rate": 4.8518333608872015e-06, "loss": 0.28764593601226807, "step": 149 }, { "epoch": 1.6129032258064515, "grad_norm": 0.6065928167636621, "learning_rate": 4.792595858935668e-06, "loss": 0.3277990221977234, "step": 150 }, { "epoch": 1.6236559139784945, "grad_norm": 0.5135483316527438, "learning_rate": 4.733387505760913e-06, "loss": 0.2962847948074341, "step": 151 }, { "epoch": 1.6344086021505375, "grad_norm": 0.6160906809336857, "learning_rate": 4.674216622561645e-06, "loss": 0.294375479221344, "step": 152 }, { "epoch": 1.6451612903225805, "grad_norm": 0.6058840168616271, "learning_rate": 4.6150915252705005e-06, "loss": 0.3054540157318115, "step": 153 }, { "epoch": 1.6559139784946235, "grad_norm": 0.5298336251451402, "learning_rate": 4.556020523385326e-06, "loss": 0.30477648973464966, "step": 154 }, { "epoch": 1.6666666666666665, "grad_norm": 0.5241054596736249, "learning_rate": 4.497011918801347e-06, "loss": 0.2855217456817627, "step": 155 }, { "epoch": 1.6774193548387095, "grad_norm": 0.48994778885097257, "learning_rate": 4.438074004644407e-06, "loss": 0.30031388998031616, "step": 156 }, { "epoch": 1.6881720430107527, "grad_norm": 0.6034882888193746, "learning_rate": 4.379215064105454e-06, "loss": 0.3072971999645233, "step": 157 }, { "epoch": 1.6989247311827957, "grad_norm": 0.5057395078651564, "learning_rate": 4.32044336927641e-06, "loss": 0.27074217796325684, "step": 158 }, { "epoch": 1.7096774193548387, "grad_norm": 0.548027010828449, "learning_rate": 4.261767179987595e-06, "loss": 0.28892627358436584, "step": 159 }, { "epoch": 1.7204301075268817, "grad_norm": 0.5949799690870136, "learning_rate": 4.203194742646893e-06, "loss": 0.31755733489990234, "step": 160 }, { "epoch": 1.7311827956989247, "grad_norm": 0.5988557766268561, "learning_rate": 4.1447342890807905e-06, "loss": 0.28983038663864136, "step": 161 }, { "epoch": 1.7419354838709677, "grad_norm": 0.5153375977084437, "learning_rate": 4.086394035377463e-06, "loss": 0.2821354269981384, "step": 162 }, { "epoch": 1.7526881720430108, "grad_norm": 0.5758453410329892, "learning_rate": 4.028182180732088e-06, "loss": 0.26659733057022095, "step": 163 }, { "epoch": 1.7634408602150538, "grad_norm": 0.5515188533069812, "learning_rate": 3.970106906294509e-06, "loss": 0.29241591691970825, "step": 164 }, { "epoch": 1.7741935483870968, "grad_norm": 0.5465960078319545, "learning_rate": 3.912176374019462e-06, "loss": 0.28333449363708496, "step": 165 }, { "epoch": 1.7849462365591398, "grad_norm": 0.5696267902501757, "learning_rate": 3.85439872551948e-06, "loss": 0.28454482555389404, "step": 166 }, { "epoch": 1.7956989247311828, "grad_norm": 0.5537225691065368, "learning_rate": 3.796782080920659e-06, "loss": 0.29573291540145874, "step": 167 }, { "epoch": 1.8064516129032258, "grad_norm": 0.5317821119388748, "learning_rate": 3.7393345377214584e-06, "loss": 0.3006817400455475, "step": 168 }, { "epoch": 1.817204301075269, "grad_norm": 0.5193143293546626, "learning_rate": 3.682064169654663e-06, "loss": 0.28759250044822693, "step": 169 }, { "epoch": 1.827956989247312, "grad_norm": 0.5024936169709733, "learning_rate": 3.6249790255526916e-06, "loss": 0.279812753200531, "step": 170 }, { "epoch": 1.838709677419355, "grad_norm": 0.522744432939167, "learning_rate": 3.568087128216414e-06, "loss": 0.2699696123600006, "step": 171 }, { "epoch": 1.849462365591398, "grad_norm": 0.6740080747223203, "learning_rate": 3.511396473287611e-06, "loss": 0.3212928771972656, "step": 172 }, { "epoch": 1.860215053763441, "grad_norm": 0.5393424030521861, "learning_rate": 3.4549150281252635e-06, "loss": 0.28264331817626953, "step": 173 }, { "epoch": 1.870967741935484, "grad_norm": 0.5389635395816308, "learning_rate": 3.398650730685813e-06, "loss": 0.2982739508152008, "step": 174 }, { "epoch": 1.881720430107527, "grad_norm": 0.5527956538218779, "learning_rate": 3.3426114884075488e-06, "loss": 0.2892792820930481, "step": 175 }, { "epoch": 1.89247311827957, "grad_norm": 0.5783362380337232, "learning_rate": 3.2868051770992935e-06, "loss": 0.2867039442062378, "step": 176 }, { "epoch": 1.903225806451613, "grad_norm": 0.4505107682619091, "learning_rate": 3.2312396398335312e-06, "loss": 0.2998119592666626, "step": 177 }, { "epoch": 1.913978494623656, "grad_norm": 0.544159970587014, "learning_rate": 3.175922685844125e-06, "loss": 0.2946498394012451, "step": 178 }, { "epoch": 1.924731182795699, "grad_norm": 0.5601933382330992, "learning_rate": 3.1208620894288105e-06, "loss": 0.30848872661590576, "step": 179 }, { "epoch": 1.935483870967742, "grad_norm": 0.5692086999688832, "learning_rate": 3.0660655888565827e-06, "loss": 0.30292433500289917, "step": 180 }, { "epoch": 1.946236559139785, "grad_norm": 0.5204759182695003, "learning_rate": 3.0115408852801535e-06, "loss": 0.2992181181907654, "step": 181 }, { "epoch": 1.956989247311828, "grad_norm": 0.547974934732699, "learning_rate": 2.9572956416536267e-06, "loss": 0.3052282929420471, "step": 182 }, { "epoch": 1.967741935483871, "grad_norm": 0.5036888640423901, "learning_rate": 2.9033374816555338e-06, "loss": 0.27619490027427673, "step": 183 }, { "epoch": 1.978494623655914, "grad_norm": 0.5857097319940581, "learning_rate": 2.8496739886173994e-06, "loss": 0.2969588041305542, "step": 184 }, { "epoch": 1.989247311827957, "grad_norm": 0.5125923920870736, "learning_rate": 2.7963127044579697e-06, "loss": 0.28236857056617737, "step": 185 }, { "epoch": 2.0, "grad_norm": 0.5460296926482203, "learning_rate": 2.743261128623269e-06, "loss": 0.2802872657775879, "step": 186 }, { "epoch": 2.010752688172043, "grad_norm": 0.5627765344162555, "learning_rate": 2.6905267170326143e-06, "loss": 0.2465861439704895, "step": 187 }, { "epoch": 2.021505376344086, "grad_norm": 0.5241819547135816, "learning_rate": 2.6381168810307536e-06, "loss": 0.2593002915382385, "step": 188 }, { "epoch": 2.032258064516129, "grad_norm": 0.5146155536713976, "learning_rate": 2.5860389863462765e-06, "loss": 0.22058740258216858, "step": 189 }, { "epoch": 2.043010752688172, "grad_norm": 0.527929699946712, "learning_rate": 2.534300352056416e-06, "loss": 0.24380052089691162, "step": 190 }, { "epoch": 2.053763440860215, "grad_norm": 0.5680981651809573, "learning_rate": 2.4829082495584244e-06, "loss": 0.2181762158870697, "step": 191 }, { "epoch": 2.064516129032258, "grad_norm": 0.5210266779766828, "learning_rate": 2.4318699015476495e-06, "loss": 0.21483556926250458, "step": 192 }, { "epoch": 2.075268817204301, "grad_norm": 0.5228849523676178, "learning_rate": 2.3811924810024385e-06, "loss": 0.2233523428440094, "step": 193 }, { "epoch": 2.086021505376344, "grad_norm": 1.2592287942902403, "learning_rate": 2.330883110176049e-06, "loss": 0.21375712752342224, "step": 194 }, { "epoch": 2.096774193548387, "grad_norm": 0.5541159570889329, "learning_rate": 2.2809488595956746e-06, "loss": 0.22336843609809875, "step": 195 }, { "epoch": 2.10752688172043, "grad_norm": 0.6275490404272853, "learning_rate": 2.2313967470687593e-06, "loss": 0.23877255618572235, "step": 196 }, { "epoch": 2.118279569892473, "grad_norm": 0.5287134027247969, "learning_rate": 2.18223373669669e-06, "loss": 0.23572032153606415, "step": 197 }, { "epoch": 2.129032258064516, "grad_norm": 0.6005092115812523, "learning_rate": 2.1334667378960642e-06, "loss": 0.25271525979042053, "step": 198 }, { "epoch": 2.139784946236559, "grad_norm": 0.5729023596023511, "learning_rate": 2.0851026044276405e-06, "loss": 0.2403365820646286, "step": 199 }, { "epoch": 2.150537634408602, "grad_norm": 0.502960648659189, "learning_rate": 2.0371481334330913e-06, "loss": 0.22194445133209229, "step": 200 }, { "epoch": 2.161290322580645, "grad_norm": 1.0792975178234447, "learning_rate": 1.9896100644797316e-06, "loss": 0.20418012142181396, "step": 201 }, { "epoch": 2.172043010752688, "grad_norm": 0.563900355756192, "learning_rate": 1.9424950786133414e-06, "loss": 0.2197030484676361, "step": 202 }, { "epoch": 2.182795698924731, "grad_norm": 0.6049813205320395, "learning_rate": 1.8958097974191909e-06, "loss": 0.23065771162509918, "step": 203 }, { "epoch": 2.193548387096774, "grad_norm": 0.5693258359321052, "learning_rate": 1.8495607820914451e-06, "loss": 0.22469733655452728, "step": 204 }, { "epoch": 2.204301075268817, "grad_norm": 0.49687778546847566, "learning_rate": 1.8037545325110506e-06, "loss": 0.22447475790977478, "step": 205 }, { "epoch": 2.21505376344086, "grad_norm": 0.6870685608162285, "learning_rate": 1.7583974863322272e-06, "loss": 0.2232748568058014, "step": 206 }, { "epoch": 2.225806451612903, "grad_norm": 0.5587791759635568, "learning_rate": 1.7134960180777171e-06, "loss": 0.21617698669433594, "step": 207 }, { "epoch": 2.236559139784946, "grad_norm": 0.5406726897347297, "learning_rate": 1.6690564382429104e-06, "loss": 0.22749440371990204, "step": 208 }, { "epoch": 2.247311827956989, "grad_norm": 0.6059659457413509, "learning_rate": 1.6250849924089485e-06, "loss": 0.24092203378677368, "step": 209 }, { "epoch": 2.258064516129032, "grad_norm": 0.5440152722934113, "learning_rate": 1.581587860364977e-06, "loss": 0.20948372781276703, "step": 210 }, { "epoch": 2.268817204301075, "grad_norm": 0.48547134095880384, "learning_rate": 1.5385711552396227e-06, "loss": 0.21325388550758362, "step": 211 }, { "epoch": 2.279569892473118, "grad_norm": 0.6324122951694761, "learning_rate": 1.4960409226418576e-06, "loss": 0.22965453565120697, "step": 212 }, { "epoch": 2.2903225806451615, "grad_norm": 0.5469586332098659, "learning_rate": 1.4540031398113335e-06, "loss": 0.22315660119056702, "step": 213 }, { "epoch": 2.3010752688172045, "grad_norm": 0.7520121411441703, "learning_rate": 1.4124637147783431e-06, "loss": 0.23476293683052063, "step": 214 }, { "epoch": 2.3118279569892475, "grad_norm": 0.535533468892341, "learning_rate": 1.371428485533498e-06, "loss": 0.21404291689395905, "step": 215 }, { "epoch": 2.3225806451612905, "grad_norm": 0.5561238179269572, "learning_rate": 1.3309032192072463e-06, "loss": 0.23378810286521912, "step": 216 }, { "epoch": 2.3333333333333335, "grad_norm": 0.5663331713554731, "learning_rate": 1.29089361125936e-06, "loss": 0.2105513960123062, "step": 217 }, { "epoch": 2.3440860215053765, "grad_norm": 0.5851084790781899, "learning_rate": 1.251405284678488e-06, "loss": 0.22501936554908752, "step": 218 }, { "epoch": 2.3548387096774195, "grad_norm": 0.5546173616985522, "learning_rate": 1.2124437891918995e-06, "loss": 0.21492747962474823, "step": 219 }, { "epoch": 2.3655913978494625, "grad_norm": 0.5542629930888266, "learning_rate": 1.1740146004855141e-06, "loss": 0.22757509350776672, "step": 220 }, { "epoch": 2.3763440860215055, "grad_norm": 0.5368948762143404, "learning_rate": 1.1361231194343436e-06, "loss": 0.20776845514774323, "step": 221 }, { "epoch": 2.3870967741935485, "grad_norm": 0.5362067849303692, "learning_rate": 1.0987746713434578e-06, "loss": 0.21643775701522827, "step": 222 }, { "epoch": 2.3978494623655915, "grad_norm": 0.7478981160056104, "learning_rate": 1.0619745051995473e-06, "loss": 0.2321006953716278, "step": 223 }, { "epoch": 2.4086021505376345, "grad_norm": 0.5589077284378241, "learning_rate": 1.0257277929332332e-06, "loss": 0.2335091233253479, "step": 224 }, { "epoch": 2.4193548387096775, "grad_norm": 0.5449990060855128, "learning_rate": 9.900396286922025e-07, "loss": 0.22067061066627502, "step": 225 }, { "epoch": 2.4301075268817205, "grad_norm": 0.545790815861183, "learning_rate": 9.549150281252633e-07, "loss": 0.24003136157989502, "step": 226 }, { "epoch": 2.4408602150537635, "grad_norm": 0.5197037503280709, "learning_rate": 9.203589276774438e-07, "loss": 0.2232522964477539, "step": 227 }, { "epoch": 2.4516129032258065, "grad_norm": 0.5783224759821376, "learning_rate": 8.86376183896226e-07, "loss": 0.2168661653995514, "step": 228 }, { "epoch": 2.4623655913978495, "grad_norm": 0.5772298663365216, "learning_rate": 8.529715727489912e-07, "loss": 0.22855976223945618, "step": 229 }, { "epoch": 2.4731182795698925, "grad_norm": 0.5901747678986362, "learning_rate": 8.201497889518073e-07, "loss": 0.23543155193328857, "step": 230 }, { "epoch": 2.4838709677419355, "grad_norm": 0.4859350321815666, "learning_rate": 7.879154453096305e-07, "loss": 0.20444487035274506, "step": 231 }, { "epoch": 2.4946236559139785, "grad_norm": 0.5733816209081586, "learning_rate": 7.562730720680111e-07, "loss": 0.22317257523536682, "step": 232 }, { "epoch": 2.5053763440860215, "grad_norm": 0.6158850373560834, "learning_rate": 7.25227116276413e-07, "loss": 0.21304547786712646, "step": 233 }, { "epoch": 2.5161290322580645, "grad_norm": 0.5303929149495401, "learning_rate": 6.947819411632223e-07, "loss": 0.22398605942726135, "step": 234 }, { "epoch": 2.5268817204301075, "grad_norm": 0.5807287890882401, "learning_rate": 6.649418255225298e-07, "loss": 0.22487413883209229, "step": 235 }, { "epoch": 2.5376344086021505, "grad_norm": 0.5151341689499322, "learning_rate": 6.357109631127889e-07, "loss": 0.20791815221309662, "step": 236 }, { "epoch": 2.5483870967741935, "grad_norm": 0.5723989318815612, "learning_rate": 6.07093462067419e-07, "loss": 0.22853252291679382, "step": 237 }, { "epoch": 2.5591397849462365, "grad_norm": 0.5151529413040788, "learning_rate": 5.79093344317449e-07, "loss": 0.23513737320899963, "step": 238 }, { "epoch": 2.5698924731182795, "grad_norm": 0.6125760680997326, "learning_rate": 5.517145450262639e-07, "loss": 0.23416566848754883, "step": 239 }, { "epoch": 2.5806451612903225, "grad_norm": 0.5169337829356161, "learning_rate": 5.249609120365579e-07, "loss": 0.2236918807029724, "step": 240 }, { "epoch": 2.5913978494623655, "grad_norm": 0.5150874552544314, "learning_rate": 4.988362053295564e-07, "loss": 0.22128576040267944, "step": 241 }, { "epoch": 2.6021505376344085, "grad_norm": 0.6254184464054292, "learning_rate": 4.733440964965791e-07, "loss": 0.23455382883548737, "step": 242 }, { "epoch": 2.6129032258064515, "grad_norm": 0.5612944471466359, "learning_rate": 4.484881682230341e-07, "loss": 0.22176575660705566, "step": 243 }, { "epoch": 2.6236559139784945, "grad_norm": 0.5178079276744024, "learning_rate": 4.242719137849077e-07, "loss": 0.21548646688461304, "step": 244 }, { "epoch": 2.6344086021505375, "grad_norm": 0.5659047109155118, "learning_rate": 4.00698736557808e-07, "loss": 0.23057113587856293, "step": 245 }, { "epoch": 2.6451612903225805, "grad_norm": 0.5098023427109231, "learning_rate": 3.777719495386567e-07, "loss": 0.2156354784965515, "step": 246 }, { "epoch": 2.6559139784946235, "grad_norm": 0.49515330466811, "learning_rate": 3.5549477488007853e-07, "loss": 0.19704020023345947, "step": 247 }, { "epoch": 2.6666666666666665, "grad_norm": 0.5297791040512543, "learning_rate": 3.3387034343755063e-07, "loss": 0.21541355550289154, "step": 248 }, { "epoch": 2.6774193548387095, "grad_norm": 0.5134148705186912, "learning_rate": 3.1290169432939556e-07, "loss": 0.20781481266021729, "step": 249 }, { "epoch": 2.688172043010753, "grad_norm": 0.6591675869543023, "learning_rate": 2.925917745096568e-07, "loss": 0.23452350497245789, "step": 250 }, { "epoch": 2.698924731182796, "grad_norm": 0.5940779586998381, "learning_rate": 2.7294343835393366e-07, "loss": 0.23721283674240112, "step": 251 }, { "epoch": 2.709677419354839, "grad_norm": 0.6904575634426214, "learning_rate": 2.539594472582213e-07, "loss": 0.23339498043060303, "step": 252 }, { "epoch": 2.720430107526882, "grad_norm": 0.549700857643845, "learning_rate": 2.3564246925082358e-07, "loss": 0.2422715723514557, "step": 253 }, { "epoch": 2.731182795698925, "grad_norm": 0.551951066945312, "learning_rate": 2.179950786173879e-07, "loss": 0.22343403100967407, "step": 254 }, { "epoch": 2.741935483870968, "grad_norm": 0.5112230065122543, "learning_rate": 2.01019755539108e-07, "loss": 0.2022184431552887, "step": 255 }, { "epoch": 2.752688172043011, "grad_norm": 0.6011326795775324, "learning_rate": 1.8471888574415953e-07, "loss": 0.21843938529491425, "step": 256 }, { "epoch": 2.763440860215054, "grad_norm": 0.5598013580182175, "learning_rate": 1.690947601724091e-07, "loss": 0.2313247174024582, "step": 257 }, { "epoch": 2.774193548387097, "grad_norm": 0.579260801737671, "learning_rate": 1.5414957465343883e-07, "loss": 0.23432737588882446, "step": 258 }, { "epoch": 2.78494623655914, "grad_norm": 0.6080486926763811, "learning_rate": 1.3988542959794627e-07, "loss": 0.2501685619354248, "step": 259 }, { "epoch": 2.795698924731183, "grad_norm": 0.6231912247704571, "learning_rate": 1.2630432970255014e-07, "loss": 0.22736617922782898, "step": 260 }, { "epoch": 2.806451612903226, "grad_norm": 0.6526633675286432, "learning_rate": 1.1340818366804728e-07, "loss": 0.2217504233121872, "step": 261 }, { "epoch": 2.817204301075269, "grad_norm": 0.5774442194442434, "learning_rate": 1.0119880393116177e-07, "loss": 0.21880705654621124, "step": 262 }, { "epoch": 2.827956989247312, "grad_norm": 0.5510924764674725, "learning_rate": 8.967790640982466e-08, "loss": 0.23220205307006836, "step": 263 }, { "epoch": 2.838709677419355, "grad_norm": 0.5940155308333281, "learning_rate": 7.884711026201586e-08, "loss": 0.22368629276752472, "step": 264 }, { "epoch": 2.849462365591398, "grad_norm": 0.48018958702032566, "learning_rate": 6.870793765820783e-08, "loss": 0.2191024124622345, "step": 265 }, { "epoch": 2.860215053763441, "grad_norm": 0.5348968055346366, "learning_rate": 5.92618135674361e-08, "loss": 0.2141711413860321, "step": 266 }, { "epoch": 2.870967741935484, "grad_norm": 0.5058952998875674, "learning_rate": 5.0510065557034526e-08, "loss": 0.21273520588874817, "step": 267 }, { "epoch": 2.881720430107527, "grad_norm": 0.5748923929140294, "learning_rate": 4.245392360605727e-08, "loss": 0.2339775264263153, "step": 268 }, { "epoch": 2.89247311827957, "grad_norm": 0.543587597117276, "learning_rate": 3.5094519932415417e-08, "loss": 0.22123411297798157, "step": 269 }, { "epoch": 2.903225806451613, "grad_norm": 0.5451715623802244, "learning_rate": 2.843288883375539e-08, "loss": 0.22225311398506165, "step": 270 }, { "epoch": 2.913978494623656, "grad_norm": 0.5639808086322913, "learning_rate": 2.2469966542096323e-08, "loss": 0.220473051071167, "step": 271 }, { "epoch": 2.924731182795699, "grad_norm": 0.5603998497829358, "learning_rate": 1.7206591092253642e-08, "loss": 0.22855965793132782, "step": 272 }, { "epoch": 2.935483870967742, "grad_norm": 0.5700123425140283, "learning_rate": 1.264350220405719e-08, "loss": 0.226323202252388, "step": 273 }, { "epoch": 2.946236559139785, "grad_norm": 0.5193108167577539, "learning_rate": 8.781341178393244e-09, "loss": 0.21938025951385498, "step": 274 }, { "epoch": 2.956989247311828, "grad_norm": 0.6419513279403505, "learning_rate": 5.620650807073857e-09, "loss": 0.24610912799835205, "step": 275 }, { "epoch": 2.967741935483871, "grad_norm": 0.6308682159730912, "learning_rate": 3.1618752965534295e-09, "loss": 0.22690746188163757, "step": 276 }, { "epoch": 2.978494623655914, "grad_norm": 0.5263395659056124, "learning_rate": 1.4053602054991954e-09, "loss": 0.20472858846187592, "step": 277 }, { "epoch": 2.989247311827957, "grad_norm": 0.5329091838420249, "learning_rate": 3.513523962256349e-10, "loss": 0.2296607792377472, "step": 278 }, { "epoch": 3.0, "grad_norm": 0.6481688008102051, "learning_rate": 0.0, "loss": 0.22974838316440582, "step": 279 } ], "logging_steps": 1, "max_steps": 279, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 142925655572480.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }