| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 279, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.010752688172043012, | |
| "grad_norm": 19.659445278083677, | |
| "learning_rate": 7.142857142857143e-07, | |
| "loss": 0.8801538944244385, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.021505376344086023, | |
| "grad_norm": 21.183622990197563, | |
| "learning_rate": 1.4285714285714286e-06, | |
| "loss": 0.9106118679046631, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.03225806451612903, | |
| "grad_norm": 9.680562864297194, | |
| "learning_rate": 2.1428571428571427e-06, | |
| "loss": 0.89518141746521, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.043010752688172046, | |
| "grad_norm": 8.463557758001873, | |
| "learning_rate": 2.8571428571428573e-06, | |
| "loss": 0.8813375234603882, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.053763440860215055, | |
| "grad_norm": 8.000585049846563, | |
| "learning_rate": 3.5714285714285718e-06, | |
| "loss": 0.8478757739067078, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.06451612903225806, | |
| "grad_norm": 6.392760393586216, | |
| "learning_rate": 4.2857142857142855e-06, | |
| "loss": 0.7342901825904846, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.07526881720430108, | |
| "grad_norm": 4.291231505071156, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6894474029541016, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.08602150537634409, | |
| "grad_norm": 3.3545639671049603, | |
| "learning_rate": 5.7142857142857145e-06, | |
| "loss": 0.6144750118255615, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0967741935483871, | |
| "grad_norm": 1.3577798099653526, | |
| "learning_rate": 6.4285714285714295e-06, | |
| "loss": 0.5370485186576843, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.10752688172043011, | |
| "grad_norm": 1.2774532761940443, | |
| "learning_rate": 7.1428571428571436e-06, | |
| "loss": 0.5807953476905823, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.11827956989247312, | |
| "grad_norm": 1.0447472651159857, | |
| "learning_rate": 7.857142857142858e-06, | |
| "loss": 0.5280188918113708, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.12903225806451613, | |
| "grad_norm": 1.3301070676426008, | |
| "learning_rate": 8.571428571428571e-06, | |
| "loss": 0.5441749691963196, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.13978494623655913, | |
| "grad_norm": 1.0565188353675214, | |
| "learning_rate": 9.285714285714288e-06, | |
| "loss": 0.5112073421478271, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.15053763440860216, | |
| "grad_norm": 1.0753192556611026, | |
| "learning_rate": 1e-05, | |
| "loss": 0.5184136629104614, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.16129032258064516, | |
| "grad_norm": 0.8209457367101312, | |
| "learning_rate": 9.999648647603774e-06, | |
| "loss": 0.5023149251937866, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.17204301075268819, | |
| "grad_norm": 0.7282985426170734, | |
| "learning_rate": 9.998594639794502e-06, | |
| "loss": 0.48974722623825073, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.1827956989247312, | |
| "grad_norm": 0.7177574512039909, | |
| "learning_rate": 9.996838124703448e-06, | |
| "loss": 0.47443556785583496, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.1935483870967742, | |
| "grad_norm": 0.7867041042254088, | |
| "learning_rate": 9.994379349192927e-06, | |
| "loss": 0.48248669505119324, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.20430107526881722, | |
| "grad_norm": 0.7192865433337082, | |
| "learning_rate": 9.991218658821609e-06, | |
| "loss": 0.46316730976104736, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.21505376344086022, | |
| "grad_norm": 0.8279557032400371, | |
| "learning_rate": 9.987356497795944e-06, | |
| "loss": 0.5312884449958801, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.22580645161290322, | |
| "grad_norm": 0.7907337063911983, | |
| "learning_rate": 9.982793408907747e-06, | |
| "loss": 0.48629915714263916, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.23655913978494625, | |
| "grad_norm": 0.634234358721607, | |
| "learning_rate": 9.977530033457906e-06, | |
| "loss": 0.47711604833602905, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.24731182795698925, | |
| "grad_norm": 0.7748372497671804, | |
| "learning_rate": 9.971567111166246e-06, | |
| "loss": 0.4741020202636719, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.25806451612903225, | |
| "grad_norm": 0.6968307669517513, | |
| "learning_rate": 9.964905480067585e-06, | |
| "loss": 0.48086652159690857, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.26881720430107525, | |
| "grad_norm": 0.6205197336090342, | |
| "learning_rate": 9.957546076393944e-06, | |
| "loss": 0.45759013295173645, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.27956989247311825, | |
| "grad_norm": 0.5942106287222008, | |
| "learning_rate": 9.949489934442966e-06, | |
| "loss": 0.454744428396225, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.2903225806451613, | |
| "grad_norm": 0.6374339549026193, | |
| "learning_rate": 9.940738186432565e-06, | |
| "loss": 0.4751560091972351, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.3010752688172043, | |
| "grad_norm": 0.5187971760627564, | |
| "learning_rate": 9.931292062341793e-06, | |
| "loss": 0.39658498764038086, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.3118279569892473, | |
| "grad_norm": 0.5360240389517962, | |
| "learning_rate": 9.921152889737985e-06, | |
| "loss": 0.46163371205329895, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.3225806451612903, | |
| "grad_norm": 0.5585203893563241, | |
| "learning_rate": 9.910322093590177e-06, | |
| "loss": 0.46062612533569336, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.3333333333333333, | |
| "grad_norm": 0.5610561072744311, | |
| "learning_rate": 9.898801196068839e-06, | |
| "loss": 0.4484734833240509, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.34408602150537637, | |
| "grad_norm": 0.5533641378752504, | |
| "learning_rate": 9.886591816331953e-06, | |
| "loss": 0.41486942768096924, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.3548387096774194, | |
| "grad_norm": 0.5365723882895206, | |
| "learning_rate": 9.87369567029745e-06, | |
| "loss": 0.43502286076545715, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.3655913978494624, | |
| "grad_norm": 0.5536007189636761, | |
| "learning_rate": 9.860114570402055e-06, | |
| "loss": 0.4295029640197754, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.3763440860215054, | |
| "grad_norm": 0.537839879636324, | |
| "learning_rate": 9.845850425346563e-06, | |
| "loss": 0.45029592514038086, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.3870967741935484, | |
| "grad_norm": 0.5299693550104672, | |
| "learning_rate": 9.830905239827592e-06, | |
| "loss": 0.38297146558761597, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.3978494623655914, | |
| "grad_norm": 0.5451092272235792, | |
| "learning_rate": 9.815281114255841e-06, | |
| "loss": 0.4300140142440796, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.40860215053763443, | |
| "grad_norm": 0.6192689983212899, | |
| "learning_rate": 9.798980244460892e-06, | |
| "loss": 0.47214269638061523, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.41935483870967744, | |
| "grad_norm": 0.5870329353069306, | |
| "learning_rate": 9.782004921382612e-06, | |
| "loss": 0.43783068656921387, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.43010752688172044, | |
| "grad_norm": 0.5903637860392952, | |
| "learning_rate": 9.764357530749178e-06, | |
| "loss": 0.4084509611129761, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.44086021505376344, | |
| "grad_norm": 0.61705449873411, | |
| "learning_rate": 9.74604055274178e-06, | |
| "loss": 0.42812418937683105, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.45161290322580644, | |
| "grad_norm": 1.156325694795339, | |
| "learning_rate": 9.727056561646067e-06, | |
| "loss": 0.4196915626525879, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.46236559139784944, | |
| "grad_norm": 0.580641483078281, | |
| "learning_rate": 9.707408225490343e-06, | |
| "loss": 0.4251587986946106, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.4731182795698925, | |
| "grad_norm": 0.5510691280763641, | |
| "learning_rate": 9.687098305670606e-06, | |
| "loss": 0.4205142855644226, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.4838709677419355, | |
| "grad_norm": 0.5225683335691166, | |
| "learning_rate": 9.66612965656245e-06, | |
| "loss": 0.37856176495552063, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.4946236559139785, | |
| "grad_norm": 0.5740926054795668, | |
| "learning_rate": 9.644505225119922e-06, | |
| "loss": 0.4447917342185974, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.5053763440860215, | |
| "grad_norm": 0.5873250280785997, | |
| "learning_rate": 9.622228050461345e-06, | |
| "loss": 0.3922080397605896, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.5161290322580645, | |
| "grad_norm": 0.6219913849412524, | |
| "learning_rate": 9.599301263442194e-06, | |
| "loss": 0.43804931640625, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.5268817204301075, | |
| "grad_norm": 0.5613998719760911, | |
| "learning_rate": 9.575728086215093e-06, | |
| "loss": 0.39639532566070557, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.5376344086021505, | |
| "grad_norm": 0.5892845271021349, | |
| "learning_rate": 9.551511831776966e-06, | |
| "loss": 0.4323637783527374, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.5483870967741935, | |
| "grad_norm": 0.6452327069861976, | |
| "learning_rate": 9.526655903503423e-06, | |
| "loss": 0.4111916422843933, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.5591397849462365, | |
| "grad_norm": 0.6139840452181067, | |
| "learning_rate": 9.501163794670445e-06, | |
| "loss": 0.41024255752563477, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.5698924731182796, | |
| "grad_norm": 0.5311229042653062, | |
| "learning_rate": 9.475039087963443e-06, | |
| "loss": 0.39047467708587646, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.5806451612903226, | |
| "grad_norm": 0.5686779350086011, | |
| "learning_rate": 9.448285454973739e-06, | |
| "loss": 0.4083505868911743, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.5913978494623656, | |
| "grad_norm": 0.5795028580373828, | |
| "learning_rate": 9.420906655682553e-06, | |
| "loss": 0.41430288553237915, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.6021505376344086, | |
| "grad_norm": 0.5426200748349074, | |
| "learning_rate": 9.392906537932582e-06, | |
| "loss": 0.39692986011505127, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.6129032258064516, | |
| "grad_norm": 0.5096666206067523, | |
| "learning_rate": 9.364289036887214e-06, | |
| "loss": 0.39456355571746826, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.6236559139784946, | |
| "grad_norm": 0.6214693543974321, | |
| "learning_rate": 9.335058174477472e-06, | |
| "loss": 0.43065640330314636, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.6344086021505376, | |
| "grad_norm": 0.6080775052566749, | |
| "learning_rate": 9.305218058836778e-06, | |
| "loss": 0.4233320653438568, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 0.5891380183146641, | |
| "learning_rate": 9.274772883723587e-06, | |
| "loss": 0.44017481803894043, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.6559139784946236, | |
| "grad_norm": 0.5317932432917034, | |
| "learning_rate": 9.24372692793199e-06, | |
| "loss": 0.39426931738853455, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.6666666666666666, | |
| "grad_norm": 0.6464375491620328, | |
| "learning_rate": 9.21208455469037e-06, | |
| "loss": 0.4302670657634735, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.6774193548387096, | |
| "grad_norm": 0.6191245668003084, | |
| "learning_rate": 9.179850211048193e-06, | |
| "loss": 0.3996105492115021, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.6881720430107527, | |
| "grad_norm": 0.5209850545783036, | |
| "learning_rate": 9.14702842725101e-06, | |
| "loss": 0.3926236629486084, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.6989247311827957, | |
| "grad_norm": 0.5469602060829455, | |
| "learning_rate": 9.113623816103775e-06, | |
| "loss": 0.4129854142665863, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.7096774193548387, | |
| "grad_norm": 0.6436176685575749, | |
| "learning_rate": 9.079641072322555e-06, | |
| "loss": 0.4262722432613373, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.7204301075268817, | |
| "grad_norm": 0.6241121753870925, | |
| "learning_rate": 9.045084971874738e-06, | |
| "loss": 0.4259442687034607, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.7311827956989247, | |
| "grad_norm": 0.520706057326722, | |
| "learning_rate": 9.009960371307798e-06, | |
| "loss": 0.40881264209747314, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.7419354838709677, | |
| "grad_norm": 0.5406305590251429, | |
| "learning_rate": 8.974272207066767e-06, | |
| "loss": 0.4136165976524353, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.7526881720430108, | |
| "grad_norm": 0.5589669382622552, | |
| "learning_rate": 8.938025494800454e-06, | |
| "loss": 0.3692484498023987, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.7634408602150538, | |
| "grad_norm": 0.5484586018961858, | |
| "learning_rate": 8.901225328656543e-06, | |
| "loss": 0.392721563577652, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.7741935483870968, | |
| "grad_norm": 0.5720189126246105, | |
| "learning_rate": 8.863876880565656e-06, | |
| "loss": 0.4131432771682739, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.7849462365591398, | |
| "grad_norm": 0.5475650690180592, | |
| "learning_rate": 8.825985399514488e-06, | |
| "loss": 0.39399850368499756, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.7956989247311828, | |
| "grad_norm": 0.571446157132029, | |
| "learning_rate": 8.787556210808101e-06, | |
| "loss": 0.3765556216239929, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.8064516129032258, | |
| "grad_norm": 0.5096610649520538, | |
| "learning_rate": 8.748594715321512e-06, | |
| "loss": 0.38086819648742676, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.8172043010752689, | |
| "grad_norm": 0.5696584948279285, | |
| "learning_rate": 8.70910638874064e-06, | |
| "loss": 0.40993741154670715, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.8279569892473119, | |
| "grad_norm": 0.5448260281274895, | |
| "learning_rate": 8.669096780792754e-06, | |
| "loss": 0.3912901282310486, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.8387096774193549, | |
| "grad_norm": 0.584810442023009, | |
| "learning_rate": 8.628571514466502e-06, | |
| "loss": 0.39722931385040283, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.8494623655913979, | |
| "grad_norm": 0.5590387314291252, | |
| "learning_rate": 8.587536285221656e-06, | |
| "loss": 0.38008394837379456, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.8602150537634409, | |
| "grad_norm": 0.5199976478034525, | |
| "learning_rate": 8.545996860188668e-06, | |
| "loss": 0.3779188394546509, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.8709677419354839, | |
| "grad_norm": 0.6237910452530793, | |
| "learning_rate": 8.503959077358143e-06, | |
| "loss": 0.40831005573272705, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.8817204301075269, | |
| "grad_norm": 0.5103230716146216, | |
| "learning_rate": 8.46142884476038e-06, | |
| "loss": 0.3830574154853821, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.8924731182795699, | |
| "grad_norm": 0.6022470754794768, | |
| "learning_rate": 8.418412139635026e-06, | |
| "loss": 0.3760561943054199, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.9032258064516129, | |
| "grad_norm": 0.5676501864622416, | |
| "learning_rate": 8.374915007591053e-06, | |
| "loss": 0.3981160521507263, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.9139784946236559, | |
| "grad_norm": 0.62733814659435, | |
| "learning_rate": 8.330943561757092e-06, | |
| "loss": 0.38736072182655334, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.9247311827956989, | |
| "grad_norm": 0.581917413328386, | |
| "learning_rate": 8.286503981922284e-06, | |
| "loss": 0.40054696798324585, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.9354838709677419, | |
| "grad_norm": 0.5069721523225358, | |
| "learning_rate": 8.241602513667775e-06, | |
| "loss": 0.36416178941726685, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.946236559139785, | |
| "grad_norm": 0.6484709424945745, | |
| "learning_rate": 8.19624546748895e-06, | |
| "loss": 0.3703634738922119, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.956989247311828, | |
| "grad_norm": 0.5102251649420729, | |
| "learning_rate": 8.150439217908557e-06, | |
| "loss": 0.39143842458724976, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.967741935483871, | |
| "grad_norm": 0.5485517368687335, | |
| "learning_rate": 8.104190202580811e-06, | |
| "loss": 0.3922935128211975, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.978494623655914, | |
| "grad_norm": 0.5680193851368315, | |
| "learning_rate": 8.057504921386661e-06, | |
| "loss": 0.3872210383415222, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.989247311827957, | |
| "grad_norm": 0.5405961708214808, | |
| "learning_rate": 8.010389935520269e-06, | |
| "loss": 0.4088854193687439, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.5234958056904553, | |
| "learning_rate": 7.962851866566912e-06, | |
| "loss": 0.3855600953102112, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 1.010752688172043, | |
| "grad_norm": 0.5751268630788401, | |
| "learning_rate": 7.914897395572362e-06, | |
| "loss": 0.32572415471076965, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 1.021505376344086, | |
| "grad_norm": 0.5181295042077261, | |
| "learning_rate": 7.866533262103937e-06, | |
| "loss": 0.3046826720237732, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 1.032258064516129, | |
| "grad_norm": 0.5508917328926465, | |
| "learning_rate": 7.817766263303312e-06, | |
| "loss": 0.29789605736732483, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 1.043010752688172, | |
| "grad_norm": 0.5870437887252276, | |
| "learning_rate": 7.768603252931243e-06, | |
| "loss": 0.3265905976295471, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 1.053763440860215, | |
| "grad_norm": 0.5133318653975599, | |
| "learning_rate": 7.719051140404327e-06, | |
| "loss": 0.31667008996009827, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 1.064516129032258, | |
| "grad_norm": 0.5679147172610466, | |
| "learning_rate": 7.669116889823955e-06, | |
| "loss": 0.2714020311832428, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 1.075268817204301, | |
| "grad_norm": 0.6225396396665326, | |
| "learning_rate": 7.6188075189975644e-06, | |
| "loss": 0.3139406740665436, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.086021505376344, | |
| "grad_norm": 0.714553736071244, | |
| "learning_rate": 7.568130098452352e-06, | |
| "loss": 0.2961091995239258, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 1.096774193548387, | |
| "grad_norm": 0.523889168005371, | |
| "learning_rate": 7.517091750441576e-06, | |
| "loss": 0.27245283126831055, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 1.10752688172043, | |
| "grad_norm": 0.639295771779283, | |
| "learning_rate": 7.465699647943586e-06, | |
| "loss": 0.29422637820243835, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 1.118279569892473, | |
| "grad_norm": 0.6150692376811179, | |
| "learning_rate": 7.413961013653725e-06, | |
| "loss": 0.29594993591308594, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 1.129032258064516, | |
| "grad_norm": 0.6901790920161586, | |
| "learning_rate": 7.361883118969248e-06, | |
| "loss": 0.3163909912109375, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 1.139784946236559, | |
| "grad_norm": 0.5947317629259291, | |
| "learning_rate": 7.309473282967387e-06, | |
| "loss": 0.2941242456436157, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 1.1505376344086022, | |
| "grad_norm": 0.5873657820382887, | |
| "learning_rate": 7.256738871376733e-06, | |
| "loss": 0.29748812317848206, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 1.1612903225806452, | |
| "grad_norm": 0.5914901548215277, | |
| "learning_rate": 7.203687295542032e-06, | |
| "loss": 0.3174619972705841, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 1.1720430107526882, | |
| "grad_norm": 0.5757880846821776, | |
| "learning_rate": 7.1503260113826035e-06, | |
| "loss": 0.31319767236709595, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 1.1827956989247312, | |
| "grad_norm": 0.7019483863871168, | |
| "learning_rate": 7.09666251834447e-06, | |
| "loss": 0.3302695155143738, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.1935483870967742, | |
| "grad_norm": 0.5613831627968954, | |
| "learning_rate": 7.042704358346375e-06, | |
| "loss": 0.30153942108154297, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 1.2043010752688172, | |
| "grad_norm": 0.5906296533769152, | |
| "learning_rate": 6.988459114719849e-06, | |
| "loss": 0.3321286141872406, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 1.2150537634408602, | |
| "grad_norm": 0.6185930436417313, | |
| "learning_rate": 6.933934411143419e-06, | |
| "loss": 0.3183019161224365, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 1.2258064516129032, | |
| "grad_norm": 0.5223193968800454, | |
| "learning_rate": 6.879137910571191e-06, | |
| "loss": 0.28522804379463196, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 1.2365591397849462, | |
| "grad_norm": 0.5306097120719284, | |
| "learning_rate": 6.824077314155877e-06, | |
| "loss": 0.27248305082321167, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 1.2473118279569892, | |
| "grad_norm": 0.5410094711890291, | |
| "learning_rate": 6.768760360166471e-06, | |
| "loss": 0.29798364639282227, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 1.2580645161290323, | |
| "grad_norm": 0.594316663926518, | |
| "learning_rate": 6.713194822900707e-06, | |
| "loss": 0.27839338779449463, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 1.2688172043010753, | |
| "grad_norm": 0.6110653082771621, | |
| "learning_rate": 6.657388511592453e-06, | |
| "loss": 0.30802613496780396, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 1.2795698924731183, | |
| "grad_norm": 0.5260300979877027, | |
| "learning_rate": 6.601349269314188e-06, | |
| "loss": 0.29552769660949707, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 1.2903225806451613, | |
| "grad_norm": 0.541255561766811, | |
| "learning_rate": 6.545084971874738e-06, | |
| "loss": 0.28333544731140137, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.3010752688172043, | |
| "grad_norm": 0.5631658544435217, | |
| "learning_rate": 6.488603526712391e-06, | |
| "loss": 0.3072783946990967, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 1.3118279569892473, | |
| "grad_norm": 0.5999697642319023, | |
| "learning_rate": 6.431912871783587e-06, | |
| "loss": 0.2956047058105469, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 1.3225806451612903, | |
| "grad_norm": 0.5550402984265157, | |
| "learning_rate": 6.3750209744473105e-06, | |
| "loss": 0.3064526319503784, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 1.3333333333333333, | |
| "grad_norm": 0.6418624402271341, | |
| "learning_rate": 6.3179358303453386e-06, | |
| "loss": 0.30682218074798584, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 1.3440860215053765, | |
| "grad_norm": 0.5305043074037712, | |
| "learning_rate": 6.260665462278544e-06, | |
| "loss": 0.3036462068557739, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.3548387096774195, | |
| "grad_norm": 0.5732839993822031, | |
| "learning_rate": 6.203217919079343e-06, | |
| "loss": 0.32605141401290894, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 1.3655913978494625, | |
| "grad_norm": 0.5272514261375651, | |
| "learning_rate": 6.145601274480521e-06, | |
| "loss": 0.2800780236721039, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 1.3763440860215055, | |
| "grad_norm": 0.5383888711112677, | |
| "learning_rate": 6.08782362598054e-06, | |
| "loss": 0.29685452580451965, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 1.3870967741935485, | |
| "grad_norm": 0.5260069329849035, | |
| "learning_rate": 6.029893093705492e-06, | |
| "loss": 0.30100512504577637, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 1.3978494623655915, | |
| "grad_norm": 0.5194176965790988, | |
| "learning_rate": 5.971817819267914e-06, | |
| "loss": 0.28969407081604004, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.4086021505376345, | |
| "grad_norm": 0.5795379145956892, | |
| "learning_rate": 5.9136059646225375e-06, | |
| "loss": 0.3088012635707855, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 1.4193548387096775, | |
| "grad_norm": 0.5605938731509191, | |
| "learning_rate": 5.855265710919211e-06, | |
| "loss": 0.2954477369785309, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 1.4301075268817205, | |
| "grad_norm": 0.5513949309182947, | |
| "learning_rate": 5.796805257353109e-06, | |
| "loss": 0.31166109442710876, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 1.4408602150537635, | |
| "grad_norm": 0.6271394444282732, | |
| "learning_rate": 5.738232820012407e-06, | |
| "loss": 0.33265459537506104, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 1.4516129032258065, | |
| "grad_norm": 0.5513926189063966, | |
| "learning_rate": 5.679556630723592e-06, | |
| "loss": 0.271618515253067, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.4623655913978495, | |
| "grad_norm": 0.5799088021416959, | |
| "learning_rate": 5.620784935894548e-06, | |
| "loss": 0.2902123034000397, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 1.4731182795698925, | |
| "grad_norm": 0.5662089313253346, | |
| "learning_rate": 5.561925995355595e-06, | |
| "loss": 0.2844197154045105, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 1.4838709677419355, | |
| "grad_norm": 0.9937555515678825, | |
| "learning_rate": 5.5029880811986546e-06, | |
| "loss": 0.2922922372817993, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 1.4946236559139785, | |
| "grad_norm": 0.591368004900944, | |
| "learning_rate": 5.443979476614674e-06, | |
| "loss": 0.28497251868247986, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 1.5053763440860215, | |
| "grad_norm": 0.6076112891374724, | |
| "learning_rate": 5.384908474729501e-06, | |
| "loss": 0.31142377853393555, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.5161290322580645, | |
| "grad_norm": 0.5765965412966652, | |
| "learning_rate": 5.325783377438357e-06, | |
| "loss": 0.3015650510787964, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 1.5268817204301075, | |
| "grad_norm": 0.5232918774810716, | |
| "learning_rate": 5.266612494239088e-06, | |
| "loss": 0.275716096162796, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 1.5376344086021505, | |
| "grad_norm": 0.5680197113944581, | |
| "learning_rate": 5.207404141064334e-06, | |
| "loss": 0.2873622179031372, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 1.5483870967741935, | |
| "grad_norm": 0.49399262225350316, | |
| "learning_rate": 5.148166639112799e-06, | |
| "loss": 0.26688140630722046, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 1.5591397849462365, | |
| "grad_norm": 0.5949244178581083, | |
| "learning_rate": 5.088908313679788e-06, | |
| "loss": 0.29102981090545654, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.5698924731182795, | |
| "grad_norm": 0.6289302836737687, | |
| "learning_rate": 5.029637492987153e-06, | |
| "loss": 0.2931191027164459, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 1.5806451612903225, | |
| "grad_norm": 0.6088833365757095, | |
| "learning_rate": 4.970362507012848e-06, | |
| "loss": 0.2844638228416443, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 1.5913978494623655, | |
| "grad_norm": 0.6971118150263104, | |
| "learning_rate": 4.911091686320213e-06, | |
| "loss": 0.3226025104522705, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 1.6021505376344085, | |
| "grad_norm": 0.5612811561354285, | |
| "learning_rate": 4.8518333608872015e-06, | |
| "loss": 0.28764593601226807, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 1.6129032258064515, | |
| "grad_norm": 0.6065928167636621, | |
| "learning_rate": 4.792595858935668e-06, | |
| "loss": 0.3277990221977234, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.6236559139784945, | |
| "grad_norm": 0.5135483316527438, | |
| "learning_rate": 4.733387505760913e-06, | |
| "loss": 0.2962847948074341, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 1.6344086021505375, | |
| "grad_norm": 0.6160906809336857, | |
| "learning_rate": 4.674216622561645e-06, | |
| "loss": 0.294375479221344, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 1.6451612903225805, | |
| "grad_norm": 0.6058840168616271, | |
| "learning_rate": 4.6150915252705005e-06, | |
| "loss": 0.3054540157318115, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 1.6559139784946235, | |
| "grad_norm": 0.5298336251451402, | |
| "learning_rate": 4.556020523385326e-06, | |
| "loss": 0.30477648973464966, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 0.5241054596736249, | |
| "learning_rate": 4.497011918801347e-06, | |
| "loss": 0.2855217456817627, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 1.6774193548387095, | |
| "grad_norm": 0.48994778885097257, | |
| "learning_rate": 4.438074004644407e-06, | |
| "loss": 0.30031388998031616, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 1.6881720430107527, | |
| "grad_norm": 0.6034882888193746, | |
| "learning_rate": 4.379215064105454e-06, | |
| "loss": 0.3072971999645233, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 1.6989247311827957, | |
| "grad_norm": 0.5057395078651564, | |
| "learning_rate": 4.32044336927641e-06, | |
| "loss": 0.27074217796325684, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 1.7096774193548387, | |
| "grad_norm": 0.548027010828449, | |
| "learning_rate": 4.261767179987595e-06, | |
| "loss": 0.28892627358436584, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 1.7204301075268817, | |
| "grad_norm": 0.5949799690870136, | |
| "learning_rate": 4.203194742646893e-06, | |
| "loss": 0.31755733489990234, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.7311827956989247, | |
| "grad_norm": 0.5988557766268561, | |
| "learning_rate": 4.1447342890807905e-06, | |
| "loss": 0.28983038663864136, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 1.7419354838709677, | |
| "grad_norm": 0.5153375977084437, | |
| "learning_rate": 4.086394035377463e-06, | |
| "loss": 0.2821354269981384, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 1.7526881720430108, | |
| "grad_norm": 0.5758453410329892, | |
| "learning_rate": 4.028182180732088e-06, | |
| "loss": 0.26659733057022095, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 1.7634408602150538, | |
| "grad_norm": 0.5515188533069812, | |
| "learning_rate": 3.970106906294509e-06, | |
| "loss": 0.29241591691970825, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 1.7741935483870968, | |
| "grad_norm": 0.5465960078319545, | |
| "learning_rate": 3.912176374019462e-06, | |
| "loss": 0.28333449363708496, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 1.7849462365591398, | |
| "grad_norm": 0.5696267902501757, | |
| "learning_rate": 3.85439872551948e-06, | |
| "loss": 0.28454482555389404, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 1.7956989247311828, | |
| "grad_norm": 0.5537225691065368, | |
| "learning_rate": 3.796782080920659e-06, | |
| "loss": 0.29573291540145874, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 1.8064516129032258, | |
| "grad_norm": 0.5317821119388748, | |
| "learning_rate": 3.7393345377214584e-06, | |
| "loss": 0.3006817400455475, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 1.817204301075269, | |
| "grad_norm": 0.5193143293546626, | |
| "learning_rate": 3.682064169654663e-06, | |
| "loss": 0.28759250044822693, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 1.827956989247312, | |
| "grad_norm": 0.5024936169709733, | |
| "learning_rate": 3.6249790255526916e-06, | |
| "loss": 0.279812753200531, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.838709677419355, | |
| "grad_norm": 0.522744432939167, | |
| "learning_rate": 3.568087128216414e-06, | |
| "loss": 0.2699696123600006, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 1.849462365591398, | |
| "grad_norm": 0.6740080747223203, | |
| "learning_rate": 3.511396473287611e-06, | |
| "loss": 0.3212928771972656, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 1.860215053763441, | |
| "grad_norm": 0.5393424030521861, | |
| "learning_rate": 3.4549150281252635e-06, | |
| "loss": 0.28264331817626953, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 1.870967741935484, | |
| "grad_norm": 0.5389635395816308, | |
| "learning_rate": 3.398650730685813e-06, | |
| "loss": 0.2982739508152008, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 1.881720430107527, | |
| "grad_norm": 0.5527956538218779, | |
| "learning_rate": 3.3426114884075488e-06, | |
| "loss": 0.2892792820930481, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 1.89247311827957, | |
| "grad_norm": 0.5783362380337232, | |
| "learning_rate": 3.2868051770992935e-06, | |
| "loss": 0.2867039442062378, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 1.903225806451613, | |
| "grad_norm": 0.4505107682619091, | |
| "learning_rate": 3.2312396398335312e-06, | |
| "loss": 0.2998119592666626, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 1.913978494623656, | |
| "grad_norm": 0.544159970587014, | |
| "learning_rate": 3.175922685844125e-06, | |
| "loss": 0.2946498394012451, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 1.924731182795699, | |
| "grad_norm": 0.5601933382330992, | |
| "learning_rate": 3.1208620894288105e-06, | |
| "loss": 0.30848872661590576, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 1.935483870967742, | |
| "grad_norm": 0.5692086999688832, | |
| "learning_rate": 3.0660655888565827e-06, | |
| "loss": 0.30292433500289917, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.946236559139785, | |
| "grad_norm": 0.5204759182695003, | |
| "learning_rate": 3.0115408852801535e-06, | |
| "loss": 0.2992181181907654, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 1.956989247311828, | |
| "grad_norm": 0.547974934732699, | |
| "learning_rate": 2.9572956416536267e-06, | |
| "loss": 0.3052282929420471, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 1.967741935483871, | |
| "grad_norm": 0.5036888640423901, | |
| "learning_rate": 2.9033374816555338e-06, | |
| "loss": 0.27619490027427673, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 1.978494623655914, | |
| "grad_norm": 0.5857097319940581, | |
| "learning_rate": 2.8496739886173994e-06, | |
| "loss": 0.2969588041305542, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 1.989247311827957, | |
| "grad_norm": 0.5125923920870736, | |
| "learning_rate": 2.7963127044579697e-06, | |
| "loss": 0.28236857056617737, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.5460296926482203, | |
| "learning_rate": 2.743261128623269e-06, | |
| "loss": 0.2802872657775879, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 2.010752688172043, | |
| "grad_norm": 0.5627765344162555, | |
| "learning_rate": 2.6905267170326143e-06, | |
| "loss": 0.2465861439704895, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 2.021505376344086, | |
| "grad_norm": 0.5241819547135816, | |
| "learning_rate": 2.6381168810307536e-06, | |
| "loss": 0.2593002915382385, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 2.032258064516129, | |
| "grad_norm": 0.5146155536713976, | |
| "learning_rate": 2.5860389863462765e-06, | |
| "loss": 0.22058740258216858, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 2.043010752688172, | |
| "grad_norm": 0.527929699946712, | |
| "learning_rate": 2.534300352056416e-06, | |
| "loss": 0.24380052089691162, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.053763440860215, | |
| "grad_norm": 0.5680981651809573, | |
| "learning_rate": 2.4829082495584244e-06, | |
| "loss": 0.2181762158870697, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 2.064516129032258, | |
| "grad_norm": 0.5210266779766828, | |
| "learning_rate": 2.4318699015476495e-06, | |
| "loss": 0.21483556926250458, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 2.075268817204301, | |
| "grad_norm": 0.5228849523676178, | |
| "learning_rate": 2.3811924810024385e-06, | |
| "loss": 0.2233523428440094, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 2.086021505376344, | |
| "grad_norm": 1.2592287942902403, | |
| "learning_rate": 2.330883110176049e-06, | |
| "loss": 0.21375712752342224, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 2.096774193548387, | |
| "grad_norm": 0.5541159570889329, | |
| "learning_rate": 2.2809488595956746e-06, | |
| "loss": 0.22336843609809875, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 2.10752688172043, | |
| "grad_norm": 0.6275490404272853, | |
| "learning_rate": 2.2313967470687593e-06, | |
| "loss": 0.23877255618572235, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 2.118279569892473, | |
| "grad_norm": 0.5287134027247969, | |
| "learning_rate": 2.18223373669669e-06, | |
| "loss": 0.23572032153606415, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 2.129032258064516, | |
| "grad_norm": 0.6005092115812523, | |
| "learning_rate": 2.1334667378960642e-06, | |
| "loss": 0.25271525979042053, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 2.139784946236559, | |
| "grad_norm": 0.5729023596023511, | |
| "learning_rate": 2.0851026044276405e-06, | |
| "loss": 0.2403365820646286, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 2.150537634408602, | |
| "grad_norm": 0.502960648659189, | |
| "learning_rate": 2.0371481334330913e-06, | |
| "loss": 0.22194445133209229, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 2.161290322580645, | |
| "grad_norm": 1.0792975178234447, | |
| "learning_rate": 1.9896100644797316e-06, | |
| "loss": 0.20418012142181396, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 2.172043010752688, | |
| "grad_norm": 0.563900355756192, | |
| "learning_rate": 1.9424950786133414e-06, | |
| "loss": 0.2197030484676361, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 2.182795698924731, | |
| "grad_norm": 0.6049813205320395, | |
| "learning_rate": 1.8958097974191909e-06, | |
| "loss": 0.23065771162509918, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 2.193548387096774, | |
| "grad_norm": 0.5693258359321052, | |
| "learning_rate": 1.8495607820914451e-06, | |
| "loss": 0.22469733655452728, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 2.204301075268817, | |
| "grad_norm": 0.49687778546847566, | |
| "learning_rate": 1.8037545325110506e-06, | |
| "loss": 0.22447475790977478, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 2.21505376344086, | |
| "grad_norm": 0.6870685608162285, | |
| "learning_rate": 1.7583974863322272e-06, | |
| "loss": 0.2232748568058014, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 2.225806451612903, | |
| "grad_norm": 0.5587791759635568, | |
| "learning_rate": 1.7134960180777171e-06, | |
| "loss": 0.21617698669433594, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 2.236559139784946, | |
| "grad_norm": 0.5406726897347297, | |
| "learning_rate": 1.6690564382429104e-06, | |
| "loss": 0.22749440371990204, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 2.247311827956989, | |
| "grad_norm": 0.6059659457413509, | |
| "learning_rate": 1.6250849924089485e-06, | |
| "loss": 0.24092203378677368, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 2.258064516129032, | |
| "grad_norm": 0.5440152722934113, | |
| "learning_rate": 1.581587860364977e-06, | |
| "loss": 0.20948372781276703, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 2.268817204301075, | |
| "grad_norm": 0.48547134095880384, | |
| "learning_rate": 1.5385711552396227e-06, | |
| "loss": 0.21325388550758362, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 2.279569892473118, | |
| "grad_norm": 0.6324122951694761, | |
| "learning_rate": 1.4960409226418576e-06, | |
| "loss": 0.22965453565120697, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 2.2903225806451615, | |
| "grad_norm": 0.5469586332098659, | |
| "learning_rate": 1.4540031398113335e-06, | |
| "loss": 0.22315660119056702, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 2.3010752688172045, | |
| "grad_norm": 0.7520121411441703, | |
| "learning_rate": 1.4124637147783431e-06, | |
| "loss": 0.23476293683052063, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 2.3118279569892475, | |
| "grad_norm": 0.535533468892341, | |
| "learning_rate": 1.371428485533498e-06, | |
| "loss": 0.21404291689395905, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 2.3225806451612905, | |
| "grad_norm": 0.5561238179269572, | |
| "learning_rate": 1.3309032192072463e-06, | |
| "loss": 0.23378810286521912, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 2.3333333333333335, | |
| "grad_norm": 0.5663331713554731, | |
| "learning_rate": 1.29089361125936e-06, | |
| "loss": 0.2105513960123062, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 2.3440860215053765, | |
| "grad_norm": 0.5851084790781899, | |
| "learning_rate": 1.251405284678488e-06, | |
| "loss": 0.22501936554908752, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 2.3548387096774195, | |
| "grad_norm": 0.5546173616985522, | |
| "learning_rate": 1.2124437891918995e-06, | |
| "loss": 0.21492747962474823, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 2.3655913978494625, | |
| "grad_norm": 0.5542629930888266, | |
| "learning_rate": 1.1740146004855141e-06, | |
| "loss": 0.22757509350776672, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 2.3763440860215055, | |
| "grad_norm": 0.5368948762143404, | |
| "learning_rate": 1.1361231194343436e-06, | |
| "loss": 0.20776845514774323, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 2.3870967741935485, | |
| "grad_norm": 0.5362067849303692, | |
| "learning_rate": 1.0987746713434578e-06, | |
| "loss": 0.21643775701522827, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 2.3978494623655915, | |
| "grad_norm": 0.7478981160056104, | |
| "learning_rate": 1.0619745051995473e-06, | |
| "loss": 0.2321006953716278, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 2.4086021505376345, | |
| "grad_norm": 0.5589077284378241, | |
| "learning_rate": 1.0257277929332332e-06, | |
| "loss": 0.2335091233253479, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 2.4193548387096775, | |
| "grad_norm": 0.5449990060855128, | |
| "learning_rate": 9.900396286922025e-07, | |
| "loss": 0.22067061066627502, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 2.4301075268817205, | |
| "grad_norm": 0.545790815861183, | |
| "learning_rate": 9.549150281252633e-07, | |
| "loss": 0.24003136157989502, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 2.4408602150537635, | |
| "grad_norm": 0.5197037503280709, | |
| "learning_rate": 9.203589276774438e-07, | |
| "loss": 0.2232522964477539, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 2.4516129032258065, | |
| "grad_norm": 0.5783224759821376, | |
| "learning_rate": 8.86376183896226e-07, | |
| "loss": 0.2168661653995514, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 2.4623655913978495, | |
| "grad_norm": 0.5772298663365216, | |
| "learning_rate": 8.529715727489912e-07, | |
| "loss": 0.22855976223945618, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 2.4731182795698925, | |
| "grad_norm": 0.5901747678986362, | |
| "learning_rate": 8.201497889518073e-07, | |
| "loss": 0.23543155193328857, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.4838709677419355, | |
| "grad_norm": 0.4859350321815666, | |
| "learning_rate": 7.879154453096305e-07, | |
| "loss": 0.20444487035274506, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 2.4946236559139785, | |
| "grad_norm": 0.5733816209081586, | |
| "learning_rate": 7.562730720680111e-07, | |
| "loss": 0.22317257523536682, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 2.5053763440860215, | |
| "grad_norm": 0.6158850373560834, | |
| "learning_rate": 7.25227116276413e-07, | |
| "loss": 0.21304547786712646, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 2.5161290322580645, | |
| "grad_norm": 0.5303929149495401, | |
| "learning_rate": 6.947819411632223e-07, | |
| "loss": 0.22398605942726135, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 2.5268817204301075, | |
| "grad_norm": 0.5807287890882401, | |
| "learning_rate": 6.649418255225298e-07, | |
| "loss": 0.22487413883209229, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 2.5376344086021505, | |
| "grad_norm": 0.5151341689499322, | |
| "learning_rate": 6.357109631127889e-07, | |
| "loss": 0.20791815221309662, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 2.5483870967741935, | |
| "grad_norm": 0.5723989318815612, | |
| "learning_rate": 6.07093462067419e-07, | |
| "loss": 0.22853252291679382, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 2.5591397849462365, | |
| "grad_norm": 0.5151529413040788, | |
| "learning_rate": 5.79093344317449e-07, | |
| "loss": 0.23513737320899963, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 2.5698924731182795, | |
| "grad_norm": 0.6125760680997326, | |
| "learning_rate": 5.517145450262639e-07, | |
| "loss": 0.23416566848754883, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 2.5806451612903225, | |
| "grad_norm": 0.5169337829356161, | |
| "learning_rate": 5.249609120365579e-07, | |
| "loss": 0.2236918807029724, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 2.5913978494623655, | |
| "grad_norm": 0.5150874552544314, | |
| "learning_rate": 4.988362053295564e-07, | |
| "loss": 0.22128576040267944, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 2.6021505376344085, | |
| "grad_norm": 0.6254184464054292, | |
| "learning_rate": 4.733440964965791e-07, | |
| "loss": 0.23455382883548737, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 2.6129032258064515, | |
| "grad_norm": 0.5612944471466359, | |
| "learning_rate": 4.484881682230341e-07, | |
| "loss": 0.22176575660705566, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 2.6236559139784945, | |
| "grad_norm": 0.5178079276744024, | |
| "learning_rate": 4.242719137849077e-07, | |
| "loss": 0.21548646688461304, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 2.6344086021505375, | |
| "grad_norm": 0.5659047109155118, | |
| "learning_rate": 4.00698736557808e-07, | |
| "loss": 0.23057113587856293, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 2.6451612903225805, | |
| "grad_norm": 0.5098023427109231, | |
| "learning_rate": 3.777719495386567e-07, | |
| "loss": 0.2156354784965515, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 2.6559139784946235, | |
| "grad_norm": 0.49515330466811, | |
| "learning_rate": 3.5549477488007853e-07, | |
| "loss": 0.19704020023345947, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 2.6666666666666665, | |
| "grad_norm": 0.5297791040512543, | |
| "learning_rate": 3.3387034343755063e-07, | |
| "loss": 0.21541355550289154, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 2.6774193548387095, | |
| "grad_norm": 0.5134148705186912, | |
| "learning_rate": 3.1290169432939556e-07, | |
| "loss": 0.20781481266021729, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 2.688172043010753, | |
| "grad_norm": 0.6591675869543023, | |
| "learning_rate": 2.925917745096568e-07, | |
| "loss": 0.23452350497245789, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.698924731182796, | |
| "grad_norm": 0.5940779586998381, | |
| "learning_rate": 2.7294343835393366e-07, | |
| "loss": 0.23721283674240112, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 2.709677419354839, | |
| "grad_norm": 0.6904575634426214, | |
| "learning_rate": 2.539594472582213e-07, | |
| "loss": 0.23339498043060303, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 2.720430107526882, | |
| "grad_norm": 0.549700857643845, | |
| "learning_rate": 2.3564246925082358e-07, | |
| "loss": 0.2422715723514557, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 2.731182795698925, | |
| "grad_norm": 0.551951066945312, | |
| "learning_rate": 2.179950786173879e-07, | |
| "loss": 0.22343403100967407, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 2.741935483870968, | |
| "grad_norm": 0.5112230065122543, | |
| "learning_rate": 2.01019755539108e-07, | |
| "loss": 0.2022184431552887, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 2.752688172043011, | |
| "grad_norm": 0.6011326795775324, | |
| "learning_rate": 1.8471888574415953e-07, | |
| "loss": 0.21843938529491425, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 2.763440860215054, | |
| "grad_norm": 0.5598013580182175, | |
| "learning_rate": 1.690947601724091e-07, | |
| "loss": 0.2313247174024582, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 2.774193548387097, | |
| "grad_norm": 0.579260801737671, | |
| "learning_rate": 1.5414957465343883e-07, | |
| "loss": 0.23432737588882446, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 2.78494623655914, | |
| "grad_norm": 0.6080486926763811, | |
| "learning_rate": 1.3988542959794627e-07, | |
| "loss": 0.2501685619354248, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 2.795698924731183, | |
| "grad_norm": 0.6231912247704571, | |
| "learning_rate": 1.2630432970255014e-07, | |
| "loss": 0.22736617922782898, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 2.806451612903226, | |
| "grad_norm": 0.6526633675286432, | |
| "learning_rate": 1.1340818366804728e-07, | |
| "loss": 0.2217504233121872, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 2.817204301075269, | |
| "grad_norm": 0.5774442194442434, | |
| "learning_rate": 1.0119880393116177e-07, | |
| "loss": 0.21880705654621124, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 2.827956989247312, | |
| "grad_norm": 0.5510924764674725, | |
| "learning_rate": 8.967790640982466e-08, | |
| "loss": 0.23220205307006836, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 2.838709677419355, | |
| "grad_norm": 0.5940155308333281, | |
| "learning_rate": 7.884711026201586e-08, | |
| "loss": 0.22368629276752472, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 2.849462365591398, | |
| "grad_norm": 0.48018958702032566, | |
| "learning_rate": 6.870793765820783e-08, | |
| "loss": 0.2191024124622345, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 2.860215053763441, | |
| "grad_norm": 0.5348968055346366, | |
| "learning_rate": 5.92618135674361e-08, | |
| "loss": 0.2141711413860321, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 2.870967741935484, | |
| "grad_norm": 0.5058952998875674, | |
| "learning_rate": 5.0510065557034526e-08, | |
| "loss": 0.21273520588874817, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 2.881720430107527, | |
| "grad_norm": 0.5748923929140294, | |
| "learning_rate": 4.245392360605727e-08, | |
| "loss": 0.2339775264263153, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 2.89247311827957, | |
| "grad_norm": 0.543587597117276, | |
| "learning_rate": 3.5094519932415417e-08, | |
| "loss": 0.22123411297798157, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 2.903225806451613, | |
| "grad_norm": 0.5451715623802244, | |
| "learning_rate": 2.843288883375539e-08, | |
| "loss": 0.22225311398506165, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.913978494623656, | |
| "grad_norm": 0.5639808086322913, | |
| "learning_rate": 2.2469966542096323e-08, | |
| "loss": 0.220473051071167, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 2.924731182795699, | |
| "grad_norm": 0.5603998497829358, | |
| "learning_rate": 1.7206591092253642e-08, | |
| "loss": 0.22855965793132782, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 2.935483870967742, | |
| "grad_norm": 0.5700123425140283, | |
| "learning_rate": 1.264350220405719e-08, | |
| "loss": 0.226323202252388, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 2.946236559139785, | |
| "grad_norm": 0.5193108167577539, | |
| "learning_rate": 8.781341178393244e-09, | |
| "loss": 0.21938025951385498, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 2.956989247311828, | |
| "grad_norm": 0.6419513279403505, | |
| "learning_rate": 5.620650807073857e-09, | |
| "loss": 0.24610912799835205, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 2.967741935483871, | |
| "grad_norm": 0.6308682159730912, | |
| "learning_rate": 3.1618752965534295e-09, | |
| "loss": 0.22690746188163757, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 2.978494623655914, | |
| "grad_norm": 0.5263395659056124, | |
| "learning_rate": 1.4053602054991954e-09, | |
| "loss": 0.20472858846187592, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 2.989247311827957, | |
| "grad_norm": 0.5329091838420249, | |
| "learning_rate": 3.513523962256349e-10, | |
| "loss": 0.2296607792377472, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 0.6481688008102051, | |
| "learning_rate": 0.0, | |
| "loss": 0.22974838316440582, | |
| "step": 279 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 279, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 142925655572480.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |