| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 200.0, |
| "global_step": 5205, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.00019212295869356388, |
| "grad_norm": 27.250099182128906, |
| "learning_rate": 1.1494252873563217e-06, |
| "loss": 12.760068893432617, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.0009606147934678194, |
| "grad_norm": 27.193613052368164, |
| "learning_rate": 5.747126436781608e-06, |
| "loss": 12.752052307128906, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.0019212295869356388, |
| "grad_norm": 23.87885093688965, |
| "learning_rate": 1.1494252873563217e-05, |
| "loss": 12.383438110351562, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.002881844380403458, |
| "grad_norm": 15.234784126281738, |
| "learning_rate": 1.7241379310344825e-05, |
| "loss": 11.199150085449219, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.0038424591738712775, |
| "grad_norm": 15.339820861816406, |
| "learning_rate": 2.2988505747126433e-05, |
| "loss": 9.962273406982423, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.004803073967339097, |
| "grad_norm": 7.6476054191589355, |
| "learning_rate": 2.8735632183908045e-05, |
| "loss": 8.961764526367187, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.005763688760806916, |
| "grad_norm": 5.635213851928711, |
| "learning_rate": 3.448275862068965e-05, |
| "loss": 8.127005004882813, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.0067243035542747355, |
| "grad_norm": 4.58963680267334, |
| "learning_rate": 4.022988505747126e-05, |
| "loss": 7.473291015625, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.007684918347742555, |
| "grad_norm": 2.4760098457336426, |
| "learning_rate": 4.5977011494252866e-05, |
| "loss": 6.995676422119141, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.008645533141210375, |
| "grad_norm": 1.5408024787902832, |
| "learning_rate": 5.172413793103448e-05, |
| "loss": 6.636921691894531, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.009606147934678195, |
| "grad_norm": 1.112014651298523, |
| "learning_rate": 5.747126436781609e-05, |
| "loss": 6.349236297607422, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.010566762728146013, |
| "grad_norm": 0.9658083319664001, |
| "learning_rate": 6.32183908045977e-05, |
| "loss": 6.117427825927734, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.011527377521613832, |
| "grad_norm": 1.8757588863372803, |
| "learning_rate": 6.89655172413793e-05, |
| "loss": 5.927629852294922, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.012487992315081652, |
| "grad_norm": 1.7964160442352295, |
| "learning_rate": 7.471264367816091e-05, |
| "loss": 5.770274353027344, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.013448607108549471, |
| "grad_norm": 2.1291167736053467, |
| "learning_rate": 8.045977011494252e-05, |
| "loss": 5.6375282287597654, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.01440922190201729, |
| "grad_norm": 3.0542609691619873, |
| "learning_rate": 8.620689655172413e-05, |
| "loss": 5.525964736938477, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.01536983669548511, |
| "grad_norm": 3.1441893577575684, |
| "learning_rate": 9.195402298850573e-05, |
| "loss": 5.427687072753907, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.01633045148895293, |
| "grad_norm": 1.3293317556381226, |
| "learning_rate": 9.770114942528733e-05, |
| "loss": 5.335831832885742, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.01729106628242075, |
| "grad_norm": 2.560171127319336, |
| "learning_rate": 0.00010344827586206896, |
| "loss": 5.258418273925781, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.01825168107588857, |
| "grad_norm": 2.012646436691284, |
| "learning_rate": 0.00010919540229885056, |
| "loss": 5.183747863769531, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.01921229586935639, |
| "grad_norm": 2.6587507724761963, |
| "learning_rate": 0.00011494252873563218, |
| "loss": 5.119210052490234, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.020172910662824207, |
| "grad_norm": 2.3047592639923096, |
| "learning_rate": 0.00012068965517241378, |
| "loss": 5.057864761352539, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.021133525456292025, |
| "grad_norm": 2.4375789165496826, |
| "learning_rate": 0.0001264367816091954, |
| "loss": 5.000635147094727, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.022094140249759846, |
| "grad_norm": 2.1402318477630615, |
| "learning_rate": 0.000132183908045977, |
| "loss": 4.947456359863281, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.023054755043227664, |
| "grad_norm": 2.815868377685547, |
| "learning_rate": 0.0001379310344827586, |
| "loss": 4.898051071166992, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.024015369836695485, |
| "grad_norm": 2.3648743629455566, |
| "learning_rate": 0.0001436781609195402, |
| "loss": 4.855741119384765, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.024975984630163303, |
| "grad_norm": 2.230957508087158, |
| "learning_rate": 0.00014942528735632183, |
| "loss": 4.808645629882813, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.025936599423631124, |
| "grad_norm": 1.3839224576950073, |
| "learning_rate": 0.00015517241379310346, |
| "loss": 4.757843780517578, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.026897214217098942, |
| "grad_norm": 1.4279674291610718, |
| "learning_rate": 0.00016091954022988503, |
| "loss": 4.719438171386718, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.027857829010566763, |
| "grad_norm": 1.4759711027145386, |
| "learning_rate": 0.00016666666666666666, |
| "loss": 4.6750732421875, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.02881844380403458, |
| "grad_norm": 1.3809517621994019, |
| "learning_rate": 0.00017241379310344826, |
| "loss": 4.63501091003418, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.029779058597502402, |
| "grad_norm": 1.6614536046981812, |
| "learning_rate": 0.00017816091954022986, |
| "loss": 4.603597640991211, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.03073967339097022, |
| "grad_norm": 2.904686689376831, |
| "learning_rate": 0.00018390804597701147, |
| "loss": 4.574514770507813, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.03170028818443804, |
| "grad_norm": 1.8000333309173584, |
| "learning_rate": 0.0001896551724137931, |
| "loss": 4.542916107177734, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.03266090297790586, |
| "grad_norm": 2.0026073455810547, |
| "learning_rate": 0.00019540229885057467, |
| "loss": 4.509420013427734, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.03362151777137368, |
| "grad_norm": 1.6088820695877075, |
| "learning_rate": 0.0002011494252873563, |
| "loss": 4.4883583068847654, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.0345821325648415, |
| "grad_norm": 3.137084722518921, |
| "learning_rate": 0.00020689655172413793, |
| "loss": 4.462603759765625, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.03554274735830932, |
| "grad_norm": 1.9390242099761963, |
| "learning_rate": 0.00021264367816091953, |
| "loss": 4.441267395019532, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.03650336215177714, |
| "grad_norm": 2.0831611156463623, |
| "learning_rate": 0.00021839080459770113, |
| "loss": 4.420868682861328, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.037463976945244955, |
| "grad_norm": 1.4936137199401855, |
| "learning_rate": 0.00022413793103448273, |
| "loss": 4.390471649169922, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.03842459173871278, |
| "grad_norm": 2.1609437465667725, |
| "learning_rate": 0.00022988505747126436, |
| "loss": 4.372611999511719, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.0393852065321806, |
| "grad_norm": 1.5894837379455566, |
| "learning_rate": 0.00023563218390804593, |
| "loss": 4.3513038635253904, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.040345821325648415, |
| "grad_norm": 2.7555136680603027, |
| "learning_rate": 0.00024137931034482756, |
| "loss": 4.324166870117187, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.04130643611911623, |
| "grad_norm": 1.7152142524719238, |
| "learning_rate": 0.00024712643678160916, |
| "loss": 4.298264694213867, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.04226705091258405, |
| "grad_norm": 2.7435386180877686, |
| "learning_rate": 0.0002528735632183908, |
| "loss": 4.284121322631836, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.043227665706051875, |
| "grad_norm": 1.9910717010498047, |
| "learning_rate": 0.00025862068965517237, |
| "loss": 4.270275497436524, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.04418828049951969, |
| "grad_norm": 1.8952627182006836, |
| "learning_rate": 0.000264367816091954, |
| "loss": 4.251060867309571, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.04514889529298751, |
| "grad_norm": 1.4182472229003906, |
| "learning_rate": 0.0002701149425287356, |
| "loss": 4.223833465576172, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.04610951008645533, |
| "grad_norm": 2.826347589492798, |
| "learning_rate": 0.0002758620689655172, |
| "loss": 4.208561706542969, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.04707012487992315, |
| "grad_norm": 2.5493690967559814, |
| "learning_rate": 0.00028160919540229883, |
| "loss": 4.194873428344726, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.04803073967339097, |
| "grad_norm": 2.2824618816375732, |
| "learning_rate": 0.0002873563218390804, |
| "loss": 4.167812347412109, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.04899135446685879, |
| "grad_norm": 1.4718077182769775, |
| "learning_rate": 0.00029310344827586203, |
| "loss": 4.150156402587891, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.049951969260326606, |
| "grad_norm": 2.2785093784332275, |
| "learning_rate": 0.00029885057471264366, |
| "loss": 4.126807403564453, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.05091258405379443, |
| "grad_norm": 1.5848008394241333, |
| "learning_rate": 0.00029999951546647263, |
| "loss": 4.111884689331054, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.05187319884726225, |
| "grad_norm": 1.6441746950149536, |
| "learning_rate": 0.0002999975470543828, |
| "loss": 4.089701843261719, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.052833813640730067, |
| "grad_norm": 1.9900462627410889, |
| "learning_rate": 0.00029999406450023966, |
| "loss": 4.0772346496582035, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.053794428434197884, |
| "grad_norm": 1.938680648803711, |
| "learning_rate": 0.0002999890678391978, |
| "loss": 4.0556285858154295, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.05475504322766571, |
| "grad_norm": 1.6333754062652588, |
| "learning_rate": 0.00029998255712169563, |
| "loss": 4.044793701171875, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.05571565802113353, |
| "grad_norm": 1.8412508964538574, |
| "learning_rate": 0.00029997453241345533, |
| "loss": 4.03118896484375, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.056676272814601344, |
| "grad_norm": 1.4252992868423462, |
| "learning_rate": 0.0002999649937954818, |
| "loss": 4.01298713684082, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.05763688760806916, |
| "grad_norm": 2.278615951538086, |
| "learning_rate": 0.0002999539413640621, |
| "loss": 3.9946434020996096, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.05859750240153699, |
| "grad_norm": 1.5454790592193604, |
| "learning_rate": 0.0002999413752307644, |
| "loss": 3.9831748962402345, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.059558117195004805, |
| "grad_norm": 1.334790587425232, |
| "learning_rate": 0.0002999272955224369, |
| "loss": 3.9761894226074217, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.06051873198847262, |
| "grad_norm": 1.9382845163345337, |
| "learning_rate": 0.0002999117023812064, |
| "loss": 3.9631324768066407, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.06147934678194044, |
| "grad_norm": 2.01582932472229, |
| "learning_rate": 0.000299894595964477, |
| "loss": 3.9521259307861327, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.06243996157540826, |
| "grad_norm": 1.8427826166152954, |
| "learning_rate": 0.0002998759764449286, |
| "loss": 3.9423927307128905, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.06340057636887608, |
| "grad_norm": 1.6123930215835571, |
| "learning_rate": 0.0002998558440105148, |
| "loss": 3.935696029663086, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.0643611911623439, |
| "grad_norm": 1.9502496719360352, |
| "learning_rate": 0.0002998341988644614, |
| "loss": 3.923968505859375, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.06532180595581172, |
| "grad_norm": 1.2648119926452637, |
| "learning_rate": 0.0002998110412252641, |
| "loss": 3.9131423950195314, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.06628242074927954, |
| "grad_norm": 1.5318907499313354, |
| "learning_rate": 0.0002997863713266866, |
| "loss": 3.9074745178222656, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.06724303554274735, |
| "grad_norm": 1.3065940141677856, |
| "learning_rate": 0.0002997601894177576, |
| "loss": 3.8965118408203123, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.06820365033621517, |
| "grad_norm": 1.2381218671798706, |
| "learning_rate": 0.00029973249576276914, |
| "loss": 3.890979766845703, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.069164265129683, |
| "grad_norm": 1.5115997791290283, |
| "learning_rate": 0.0002997032906412732, |
| "loss": 3.8882400512695314, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.07012487992315082, |
| "grad_norm": 1.6171698570251465, |
| "learning_rate": 0.0002996725743480793, |
| "loss": 3.876806640625, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.07108549471661864, |
| "grad_norm": 1.0753288269042969, |
| "learning_rate": 0.00029964034719325147, |
| "loss": 3.860057067871094, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.07204610951008646, |
| "grad_norm": 1.4012713432312012, |
| "learning_rate": 0.0002996066095021048, |
| "loss": 3.857683563232422, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.07300672430355427, |
| "grad_norm": 1.765068769454956, |
| "learning_rate": 0.0002995713616152028, |
| "loss": 3.8544296264648437, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.07396733909702209, |
| "grad_norm": 1.5474210977554321, |
| "learning_rate": 0.0002995346038883532, |
| "loss": 3.849739837646484, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.07492795389048991, |
| "grad_norm": 1.443384051322937, |
| "learning_rate": 0.0002994963366926048, |
| "loss": 3.845722961425781, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.07588856868395773, |
| "grad_norm": 1.4993858337402344, |
| "learning_rate": 0.0002994565604142439, |
| "loss": 3.8389007568359377, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.07684918347742556, |
| "grad_norm": 1.4684888124465942, |
| "learning_rate": 0.00029941527545478976, |
| "loss": 3.8267383575439453, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.07780979827089338, |
| "grad_norm": 1.2325655221939087, |
| "learning_rate": 0.00029937248223099136, |
| "loss": 3.818804168701172, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.0787704130643612, |
| "grad_norm": 1.0254677534103394, |
| "learning_rate": 0.00029932818117482245, |
| "loss": 3.812041473388672, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.07973102785782901, |
| "grad_norm": 1.3646955490112305, |
| "learning_rate": 0.0002992823727334776, |
| "loss": 3.810979461669922, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.08069164265129683, |
| "grad_norm": 1.6398770809173584, |
| "learning_rate": 0.00029923505736936774, |
| "loss": 3.8059127807617186, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.08165225744476465, |
| "grad_norm": 1.2734476327896118, |
| "learning_rate": 0.0002991862355601151, |
| "loss": 3.798663330078125, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.08261287223823247, |
| "grad_norm": 1.4652912616729736, |
| "learning_rate": 0.00029913590779854886, |
| "loss": 3.7902076721191404, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.08357348703170028, |
| "grad_norm": 1.0962094068527222, |
| "learning_rate": 0.00029908407459269977, |
| "loss": 3.786347198486328, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.0845341018251681, |
| "grad_norm": 1.054427981376648, |
| "learning_rate": 0.0002990307364657954, |
| "loss": 3.780842590332031, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.08549471661863593, |
| "grad_norm": 1.2044267654418945, |
| "learning_rate": 0.0002989758939562545, |
| "loss": 3.7756649017333985, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.08645533141210375, |
| "grad_norm": 1.3676034212112427, |
| "learning_rate": 0.0002989195476176818, |
| "loss": 3.7702369689941406, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.08741594620557157, |
| "grad_norm": 1.455597996711731, |
| "learning_rate": 0.00029886169801886237, |
| "loss": 3.76458740234375, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.08837656099903939, |
| "grad_norm": 1.4684407711029053, |
| "learning_rate": 0.00029880234574375576, |
| "loss": 3.7579875946044923, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.0893371757925072, |
| "grad_norm": 1.4192808866500854, |
| "learning_rate": 0.00029874149139149037, |
| "loss": 3.757953643798828, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.09029779058597502, |
| "grad_norm": 0.8747047781944275, |
| "learning_rate": 0.00029867913557635704, |
| "loss": 3.7510101318359377, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.09125840537944284, |
| "grad_norm": 1.4938651323318481, |
| "learning_rate": 0.0002986152789278031, |
| "loss": 3.746034622192383, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.09221902017291066, |
| "grad_norm": 1.2561233043670654, |
| "learning_rate": 0.00029854992209042626, |
| "loss": 3.743701171875, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.09317963496637849, |
| "grad_norm": 1.431540846824646, |
| "learning_rate": 0.0002984830657239673, |
| "loss": 3.736903762817383, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.0941402497598463, |
| "grad_norm": 1.0810083150863647, |
| "learning_rate": 0.00029841471050330424, |
| "loss": 3.732843017578125, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.09510086455331412, |
| "grad_norm": 0.9057182669639587, |
| "learning_rate": 0.00029834485711844515, |
| "loss": 3.7288368225097654, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.09606147934678194, |
| "grad_norm": 1.559844970703125, |
| "learning_rate": 0.00029827350627452116, |
| "loss": 3.72186393737793, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.09702209414024976, |
| "grad_norm": 1.3021472692489624, |
| "learning_rate": 0.00029820065869177954, |
| "loss": 3.7243560791015624, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.09798270893371758, |
| "grad_norm": 0.9313806891441345, |
| "learning_rate": 0.0002981263151055762, |
| "loss": 3.7176589965820312, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.0989433237271854, |
| "grad_norm": 1.2589117288589478, |
| "learning_rate": 0.0002980504762663683, |
| "loss": 3.712178039550781, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.09990393852065321, |
| "grad_norm": 1.3793448209762573, |
| "learning_rate": 0.0002979731429397071, |
| "loss": 3.707489013671875, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.10086455331412104, |
| "grad_norm": 1.6294807195663452, |
| "learning_rate": 0.0002978943159062295, |
| "loss": 3.7033920288085938, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.10182516810758886, |
| "grad_norm": 1.3319975137710571, |
| "learning_rate": 0.0002978139959616507, |
| "loss": 3.7044357299804687, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.10278578290105668, |
| "grad_norm": 1.1249921321868896, |
| "learning_rate": 0.00029773218391675594, |
| "loss": 3.6977405548095703, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.1037463976945245, |
| "grad_norm": 1.7312142848968506, |
| "learning_rate": 0.00029764888059739255, |
| "loss": 3.698072814941406, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.10470701248799232, |
| "grad_norm": 1.3000468015670776, |
| "learning_rate": 0.00029756408684446136, |
| "loss": 3.7032306671142576, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.10566762728146013, |
| "grad_norm": 1.2404688596725464, |
| "learning_rate": 0.0002974778035139081, |
| "loss": 3.693822479248047, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.10662824207492795, |
| "grad_norm": 0.9569908380508423, |
| "learning_rate": 0.00029739003147671536, |
| "loss": 3.6903202056884767, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.10758885686839577, |
| "grad_norm": 1.3814141750335693, |
| "learning_rate": 0.00029730077161889304, |
| "loss": 3.6874603271484374, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.10854947166186359, |
| "grad_norm": 1.3034733533859253, |
| "learning_rate": 0.00029721002484147, |
| "loss": 3.6825042724609376, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.10951008645533142, |
| "grad_norm": 1.6188613176345825, |
| "learning_rate": 0.00029711779206048454, |
| "loss": 3.6817798614501953, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.11047070124879924, |
| "grad_norm": 1.411993384361267, |
| "learning_rate": 0.0002970240742069755, |
| "loss": 3.677989959716797, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.11143131604226705, |
| "grad_norm": 1.6159188747406006, |
| "learning_rate": 0.0002969288722269726, |
| "loss": 3.673622894287109, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.11239193083573487, |
| "grad_norm": 1.2981064319610596, |
| "learning_rate": 0.000296832187081487, |
| "loss": 3.674951934814453, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.11335254562920269, |
| "grad_norm": 1.0490260124206543, |
| "learning_rate": 0.0002967340197465017, |
| "loss": 3.665152740478516, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.1143131604226705, |
| "grad_norm": 1.5031790733337402, |
| "learning_rate": 0.00029663437121296146, |
| "loss": 3.6685272216796876, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.11527377521613832, |
| "grad_norm": 1.6380650997161865, |
| "learning_rate": 0.000296533242486763, |
| "loss": 3.6649364471435546, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.11623439000960614, |
| "grad_norm": 1.1084094047546387, |
| "learning_rate": 0.0002964306345887447, |
| "loss": 3.658610153198242, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.11719500480307397, |
| "grad_norm": 1.454255223274231, |
| "learning_rate": 0.0002963265485546764, |
| "loss": 3.655583953857422, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.11815561959654179, |
| "grad_norm": 1.5477410554885864, |
| "learning_rate": 0.00029622098543524884, |
| "loss": 3.6575542449951173, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.11911623439000961, |
| "grad_norm": 1.0675349235534668, |
| "learning_rate": 0.00029611394629606324, |
| "loss": 3.6511856079101563, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.12007684918347743, |
| "grad_norm": 1.7654082775115967, |
| "learning_rate": 0.0002960054322176204, |
| "loss": 3.6485317230224608, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.12103746397694524, |
| "grad_norm": 1.3815727233886719, |
| "learning_rate": 0.0002958954442953096, |
| "loss": 3.644425964355469, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.12199807877041306, |
| "grad_norm": 1.0989855527877808, |
| "learning_rate": 0.000295783983639398, |
| "loss": 3.6440811157226562, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.12295869356388088, |
| "grad_norm": 1.3829193115234375, |
| "learning_rate": 0.00029567105137501916, |
| "loss": 3.640919876098633, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.1239193083573487, |
| "grad_norm": 1.3325438499450684, |
| "learning_rate": 0.00029555664864216156, |
| "loss": 3.6379947662353516, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.12487992315081652, |
| "grad_norm": 1.112707257270813, |
| "learning_rate": 0.00029544077659565747, |
| "loss": 3.636188507080078, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.12584053794428435, |
| "grad_norm": 1.1048657894134521, |
| "learning_rate": 0.0002953234364051708, |
| "loss": 3.637079620361328, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.12680115273775217, |
| "grad_norm": 1.0728777647018433, |
| "learning_rate": 0.00029520462925518575, |
| "loss": 3.6303642272949217, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.12776176753121998, |
| "grad_norm": 0.9134336113929749, |
| "learning_rate": 0.00029508435634499467, |
| "loss": 3.625618743896484, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.1287223823246878, |
| "grad_norm": 1.0948387384414673, |
| "learning_rate": 0.00029496261888868586, |
| "loss": 3.6273883819580077, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.12968299711815562, |
| "grad_norm": 1.1168876886367798, |
| "learning_rate": 0.0002948394181151314, |
| "loss": 3.6221595764160157, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.13064361191162344, |
| "grad_norm": 1.1925864219665527, |
| "learning_rate": 0.0002947147552679748, |
| "loss": 3.6182037353515626, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.13160422670509125, |
| "grad_norm": 1.2935006618499756, |
| "learning_rate": 0.00029458863160561837, |
| "loss": 3.6180152893066406, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.13256484149855907, |
| "grad_norm": 1.0551207065582275, |
| "learning_rate": 0.0002944610484012105, |
| "loss": 3.614363098144531, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.1335254562920269, |
| "grad_norm": 1.2916655540466309, |
| "learning_rate": 0.0002943320069426329, |
| "loss": 3.611812210083008, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.1344860710854947, |
| "grad_norm": 1.084269404411316, |
| "learning_rate": 0.00029420150853248756, |
| "loss": 3.609844970703125, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.13544668587896252, |
| "grad_norm": 1.6168688535690308, |
| "learning_rate": 0.0002940695544880836, |
| "loss": 3.6106407165527346, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.13640730067243034, |
| "grad_norm": 1.294932246208191, |
| "learning_rate": 0.0002939361461414238, |
| "loss": 3.607910919189453, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.1373679154658982, |
| "grad_norm": 0.9968737363815308, |
| "learning_rate": 0.0002938012848391915, |
| "loss": 3.6028270721435547, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.138328530259366, |
| "grad_norm": 1.143962025642395, |
| "learning_rate": 0.0002936649719427367, |
| "loss": 3.6049777984619142, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.13928914505283382, |
| "grad_norm": 1.4859670400619507, |
| "learning_rate": 0.00029352720882806267, |
| "loss": 3.6023681640625, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.14024975984630164, |
| "grad_norm": 1.383272647857666, |
| "learning_rate": 0.00029338799688581146, |
| "loss": 3.5986564636230467, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.14121037463976946, |
| "grad_norm": 1.3845287561416626, |
| "learning_rate": 0.00029324733752125054, |
| "loss": 3.597808074951172, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.14217098943323728, |
| "grad_norm": 1.5982863903045654, |
| "learning_rate": 0.0002931052321542581, |
| "loss": 3.5963520050048827, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.1431316042267051, |
| "grad_norm": 1.0727800130844116, |
| "learning_rate": 0.00029296168221930904, |
| "loss": 3.592487335205078, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.1440922190201729, |
| "grad_norm": 1.4507306814193726, |
| "learning_rate": 0.0002928166891654604, |
| "loss": 3.595014190673828, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.14505283381364073, |
| "grad_norm": 1.4457939863204956, |
| "learning_rate": 0.00029267025445633667, |
| "loss": 3.590290069580078, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.14601344860710855, |
| "grad_norm": 1.1515284776687622, |
| "learning_rate": 0.0002925223795701149, |
| "loss": 3.5872840881347656, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.14697406340057637, |
| "grad_norm": 1.0049299001693726, |
| "learning_rate": 0.00029237306599951007, |
| "loss": 3.5840923309326174, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.14793467819404418, |
| "grad_norm": 1.628124475479126, |
| "learning_rate": 0.00029222231525176005, |
| "loss": 3.5856204986572267, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.148895292987512, |
| "grad_norm": 1.5615901947021484, |
| "learning_rate": 0.0002920701288486099, |
| "loss": 3.5824256896972657, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.14985590778097982, |
| "grad_norm": 1.0828890800476074, |
| "learning_rate": 0.00029191650832629694, |
| "loss": 3.5833717346191407, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.15081652257444764, |
| "grad_norm": 0.8812053799629211, |
| "learning_rate": 0.00029176145523553517, |
| "loss": 3.581065368652344, |
| "step": 785 |
| }, |
| { |
| "epoch": 0.15177713736791545, |
| "grad_norm": 1.3092104196548462, |
| "learning_rate": 0.0002916049711414996, |
| "loss": 3.5770198822021486, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.15273775216138327, |
| "grad_norm": 1.4702849388122559, |
| "learning_rate": 0.00029144705762381036, |
| "loss": 3.5817501068115236, |
| "step": 795 |
| }, |
| { |
| "epoch": 0.15369836695485112, |
| "grad_norm": 1.110189437866211, |
| "learning_rate": 0.0002912877162765169, |
| "loss": 3.5730361938476562, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.15465898174831894, |
| "grad_norm": 0.9270662665367126, |
| "learning_rate": 0.00029112694870808155, |
| "loss": 3.5725852966308596, |
| "step": 805 |
| }, |
| { |
| "epoch": 0.15561959654178675, |
| "grad_norm": 1.3707542419433594, |
| "learning_rate": 0.00029096475654136395, |
| "loss": 3.5693943023681642, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.15658021133525457, |
| "grad_norm": 1.4934515953063965, |
| "learning_rate": 0.000290801141413604, |
| "loss": 3.565943145751953, |
| "step": 815 |
| }, |
| { |
| "epoch": 0.1575408261287224, |
| "grad_norm": 1.0972487926483154, |
| "learning_rate": 0.00029063610497640576, |
| "loss": 3.5643775939941404, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.1585014409221902, |
| "grad_norm": 1.2283886671066284, |
| "learning_rate": 0.0002904696488957204, |
| "loss": 3.5635589599609374, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.15946205571565802, |
| "grad_norm": 1.164687156677246, |
| "learning_rate": 0.0002903017748518298, |
| "loss": 3.555510711669922, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.16042267050912584, |
| "grad_norm": 0.9848024845123291, |
| "learning_rate": 0.0002901324845393294, |
| "loss": 3.5610126495361327, |
| "step": 835 |
| }, |
| { |
| "epoch": 0.16138328530259366, |
| "grad_norm": 1.4761382341384888, |
| "learning_rate": 0.00028996177966711097, |
| "loss": 3.5633079528808596, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.16234390009606148, |
| "grad_norm": 1.265584945678711, |
| "learning_rate": 0.0002897896619583455, |
| "loss": 3.5593791961669923, |
| "step": 845 |
| }, |
| { |
| "epoch": 0.1633045148895293, |
| "grad_norm": 1.2358146905899048, |
| "learning_rate": 0.0002896161331504659, |
| "loss": 3.5586807250976564, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.1642651296829971, |
| "grad_norm": 1.2075212001800537, |
| "learning_rate": 0.00028944119499514913, |
| "loss": 3.5556678771972656, |
| "step": 855 |
| }, |
| { |
| "epoch": 0.16522574447646493, |
| "grad_norm": 1.3434778451919556, |
| "learning_rate": 0.0002892648492582989, |
| "loss": 3.548540496826172, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.16618635926993275, |
| "grad_norm": 0.7819045186042786, |
| "learning_rate": 0.00028908709772002765, |
| "loss": 3.5495044708251955, |
| "step": 865 |
| }, |
| { |
| "epoch": 0.16714697406340057, |
| "grad_norm": 1.3979746103286743, |
| "learning_rate": 0.00028890794217463863, |
| "loss": 3.5515655517578124, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.16810758885686838, |
| "grad_norm": 1.0144678354263306, |
| "learning_rate": 0.0002887273844306076, |
| "loss": 3.547974395751953, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.1690682036503362, |
| "grad_norm": 0.8473827242851257, |
| "learning_rate": 0.00028854542631056494, |
| "loss": 3.543077087402344, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.17002881844380405, |
| "grad_norm": 0.9643390774726868, |
| "learning_rate": 0.0002883620696512769, |
| "loss": 3.546183776855469, |
| "step": 885 |
| }, |
| { |
| "epoch": 0.17098943323727187, |
| "grad_norm": 1.3916168212890625, |
| "learning_rate": 0.0002881773163036273, |
| "loss": 3.5453559875488283, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.17195004803073968, |
| "grad_norm": 0.9916486740112305, |
| "learning_rate": 0.00028799116813259875, |
| "loss": 3.543218994140625, |
| "step": 895 |
| }, |
| { |
| "epoch": 0.1729106628242075, |
| "grad_norm": 0.9322313666343689, |
| "learning_rate": 0.0002878036270172538, |
| "loss": 3.5352169036865235, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.17387127761767532, |
| "grad_norm": 1.0106613636016846, |
| "learning_rate": 0.000287614694850716, |
| "loss": 3.539842224121094, |
| "step": 905 |
| }, |
| { |
| "epoch": 0.17483189241114314, |
| "grad_norm": 0.9181408286094666, |
| "learning_rate": 0.00028742437354015073, |
| "loss": 3.53677978515625, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.17579250720461095, |
| "grad_norm": 1.0180490016937256, |
| "learning_rate": 0.0002872326650067462, |
| "loss": 3.536122131347656, |
| "step": 915 |
| }, |
| { |
| "epoch": 0.17675312199807877, |
| "grad_norm": 1.140866994857788, |
| "learning_rate": 0.00028703957118569363, |
| "loss": 3.5390419006347655, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.1777137367915466, |
| "grad_norm": 1.3825160264968872, |
| "learning_rate": 0.000286845094026168, |
| "loss": 3.534088897705078, |
| "step": 925 |
| }, |
| { |
| "epoch": 0.1786743515850144, |
| "grad_norm": 0.9901686310768127, |
| "learning_rate": 0.0002866492354913086, |
| "loss": 3.5334190368652343, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.17963496637848222, |
| "grad_norm": 1.3177112340927124, |
| "learning_rate": 0.0002864519975581986, |
| "loss": 3.531676483154297, |
| "step": 935 |
| }, |
| { |
| "epoch": 0.18059558117195004, |
| "grad_norm": 1.5046802759170532, |
| "learning_rate": 0.0002862533822178456, |
| "loss": 3.5300270080566407, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.18155619596541786, |
| "grad_norm": 0.8187685012817383, |
| "learning_rate": 0.00028605339147516113, |
| "loss": 3.527945709228516, |
| "step": 945 |
| }, |
| { |
| "epoch": 0.18251681075888568, |
| "grad_norm": 1.0654352903366089, |
| "learning_rate": 0.00028585202734894105, |
| "loss": 3.5284492492675783, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.1834774255523535, |
| "grad_norm": 1.0617858171463013, |
| "learning_rate": 0.00028564929187184447, |
| "loss": 3.526523208618164, |
| "step": 955 |
| }, |
| { |
| "epoch": 0.1844380403458213, |
| "grad_norm": 0.9183857440948486, |
| "learning_rate": 0.00028544518709037363, |
| "loss": 3.5269630432128904, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.18539865513928913, |
| "grad_norm": 1.3353052139282227, |
| "learning_rate": 0.000285239715064853, |
| "loss": 3.5191162109375, |
| "step": 965 |
| }, |
| { |
| "epoch": 0.18635926993275698, |
| "grad_norm": 1.1203348636627197, |
| "learning_rate": 0.0002850328778694088, |
| "loss": 3.5222145080566407, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.1873198847262248, |
| "grad_norm": 1.1545228958129883, |
| "learning_rate": 0.0002848246775919478, |
| "loss": 3.518872833251953, |
| "step": 975 |
| }, |
| { |
| "epoch": 0.1882804995196926, |
| "grad_norm": 0.9494602084159851, |
| "learning_rate": 0.0002846151163341364, |
| "loss": 3.519129180908203, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.18924111431316043, |
| "grad_norm": 0.8861366510391235, |
| "learning_rate": 0.0002844041962113792, |
| "loss": 3.517262268066406, |
| "step": 985 |
| }, |
| { |
| "epoch": 0.19020172910662825, |
| "grad_norm": 1.097005844116211, |
| "learning_rate": 0.00028419191935279793, |
| "loss": 3.5180564880371095, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.19116234390009607, |
| "grad_norm": 0.9788710474967957, |
| "learning_rate": 0.00028397828790120965, |
| "loss": 3.5166107177734376, |
| "step": 995 |
| }, |
| { |
| "epoch": 0.19212295869356388, |
| "grad_norm": 1.3244202136993408, |
| "learning_rate": 0.0002837633040131055, |
| "loss": 3.5119117736816405, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.1930835734870317, |
| "grad_norm": 1.2634823322296143, |
| "learning_rate": 0.00028354696985862865, |
| "loss": 3.5120399475097654, |
| "step": 1005 |
| }, |
| { |
| "epoch": 0.19404418828049952, |
| "grad_norm": 1.3096829652786255, |
| "learning_rate": 0.00028332928762155225, |
| "loss": 3.5108917236328123, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.19500480307396734, |
| "grad_norm": 1.0718635320663452, |
| "learning_rate": 0.0002831102594992579, |
| "loss": 3.5096744537353515, |
| "step": 1015 |
| }, |
| { |
| "epoch": 0.19596541786743515, |
| "grad_norm": 1.0316826105117798, |
| "learning_rate": 0.00028288988770271297, |
| "loss": 3.5067817687988283, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.19692603266090297, |
| "grad_norm": 1.7893810272216797, |
| "learning_rate": 0.00028266817445644855, |
| "loss": 3.5084625244140626, |
| "step": 1025 |
| }, |
| { |
| "epoch": 0.1978866474543708, |
| "grad_norm": 1.3311545848846436, |
| "learning_rate": 0.0002824451219985369, |
| "loss": 3.5091777801513673, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.1988472622478386, |
| "grad_norm": 1.0418428182601929, |
| "learning_rate": 0.0002822207325805688, |
| "loss": 3.5157501220703127, |
| "step": 1035 |
| }, |
| { |
| "epoch": 0.19980787704130643, |
| "grad_norm": 1.1201485395431519, |
| "learning_rate": 0.00028199500846763116, |
| "loss": 3.5064178466796876, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.20076849183477424, |
| "grad_norm": 1.0694836378097534, |
| "learning_rate": 0.0002817679519382836, |
| "loss": 3.507163238525391, |
| "step": 1045 |
| }, |
| { |
| "epoch": 0.2017291066282421, |
| "grad_norm": 1.2998982667922974, |
| "learning_rate": 0.0002815395652845359, |
| "loss": 3.5081203460693358, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.2026897214217099, |
| "grad_norm": 1.3336325883865356, |
| "learning_rate": 0.0002813098508118247, |
| "loss": 3.499872589111328, |
| "step": 1055 |
| }, |
| { |
| "epoch": 0.20365033621517772, |
| "grad_norm": 0.7667920589447021, |
| "learning_rate": 0.0002810788108389901, |
| "loss": 3.499583435058594, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.20461095100864554, |
| "grad_norm": 1.072025179862976, |
| "learning_rate": 0.0002808464476982526, |
| "loss": 3.495106506347656, |
| "step": 1065 |
| }, |
| { |
| "epoch": 0.20557156580211336, |
| "grad_norm": 0.964080274105072, |
| "learning_rate": 0.0002806127637351892, |
| "loss": 3.4985504150390625, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.20653218059558118, |
| "grad_norm": 1.0775015354156494, |
| "learning_rate": 0.00028037776130871, |
| "loss": 3.4976402282714845, |
| "step": 1075 |
| }, |
| { |
| "epoch": 0.207492795389049, |
| "grad_norm": 1.0506839752197266, |
| "learning_rate": 0.00028014144279103406, |
| "loss": 3.4984210968017577, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.2084534101825168, |
| "grad_norm": 0.8290379643440247, |
| "learning_rate": 0.0002799038105676658, |
| "loss": 3.494451141357422, |
| "step": 1085 |
| }, |
| { |
| "epoch": 0.20941402497598463, |
| "grad_norm": 1.1757789850234985, |
| "learning_rate": 0.00027966486703737066, |
| "loss": 3.4927711486816406, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.21037463976945245, |
| "grad_norm": 1.0375887155532837, |
| "learning_rate": 0.0002794246146121512, |
| "loss": 3.4952877044677733, |
| "step": 1095 |
| }, |
| { |
| "epoch": 0.21133525456292027, |
| "grad_norm": 0.9264093637466431, |
| "learning_rate": 0.0002791830557172224, |
| "loss": 3.49503173828125, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.21229586935638808, |
| "grad_norm": 0.9268721342086792, |
| "learning_rate": 0.00027894019279098726, |
| "loss": 3.4857433319091795, |
| "step": 1105 |
| }, |
| { |
| "epoch": 0.2132564841498559, |
| "grad_norm": 1.1226038932800293, |
| "learning_rate": 0.00027869602828501234, |
| "loss": 3.488873291015625, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.21421709894332372, |
| "grad_norm": 1.1734631061553955, |
| "learning_rate": 0.00027845056466400297, |
| "loss": 3.487217330932617, |
| "step": 1115 |
| }, |
| { |
| "epoch": 0.21517771373679154, |
| "grad_norm": 0.9604414701461792, |
| "learning_rate": 0.0002782038044057783, |
| "loss": 3.487082672119141, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.21613832853025935, |
| "grad_norm": 0.9780157208442688, |
| "learning_rate": 0.0002779557500012462, |
| "loss": 3.48626708984375, |
| "step": 1125 |
| }, |
| { |
| "epoch": 0.21709894332372717, |
| "grad_norm": 1.1477540731430054, |
| "learning_rate": 0.0002777064039543784, |
| "loss": 3.4851287841796874, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.21805955811719502, |
| "grad_norm": 1.1044639348983765, |
| "learning_rate": 0.00027745576878218496, |
| "loss": 3.483640670776367, |
| "step": 1135 |
| }, |
| { |
| "epoch": 0.21902017291066284, |
| "grad_norm": 0.8308424949645996, |
| "learning_rate": 0.0002772038470146888, |
| "loss": 3.4800315856933595, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.21998078770413065, |
| "grad_norm": 1.204115390777588, |
| "learning_rate": 0.0002769506411949007, |
| "loss": 3.481386184692383, |
| "step": 1145 |
| }, |
| { |
| "epoch": 0.22094140249759847, |
| "grad_norm": 1.0885634422302246, |
| "learning_rate": 0.00027669615387879284, |
| "loss": 3.4802738189697267, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.2219020172910663, |
| "grad_norm": 1.1829088926315308, |
| "learning_rate": 0.0002764403876352736, |
| "loss": 3.4813682556152346, |
| "step": 1155 |
| }, |
| { |
| "epoch": 0.2228626320845341, |
| "grad_norm": 1.4106764793395996, |
| "learning_rate": 0.0002761833450461613, |
| "loss": 3.4792640686035154, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.22382324687800192, |
| "grad_norm": 1.074035406112671, |
| "learning_rate": 0.0002759250287061583, |
| "loss": 3.478520965576172, |
| "step": 1165 |
| }, |
| { |
| "epoch": 0.22478386167146974, |
| "grad_norm": 1.211889624595642, |
| "learning_rate": 0.00027566544122282496, |
| "loss": 3.478863525390625, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.22574447646493756, |
| "grad_norm": 0.9881852865219116, |
| "learning_rate": 0.0002754045852165529, |
| "loss": 3.4755462646484374, |
| "step": 1175 |
| }, |
| { |
| "epoch": 0.22670509125840538, |
| "grad_norm": 1.0194916725158691, |
| "learning_rate": 0.00027514246332053876, |
| "loss": 3.476274108886719, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.2276657060518732, |
| "grad_norm": 1.0759530067443848, |
| "learning_rate": 0.0002748790781807577, |
| "loss": 3.4741195678710937, |
| "step": 1185 |
| }, |
| { |
| "epoch": 0.228626320845341, |
| "grad_norm": 1.181774377822876, |
| "learning_rate": 0.0002746144324559368, |
| "loss": 3.4723861694335936, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.22958693563880883, |
| "grad_norm": 0.9129538536071777, |
| "learning_rate": 0.00027434852881752774, |
| "loss": 3.472820281982422, |
| "step": 1195 |
| }, |
| { |
| "epoch": 0.23054755043227665, |
| "grad_norm": 1.01265287399292, |
| "learning_rate": 0.0002740813699496804, |
| "loss": 3.472024917602539, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.23150816522574447, |
| "grad_norm": 1.2006109952926636, |
| "learning_rate": 0.0002738129585492153, |
| "loss": 3.4669479370117187, |
| "step": 1205 |
| }, |
| { |
| "epoch": 0.23246878001921228, |
| "grad_norm": 0.8117277026176453, |
| "learning_rate": 0.0002735432973255967, |
| "loss": 3.46884765625, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.2334293948126801, |
| "grad_norm": 0.9207207560539246, |
| "learning_rate": 0.0002732723890009051, |
| "loss": 3.4688953399658202, |
| "step": 1215 |
| }, |
| { |
| "epoch": 0.23439000960614795, |
| "grad_norm": 1.1179691553115845, |
| "learning_rate": 0.00027300023630980985, |
| "loss": 3.472461700439453, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.23535062439961577, |
| "grad_norm": 0.971973180770874, |
| "learning_rate": 0.00027272684199954137, |
| "loss": 3.467519760131836, |
| "step": 1225 |
| }, |
| { |
| "epoch": 0.23631123919308358, |
| "grad_norm": 0.8591349720954895, |
| "learning_rate": 0.0002724522088298637, |
| "loss": 3.4663330078125, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.2372718539865514, |
| "grad_norm": 1.1340241432189941, |
| "learning_rate": 0.0002721763395730462, |
| "loss": 3.466988372802734, |
| "step": 1235 |
| }, |
| { |
| "epoch": 0.23823246878001922, |
| "grad_norm": 0.7911289930343628, |
| "learning_rate": 0.00027189923701383627, |
| "loss": 3.463814544677734, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.23919308357348704, |
| "grad_norm": 1.2324531078338623, |
| "learning_rate": 0.0002716209039494304, |
| "loss": 3.46383056640625, |
| "step": 1245 |
| }, |
| { |
| "epoch": 0.24015369836695485, |
| "grad_norm": 1.0037308931350708, |
| "learning_rate": 0.0002713413431894466, |
| "loss": 3.4620521545410154, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.24111431316042267, |
| "grad_norm": 1.3981319665908813, |
| "learning_rate": 0.00027106055755589566, |
| "loss": 3.4639919281005858, |
| "step": 1255 |
| }, |
| { |
| "epoch": 0.2420749279538905, |
| "grad_norm": 1.0288619995117188, |
| "learning_rate": 0.00027077854988315285, |
| "loss": 3.4653533935546874, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.2430355427473583, |
| "grad_norm": 1.0437750816345215, |
| "learning_rate": 0.00027049532301792924, |
| "loss": 3.4617347717285156, |
| "step": 1265 |
| }, |
| { |
| "epoch": 0.24399615754082613, |
| "grad_norm": 1.018276572227478, |
| "learning_rate": 0.00027021087981924296, |
| "loss": 3.4626544952392577, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.24495677233429394, |
| "grad_norm": 1.2246990203857422, |
| "learning_rate": 0.0002699252231583904, |
| "loss": 3.458060073852539, |
| "step": 1275 |
| }, |
| { |
| "epoch": 0.24591738712776176, |
| "grad_norm": 1.440726637840271, |
| "learning_rate": 0.000269638355918917, |
| "loss": 3.4551132202148436, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.24687800192122958, |
| "grad_norm": 1.0400745868682861, |
| "learning_rate": 0.00026935028099658864, |
| "loss": 3.455486297607422, |
| "step": 1285 |
| }, |
| { |
| "epoch": 0.2478386167146974, |
| "grad_norm": 1.271278738975525, |
| "learning_rate": 0.00026906100129936173, |
| "loss": 3.4623786926269533, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.24879923150816521, |
| "grad_norm": 1.0657492876052856, |
| "learning_rate": 0.0002687705197473545, |
| "loss": 3.455769348144531, |
| "step": 1295 |
| }, |
| { |
| "epoch": 0.24975984630163303, |
| "grad_norm": 0.913031816482544, |
| "learning_rate": 0.00026847883927281715, |
| "loss": 3.4589412689208983, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.2507204610951009, |
| "grad_norm": 0.8772344589233398, |
| "learning_rate": 0.00026818596282010223, |
| "loss": 3.4558467864990234, |
| "step": 1305 |
| }, |
| { |
| "epoch": 0.2516810758885687, |
| "grad_norm": 1.1004761457443237, |
| "learning_rate": 0.00026789189334563507, |
| "loss": 3.455731964111328, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.2526416906820365, |
| "grad_norm": 1.0385078191757202, |
| "learning_rate": 0.00026759663381788407, |
| "loss": 3.4509082794189454, |
| "step": 1315 |
| }, |
| { |
| "epoch": 0.25360230547550433, |
| "grad_norm": 0.8879292011260986, |
| "learning_rate": 0.00026730018721733034, |
| "loss": 3.447312927246094, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.25456292026897215, |
| "grad_norm": 1.32045316696167, |
| "learning_rate": 0.0002670025565364379, |
| "loss": 3.4483909606933594, |
| "step": 1325 |
| }, |
| { |
| "epoch": 0.25552353506243997, |
| "grad_norm": 1.0469266176223755, |
| "learning_rate": 0.0002667037447796234, |
| "loss": 3.44598388671875, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.2564841498559078, |
| "grad_norm": 0.9312946796417236, |
| "learning_rate": 0.0002664037549632259, |
| "loss": 3.449735641479492, |
| "step": 1335 |
| }, |
| { |
| "epoch": 0.2574447646493756, |
| "grad_norm": 1.0092568397521973, |
| "learning_rate": 0.00026610259011547617, |
| "loss": 3.4483173370361326, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.2584053794428434, |
| "grad_norm": 0.905603289604187, |
| "learning_rate": 0.0002658002532764663, |
| "loss": 3.4480667114257812, |
| "step": 1345 |
| }, |
| { |
| "epoch": 0.25936599423631124, |
| "grad_norm": 1.152783989906311, |
| "learning_rate": 0.00026549674749811917, |
| "loss": 3.4437828063964844, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.26032660902977905, |
| "grad_norm": 0.9953619241714478, |
| "learning_rate": 0.00026519207584415705, |
| "loss": 3.444900131225586, |
| "step": 1355 |
| }, |
| { |
| "epoch": 0.2612872238232469, |
| "grad_norm": 0.8666988611221313, |
| "learning_rate": 0.00026488624139007154, |
| "loss": 3.4405364990234375, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.2622478386167147, |
| "grad_norm": 1.0244816541671753, |
| "learning_rate": 0.0002645792472230917, |
| "loss": 3.4408805847167967, |
| "step": 1365 |
| }, |
| { |
| "epoch": 0.2632084534101825, |
| "grad_norm": 1.211923599243164, |
| "learning_rate": 0.0002642710964421535, |
| "loss": 3.441836929321289, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.2641690682036503, |
| "grad_norm": 0.9250677824020386, |
| "learning_rate": 0.0002639617921578681, |
| "loss": 3.4430908203125, |
| "step": 1375 |
| }, |
| { |
| "epoch": 0.26512968299711814, |
| "grad_norm": 0.9860612154006958, |
| "learning_rate": 0.0002636513374924908, |
| "loss": 3.4425369262695313, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.26609029779058596, |
| "grad_norm": 1.1464399099349976, |
| "learning_rate": 0.00026333973557988923, |
| "loss": 3.4433387756347655, |
| "step": 1385 |
| }, |
| { |
| "epoch": 0.2670509125840538, |
| "grad_norm": 0.9755972027778625, |
| "learning_rate": 0.0002630269895655119, |
| "loss": 3.4360820770263674, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.2680115273775216, |
| "grad_norm": 0.7360960245132446, |
| "learning_rate": 0.00026271310260635633, |
| "loss": 3.4424560546875, |
| "step": 1395 |
| }, |
| { |
| "epoch": 0.2689721421709894, |
| "grad_norm": 1.3588593006134033, |
| "learning_rate": 0.0002623980778709374, |
| "loss": 3.4407089233398436, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.26993275696445723, |
| "grad_norm": 1.060451626777649, |
| "learning_rate": 0.0002620819185392551, |
| "loss": 3.4387184143066407, |
| "step": 1405 |
| }, |
| { |
| "epoch": 0.27089337175792505, |
| "grad_norm": 1.108783483505249, |
| "learning_rate": 0.00026176462780276246, |
| "loss": 3.440538787841797, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.27185398655139287, |
| "grad_norm": 1.033456802368164, |
| "learning_rate": 0.0002614462088643336, |
| "loss": 3.438457489013672, |
| "step": 1415 |
| }, |
| { |
| "epoch": 0.2728146013448607, |
| "grad_norm": 1.1797153949737549, |
| "learning_rate": 0.00026112666493823103, |
| "loss": 3.4355682373046874, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.2737752161383285, |
| "grad_norm": 0.8781239986419678, |
| "learning_rate": 0.00026080599925007355, |
| "loss": 3.437305450439453, |
| "step": 1425 |
| }, |
| { |
| "epoch": 0.2747358309317964, |
| "grad_norm": 0.812835156917572, |
| "learning_rate": 0.00026048421503680337, |
| "loss": 3.4358146667480467, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.2756964457252642, |
| "grad_norm": 1.6010043621063232, |
| "learning_rate": 0.00026016131554665377, |
| "loss": 3.4342422485351562, |
| "step": 1435 |
| }, |
| { |
| "epoch": 0.276657060518732, |
| "grad_norm": 1.0615707635879517, |
| "learning_rate": 0.000259837304039116, |
| "loss": 3.434612274169922, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.27761767531219983, |
| "grad_norm": 0.9948373436927795, |
| "learning_rate": 0.0002595121837849065, |
| "loss": 3.434070587158203, |
| "step": 1445 |
| }, |
| { |
| "epoch": 0.27857829010566765, |
| "grad_norm": 1.335551142692566, |
| "learning_rate": 0.000259185958065934, |
| "loss": 3.4336421966552733, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.27953890489913547, |
| "grad_norm": 1.2016087770462036, |
| "learning_rate": 0.00025885863017526613, |
| "loss": 3.432381439208984, |
| "step": 1455 |
| }, |
| { |
| "epoch": 0.2804995196926033, |
| "grad_norm": 1.0571645498275757, |
| "learning_rate": 0.00025853020341709646, |
| "loss": 3.4373619079589846, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.2814601344860711, |
| "grad_norm": 0.9469745755195618, |
| "learning_rate": 0.000258200681106711, |
| "loss": 3.433266448974609, |
| "step": 1465 |
| }, |
| { |
| "epoch": 0.2824207492795389, |
| "grad_norm": 1.0642951726913452, |
| "learning_rate": 0.00025787006657045477, |
| "loss": 3.4274715423583983, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.28338136407300674, |
| "grad_norm": 1.117781162261963, |
| "learning_rate": 0.000257538363145698, |
| "loss": 3.430649185180664, |
| "step": 1475 |
| }, |
| { |
| "epoch": 0.28434197886647455, |
| "grad_norm": 1.1161621809005737, |
| "learning_rate": 0.00025720557418080304, |
| "loss": 3.43245849609375, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.28530259365994237, |
| "grad_norm": 1.0228129625320435, |
| "learning_rate": 0.00025687170303508977, |
| "loss": 3.4293182373046873, |
| "step": 1485 |
| }, |
| { |
| "epoch": 0.2862632084534102, |
| "grad_norm": 0.8712406754493713, |
| "learning_rate": 0.00025653675307880225, |
| "loss": 3.424097442626953, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.287223823246878, |
| "grad_norm": 0.8777533173561096, |
| "learning_rate": 0.00025620072769307463, |
| "loss": 3.424530029296875, |
| "step": 1495 |
| }, |
| { |
| "epoch": 0.2881844380403458, |
| "grad_norm": 0.9582040309906006, |
| "learning_rate": 0.00025586363026989677, |
| "loss": 3.4286567687988283, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.28914505283381364, |
| "grad_norm": 1.0138081312179565, |
| "learning_rate": 0.0002555254642120802, |
| "loss": 3.4281455993652346, |
| "step": 1505 |
| }, |
| { |
| "epoch": 0.29010566762728146, |
| "grad_norm": 1.2888671159744263, |
| "learning_rate": 0.0002551862329332238, |
| "loss": 3.425878143310547, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.2910662824207493, |
| "grad_norm": 0.9225014448165894, |
| "learning_rate": 0.0002548459398576791, |
| "loss": 3.426416015625, |
| "step": 1515 |
| }, |
| { |
| "epoch": 0.2920268972142171, |
| "grad_norm": 0.7408014535903931, |
| "learning_rate": 0.00025450458842051616, |
| "loss": 3.4226699829101563, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.2929875120076849, |
| "grad_norm": 1.099754810333252, |
| "learning_rate": 0.0002541621820674882, |
| "loss": 3.42437858581543, |
| "step": 1525 |
| }, |
| { |
| "epoch": 0.29394812680115273, |
| "grad_norm": 1.210573434829712, |
| "learning_rate": 0.0002538187242549976, |
| "loss": 3.422506332397461, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.29490874159462055, |
| "grad_norm": 1.1748319864273071, |
| "learning_rate": 0.00025347421845006056, |
| "loss": 3.4207489013671877, |
| "step": 1535 |
| }, |
| { |
| "epoch": 0.29586935638808837, |
| "grad_norm": 1.2270821332931519, |
| "learning_rate": 0.00025312866813027195, |
| "loss": 3.4193565368652346, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.2968299711815562, |
| "grad_norm": 1.1345160007476807, |
| "learning_rate": 0.0002527820767837708, |
| "loss": 3.4215885162353517, |
| "step": 1545 |
| }, |
| { |
| "epoch": 0.297790585975024, |
| "grad_norm": 1.4022908210754395, |
| "learning_rate": 0.00025243444790920447, |
| "loss": 3.4179046630859373, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.2987512007684918, |
| "grad_norm": 0.9985158443450928, |
| "learning_rate": 0.0002520857850156936, |
| "loss": 3.4201751708984376, |
| "step": 1555 |
| }, |
| { |
| "epoch": 0.29971181556195964, |
| "grad_norm": 1.081823468208313, |
| "learning_rate": 0.0002517360916227968, |
| "loss": 3.416116714477539, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.30067243035542746, |
| "grad_norm": 1.2369301319122314, |
| "learning_rate": 0.000251385371260475, |
| "loss": 3.4199737548828124, |
| "step": 1565 |
| }, |
| { |
| "epoch": 0.3016330451488953, |
| "grad_norm": 0.8743671178817749, |
| "learning_rate": 0.0002510336274690557, |
| "loss": 3.4136940002441407, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.3025936599423631, |
| "grad_norm": 1.0848971605300903, |
| "learning_rate": 0.0002506808637991974, |
| "loss": 3.417308044433594, |
| "step": 1575 |
| }, |
| { |
| "epoch": 0.3035542747358309, |
| "grad_norm": 1.450103521347046, |
| "learning_rate": 0.0002503270838118537, |
| "loss": 3.41820068359375, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.3045148895292987, |
| "grad_norm": 0.9271462559700012, |
| "learning_rate": 0.0002499722910782374, |
| "loss": 3.4144283294677735, |
| "step": 1585 |
| }, |
| { |
| "epoch": 0.30547550432276654, |
| "grad_norm": 0.9813090562820435, |
| "learning_rate": 0.0002496164891797844, |
| "loss": 3.41821403503418, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.30643611911623436, |
| "grad_norm": 0.9819661378860474, |
| "learning_rate": 0.0002492596817081175, |
| "loss": 3.4166328430175783, |
| "step": 1595 |
| }, |
| { |
| "epoch": 0.30739673390970224, |
| "grad_norm": 1.1119061708450317, |
| "learning_rate": 0.0002489018722650103, |
| "loss": 3.4129749298095704, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.30835734870317005, |
| "grad_norm": 1.0902990102767944, |
| "learning_rate": 0.0002485430644623507, |
| "loss": 3.415445327758789, |
| "step": 1605 |
| }, |
| { |
| "epoch": 0.30931796349663787, |
| "grad_norm": 1.1950386762619019, |
| "learning_rate": 0.00024818326192210447, |
| "loss": 3.411751937866211, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.3102785782901057, |
| "grad_norm": 0.8721151351928711, |
| "learning_rate": 0.0002478224682762787, |
| "loss": 3.409122085571289, |
| "step": 1615 |
| }, |
| { |
| "epoch": 0.3112391930835735, |
| "grad_norm": 1.1784799098968506, |
| "learning_rate": 0.0002474606871668852, |
| "loss": 3.411494827270508, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.3121998078770413, |
| "grad_norm": 1.107187271118164, |
| "learning_rate": 0.00024709792224590356, |
| "loss": 3.4105316162109376, |
| "step": 1625 |
| }, |
| { |
| "epoch": 0.31316042267050914, |
| "grad_norm": 0.8363384008407593, |
| "learning_rate": 0.0002467341771752446, |
| "loss": 3.406443786621094, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.31412103746397696, |
| "grad_norm": 0.8663597702980042, |
| "learning_rate": 0.000246369455626713, |
| "loss": 3.409493637084961, |
| "step": 1635 |
| }, |
| { |
| "epoch": 0.3150816522574448, |
| "grad_norm": 0.7546355724334717, |
| "learning_rate": 0.00024600376128197047, |
| "loss": 3.403882598876953, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.3160422670509126, |
| "grad_norm": 1.0707752704620361, |
| "learning_rate": 0.00024563709783249877, |
| "loss": 3.405009460449219, |
| "step": 1645 |
| }, |
| { |
| "epoch": 0.3170028818443804, |
| "grad_norm": 1.0536484718322754, |
| "learning_rate": 0.00024526946897956194, |
| "loss": 3.4075836181640624, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.31796349663784823, |
| "grad_norm": 1.0901403427124023, |
| "learning_rate": 0.00024490087843416947, |
| "loss": 3.406299591064453, |
| "step": 1655 |
| }, |
| { |
| "epoch": 0.31892411143131605, |
| "grad_norm": 1.224731683731079, |
| "learning_rate": 0.00024453132991703844, |
| "loss": 3.408635711669922, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.31988472622478387, |
| "grad_norm": 0.9345903396606445, |
| "learning_rate": 0.00024416082715855627, |
| "loss": 3.4082077026367186, |
| "step": 1665 |
| }, |
| { |
| "epoch": 0.3208453410182517, |
| "grad_norm": 1.0254333019256592, |
| "learning_rate": 0.00024378937389874276, |
| "loss": 3.409267807006836, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.3218059558117195, |
| "grad_norm": 1.0035061836242676, |
| "learning_rate": 0.0002434169738872126, |
| "loss": 3.4060768127441405, |
| "step": 1675 |
| }, |
| { |
| "epoch": 0.3227665706051873, |
| "grad_norm": 0.8024983406066895, |
| "learning_rate": 0.0002430436308831374, |
| "loss": 3.4000919342041014, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.32372718539865514, |
| "grad_norm": 0.8286018371582031, |
| "learning_rate": 0.00024266934865520767, |
| "loss": 3.405823516845703, |
| "step": 1685 |
| }, |
| { |
| "epoch": 0.32468780019212296, |
| "grad_norm": 0.9187076091766357, |
| "learning_rate": 0.00024229413098159506, |
| "loss": 3.402732086181641, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.3256484149855908, |
| "grad_norm": 0.9258525967597961, |
| "learning_rate": 0.00024191798164991378, |
| "loss": 3.4013343811035157, |
| "step": 1695 |
| }, |
| { |
| "epoch": 0.3266090297790586, |
| "grad_norm": 1.202602744102478, |
| "learning_rate": 0.0002415409044571828, |
| "loss": 3.4034278869628904, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.3275696445725264, |
| "grad_norm": 1.0954785346984863, |
| "learning_rate": 0.00024116290320978724, |
| "loss": 3.4010189056396483, |
| "step": 1705 |
| }, |
| { |
| "epoch": 0.3285302593659942, |
| "grad_norm": 0.8299508690834045, |
| "learning_rate": 0.00024078398172344006, |
| "loss": 3.402983856201172, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.32949087415946204, |
| "grad_norm": 1.2064135074615479, |
| "learning_rate": 0.00024040414382314358, |
| "loss": 3.397635650634766, |
| "step": 1715 |
| }, |
| { |
| "epoch": 0.33045148895292986, |
| "grad_norm": 1.0353888273239136, |
| "learning_rate": 0.00024002339334315066, |
| "loss": 3.4012256622314454, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.3314121037463977, |
| "grad_norm": 0.9317356944084167, |
| "learning_rate": 0.00023964173412692631, |
| "loss": 3.3994606018066404, |
| "step": 1725 |
| }, |
| { |
| "epoch": 0.3323727185398655, |
| "grad_norm": 1.0982980728149414, |
| "learning_rate": 0.00023925917002710865, |
| "loss": 3.398478698730469, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.3333333333333333, |
| "grad_norm": 0.8598034977912903, |
| "learning_rate": 0.0002388757049054701, |
| "loss": 3.400804138183594, |
| "step": 1735 |
| }, |
| { |
| "epoch": 0.33429394812680113, |
| "grad_norm": 0.9004125595092773, |
| "learning_rate": 0.00023849134263287836, |
| "loss": 3.3936607360839846, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.33525456292026895, |
| "grad_norm": 1.0058287382125854, |
| "learning_rate": 0.00023810608708925755, |
| "loss": 3.397069549560547, |
| "step": 1745 |
| }, |
| { |
| "epoch": 0.33621517771373677, |
| "grad_norm": 1.0947821140289307, |
| "learning_rate": 0.00023771994216354857, |
| "loss": 3.398267364501953, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.3371757925072046, |
| "grad_norm": 1.1534372568130493, |
| "learning_rate": 0.00023733291175367046, |
| "loss": 3.3978240966796873, |
| "step": 1755 |
| }, |
| { |
| "epoch": 0.3381364073006724, |
| "grad_norm": 0.9231524467468262, |
| "learning_rate": 0.00023694499976648043, |
| "loss": 3.4001846313476562, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.3390970220941403, |
| "grad_norm": 1.1671085357666016, |
| "learning_rate": 0.0002365562101177349, |
| "loss": 3.3949043273925783, |
| "step": 1765 |
| }, |
| { |
| "epoch": 0.3400576368876081, |
| "grad_norm": 1.1204719543457031, |
| "learning_rate": 0.00023616654673204983, |
| "loss": 3.3966522216796875, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.3410182516810759, |
| "grad_norm": 0.9162036776542664, |
| "learning_rate": 0.00023577601354286094, |
| "loss": 3.3971607208251955, |
| "step": 1775 |
| }, |
| { |
| "epoch": 0.34197886647454373, |
| "grad_norm": 0.8886377811431885, |
| "learning_rate": 0.00023538461449238422, |
| "loss": 3.3945545196533202, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.34293948126801155, |
| "grad_norm": 0.8314016461372375, |
| "learning_rate": 0.00023499235353157603, |
| "loss": 3.397966766357422, |
| "step": 1785 |
| }, |
| { |
| "epoch": 0.34390009606147937, |
| "grad_norm": 0.9351712465286255, |
| "learning_rate": 0.0002345992346200932, |
| "loss": 3.3912559509277345, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.3448607108549472, |
| "grad_norm": 1.2932080030441284, |
| "learning_rate": 0.00023420526172625316, |
| "loss": 3.399237060546875, |
| "step": 1795 |
| }, |
| { |
| "epoch": 0.345821325648415, |
| "grad_norm": 1.092822551727295, |
| "learning_rate": 0.00023381043882699377, |
| "loss": 3.3947265625, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.3467819404418828, |
| "grad_norm": 1.2571241855621338, |
| "learning_rate": 0.0002334147699078333, |
| "loss": 3.3920978546142577, |
| "step": 1805 |
| }, |
| { |
| "epoch": 0.34774255523535064, |
| "grad_norm": 0.7598462104797363, |
| "learning_rate": 0.00023301825896282992, |
| "loss": 3.3873401641845704, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.34870317002881845, |
| "grad_norm": 0.9965523481369019, |
| "learning_rate": 0.00023262090999454194, |
| "loss": 3.390736389160156, |
| "step": 1815 |
| }, |
| { |
| "epoch": 0.34966378482228627, |
| "grad_norm": 0.9940029978752136, |
| "learning_rate": 0.00023222272701398664, |
| "loss": 3.387689208984375, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.3506243996157541, |
| "grad_norm": 1.051283597946167, |
| "learning_rate": 0.00023182371404060047, |
| "loss": 3.3928714752197267, |
| "step": 1825 |
| }, |
| { |
| "epoch": 0.3515850144092219, |
| "grad_norm": 0.9655299186706543, |
| "learning_rate": 0.00023142387510219814, |
| "loss": 3.3853427886962892, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.3525456292026897, |
| "grad_norm": 0.912288248538971, |
| "learning_rate": 0.00023102321423493192, |
| "loss": 3.3889575958251954, |
| "step": 1835 |
| }, |
| { |
| "epoch": 0.35350624399615754, |
| "grad_norm": 0.890324592590332, |
| "learning_rate": 0.00023062173548325112, |
| "loss": 3.388275146484375, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.35446685878962536, |
| "grad_norm": 0.819926381111145, |
| "learning_rate": 0.0002302194428998611, |
| "loss": 3.3917514801025392, |
| "step": 1845 |
| }, |
| { |
| "epoch": 0.3554274735830932, |
| "grad_norm": 1.0578113794326782, |
| "learning_rate": 0.0002298163405456824, |
| "loss": 3.384230041503906, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.356388088376561, |
| "grad_norm": 1.0350604057312012, |
| "learning_rate": 0.00022941243248980966, |
| "loss": 3.388728713989258, |
| "step": 1855 |
| }, |
| { |
| "epoch": 0.3573487031700288, |
| "grad_norm": 1.0110005140304565, |
| "learning_rate": 0.0002290077228094708, |
| "loss": 3.386165237426758, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.35830931796349663, |
| "grad_norm": 0.7671937346458435, |
| "learning_rate": 0.00022860221558998554, |
| "loss": 3.386688995361328, |
| "step": 1865 |
| }, |
| { |
| "epoch": 0.35926993275696445, |
| "grad_norm": 1.032853603363037, |
| "learning_rate": 0.00022819591492472438, |
| "loss": 3.385240173339844, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.36023054755043227, |
| "grad_norm": 0.8370668888092041, |
| "learning_rate": 0.00022778882491506725, |
| "loss": 3.3829574584960938, |
| "step": 1875 |
| }, |
| { |
| "epoch": 0.3611911623439001, |
| "grad_norm": 1.0862385034561157, |
| "learning_rate": 0.00022738094967036208, |
| "loss": 3.3890167236328126, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.3621517771373679, |
| "grad_norm": 1.2748796939849854, |
| "learning_rate": 0.00022697229330788312, |
| "loss": 3.384657287597656, |
| "step": 1885 |
| }, |
| { |
| "epoch": 0.3631123919308357, |
| "grad_norm": 1.163509726524353, |
| "learning_rate": 0.00022656285995278984, |
| "loss": 3.3868003845214845, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.36407300672430354, |
| "grad_norm": 0.832695484161377, |
| "learning_rate": 0.00022615265373808488, |
| "loss": 3.38619384765625, |
| "step": 1895 |
| }, |
| { |
| "epoch": 0.36503362151777136, |
| "grad_norm": 1.1879743337631226, |
| "learning_rate": 0.00022574167880457245, |
| "loss": 3.384031295776367, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.3659942363112392, |
| "grad_norm": 0.9494090676307678, |
| "learning_rate": 0.00022532993930081668, |
| "loss": 3.379372406005859, |
| "step": 1905 |
| }, |
| { |
| "epoch": 0.366954851104707, |
| "grad_norm": 0.7896311283111572, |
| "learning_rate": 0.00022491743938309936, |
| "loss": 3.3832141876220705, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.3679154658981748, |
| "grad_norm": 0.911125898361206, |
| "learning_rate": 0.0002245041832153786, |
| "loss": 3.3824745178222657, |
| "step": 1915 |
| }, |
| { |
| "epoch": 0.3688760806916426, |
| "grad_norm": 1.241249680519104, |
| "learning_rate": 0.0002240901749692461, |
| "loss": 3.3809112548828124, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.36983669548511044, |
| "grad_norm": 0.8434759974479675, |
| "learning_rate": 0.00022367541882388554, |
| "loss": 3.3806549072265626, |
| "step": 1925 |
| }, |
| { |
| "epoch": 0.37079731027857826, |
| "grad_norm": 1.0567084550857544, |
| "learning_rate": 0.00022325991896603018, |
| "loss": 3.380388641357422, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.37175792507204614, |
| "grad_norm": 1.1255742311477661, |
| "learning_rate": 0.00022284367958992065, |
| "loss": 3.3810966491699217, |
| "step": 1935 |
| }, |
| { |
| "epoch": 0.37271853986551395, |
| "grad_norm": 0.8697860240936279, |
| "learning_rate": 0.0002224267048972627, |
| "loss": 3.381666564941406, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.37367915465898177, |
| "grad_norm": 1.1864638328552246, |
| "learning_rate": 0.00022200899909718456, |
| "loss": 3.3785327911376952, |
| "step": 1945 |
| }, |
| { |
| "epoch": 0.3746397694524496, |
| "grad_norm": 0.8901607394218445, |
| "learning_rate": 0.00022159056640619457, |
| "loss": 3.3746604919433594, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.3756003842459174, |
| "grad_norm": 0.940137505531311, |
| "learning_rate": 0.00022117141104813876, |
| "loss": 3.3786911010742187, |
| "step": 1955 |
| }, |
| { |
| "epoch": 0.3765609990393852, |
| "grad_norm": 0.9261166453361511, |
| "learning_rate": 0.000220751537254158, |
| "loss": 3.377775192260742, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.37752161383285304, |
| "grad_norm": 0.9173313975334167, |
| "learning_rate": 0.00022033094926264548, |
| "loss": 3.375761795043945, |
| "step": 1965 |
| }, |
| { |
| "epoch": 0.37848222862632086, |
| "grad_norm": 1.0608283281326294, |
| "learning_rate": 0.00021990965131920358, |
| "loss": 3.3748458862304687, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.3794428434197887, |
| "grad_norm": 0.9284070134162903, |
| "learning_rate": 0.0002194876476766015, |
| "loss": 3.374094772338867, |
| "step": 1975 |
| }, |
| { |
| "epoch": 0.3804034582132565, |
| "grad_norm": 0.8284513354301453, |
| "learning_rate": 0.00021906494259473196, |
| "loss": 3.3742515563964846, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.3813640730067243, |
| "grad_norm": 1.1894856691360474, |
| "learning_rate": 0.00021864154034056832, |
| "loss": 3.3718952178955077, |
| "step": 1985 |
| }, |
| { |
| "epoch": 0.38232468780019213, |
| "grad_norm": 0.9039756655693054, |
| "learning_rate": 0.00021821744518812154, |
| "loss": 3.3747280120849608, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.38328530259365995, |
| "grad_norm": 0.8699926137924194, |
| "learning_rate": 0.00021779266141839699, |
| "loss": 3.372933197021484, |
| "step": 1995 |
| }, |
| { |
| "epoch": 0.38424591738712777, |
| "grad_norm": 0.9204120635986328, |
| "learning_rate": 0.00021736719331935127, |
| "loss": 3.370801544189453, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.3852065321805956, |
| "grad_norm": 1.0923502445220947, |
| "learning_rate": 0.00021694104518584886, |
| "loss": 3.374231719970703, |
| "step": 2005 |
| }, |
| { |
| "epoch": 0.3861671469740634, |
| "grad_norm": 0.8937914967536926, |
| "learning_rate": 0.00021651422131961884, |
| "loss": 3.374364471435547, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.3871277617675312, |
| "grad_norm": 0.8756441473960876, |
| "learning_rate": 0.0002160867260292115, |
| "loss": 3.3731971740722657, |
| "step": 2015 |
| }, |
| { |
| "epoch": 0.38808837656099904, |
| "grad_norm": 0.7807921171188354, |
| "learning_rate": 0.00021565856362995464, |
| "loss": 3.3712364196777345, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.38904899135446686, |
| "grad_norm": 1.0138747692108154, |
| "learning_rate": 0.00021522973844391024, |
| "loss": 3.370909881591797, |
| "step": 2025 |
| }, |
| { |
| "epoch": 0.3900096061479347, |
| "grad_norm": 0.9295784831047058, |
| "learning_rate": 0.00021480025479983077, |
| "loss": 3.369549560546875, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.3909702209414025, |
| "grad_norm": 0.7283740043640137, |
| "learning_rate": 0.00021437011703311545, |
| "loss": 3.3707386016845704, |
| "step": 2035 |
| }, |
| { |
| "epoch": 0.3919308357348703, |
| "grad_norm": 0.9124945402145386, |
| "learning_rate": 0.0002139393294857665, |
| "loss": 3.3750667572021484, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.3928914505283381, |
| "grad_norm": 1.0313608646392822, |
| "learning_rate": 0.00021350789650634525, |
| "loss": 3.369192123413086, |
| "step": 2045 |
| }, |
| { |
| "epoch": 0.39385206532180594, |
| "grad_norm": 0.904303252696991, |
| "learning_rate": 0.00021307582244992838, |
| "loss": 3.3705421447753907, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.39481268011527376, |
| "grad_norm": 0.9359715580940247, |
| "learning_rate": 0.0002126431116780639, |
| "loss": 3.3716297149658203, |
| "step": 2055 |
| }, |
| { |
| "epoch": 0.3957732949087416, |
| "grad_norm": 0.9251255989074707, |
| "learning_rate": 0.00021220976855872712, |
| "loss": 3.3701812744140627, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.3967339097022094, |
| "grad_norm": 0.8808803558349609, |
| "learning_rate": 0.00021177579746627643, |
| "loss": 3.369044876098633, |
| "step": 2065 |
| }, |
| { |
| "epoch": 0.3976945244956772, |
| "grad_norm": 0.8854672908782959, |
| "learning_rate": 0.0002113412027814094, |
| "loss": 3.3659637451171873, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.39865513928914503, |
| "grad_norm": 0.8999817371368408, |
| "learning_rate": 0.0002109059888911183, |
| "loss": 3.365467071533203, |
| "step": 2075 |
| }, |
| { |
| "epoch": 0.39961575408261285, |
| "grad_norm": 0.9394131898880005, |
| "learning_rate": 0.00021047016018864602, |
| "loss": 3.372660827636719, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.40057636887608067, |
| "grad_norm": 0.7608519792556763, |
| "learning_rate": 0.00021003372107344167, |
| "loss": 3.367539978027344, |
| "step": 2085 |
| }, |
| { |
| "epoch": 0.4015369836695485, |
| "grad_norm": 1.1321345567703247, |
| "learning_rate": 0.00020959667595111603, |
| "loss": 3.3716583251953125, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.4024975984630163, |
| "grad_norm": 0.9030200242996216, |
| "learning_rate": 0.00020915902923339722, |
| "loss": 3.3653472900390624, |
| "step": 2095 |
| }, |
| { |
| "epoch": 0.4034582132564842, |
| "grad_norm": 0.9246105551719666, |
| "learning_rate": 0.0002087207853380862, |
| "loss": 3.3666152954101562, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.404418828049952, |
| "grad_norm": 0.9399805665016174, |
| "learning_rate": 0.00020828194868901205, |
| "loss": 3.3659893035888673, |
| "step": 2105 |
| }, |
| { |
| "epoch": 0.4053794428434198, |
| "grad_norm": 0.8061392307281494, |
| "learning_rate": 0.00020784252371598732, |
| "loss": 3.366693115234375, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.40634005763688763, |
| "grad_norm": 1.1672389507293701, |
| "learning_rate": 0.00020740251485476345, |
| "loss": 3.3655033111572266, |
| "step": 2115 |
| }, |
| { |
| "epoch": 0.40730067243035545, |
| "grad_norm": 0.7972300052642822, |
| "learning_rate": 0.00020696192654698592, |
| "loss": 3.3643836975097656, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.40826128722382327, |
| "grad_norm": 0.6781191825866699, |
| "learning_rate": 0.00020652076324014927, |
| "loss": 3.363838958740234, |
| "step": 2125 |
| }, |
| { |
| "epoch": 0.4092219020172911, |
| "grad_norm": 1.1469253301620483, |
| "learning_rate": 0.00020607902938755252, |
| "loss": 3.3665115356445314, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.4101825168107589, |
| "grad_norm": 0.9588373303413391, |
| "learning_rate": 0.00020563672944825392, |
| "loss": 3.3640487670898436, |
| "step": 2135 |
| }, |
| { |
| "epoch": 0.4111431316042267, |
| "grad_norm": 0.8890582323074341, |
| "learning_rate": 0.00020519386788702602, |
| "loss": 3.361592102050781, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.41210374639769454, |
| "grad_norm": 0.9977627396583557, |
| "learning_rate": 0.0002047504491743107, |
| "loss": 3.3624588012695313, |
| "step": 2145 |
| }, |
| { |
| "epoch": 0.41306436119116235, |
| "grad_norm": 1.0691829919815063, |
| "learning_rate": 0.000204306477786174, |
| "loss": 3.3603302001953126, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.4140249759846302, |
| "grad_norm": 1.0000925064086914, |
| "learning_rate": 0.00020386195820426082, |
| "loss": 3.364809036254883, |
| "step": 2155 |
| }, |
| { |
| "epoch": 0.414985590778098, |
| "grad_norm": 1.147975206375122, |
| "learning_rate": 0.00020341689491574984, |
| "loss": 3.3625713348388673, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.4159462055715658, |
| "grad_norm": 0.9693566560745239, |
| "learning_rate": 0.00020297129241330817, |
| "loss": 3.3619319915771486, |
| "step": 2165 |
| }, |
| { |
| "epoch": 0.4169068203650336, |
| "grad_norm": 0.8027826547622681, |
| "learning_rate": 0.00020252515519504592, |
| "loss": 3.3606258392333985, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.41786743515850144, |
| "grad_norm": 0.7421839833259583, |
| "learning_rate": 0.0002020784877644709, |
| "loss": 3.3613494873046874, |
| "step": 2175 |
| }, |
| { |
| "epoch": 0.41882804995196926, |
| "grad_norm": 0.7940289378166199, |
| "learning_rate": 0.00020163129463044308, |
| "loss": 3.360422897338867, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.4197886647454371, |
| "grad_norm": 1.0238333940505981, |
| "learning_rate": 0.0002011835803071292, |
| "loss": 3.3607128143310545, |
| "step": 2185 |
| }, |
| { |
| "epoch": 0.4207492795389049, |
| "grad_norm": 1.0974775552749634, |
| "learning_rate": 0.00020073534931395697, |
| "loss": 3.3562847137451173, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.4217098943323727, |
| "grad_norm": 0.9029604196548462, |
| "learning_rate": 0.00020028660617556985, |
| "loss": 3.3622852325439454, |
| "step": 2195 |
| }, |
| { |
| "epoch": 0.42267050912584053, |
| "grad_norm": 1.0699923038482666, |
| "learning_rate": 0.00019983735542178086, |
| "loss": 3.359086608886719, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.42363112391930835, |
| "grad_norm": 0.9210760593414307, |
| "learning_rate": 0.00019938760158752725, |
| "loss": 3.360548400878906, |
| "step": 2205 |
| }, |
| { |
| "epoch": 0.42459173871277617, |
| "grad_norm": 1.0080957412719727, |
| "learning_rate": 0.00019893734921282448, |
| "loss": 3.3639129638671874, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.425552353506244, |
| "grad_norm": 0.668106734752655, |
| "learning_rate": 0.0001984866028427207, |
| "loss": 3.3550628662109374, |
| "step": 2215 |
| }, |
| { |
| "epoch": 0.4265129682997118, |
| "grad_norm": 0.9900498390197754, |
| "learning_rate": 0.00019803536702725044, |
| "loss": 3.3611160278320313, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.4274735830931796, |
| "grad_norm": 0.6974493861198425, |
| "learning_rate": 0.00019758364632138908, |
| "loss": 3.354916000366211, |
| "step": 2225 |
| }, |
| { |
| "epoch": 0.42843419788664744, |
| "grad_norm": 0.8031367063522339, |
| "learning_rate": 0.0001971314452850066, |
| "loss": 3.360482406616211, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.42939481268011526, |
| "grad_norm": 0.9039629697799683, |
| "learning_rate": 0.00019667876848282167, |
| "loss": 3.3570487976074217, |
| "step": 2235 |
| }, |
| { |
| "epoch": 0.4303554274735831, |
| "grad_norm": 0.7340952157974243, |
| "learning_rate": 0.0001962256204843556, |
| "loss": 3.3560821533203127, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.4313160422670509, |
| "grad_norm": 1.069008469581604, |
| "learning_rate": 0.00019577200586388618, |
| "loss": 3.358184814453125, |
| "step": 2245 |
| }, |
| { |
| "epoch": 0.4322766570605187, |
| "grad_norm": 0.8336465358734131, |
| "learning_rate": 0.00019531792920040133, |
| "loss": 3.359252166748047, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.4332372718539865, |
| "grad_norm": 1.0504704713821411, |
| "learning_rate": 0.0001948633950775532, |
| "loss": 3.3570930480957033, |
| "step": 2255 |
| }, |
| { |
| "epoch": 0.43419788664745435, |
| "grad_norm": 0.879917562007904, |
| "learning_rate": 0.00019440840808361174, |
| "loss": 3.3541915893554686, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.43515850144092216, |
| "grad_norm": 0.9227856397628784, |
| "learning_rate": 0.00019395297281141828, |
| "loss": 3.3502120971679688, |
| "step": 2265 |
| }, |
| { |
| "epoch": 0.43611911623439004, |
| "grad_norm": 0.7166210412979126, |
| "learning_rate": 0.0001934970938583393, |
| "loss": 3.3565288543701173, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.43707973102785785, |
| "grad_norm": 0.8305450677871704, |
| "learning_rate": 0.00019304077582622003, |
| "loss": 3.352977752685547, |
| "step": 2275 |
| }, |
| { |
| "epoch": 0.43804034582132567, |
| "grad_norm": 0.9930862188339233, |
| "learning_rate": 0.00019258402332133798, |
| "loss": 3.360272979736328, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.4390009606147935, |
| "grad_norm": 1.1689637899398804, |
| "learning_rate": 0.00019212684095435637, |
| "loss": 3.3537288665771485, |
| "step": 2285 |
| }, |
| { |
| "epoch": 0.4399615754082613, |
| "grad_norm": 0.7753242254257202, |
| "learning_rate": 0.00019166923334027765, |
| "loss": 3.356067657470703, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.4409221902017291, |
| "grad_norm": 0.8336865901947021, |
| "learning_rate": 0.00019121120509839692, |
| "loss": 3.3564964294433595, |
| "step": 2295 |
| }, |
| { |
| "epoch": 0.44188280499519694, |
| "grad_norm": 0.8085048198699951, |
| "learning_rate": 0.0001907527608522552, |
| "loss": 3.3494583129882813, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.44284341978866476, |
| "grad_norm": 1.051607370376587, |
| "learning_rate": 0.000190293905229593, |
| "loss": 3.351207733154297, |
| "step": 2305 |
| }, |
| { |
| "epoch": 0.4438040345821326, |
| "grad_norm": 0.7590753436088562, |
| "learning_rate": 0.00018983464286230327, |
| "loss": 3.3514862060546875, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.4447646493756004, |
| "grad_norm": 0.9674370288848877, |
| "learning_rate": 0.00018937497838638509, |
| "loss": 3.3499351501464845, |
| "step": 2315 |
| }, |
| { |
| "epoch": 0.4457252641690682, |
| "grad_norm": 0.8794429898262024, |
| "learning_rate": 0.0001889149164418963, |
| "loss": 3.3523765563964845, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.44668587896253603, |
| "grad_norm": 0.7959157228469849, |
| "learning_rate": 0.00018845446167290705, |
| "loss": 3.350267791748047, |
| "step": 2325 |
| }, |
| { |
| "epoch": 0.44764649375600385, |
| "grad_norm": 0.625482976436615, |
| "learning_rate": 0.0001879936187274529, |
| "loss": 3.348467254638672, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.44860710854947167, |
| "grad_norm": 1.1344565153121948, |
| "learning_rate": 0.00018753239225748796, |
| "loss": 3.35140380859375, |
| "step": 2335 |
| }, |
| { |
| "epoch": 0.4495677233429395, |
| "grad_norm": 1.0100090503692627, |
| "learning_rate": 0.0001870707869188375, |
| "loss": 3.3482070922851563, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.4505283381364073, |
| "grad_norm": 0.9724920988082886, |
| "learning_rate": 0.00018660880737115146, |
| "loss": 3.3504077911376955, |
| "step": 2345 |
| }, |
| { |
| "epoch": 0.4514889529298751, |
| "grad_norm": 0.8876535296440125, |
| "learning_rate": 0.0001861464582778572, |
| "loss": 3.352552032470703, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.45244956772334294, |
| "grad_norm": 1.3819876909255981, |
| "learning_rate": 0.00018568374430611242, |
| "loss": 3.3520408630371095, |
| "step": 2355 |
| }, |
| { |
| "epoch": 0.45341018251681076, |
| "grad_norm": 1.1260014772415161, |
| "learning_rate": 0.00018522067012675798, |
| "loss": 3.353636932373047, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.4543707973102786, |
| "grad_norm": 0.8487617373466492, |
| "learning_rate": 0.00018475724041427106, |
| "loss": 3.3487457275390624, |
| "step": 2365 |
| }, |
| { |
| "epoch": 0.4553314121037464, |
| "grad_norm": 1.024625539779663, |
| "learning_rate": 0.00018429345984671743, |
| "loss": 3.3499549865722655, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.4562920268972142, |
| "grad_norm": 0.6985915303230286, |
| "learning_rate": 0.0001838293331057048, |
| "loss": 3.349627685546875, |
| "step": 2375 |
| }, |
| { |
| "epoch": 0.457252641690682, |
| "grad_norm": 0.6404640674591064, |
| "learning_rate": 0.00018336486487633528, |
| "loss": 3.347215270996094, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.45821325648414984, |
| "grad_norm": 1.0900517702102661, |
| "learning_rate": 0.000182900059847158, |
| "loss": 3.346749114990234, |
| "step": 2385 |
| }, |
| { |
| "epoch": 0.45917387127761766, |
| "grad_norm": 0.9426947832107544, |
| "learning_rate": 0.00018243492271012202, |
| "loss": 3.3453174591064454, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.4601344860710855, |
| "grad_norm": 0.8978161215782166, |
| "learning_rate": 0.00018196945816052867, |
| "loss": 3.3470741271972657, |
| "step": 2395 |
| }, |
| { |
| "epoch": 0.4610951008645533, |
| "grad_norm": 1.0698800086975098, |
| "learning_rate": 0.00018150367089698452, |
| "loss": 3.3448417663574217, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.4620557156580211, |
| "grad_norm": 0.9989749193191528, |
| "learning_rate": 0.00018103756562135373, |
| "loss": 3.3498584747314455, |
| "step": 2405 |
| }, |
| { |
| "epoch": 0.46301633045148893, |
| "grad_norm": 1.1512004137039185, |
| "learning_rate": 0.0001805711470387105, |
| "loss": 3.344708251953125, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.46397694524495675, |
| "grad_norm": 0.7708789706230164, |
| "learning_rate": 0.00018010441985729183, |
| "loss": 3.341356658935547, |
| "step": 2415 |
| }, |
| { |
| "epoch": 0.46493756003842457, |
| "grad_norm": 0.8291172385215759, |
| "learning_rate": 0.00017963738878844966, |
| "loss": 3.3480514526367187, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.4658981748318924, |
| "grad_norm": 0.8072155714035034, |
| "learning_rate": 0.00017917005854660374, |
| "loss": 3.3407440185546875, |
| "step": 2425 |
| }, |
| { |
| "epoch": 0.4668587896253602, |
| "grad_norm": 1.1140457391738892, |
| "learning_rate": 0.00017870243384919364, |
| "loss": 3.3480880737304686, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.4678194044188281, |
| "grad_norm": 1.1302070617675781, |
| "learning_rate": 0.0001782345194166314, |
| "loss": 3.3469482421875, |
| "step": 2435 |
| }, |
| { |
| "epoch": 0.4687800192122959, |
| "grad_norm": 0.835529625415802, |
| "learning_rate": 0.00017776631997225365, |
| "loss": 3.341139221191406, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.4697406340057637, |
| "grad_norm": 0.8538889288902283, |
| "learning_rate": 0.0001772978402422742, |
| "loss": 3.342560958862305, |
| "step": 2445 |
| }, |
| { |
| "epoch": 0.47070124879923153, |
| "grad_norm": 0.904460608959198, |
| "learning_rate": 0.0001768290849557361, |
| "loss": 3.3424705505371093, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.47166186359269935, |
| "grad_norm": 0.8424951434135437, |
| "learning_rate": 0.00017636005884446397, |
| "loss": 3.3413028717041016, |
| "step": 2455 |
| }, |
| { |
| "epoch": 0.47262247838616717, |
| "grad_norm": 0.924486517906189, |
| "learning_rate": 0.00017589076664301637, |
| "loss": 3.3444580078125, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.473583093179635, |
| "grad_norm": 0.9826391935348511, |
| "learning_rate": 0.00017542121308863776, |
| "loss": 3.3431236267089846, |
| "step": 2465 |
| }, |
| { |
| "epoch": 0.4745437079731028, |
| "grad_norm": 1.2090203762054443, |
| "learning_rate": 0.00017495140292121084, |
| "loss": 3.344011688232422, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.4755043227665706, |
| "grad_norm": 0.7930108904838562, |
| "learning_rate": 0.0001744813408832088, |
| "loss": 3.3414871215820314, |
| "step": 2475 |
| }, |
| { |
| "epoch": 0.47646493756003844, |
| "grad_norm": 0.7430881857872009, |
| "learning_rate": 0.0001740110317196472, |
| "loss": 3.3419719696044923, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.47742555235350626, |
| "grad_norm": 1.0060484409332275, |
| "learning_rate": 0.0001735404801780362, |
| "loss": 3.3399085998535156, |
| "step": 2485 |
| }, |
| { |
| "epoch": 0.4783861671469741, |
| "grad_norm": 1.0594747066497803, |
| "learning_rate": 0.0001730696910083326, |
| "loss": 3.3429637908935548, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.4793467819404419, |
| "grad_norm": 0.9936187863349915, |
| "learning_rate": 0.0001725986689628921, |
| "loss": 3.3392757415771483, |
| "step": 2495 |
| }, |
| { |
| "epoch": 0.4803073967339097, |
| "grad_norm": 0.6973989009857178, |
| "learning_rate": 0.00017212741879642096, |
| "loss": 3.3416038513183595, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.4812680115273775, |
| "grad_norm": 0.5994371771812439, |
| "learning_rate": 0.00017165594526592833, |
| "loss": 3.3429832458496094, |
| "step": 2505 |
| }, |
| { |
| "epoch": 0.48222862632084534, |
| "grad_norm": 0.8716941475868225, |
| "learning_rate": 0.000171184253130678, |
| "loss": 3.339080810546875, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.48318924111431316, |
| "grad_norm": 0.7032945156097412, |
| "learning_rate": 0.00017071234715214045, |
| "loss": 3.340536880493164, |
| "step": 2515 |
| }, |
| { |
| "epoch": 0.484149855907781, |
| "grad_norm": 0.7946033477783203, |
| "learning_rate": 0.0001702402320939449, |
| "loss": 3.3366943359375, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.4851104707012488, |
| "grad_norm": 0.8639918565750122, |
| "learning_rate": 0.00016976791272183098, |
| "loss": 3.3365756988525392, |
| "step": 2525 |
| }, |
| { |
| "epoch": 0.4860710854947166, |
| "grad_norm": 0.7089393734931946, |
| "learning_rate": 0.0001692953938036008, |
| "loss": 3.3373767852783205, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.48703170028818443, |
| "grad_norm": 0.745219349861145, |
| "learning_rate": 0.00016882268010907087, |
| "loss": 3.339314270019531, |
| "step": 2535 |
| }, |
| { |
| "epoch": 0.48799231508165225, |
| "grad_norm": 0.733579695224762, |
| "learning_rate": 0.00016834977641002377, |
| "loss": 3.3369155883789063, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.48895292987512007, |
| "grad_norm": 0.808840274810791, |
| "learning_rate": 0.00016787668748016008, |
| "loss": 3.336173248291016, |
| "step": 2545 |
| }, |
| { |
| "epoch": 0.4899135446685879, |
| "grad_norm": 0.885305643081665, |
| "learning_rate": 0.00016740341809505017, |
| "loss": 3.34091796875, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.4908741594620557, |
| "grad_norm": 0.8833165764808655, |
| "learning_rate": 0.000166929973032086, |
| "loss": 3.339965057373047, |
| "step": 2555 |
| }, |
| { |
| "epoch": 0.4918347742555235, |
| "grad_norm": 0.948832094669342, |
| "learning_rate": 0.0001664563570704329, |
| "loss": 3.341967010498047, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.49279538904899134, |
| "grad_norm": 0.7376810908317566, |
| "learning_rate": 0.0001659825749909814, |
| "loss": 3.339089584350586, |
| "step": 2565 |
| }, |
| { |
| "epoch": 0.49375600384245916, |
| "grad_norm": 0.6202611327171326, |
| "learning_rate": 0.00016550863157629888, |
| "loss": 3.337108612060547, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.494716618635927, |
| "grad_norm": 1.3193303346633911, |
| "learning_rate": 0.00016503453161058123, |
| "loss": 3.337067413330078, |
| "step": 2575 |
| }, |
| { |
| "epoch": 0.4956772334293948, |
| "grad_norm": 0.9450001120567322, |
| "learning_rate": 0.00016456027987960466, |
| "loss": 3.3382850646972657, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.4966378482228626, |
| "grad_norm": 0.9356205463409424, |
| "learning_rate": 0.00016408588117067743, |
| "loss": 3.334228515625, |
| "step": 2585 |
| }, |
| { |
| "epoch": 0.49759846301633043, |
| "grad_norm": 0.9595420360565186, |
| "learning_rate": 0.00016361134027259136, |
| "loss": 3.3412109375, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.49855907780979825, |
| "grad_norm": 0.9596777558326721, |
| "learning_rate": 0.00016313666197557373, |
| "loss": 3.337894058227539, |
| "step": 2595 |
| }, |
| { |
| "epoch": 0.49951969260326606, |
| "grad_norm": 0.6967231631278992, |
| "learning_rate": 0.00016266185107123864, |
| "loss": 3.337984085083008, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.5004803073967339, |
| "grad_norm": 0.638558030128479, |
| "learning_rate": 0.00016218691235253893, |
| "loss": 3.3327606201171873, |
| "step": 2605 |
| }, |
| { |
| "epoch": 0.5014409221902018, |
| "grad_norm": 1.0894724130630493, |
| "learning_rate": 0.0001617118506137175, |
| "loss": 3.3385719299316405, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.5024015369836695, |
| "grad_norm": 0.9641280174255371, |
| "learning_rate": 0.00016123667065025914, |
| "loss": 3.3361286163330077, |
| "step": 2615 |
| }, |
| { |
| "epoch": 0.5033621517771374, |
| "grad_norm": 0.8381847143173218, |
| "learning_rate": 0.00016076137725884218, |
| "loss": 3.334253692626953, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.5043227665706052, |
| "grad_norm": 0.8258698582649231, |
| "learning_rate": 0.0001602859752372897, |
| "loss": 3.3326160430908205, |
| "step": 2625 |
| }, |
| { |
| "epoch": 0.505283381364073, |
| "grad_norm": 0.6313220262527466, |
| "learning_rate": 0.00015981046938452146, |
| "loss": 3.3321769714355467, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.5062439961575408, |
| "grad_norm": 0.6548064947128296, |
| "learning_rate": 0.0001593348645005054, |
| "loss": 3.3319812774658204, |
| "step": 2635 |
| }, |
| { |
| "epoch": 0.5072046109510087, |
| "grad_norm": 0.9846540093421936, |
| "learning_rate": 0.00015885916538620906, |
| "loss": 3.3337692260742187, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.5081652257444764, |
| "grad_norm": 0.9483662843704224, |
| "learning_rate": 0.00015838337684355121, |
| "loss": 3.335358810424805, |
| "step": 2645 |
| }, |
| { |
| "epoch": 0.5091258405379443, |
| "grad_norm": 0.7134802341461182, |
| "learning_rate": 0.0001579075036753533, |
| "loss": 3.333547592163086, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.5100864553314121, |
| "grad_norm": 0.6113592386245728, |
| "learning_rate": 0.00015743155068529102, |
| "loss": 3.3339920043945312, |
| "step": 2655 |
| }, |
| { |
| "epoch": 0.5110470701248799, |
| "grad_norm": 0.7905352115631104, |
| "learning_rate": 0.0001569555226778459, |
| "loss": 3.3328788757324217, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.5120076849183477, |
| "grad_norm": 0.7155654430389404, |
| "learning_rate": 0.0001564794244582567, |
| "loss": 3.3339805603027344, |
| "step": 2665 |
| }, |
| { |
| "epoch": 0.5129682997118156, |
| "grad_norm": 0.6985665559768677, |
| "learning_rate": 0.0001560032608324709, |
| "loss": 3.332271194458008, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.5139289145052833, |
| "grad_norm": 0.9106906056404114, |
| "learning_rate": 0.00015552703660709618, |
| "loss": 3.3307167053222657, |
| "step": 2675 |
| }, |
| { |
| "epoch": 0.5148895292987512, |
| "grad_norm": 1.00930917263031, |
| "learning_rate": 0.00015505075658935207, |
| "loss": 3.3333648681640624, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.515850144092219, |
| "grad_norm": 0.740381121635437, |
| "learning_rate": 0.0001545744255870212, |
| "loss": 3.335179901123047, |
| "step": 2685 |
| }, |
| { |
| "epoch": 0.5168107588856868, |
| "grad_norm": 0.5899009704589844, |
| "learning_rate": 0.00015409804840840088, |
| "loss": 3.3299629211425783, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.5177713736791547, |
| "grad_norm": 1.0662180185317993, |
| "learning_rate": 0.0001536216298622545, |
| "loss": 3.332952880859375, |
| "step": 2695 |
| }, |
| { |
| "epoch": 0.5187319884726225, |
| "grad_norm": 0.8961836099624634, |
| "learning_rate": 0.00015314517475776318, |
| "loss": 3.331147003173828, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.5196926032660903, |
| "grad_norm": 0.8946348428726196, |
| "learning_rate": 0.00015266868790447685, |
| "loss": 3.3304420471191407, |
| "step": 2705 |
| }, |
| { |
| "epoch": 0.5206532180595581, |
| "grad_norm": 1.0596193075180054, |
| "learning_rate": 0.0001521921741122661, |
| "loss": 3.3243175506591798, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.521613832853026, |
| "grad_norm": 1.1533540487289429, |
| "learning_rate": 0.00015171563819127342, |
| "loss": 3.329667663574219, |
| "step": 2715 |
| }, |
| { |
| "epoch": 0.5225744476464937, |
| "grad_norm": 0.8710746765136719, |
| "learning_rate": 0.00015123908495186464, |
| "loss": 3.3310035705566405, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.5235350624399616, |
| "grad_norm": 0.7002533078193665, |
| "learning_rate": 0.0001507625192045804, |
| "loss": 3.3324230194091795, |
| "step": 2725 |
| }, |
| { |
| "epoch": 0.5244956772334294, |
| "grad_norm": 1.0313575267791748, |
| "learning_rate": 0.00015028594576008773, |
| "loss": 3.3279163360595705, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.5254562920268973, |
| "grad_norm": 0.832625687122345, |
| "learning_rate": 0.00014980936942913113, |
| "loss": 3.331475830078125, |
| "step": 2735 |
| }, |
| { |
| "epoch": 0.526416906820365, |
| "grad_norm": 1.0033438205718994, |
| "learning_rate": 0.00014933279502248444, |
| "loss": 3.326156234741211, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.5273775216138329, |
| "grad_norm": 0.5835141539573669, |
| "learning_rate": 0.000148856227350902, |
| "loss": 3.325116729736328, |
| "step": 2745 |
| }, |
| { |
| "epoch": 0.5283381364073007, |
| "grad_norm": 0.5704639554023743, |
| "learning_rate": 0.00014837967122507015, |
| "loss": 3.3299407958984375, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.5292987512007685, |
| "grad_norm": 1.1036860942840576, |
| "learning_rate": 0.00014790313145555863, |
| "loss": 3.327469253540039, |
| "step": 2755 |
| }, |
| { |
| "epoch": 0.5302593659942363, |
| "grad_norm": 0.8866048455238342, |
| "learning_rate": 0.00014742661285277228, |
| "loss": 3.3271484375, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.5312199807877042, |
| "grad_norm": 0.8478221297264099, |
| "learning_rate": 0.00014695012022690205, |
| "loss": 3.3288875579833985, |
| "step": 2765 |
| }, |
| { |
| "epoch": 0.5321805955811719, |
| "grad_norm": 0.6032402515411377, |
| "learning_rate": 0.00014647365838787673, |
| "loss": 3.3231765747070314, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.5331412103746398, |
| "grad_norm": 1.1227121353149414, |
| "learning_rate": 0.00014599723214531434, |
| "loss": 3.326404571533203, |
| "step": 2775 |
| }, |
| { |
| "epoch": 0.5341018251681076, |
| "grad_norm": 0.9632681012153625, |
| "learning_rate": 0.0001455208463084737, |
| "loss": 3.32501106262207, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.5350624399615754, |
| "grad_norm": 0.6381242871284485, |
| "learning_rate": 0.00014504450568620557, |
| "loss": 3.328451919555664, |
| "step": 2785 |
| }, |
| { |
| "epoch": 0.5360230547550432, |
| "grad_norm": 0.818413496017456, |
| "learning_rate": 0.00014456821508690432, |
| "loss": 3.325485610961914, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.5369836695485111, |
| "grad_norm": 0.9857754707336426, |
| "learning_rate": 0.0001440919793184595, |
| "loss": 3.323963928222656, |
| "step": 2795 |
| }, |
| { |
| "epoch": 0.5379442843419788, |
| "grad_norm": 0.7225484848022461, |
| "learning_rate": 0.00014361580318820696, |
| "loss": 3.3251392364501955, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.5389048991354467, |
| "grad_norm": 0.8891339898109436, |
| "learning_rate": 0.00014313969150288083, |
| "loss": 3.3246475219726563, |
| "step": 2805 |
| }, |
| { |
| "epoch": 0.5398655139289145, |
| "grad_norm": 1.0610110759735107, |
| "learning_rate": 0.00014266364906856442, |
| "loss": 3.3243381500244142, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.5408261287223823, |
| "grad_norm": 0.7058207392692566, |
| "learning_rate": 0.00014218768069064214, |
| "loss": 3.32562141418457, |
| "step": 2815 |
| }, |
| { |
| "epoch": 0.5417867435158501, |
| "grad_norm": 1.1421717405319214, |
| "learning_rate": 0.00014171179117375082, |
| "loss": 3.3252777099609374, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.542747358309318, |
| "grad_norm": 0.7852960228919983, |
| "learning_rate": 0.00014123598532173113, |
| "loss": 3.3269309997558594, |
| "step": 2825 |
| }, |
| { |
| "epoch": 0.5437079731027857, |
| "grad_norm": 0.9386373162269592, |
| "learning_rate": 0.00014076026793757943, |
| "loss": 3.325482940673828, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.5446685878962536, |
| "grad_norm": 0.9038012027740479, |
| "learning_rate": 0.00014028464382339877, |
| "loss": 3.3256744384765624, |
| "step": 2835 |
| }, |
| { |
| "epoch": 0.5456292026897214, |
| "grad_norm": 0.6614556312561035, |
| "learning_rate": 0.0001398091177803509, |
| "loss": 3.3238311767578126, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.5465898174831892, |
| "grad_norm": 0.7346850037574768, |
| "learning_rate": 0.00013933369460860748, |
| "loss": 3.3215705871582033, |
| "step": 2845 |
| }, |
| { |
| "epoch": 0.547550432276657, |
| "grad_norm": 0.9192922115325928, |
| "learning_rate": 0.00013885837910730168, |
| "loss": 3.3206497192382813, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.5485110470701249, |
| "grad_norm": 0.8936217427253723, |
| "learning_rate": 0.00013838317607448004, |
| "loss": 3.325290298461914, |
| "step": 2855 |
| }, |
| { |
| "epoch": 0.5494716618635928, |
| "grad_norm": 0.9450289607048035, |
| "learning_rate": 0.00013790809030705354, |
| "loss": 3.324379730224609, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.5504322766570605, |
| "grad_norm": 0.868273913860321, |
| "learning_rate": 0.00013743312660074962, |
| "loss": 3.3241962432861327, |
| "step": 2865 |
| }, |
| { |
| "epoch": 0.5513928914505284, |
| "grad_norm": 0.6488667130470276, |
| "learning_rate": 0.00013695828975006336, |
| "loss": 3.3240901947021486, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.5523535062439962, |
| "grad_norm": 0.4751059114933014, |
| "learning_rate": 0.00013648358454820957, |
| "loss": 3.3235042572021483, |
| "step": 2875 |
| }, |
| { |
| "epoch": 0.553314121037464, |
| "grad_norm": 1.1512478590011597, |
| "learning_rate": 0.00013600901578707402, |
| "loss": 3.3230491638183595, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.5542747358309318, |
| "grad_norm": 0.8698762059211731, |
| "learning_rate": 0.00013553458825716502, |
| "loss": 3.324365997314453, |
| "step": 2885 |
| }, |
| { |
| "epoch": 0.5552353506243997, |
| "grad_norm": 0.8694410920143127, |
| "learning_rate": 0.0001350603067475655, |
| "loss": 3.324441146850586, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.5561959654178674, |
| "grad_norm": 0.8007134199142456, |
| "learning_rate": 0.00013458617604588418, |
| "loss": 3.3229400634765627, |
| "step": 2895 |
| }, |
| { |
| "epoch": 0.5571565802113353, |
| "grad_norm": 0.765385091304779, |
| "learning_rate": 0.00013411220093820773, |
| "loss": 3.323167419433594, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.5581171950048031, |
| "grad_norm": 0.4617004096508026, |
| "learning_rate": 0.00013363838620905188, |
| "loss": 3.320812225341797, |
| "step": 2905 |
| }, |
| { |
| "epoch": 0.5590778097982709, |
| "grad_norm": 0.7083408832550049, |
| "learning_rate": 0.00013316473664131347, |
| "loss": 3.324010467529297, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.5600384245917387, |
| "grad_norm": 0.8876819014549255, |
| "learning_rate": 0.00013269125701622243, |
| "loss": 3.3241298675537108, |
| "step": 2915 |
| }, |
| { |
| "epoch": 0.5609990393852066, |
| "grad_norm": 0.806339681148529, |
| "learning_rate": 0.00013221795211329281, |
| "loss": 3.319907379150391, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.5619596541786743, |
| "grad_norm": 0.6743384599685669, |
| "learning_rate": 0.00013174482671027526, |
| "loss": 3.319793701171875, |
| "step": 2925 |
| }, |
| { |
| "epoch": 0.5629202689721422, |
| "grad_norm": 0.6159988641738892, |
| "learning_rate": 0.00013127188558310823, |
| "loss": 3.323670196533203, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.56388088376561, |
| "grad_norm": 0.9810566902160645, |
| "learning_rate": 0.0001307991335058702, |
| "loss": 3.3209304809570312, |
| "step": 2935 |
| }, |
| { |
| "epoch": 0.5648414985590778, |
| "grad_norm": 0.913245677947998, |
| "learning_rate": 0.00013032657525073122, |
| "loss": 3.320903778076172, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.5658021133525456, |
| "grad_norm": 0.6053956151008606, |
| "learning_rate": 0.00012985421558790473, |
| "loss": 3.3186370849609377, |
| "step": 2945 |
| }, |
| { |
| "epoch": 0.5667627281460135, |
| "grad_norm": 0.6618540287017822, |
| "learning_rate": 0.00012938205928559964, |
| "loss": 3.3157257080078124, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.5677233429394812, |
| "grad_norm": 0.5338152647018433, |
| "learning_rate": 0.00012891011110997187, |
| "loss": 3.3171661376953123, |
| "step": 2955 |
| }, |
| { |
| "epoch": 0.5686839577329491, |
| "grad_norm": 0.7981716394424438, |
| "learning_rate": 0.0001284383758250767, |
| "loss": 3.3200511932373047, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.5696445725264169, |
| "grad_norm": 0.8445151448249817, |
| "learning_rate": 0.00012796685819282009, |
| "loss": 3.319135284423828, |
| "step": 2965 |
| }, |
| { |
| "epoch": 0.5706051873198847, |
| "grad_norm": 0.6141214370727539, |
| "learning_rate": 0.0001274955629729111, |
| "loss": 3.3177207946777343, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.5715658021133525, |
| "grad_norm": 1.2424954175949097, |
| "learning_rate": 0.0001270244949228136, |
| "loss": 3.322146987915039, |
| "step": 2975 |
| }, |
| { |
| "epoch": 0.5725264169068204, |
| "grad_norm": 0.7265079021453857, |
| "learning_rate": 0.00012655365879769826, |
| "loss": 3.3195030212402346, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.5734870317002881, |
| "grad_norm": 0.5697556734085083, |
| "learning_rate": 0.00012608305935039475, |
| "loss": 3.319179153442383, |
| "step": 2985 |
| }, |
| { |
| "epoch": 0.574447646493756, |
| "grad_norm": 0.6604142189025879, |
| "learning_rate": 0.00012561270133134344, |
| "loss": 3.317556381225586, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.5754082612872238, |
| "grad_norm": 0.6255987882614136, |
| "learning_rate": 0.00012514258948854773, |
| "loss": 3.3171707153320313, |
| "step": 2995 |
| }, |
| { |
| "epoch": 0.5763688760806917, |
| "grad_norm": 0.6028192639350891, |
| "learning_rate": 0.00012467272856752593, |
| "loss": 3.317168426513672, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.5773294908741594, |
| "grad_norm": 0.8794112801551819, |
| "learning_rate": 0.0001242031233112634, |
| "loss": 3.3190616607666015, |
| "step": 3005 |
| }, |
| { |
| "epoch": 0.5782901056676273, |
| "grad_norm": 0.6320680975914001, |
| "learning_rate": 0.00012373377846016493, |
| "loss": 3.315277099609375, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.579250720461095, |
| "grad_norm": 0.5714786052703857, |
| "learning_rate": 0.0001232646987520064, |
| "loss": 3.3192188262939455, |
| "step": 3015 |
| }, |
| { |
| "epoch": 0.5802113352545629, |
| "grad_norm": 0.7069401144981384, |
| "learning_rate": 0.00012279588892188736, |
| "loss": 3.317308807373047, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.5811719500480308, |
| "grad_norm": 0.7392922639846802, |
| "learning_rate": 0.000122327353702183, |
| "loss": 3.317064666748047, |
| "step": 3025 |
| }, |
| { |
| "epoch": 0.5821325648414986, |
| "grad_norm": 1.0391535758972168, |
| "learning_rate": 0.0001218590978224966, |
| "loss": 3.313740539550781, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.5830931796349664, |
| "grad_norm": 0.7290921211242676, |
| "learning_rate": 0.00012139112600961159, |
| "loss": 3.317361831665039, |
| "step": 3035 |
| }, |
| { |
| "epoch": 0.5840537944284342, |
| "grad_norm": 0.5767372250556946, |
| "learning_rate": 0.00012092344298744383, |
| "loss": 3.3169986724853517, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.5850144092219021, |
| "grad_norm": 1.0307767391204834, |
| "learning_rate": 0.00012045605347699411, |
| "loss": 3.3194618225097656, |
| "step": 3045 |
| }, |
| { |
| "epoch": 0.5859750240153698, |
| "grad_norm": 0.622552752494812, |
| "learning_rate": 0.00011998896219630029, |
| "loss": 3.316600799560547, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.5869356388088377, |
| "grad_norm": 0.5956411957740784, |
| "learning_rate": 0.0001195221738603899, |
| "loss": 3.312654495239258, |
| "step": 3055 |
| }, |
| { |
| "epoch": 0.5878962536023055, |
| "grad_norm": 0.7186912298202515, |
| "learning_rate": 0.00011905569318123223, |
| "loss": 3.316633605957031, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.5888568683957733, |
| "grad_norm": 0.6377474069595337, |
| "learning_rate": 0.00011858952486769114, |
| "loss": 3.315812683105469, |
| "step": 3065 |
| }, |
| { |
| "epoch": 0.5898174831892411, |
| "grad_norm": 0.6208492517471313, |
| "learning_rate": 0.00011812367362547716, |
| "loss": 3.3148754119873045, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.590778097982709, |
| "grad_norm": 0.600408673286438, |
| "learning_rate": 0.0001176581441571002, |
| "loss": 3.3176692962646483, |
| "step": 3075 |
| }, |
| { |
| "epoch": 0.5917387127761767, |
| "grad_norm": 0.8082125782966614, |
| "learning_rate": 0.00011719294116182217, |
| "loss": 3.314889144897461, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.5926993275696446, |
| "grad_norm": 0.6132190823554993, |
| "learning_rate": 0.00011672806933560925, |
| "loss": 3.3137298583984376, |
| "step": 3085 |
| }, |
| { |
| "epoch": 0.5936599423631124, |
| "grad_norm": 0.6647592782974243, |
| "learning_rate": 0.00011626353337108476, |
| "loss": 3.3115211486816407, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.5946205571565802, |
| "grad_norm": 0.5742537379264832, |
| "learning_rate": 0.00011579933795748164, |
| "loss": 3.3127456665039063, |
| "step": 3095 |
| }, |
| { |
| "epoch": 0.595581171950048, |
| "grad_norm": 0.6123734712600708, |
| "learning_rate": 0.00011533548778059508, |
| "loss": 3.315221405029297, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.5965417867435159, |
| "grad_norm": 0.8112967014312744, |
| "learning_rate": 0.00011487198752273552, |
| "loss": 3.315155792236328, |
| "step": 3105 |
| }, |
| { |
| "epoch": 0.5975024015369836, |
| "grad_norm": 1.0377261638641357, |
| "learning_rate": 0.0001144088418626809, |
| "loss": 3.31671142578125, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.5984630163304515, |
| "grad_norm": 0.6357848644256592, |
| "learning_rate": 0.00011394605547562989, |
| "loss": 3.314447784423828, |
| "step": 3115 |
| }, |
| { |
| "epoch": 0.5994236311239193, |
| "grad_norm": 0.5305024981498718, |
| "learning_rate": 0.00011348363303315434, |
| "loss": 3.313646697998047, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.6003842459173871, |
| "grad_norm": 0.6368167400360107, |
| "learning_rate": 0.00011302157920315244, |
| "loss": 3.3107112884521483, |
| "step": 3125 |
| }, |
| { |
| "epoch": 0.6013448607108549, |
| "grad_norm": 0.6398528218269348, |
| "learning_rate": 0.00011255989864980133, |
| "loss": 3.3126060485839846, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.6023054755043228, |
| "grad_norm": 0.7452788352966309, |
| "learning_rate": 0.00011209859603351015, |
| "loss": 3.3152156829833985, |
| "step": 3135 |
| }, |
| { |
| "epoch": 0.6032660902977905, |
| "grad_norm": 0.6287251710891724, |
| "learning_rate": 0.00011163767601087301, |
| "loss": 3.3126220703125, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.6042267050912584, |
| "grad_norm": 0.6889411807060242, |
| "learning_rate": 0.00011117714323462186, |
| "loss": 3.3144298553466798, |
| "step": 3145 |
| }, |
| { |
| "epoch": 0.6051873198847262, |
| "grad_norm": 0.6560638546943665, |
| "learning_rate": 0.00011071700235357979, |
| "loss": 3.311948776245117, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.6061479346781941, |
| "grad_norm": 0.4844714403152466, |
| "learning_rate": 0.00011025725801261373, |
| "loss": 3.3093055725097655, |
| "step": 3155 |
| }, |
| { |
| "epoch": 0.6071085494716618, |
| "grad_norm": 0.8295413255691528, |
| "learning_rate": 0.00010979791485258788, |
| "loss": 3.3112449645996094, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.6080691642651297, |
| "grad_norm": 0.8754414916038513, |
| "learning_rate": 0.00010933897751031671, |
| "loss": 3.311637115478516, |
| "step": 3165 |
| }, |
| { |
| "epoch": 0.6090297790585975, |
| "grad_norm": 0.5046402812004089, |
| "learning_rate": 0.00010888045061851807, |
| "loss": 3.309844207763672, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.6099903938520653, |
| "grad_norm": 0.6311172246932983, |
| "learning_rate": 0.00010842233880576681, |
| "loss": 3.3105926513671875, |
| "step": 3175 |
| }, |
| { |
| "epoch": 0.6109510086455331, |
| "grad_norm": 0.6714319586753845, |
| "learning_rate": 0.0001079646466964475, |
| "loss": 3.308180236816406, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.611911623439001, |
| "grad_norm": 0.6159961223602295, |
| "learning_rate": 0.00010750737891070824, |
| "loss": 3.3096988677978514, |
| "step": 3185 |
| }, |
| { |
| "epoch": 0.6128722382324687, |
| "grad_norm": 0.5705515742301941, |
| "learning_rate": 0.00010705054006441371, |
| "loss": 3.3115421295166017, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.6138328530259366, |
| "grad_norm": 0.49575576186180115, |
| "learning_rate": 0.00010659413476909865, |
| "loss": 3.3115379333496096, |
| "step": 3195 |
| }, |
| { |
| "epoch": 0.6147934678194045, |
| "grad_norm": 1.0461304187774658, |
| "learning_rate": 0.00010613816763192152, |
| "loss": 3.3100265502929687, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.6157540826128722, |
| "grad_norm": 0.6978023648262024, |
| "learning_rate": 0.00010568264325561763, |
| "loss": 3.308924102783203, |
| "step": 3205 |
| }, |
| { |
| "epoch": 0.6167146974063401, |
| "grad_norm": 0.7631133198738098, |
| "learning_rate": 0.000105227566238453, |
| "loss": 3.3145515441894533, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.6176753121998079, |
| "grad_norm": 0.8663102388381958, |
| "learning_rate": 0.00010477294117417762, |
| "loss": 3.3145401000976564, |
| "step": 3215 |
| }, |
| { |
| "epoch": 0.6186359269932757, |
| "grad_norm": 0.4759737551212311, |
| "learning_rate": 0.00010431877265197955, |
| "loss": 3.311273193359375, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.6195965417867435, |
| "grad_norm": 0.47145599126815796, |
| "learning_rate": 0.00010386506525643808, |
| "loss": 3.306778335571289, |
| "step": 3225 |
| }, |
| { |
| "epoch": 0.6205571565802114, |
| "grad_norm": 0.6582973599433899, |
| "learning_rate": 0.00010341182356747771, |
| "loss": 3.3100128173828125, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.6215177713736791, |
| "grad_norm": 0.6123160123825073, |
| "learning_rate": 0.00010295905216032203, |
| "loss": 3.3093284606933593, |
| "step": 3235 |
| }, |
| { |
| "epoch": 0.622478386167147, |
| "grad_norm": 0.5266756415367126, |
| "learning_rate": 0.00010250675560544717, |
| "loss": 3.3095718383789063, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.6234390009606148, |
| "grad_norm": 0.7521538138389587, |
| "learning_rate": 0.00010205493846853618, |
| "loss": 3.307404327392578, |
| "step": 3245 |
| }, |
| { |
| "epoch": 0.6243996157540826, |
| "grad_norm": 0.5699673891067505, |
| "learning_rate": 0.00010160360531043239, |
| "loss": 3.3089645385742186, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.6253602305475504, |
| "grad_norm": 0.7614957690238953, |
| "learning_rate": 0.00010115276068709377, |
| "loss": 3.3093006134033205, |
| "step": 3255 |
| }, |
| { |
| "epoch": 0.6263208453410183, |
| "grad_norm": 0.652992844581604, |
| "learning_rate": 0.00010070240914954676, |
| "loss": 3.306406784057617, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.627281460134486, |
| "grad_norm": 0.5600461363792419, |
| "learning_rate": 0.00010025255524384033, |
| "loss": 3.3075355529785155, |
| "step": 3265 |
| }, |
| { |
| "epoch": 0.6282420749279539, |
| "grad_norm": 0.7442777156829834, |
| "learning_rate": 9.980320351100028e-05, |
| "loss": 3.308829498291016, |
| "step": 3270 |
| }, |
| { |
| "epoch": 0.6292026897214217, |
| "grad_norm": 0.8243304491043091, |
| "learning_rate": 9.935435848698307e-05, |
| "loss": 3.307141876220703, |
| "step": 3275 |
| }, |
| { |
| "epoch": 0.6301633045148896, |
| "grad_norm": 0.6337246298789978, |
| "learning_rate": 9.890602470263037e-05, |
| "loss": 3.311864471435547, |
| "step": 3280 |
| }, |
| { |
| "epoch": 0.6311239193083573, |
| "grad_norm": 0.6521264910697937, |
| "learning_rate": 9.845820668362308e-05, |
| "loss": 3.308788299560547, |
| "step": 3285 |
| }, |
| { |
| "epoch": 0.6320845341018252, |
| "grad_norm": 0.6866946220397949, |
| "learning_rate": 9.801090895043566e-05, |
| "loss": 3.307585906982422, |
| "step": 3290 |
| }, |
| { |
| "epoch": 0.633045148895293, |
| "grad_norm": 0.6082757711410522, |
| "learning_rate": 9.756413601829083e-05, |
| "loss": 3.307001495361328, |
| "step": 3295 |
| }, |
| { |
| "epoch": 0.6340057636887608, |
| "grad_norm": 0.515135645866394, |
| "learning_rate": 9.711789239711344e-05, |
| "loss": 3.308676528930664, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.6349663784822286, |
| "grad_norm": 0.503957211971283, |
| "learning_rate": 9.667218259148547e-05, |
| "loss": 3.3088623046875, |
| "step": 3305 |
| }, |
| { |
| "epoch": 0.6359269932756965, |
| "grad_norm": 0.663313627243042, |
| "learning_rate": 9.62270111006001e-05, |
| "loss": 3.305764007568359, |
| "step": 3310 |
| }, |
| { |
| "epoch": 0.6368876080691642, |
| "grad_norm": 0.7052382826805115, |
| "learning_rate": 9.57823824182168e-05, |
| "loss": 3.305963134765625, |
| "step": 3315 |
| }, |
| { |
| "epoch": 0.6378482228626321, |
| "grad_norm": 0.5659816861152649, |
| "learning_rate": 9.53383010326155e-05, |
| "loss": 3.306070327758789, |
| "step": 3320 |
| }, |
| { |
| "epoch": 0.6388088376560999, |
| "grad_norm": 0.46071678400039673, |
| "learning_rate": 9.489477142655147e-05, |
| "loss": 3.3049732208251954, |
| "step": 3325 |
| }, |
| { |
| "epoch": 0.6397694524495677, |
| "grad_norm": 0.9557482004165649, |
| "learning_rate": 9.445179807721012e-05, |
| "loss": 3.3062675476074217, |
| "step": 3330 |
| }, |
| { |
| "epoch": 0.6407300672430355, |
| "grad_norm": 0.8419716954231262, |
| "learning_rate": 9.400938545616173e-05, |
| "loss": 3.308432769775391, |
| "step": 3335 |
| }, |
| { |
| "epoch": 0.6416906820365034, |
| "grad_norm": 0.5912371873855591, |
| "learning_rate": 9.356753802931646e-05, |
| "loss": 3.3038196563720703, |
| "step": 3340 |
| }, |
| { |
| "epoch": 0.6426512968299711, |
| "grad_norm": 0.4956425726413727, |
| "learning_rate": 9.312626025687897e-05, |
| "loss": 3.3066734313964843, |
| "step": 3345 |
| }, |
| { |
| "epoch": 0.643611911623439, |
| "grad_norm": 0.5250487923622131, |
| "learning_rate": 9.268555659330364e-05, |
| "loss": 3.307097625732422, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.6445725264169068, |
| "grad_norm": 0.5282003283500671, |
| "learning_rate": 9.22454314872496e-05, |
| "loss": 3.305289459228516, |
| "step": 3355 |
| }, |
| { |
| "epoch": 0.6455331412103746, |
| "grad_norm": 0.4647374749183655, |
| "learning_rate": 9.180588938153566e-05, |
| "loss": 3.3030609130859374, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.6464937560038425, |
| "grad_norm": 1.0786409378051758, |
| "learning_rate": 9.136693471309568e-05, |
| "loss": 3.311345672607422, |
| "step": 3365 |
| }, |
| { |
| "epoch": 0.6474543707973103, |
| "grad_norm": 0.6998493075370789, |
| "learning_rate": 9.092857191293356e-05, |
| "loss": 3.3091506958007812, |
| "step": 3370 |
| }, |
| { |
| "epoch": 0.6484149855907781, |
| "grad_norm": 0.446397066116333, |
| "learning_rate": 9.049080540607875e-05, |
| "loss": 3.3034263610839845, |
| "step": 3375 |
| }, |
| { |
| "epoch": 0.6493756003842459, |
| "grad_norm": 0.6110076308250427, |
| "learning_rate": 9.005363961154126e-05, |
| "loss": 3.3031074523925783, |
| "step": 3380 |
| }, |
| { |
| "epoch": 0.6503362151777138, |
| "grad_norm": 0.4611775279045105, |
| "learning_rate": 8.961707894226735e-05, |
| "loss": 3.3057632446289062, |
| "step": 3385 |
| }, |
| { |
| "epoch": 0.6512968299711815, |
| "grad_norm": 0.8725507855415344, |
| "learning_rate": 8.918112780509494e-05, |
| "loss": 3.3007755279541016, |
| "step": 3390 |
| }, |
| { |
| "epoch": 0.6522574447646494, |
| "grad_norm": 0.8378633856773376, |
| "learning_rate": 8.874579060070894e-05, |
| "loss": 3.3028465270996095, |
| "step": 3395 |
| }, |
| { |
| "epoch": 0.6532180595581172, |
| "grad_norm": 0.6364403367042542, |
| "learning_rate": 8.831107172359707e-05, |
| "loss": 3.3054805755615235, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.654178674351585, |
| "grad_norm": 0.46224355697631836, |
| "learning_rate": 8.787697556200519e-05, |
| "loss": 3.304944610595703, |
| "step": 3405 |
| }, |
| { |
| "epoch": 0.6551392891450528, |
| "grad_norm": 0.5552151799201965, |
| "learning_rate": 8.744350649789347e-05, |
| "loss": 3.301805114746094, |
| "step": 3410 |
| }, |
| { |
| "epoch": 0.6560999039385207, |
| "grad_norm": 0.6075546741485596, |
| "learning_rate": 8.701066890689166e-05, |
| "loss": 3.300829315185547, |
| "step": 3415 |
| }, |
| { |
| "epoch": 0.6570605187319885, |
| "grad_norm": 0.46151185035705566, |
| "learning_rate": 8.657846715825508e-05, |
| "loss": 3.30389289855957, |
| "step": 3420 |
| }, |
| { |
| "epoch": 0.6580211335254563, |
| "grad_norm": 0.5615620017051697, |
| "learning_rate": 8.61469056148209e-05, |
| "loss": 3.302476501464844, |
| "step": 3425 |
| }, |
| { |
| "epoch": 0.6589817483189241, |
| "grad_norm": 0.47145524621009827, |
| "learning_rate": 8.571598863296342e-05, |
| "loss": 3.303362274169922, |
| "step": 3430 |
| }, |
| { |
| "epoch": 0.659942363112392, |
| "grad_norm": 0.4290817081928253, |
| "learning_rate": 8.528572056255065e-05, |
| "loss": 3.3021663665771483, |
| "step": 3435 |
| }, |
| { |
| "epoch": 0.6609029779058597, |
| "grad_norm": 0.5213423371315002, |
| "learning_rate": 8.485610574690021e-05, |
| "loss": 3.3011238098144533, |
| "step": 3440 |
| }, |
| { |
| "epoch": 0.6618635926993276, |
| "grad_norm": 0.661228358745575, |
| "learning_rate": 8.442714852273523e-05, |
| "loss": 3.3015186309814455, |
| "step": 3445 |
| }, |
| { |
| "epoch": 0.6628242074927954, |
| "grad_norm": 0.9645004868507385, |
| "learning_rate": 8.399885322014123e-05, |
| "loss": 3.3042228698730467, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.6637848222862632, |
| "grad_norm": 0.5251511335372925, |
| "learning_rate": 8.35712241625216e-05, |
| "loss": 3.3040817260742186, |
| "step": 3455 |
| }, |
| { |
| "epoch": 0.664745437079731, |
| "grad_norm": 0.45383018255233765, |
| "learning_rate": 8.314426566655458e-05, |
| "loss": 3.3012584686279296, |
| "step": 3460 |
| }, |
| { |
| "epoch": 0.6657060518731989, |
| "grad_norm": 0.3945951461791992, |
| "learning_rate": 8.271798204214942e-05, |
| "loss": 3.3004776000976563, |
| "step": 3465 |
| }, |
| { |
| "epoch": 0.6666666666666666, |
| "grad_norm": 0.4676642119884491, |
| "learning_rate": 8.229237759240289e-05, |
| "loss": 3.2975730895996094, |
| "step": 3470 |
| }, |
| { |
| "epoch": 0.6676272814601345, |
| "grad_norm": 0.5338938236236572, |
| "learning_rate": 8.186745661355595e-05, |
| "loss": 3.297288513183594, |
| "step": 3475 |
| }, |
| { |
| "epoch": 0.6685878962536023, |
| "grad_norm": 0.6830979585647583, |
| "learning_rate": 8.144322339495012e-05, |
| "loss": 3.3035240173339844, |
| "step": 3480 |
| }, |
| { |
| "epoch": 0.6695485110470701, |
| "grad_norm": 0.4912271201610565, |
| "learning_rate": 8.101968221898453e-05, |
| "loss": 3.300017547607422, |
| "step": 3485 |
| }, |
| { |
| "epoch": 0.6705091258405379, |
| "grad_norm": 0.5433835387229919, |
| "learning_rate": 8.059683736107245e-05, |
| "loss": 3.304372787475586, |
| "step": 3490 |
| }, |
| { |
| "epoch": 0.6714697406340058, |
| "grad_norm": 0.43095162510871887, |
| "learning_rate": 8.017469308959823e-05, |
| "loss": 3.302097702026367, |
| "step": 3495 |
| }, |
| { |
| "epoch": 0.6724303554274735, |
| "grad_norm": 0.5721524953842163, |
| "learning_rate": 7.97532536658742e-05, |
| "loss": 3.302288055419922, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.6733909702209414, |
| "grad_norm": 0.8145022392272949, |
| "learning_rate": 7.933252334409766e-05, |
| "loss": 3.302367401123047, |
| "step": 3505 |
| }, |
| { |
| "epoch": 0.6743515850144092, |
| "grad_norm": 0.41017457842826843, |
| "learning_rate": 7.891250637130779e-05, |
| "loss": 3.3008522033691405, |
| "step": 3510 |
| }, |
| { |
| "epoch": 0.675312199807877, |
| "grad_norm": 0.4966427981853485, |
| "learning_rate": 7.849320698734306e-05, |
| "loss": 3.305078887939453, |
| "step": 3515 |
| }, |
| { |
| "epoch": 0.6762728146013448, |
| "grad_norm": 0.44679656624794006, |
| "learning_rate": 7.80746294247982e-05, |
| "loss": 3.297787094116211, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.6772334293948127, |
| "grad_norm": 0.5052957534790039, |
| "learning_rate": 7.765677790898155e-05, |
| "loss": 3.2971946716308596, |
| "step": 3525 |
| }, |
| { |
| "epoch": 0.6781940441882806, |
| "grad_norm": 0.5869050025939941, |
| "learning_rate": 7.723965665787255e-05, |
| "loss": 3.3033409118652344, |
| "step": 3530 |
| }, |
| { |
| "epoch": 0.6791546589817483, |
| "grad_norm": 0.5643454194068909, |
| "learning_rate": 7.682326988207877e-05, |
| "loss": 3.3006324768066406, |
| "step": 3535 |
| }, |
| { |
| "epoch": 0.6801152737752162, |
| "grad_norm": 0.45685872435569763, |
| "learning_rate": 7.640762178479382e-05, |
| "loss": 3.2982528686523436, |
| "step": 3540 |
| }, |
| { |
| "epoch": 0.681075888568684, |
| "grad_norm": 0.42118874192237854, |
| "learning_rate": 7.599271656175476e-05, |
| "loss": 3.303014373779297, |
| "step": 3545 |
| }, |
| { |
| "epoch": 0.6820365033621518, |
| "grad_norm": 0.482282817363739, |
| "learning_rate": 7.557855840119976e-05, |
| "loss": 3.2992012023925783, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.6829971181556196, |
| "grad_norm": 0.4499205946922302, |
| "learning_rate": 7.516515148382576e-05, |
| "loss": 3.2981277465820313, |
| "step": 3555 |
| }, |
| { |
| "epoch": 0.6839577329490875, |
| "grad_norm": 0.6683313846588135, |
| "learning_rate": 7.475249998274621e-05, |
| "loss": 3.298491668701172, |
| "step": 3560 |
| }, |
| { |
| "epoch": 0.6849183477425552, |
| "grad_norm": 0.5663706064224243, |
| "learning_rate": 7.434060806344936e-05, |
| "loss": 3.301411819458008, |
| "step": 3565 |
| }, |
| { |
| "epoch": 0.6858789625360231, |
| "grad_norm": 0.43267935514450073, |
| "learning_rate": 7.392947988375555e-05, |
| "loss": 3.29864616394043, |
| "step": 3570 |
| }, |
| { |
| "epoch": 0.6868395773294909, |
| "grad_norm": 0.43747061491012573, |
| "learning_rate": 7.351911959377585e-05, |
| "loss": 3.298960876464844, |
| "step": 3575 |
| }, |
| { |
| "epoch": 0.6878001921229587, |
| "grad_norm": 0.5621949434280396, |
| "learning_rate": 7.310953133586981e-05, |
| "loss": 3.2972442626953127, |
| "step": 3580 |
| }, |
| { |
| "epoch": 0.6887608069164265, |
| "grad_norm": 0.4505578279495239, |
| "learning_rate": 7.27007192446036e-05, |
| "loss": 3.2979637145996095, |
| "step": 3585 |
| }, |
| { |
| "epoch": 0.6897214217098944, |
| "grad_norm": 0.7796880602836609, |
| "learning_rate": 7.229268744670883e-05, |
| "loss": 3.297809600830078, |
| "step": 3590 |
| }, |
| { |
| "epoch": 0.6906820365033621, |
| "grad_norm": 0.6508268713951111, |
| "learning_rate": 7.188544006104e-05, |
| "loss": 3.2982086181640624, |
| "step": 3595 |
| }, |
| { |
| "epoch": 0.69164265129683, |
| "grad_norm": 0.44308528304100037, |
| "learning_rate": 7.147898119853367e-05, |
| "loss": 3.2976150512695312, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.6926032660902978, |
| "grad_norm": 0.4720350205898285, |
| "learning_rate": 7.107331496216676e-05, |
| "loss": 3.2995452880859375, |
| "step": 3605 |
| }, |
| { |
| "epoch": 0.6935638808837656, |
| "grad_norm": 0.45068585872650146, |
| "learning_rate": 7.066844544691474e-05, |
| "loss": 3.301127243041992, |
| "step": 3610 |
| }, |
| { |
| "epoch": 0.6945244956772334, |
| "grad_norm": 0.42395853996276855, |
| "learning_rate": 7.026437673971107e-05, |
| "loss": 3.2964126586914064, |
| "step": 3615 |
| }, |
| { |
| "epoch": 0.6954851104707013, |
| "grad_norm": 0.4400024712085724, |
| "learning_rate": 6.986111291940511e-05, |
| "loss": 3.2967502593994142, |
| "step": 3620 |
| }, |
| { |
| "epoch": 0.696445725264169, |
| "grad_norm": 0.410159170627594, |
| "learning_rate": 6.945865805672154e-05, |
| "loss": 3.2917686462402345, |
| "step": 3625 |
| }, |
| { |
| "epoch": 0.6974063400576369, |
| "grad_norm": 0.6144173741340637, |
| "learning_rate": 6.905701621421904e-05, |
| "loss": 3.300373077392578, |
| "step": 3630 |
| }, |
| { |
| "epoch": 0.6983669548511047, |
| "grad_norm": 0.47794288396835327, |
| "learning_rate": 6.865619144624914e-05, |
| "loss": 3.297341156005859, |
| "step": 3635 |
| }, |
| { |
| "epoch": 0.6993275696445725, |
| "grad_norm": 0.520939826965332, |
| "learning_rate": 6.825618779891577e-05, |
| "loss": 3.298029327392578, |
| "step": 3640 |
| }, |
| { |
| "epoch": 0.7002881844380403, |
| "grad_norm": 0.5281940698623657, |
| "learning_rate": 6.785700931003381e-05, |
| "loss": 3.291861724853516, |
| "step": 3645 |
| }, |
| { |
| "epoch": 0.7012487992315082, |
| "grad_norm": 0.3904149532318115, |
| "learning_rate": 6.745866000908874e-05, |
| "loss": 3.296683502197266, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.7022094140249759, |
| "grad_norm": 0.47106507420539856, |
| "learning_rate": 6.706114391719586e-05, |
| "loss": 3.29759521484375, |
| "step": 3655 |
| }, |
| { |
| "epoch": 0.7031700288184438, |
| "grad_norm": 0.4817742705345154, |
| "learning_rate": 6.66644650470597e-05, |
| "loss": 3.299455261230469, |
| "step": 3660 |
| }, |
| { |
| "epoch": 0.7041306436119116, |
| "grad_norm": 0.473783940076828, |
| "learning_rate": 6.626862740293338e-05, |
| "loss": 3.2957839965820312, |
| "step": 3665 |
| }, |
| { |
| "epoch": 0.7050912584053795, |
| "grad_norm": 0.5512199401855469, |
| "learning_rate": 6.587363498057845e-05, |
| "loss": 3.297314453125, |
| "step": 3670 |
| }, |
| { |
| "epoch": 0.7060518731988472, |
| "grad_norm": 0.6732028722763062, |
| "learning_rate": 6.547949176722437e-05, |
| "loss": 3.2939273834228517, |
| "step": 3675 |
| }, |
| { |
| "epoch": 0.7070124879923151, |
| "grad_norm": 0.5238297581672668, |
| "learning_rate": 6.508620174152826e-05, |
| "loss": 3.298822784423828, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.7079731027857828, |
| "grad_norm": 0.40396222472190857, |
| "learning_rate": 6.469376887353491e-05, |
| "loss": 3.29783935546875, |
| "step": 3685 |
| }, |
| { |
| "epoch": 0.7089337175792507, |
| "grad_norm": 0.4132029116153717, |
| "learning_rate": 6.430219712463634e-05, |
| "loss": 3.2966102600097655, |
| "step": 3690 |
| }, |
| { |
| "epoch": 0.7098943323727186, |
| "grad_norm": 0.4176599085330963, |
| "learning_rate": 6.391149044753238e-05, |
| "loss": 3.2970817565917967, |
| "step": 3695 |
| }, |
| { |
| "epoch": 0.7108549471661864, |
| "grad_norm": 0.3097979426383972, |
| "learning_rate": 6.352165278619012e-05, |
| "loss": 3.294615936279297, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.7118155619596542, |
| "grad_norm": 0.4419673979282379, |
| "learning_rate": 6.313268807580462e-05, |
| "loss": 3.2989078521728517, |
| "step": 3705 |
| }, |
| { |
| "epoch": 0.712776176753122, |
| "grad_norm": 0.5254682302474976, |
| "learning_rate": 6.274460024275895e-05, |
| "loss": 3.2966861724853516, |
| "step": 3710 |
| }, |
| { |
| "epoch": 0.7137367915465899, |
| "grad_norm": 0.5654774308204651, |
| "learning_rate": 6.235739320458442e-05, |
| "loss": 3.293225860595703, |
| "step": 3715 |
| }, |
| { |
| "epoch": 0.7146974063400576, |
| "grad_norm": 0.35445114970207214, |
| "learning_rate": 6.197107086992156e-05, |
| "loss": 3.2962112426757812, |
| "step": 3720 |
| }, |
| { |
| "epoch": 0.7156580211335255, |
| "grad_norm": 0.43310287594795227, |
| "learning_rate": 6.158563713847994e-05, |
| "loss": 3.2953845977783205, |
| "step": 3725 |
| }, |
| { |
| "epoch": 0.7166186359269933, |
| "grad_norm": 0.4213480055332184, |
| "learning_rate": 6.12010959009994e-05, |
| "loss": 3.2970962524414062, |
| "step": 3730 |
| }, |
| { |
| "epoch": 0.7175792507204611, |
| "grad_norm": 0.3784043490886688, |
| "learning_rate": 6.081745103921047e-05, |
| "loss": 3.2936870574951174, |
| "step": 3735 |
| }, |
| { |
| "epoch": 0.7185398655139289, |
| "grad_norm": 0.6028347611427307, |
| "learning_rate": 6.043470642579516e-05, |
| "loss": 3.2977680206298827, |
| "step": 3740 |
| }, |
| { |
| "epoch": 0.7195004803073968, |
| "grad_norm": 0.44696855545043945, |
| "learning_rate": 6.005286592434828e-05, |
| "loss": 3.293193817138672, |
| "step": 3745 |
| }, |
| { |
| "epoch": 0.7204610951008645, |
| "grad_norm": 0.4338856339454651, |
| "learning_rate": 5.967193338933778e-05, |
| "loss": 3.294358825683594, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.7214217098943324, |
| "grad_norm": 0.4633798599243164, |
| "learning_rate": 5.9291912666066405e-05, |
| "loss": 3.2951465606689454, |
| "step": 3755 |
| }, |
| { |
| "epoch": 0.7223823246878002, |
| "grad_norm": 0.42019450664520264, |
| "learning_rate": 5.891280759063265e-05, |
| "loss": 3.292225646972656, |
| "step": 3760 |
| }, |
| { |
| "epoch": 0.723342939481268, |
| "grad_norm": 0.418599396944046, |
| "learning_rate": 5.853462198989184e-05, |
| "loss": 3.2958885192871095, |
| "step": 3765 |
| }, |
| { |
| "epoch": 0.7243035542747358, |
| "grad_norm": 0.35753360390663147, |
| "learning_rate": 5.815735968141813e-05, |
| "loss": 3.2968238830566405, |
| "step": 3770 |
| }, |
| { |
| "epoch": 0.7252641690682037, |
| "grad_norm": 0.4825795888900757, |
| "learning_rate": 5.778102447346514e-05, |
| "loss": 3.2957897186279297, |
| "step": 3775 |
| }, |
| { |
| "epoch": 0.7262247838616714, |
| "grad_norm": 0.41493552923202515, |
| "learning_rate": 5.740562016492811e-05, |
| "loss": 3.2921653747558595, |
| "step": 3780 |
| }, |
| { |
| "epoch": 0.7271853986551393, |
| "grad_norm": 0.5309422612190247, |
| "learning_rate": 5.703115054530537e-05, |
| "loss": 3.293100357055664, |
| "step": 3785 |
| }, |
| { |
| "epoch": 0.7281460134486071, |
| "grad_norm": 0.36943376064300537, |
| "learning_rate": 5.665761939466008e-05, |
| "loss": 3.294559860229492, |
| "step": 3790 |
| }, |
| { |
| "epoch": 0.729106628242075, |
| "grad_norm": 0.5610998272895813, |
| "learning_rate": 5.628503048358207e-05, |
| "loss": 3.292363739013672, |
| "step": 3795 |
| }, |
| { |
| "epoch": 0.7300672430355427, |
| "grad_norm": 0.5705908536911011, |
| "learning_rate": 5.591338757314968e-05, |
| "loss": 3.2965774536132812, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.7310278578290106, |
| "grad_norm": 0.5056052803993225, |
| "learning_rate": 5.554269441489204e-05, |
| "loss": 3.295191192626953, |
| "step": 3805 |
| }, |
| { |
| "epoch": 0.7319884726224783, |
| "grad_norm": 0.4820818603038788, |
| "learning_rate": 5.517295475075102e-05, |
| "loss": 3.2937217712402345, |
| "step": 3810 |
| }, |
| { |
| "epoch": 0.7329490874159462, |
| "grad_norm": 0.3636917471885681, |
| "learning_rate": 5.4804172313043465e-05, |
| "loss": 3.291709136962891, |
| "step": 3815 |
| }, |
| { |
| "epoch": 0.733909702209414, |
| "grad_norm": 0.3684285283088684, |
| "learning_rate": 5.443635082442363e-05, |
| "loss": 3.293587493896484, |
| "step": 3820 |
| }, |
| { |
| "epoch": 0.7348703170028819, |
| "grad_norm": 0.741382896900177, |
| "learning_rate": 5.4069493997845356e-05, |
| "loss": 3.2945404052734375, |
| "step": 3825 |
| }, |
| { |
| "epoch": 0.7358309317963496, |
| "grad_norm": 0.41890212893486023, |
| "learning_rate": 5.3703605536524905e-05, |
| "loss": 3.2955181121826174, |
| "step": 3830 |
| }, |
| { |
| "epoch": 0.7367915465898175, |
| "grad_norm": 0.3295837938785553, |
| "learning_rate": 5.333868913390338e-05, |
| "loss": 3.290005111694336, |
| "step": 3835 |
| }, |
| { |
| "epoch": 0.7377521613832853, |
| "grad_norm": 0.3231881558895111, |
| "learning_rate": 5.2974748473609505e-05, |
| "loss": 3.291563034057617, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.7387127761767531, |
| "grad_norm": 0.44054171442985535, |
| "learning_rate": 5.261178722942242e-05, |
| "loss": 3.2941848754882814, |
| "step": 3845 |
| }, |
| { |
| "epoch": 0.7396733909702209, |
| "grad_norm": 0.3981585204601288, |
| "learning_rate": 5.224980906523462e-05, |
| "loss": 3.292535400390625, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.7406340057636888, |
| "grad_norm": 0.36969125270843506, |
| "learning_rate": 5.188881763501486e-05, |
| "loss": 3.2916770935058595, |
| "step": 3855 |
| }, |
| { |
| "epoch": 0.7415946205571565, |
| "grad_norm": 0.414196252822876, |
| "learning_rate": 5.152881658277147e-05, |
| "loss": 3.293859100341797, |
| "step": 3860 |
| }, |
| { |
| "epoch": 0.7425552353506244, |
| "grad_norm": 0.3872639536857605, |
| "learning_rate": 5.1169809542515404e-05, |
| "loss": 3.296031188964844, |
| "step": 3865 |
| }, |
| { |
| "epoch": 0.7435158501440923, |
| "grad_norm": 0.4767828583717346, |
| "learning_rate": 5.081180013822368e-05, |
| "loss": 3.292156219482422, |
| "step": 3870 |
| }, |
| { |
| "epoch": 0.74447646493756, |
| "grad_norm": 0.41197365522384644, |
| "learning_rate": 5.045479198380272e-05, |
| "loss": 3.292400360107422, |
| "step": 3875 |
| }, |
| { |
| "epoch": 0.7454370797310279, |
| "grad_norm": 0.40532881021499634, |
| "learning_rate": 5.009878868305171e-05, |
| "loss": 3.2940216064453125, |
| "step": 3880 |
| }, |
| { |
| "epoch": 0.7463976945244957, |
| "grad_norm": 0.3654780685901642, |
| "learning_rate": 4.9743793829626736e-05, |
| "loss": 3.293231964111328, |
| "step": 3885 |
| }, |
| { |
| "epoch": 0.7473583093179635, |
| "grad_norm": 0.2990121841430664, |
| "learning_rate": 4.9389811007003834e-05, |
| "loss": 3.292677307128906, |
| "step": 3890 |
| }, |
| { |
| "epoch": 0.7483189241114313, |
| "grad_norm": 0.3309316039085388, |
| "learning_rate": 4.903684378844333e-05, |
| "loss": 3.294384765625, |
| "step": 3895 |
| }, |
| { |
| "epoch": 0.7492795389048992, |
| "grad_norm": 0.2731437683105469, |
| "learning_rate": 4.86848957369536e-05, |
| "loss": 3.292652893066406, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.7502401536983669, |
| "grad_norm": 0.3116244971752167, |
| "learning_rate": 4.8333970405254904e-05, |
| "loss": 3.288296127319336, |
| "step": 3905 |
| }, |
| { |
| "epoch": 0.7512007684918348, |
| "grad_norm": 0.42631855607032776, |
| "learning_rate": 4.798407133574405e-05, |
| "loss": 3.2939361572265624, |
| "step": 3910 |
| }, |
| { |
| "epoch": 0.7521613832853026, |
| "grad_norm": 0.35764989256858826, |
| "learning_rate": 4.7635202060457945e-05, |
| "loss": 3.2916053771972655, |
| "step": 3915 |
| }, |
| { |
| "epoch": 0.7531219980787704, |
| "grad_norm": 0.2865849733352661, |
| "learning_rate": 4.72873661010385e-05, |
| "loss": 3.2932079315185545, |
| "step": 3920 |
| }, |
| { |
| "epoch": 0.7540826128722382, |
| "grad_norm": 0.3286271095275879, |
| "learning_rate": 4.694056696869688e-05, |
| "loss": 3.295194625854492, |
| "step": 3925 |
| }, |
| { |
| "epoch": 0.7550432276657061, |
| "grad_norm": 0.346123605966568, |
| "learning_rate": 4.659480816417785e-05, |
| "loss": 3.288124847412109, |
| "step": 3930 |
| }, |
| { |
| "epoch": 0.7560038424591738, |
| "grad_norm": 0.4188701808452606, |
| "learning_rate": 4.6250093177725e-05, |
| "loss": 3.2938987731933596, |
| "step": 3935 |
| }, |
| { |
| "epoch": 0.7569644572526417, |
| "grad_norm": 0.44901224970817566, |
| "learning_rate": 4.590642548904479e-05, |
| "loss": 3.2915252685546874, |
| "step": 3940 |
| }, |
| { |
| "epoch": 0.7579250720461095, |
| "grad_norm": 0.4020468592643738, |
| "learning_rate": 4.5563808567272e-05, |
| "loss": 3.290658187866211, |
| "step": 3945 |
| }, |
| { |
| "epoch": 0.7588856868395774, |
| "grad_norm": 0.32219162583351135, |
| "learning_rate": 4.52222458709345e-05, |
| "loss": 3.2903762817382813, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.7598463016330451, |
| "grad_norm": 0.25904515385627747, |
| "learning_rate": 4.4881740847918155e-05, |
| "loss": 3.288920593261719, |
| "step": 3955 |
| }, |
| { |
| "epoch": 0.760806916426513, |
| "grad_norm": 0.29081347584724426, |
| "learning_rate": 4.454229693543251e-05, |
| "loss": 3.293811798095703, |
| "step": 3960 |
| }, |
| { |
| "epoch": 0.7617675312199808, |
| "grad_norm": 0.24549178779125214, |
| "learning_rate": 4.420391755997548e-05, |
| "loss": 3.2932552337646483, |
| "step": 3965 |
| }, |
| { |
| "epoch": 0.7627281460134486, |
| "grad_norm": 0.459926038980484, |
| "learning_rate": 4.386660613729925e-05, |
| "loss": 3.2938629150390626, |
| "step": 3970 |
| }, |
| { |
| "epoch": 0.7636887608069164, |
| "grad_norm": 0.4720708429813385, |
| "learning_rate": 4.35303660723756e-05, |
| "loss": 3.2878982543945314, |
| "step": 3975 |
| }, |
| { |
| "epoch": 0.7646493756003843, |
| "grad_norm": 0.28443172574043274, |
| "learning_rate": 4.3195200759361455e-05, |
| "loss": 3.2919075012207033, |
| "step": 3980 |
| }, |
| { |
| "epoch": 0.765609990393852, |
| "grad_norm": 0.33102986216545105, |
| "learning_rate": 4.2861113581564884e-05, |
| "loss": 3.2893836975097654, |
| "step": 3985 |
| }, |
| { |
| "epoch": 0.7665706051873199, |
| "grad_norm": 0.348117858171463, |
| "learning_rate": 4.252810791141054e-05, |
| "loss": 3.293404769897461, |
| "step": 3990 |
| }, |
| { |
| "epoch": 0.7675312199807877, |
| "grad_norm": 0.3010897636413574, |
| "learning_rate": 4.2196187110406054e-05, |
| "loss": 3.2900314331054688, |
| "step": 3995 |
| }, |
| { |
| "epoch": 0.7684918347742555, |
| "grad_norm": 0.274684339761734, |
| "learning_rate": 4.186535452910784e-05, |
| "loss": 3.2879261016845702, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.7694524495677233, |
| "grad_norm": 0.29900944232940674, |
| "learning_rate": 4.153561350708732e-05, |
| "loss": 3.292841339111328, |
| "step": 4005 |
| }, |
| { |
| "epoch": 0.7704130643611912, |
| "grad_norm": 0.30298614501953125, |
| "learning_rate": 4.12069673728973e-05, |
| "loss": 3.2894565582275392, |
| "step": 4010 |
| }, |
| { |
| "epoch": 0.7713736791546589, |
| "grad_norm": 0.31402137875556946, |
| "learning_rate": 4.087941944403815e-05, |
| "loss": 3.2861660003662108, |
| "step": 4015 |
| }, |
| { |
| "epoch": 0.7723342939481268, |
| "grad_norm": 0.4134189784526825, |
| "learning_rate": 4.0552973026924625e-05, |
| "loss": 3.289139175415039, |
| "step": 4020 |
| }, |
| { |
| "epoch": 0.7732949087415946, |
| "grad_norm": 0.3398532569408417, |
| "learning_rate": 4.022763141685226e-05, |
| "loss": 3.2890396118164062, |
| "step": 4025 |
| }, |
| { |
| "epoch": 0.7742555235350624, |
| "grad_norm": 0.33045315742492676, |
| "learning_rate": 3.990339789796418e-05, |
| "loss": 3.2925796508789062, |
| "step": 4030 |
| }, |
| { |
| "epoch": 0.7752161383285303, |
| "grad_norm": 0.28469014167785645, |
| "learning_rate": 3.958027574321794e-05, |
| "loss": 3.2899627685546875, |
| "step": 4035 |
| }, |
| { |
| "epoch": 0.7761767531219981, |
| "grad_norm": 0.23531687259674072, |
| "learning_rate": 3.9258268214352566e-05, |
| "loss": 3.286402130126953, |
| "step": 4040 |
| }, |
| { |
| "epoch": 0.777137367915466, |
| "grad_norm": 0.3813510239124298, |
| "learning_rate": 3.893737856185538e-05, |
| "loss": 3.2916938781738283, |
| "step": 4045 |
| }, |
| { |
| "epoch": 0.7780979827089337, |
| "grad_norm": 0.3439390957355499, |
| "learning_rate": 3.861761002492952e-05, |
| "loss": 3.288800811767578, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.7790585975024016, |
| "grad_norm": 0.2623279392719269, |
| "learning_rate": 3.8298965831461024e-05, |
| "loss": 3.288500213623047, |
| "step": 4055 |
| }, |
| { |
| "epoch": 0.7800192122958693, |
| "grad_norm": 0.3051926791667938, |
| "learning_rate": 3.798144919798631e-05, |
| "loss": 3.2871044158935545, |
| "step": 4060 |
| }, |
| { |
| "epoch": 0.7809798270893372, |
| "grad_norm": 0.2598731815814972, |
| "learning_rate": 3.766506332965976e-05, |
| "loss": 3.2878665924072266, |
| "step": 4065 |
| }, |
| { |
| "epoch": 0.781940441882805, |
| "grad_norm": 0.3285099267959595, |
| "learning_rate": 3.734981142022117e-05, |
| "loss": 3.287023162841797, |
| "step": 4070 |
| }, |
| { |
| "epoch": 0.7829010566762729, |
| "grad_norm": 0.2698347568511963, |
| "learning_rate": 3.70356966519638e-05, |
| "loss": 3.2910301208496096, |
| "step": 4075 |
| }, |
| { |
| "epoch": 0.7838616714697406, |
| "grad_norm": 0.2935537099838257, |
| "learning_rate": 3.672272219570199e-05, |
| "loss": 3.2882232666015625, |
| "step": 4080 |
| }, |
| { |
| "epoch": 0.7848222862632085, |
| "grad_norm": 0.2180212140083313, |
| "learning_rate": 3.641089121073934e-05, |
| "loss": 3.288380432128906, |
| "step": 4085 |
| }, |
| { |
| "epoch": 0.7857829010566763, |
| "grad_norm": 0.24223865568637848, |
| "learning_rate": 3.610020684483674e-05, |
| "loss": 3.2879989624023436, |
| "step": 4090 |
| }, |
| { |
| "epoch": 0.7867435158501441, |
| "grad_norm": 0.1990312784910202, |
| "learning_rate": 3.579067223418046e-05, |
| "loss": 3.2866798400878907, |
| "step": 4095 |
| }, |
| { |
| "epoch": 0.7877041306436119, |
| "grad_norm": 0.4838339388370514, |
| "learning_rate": 3.548229050335089e-05, |
| "loss": 3.290237808227539, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.7886647454370798, |
| "grad_norm": 0.27035945653915405, |
| "learning_rate": 3.517506476529045e-05, |
| "loss": 3.2898136138916017, |
| "step": 4105 |
| }, |
| { |
| "epoch": 0.7896253602305475, |
| "grad_norm": 0.2790147066116333, |
| "learning_rate": 3.486899812127264e-05, |
| "loss": 3.2898269653320313, |
| "step": 4110 |
| }, |
| { |
| "epoch": 0.7905859750240154, |
| "grad_norm": 0.2595069110393524, |
| "learning_rate": 3.456409366087054e-05, |
| "loss": 3.28946533203125, |
| "step": 4115 |
| }, |
| { |
| "epoch": 0.7915465898174832, |
| "grad_norm": 0.3146030306816101, |
| "learning_rate": 3.426035446192546e-05, |
| "loss": 3.2875953674316407, |
| "step": 4120 |
| }, |
| { |
| "epoch": 0.792507204610951, |
| "grad_norm": 0.28389808535575867, |
| "learning_rate": 3.395778359051634e-05, |
| "loss": 3.290841293334961, |
| "step": 4125 |
| }, |
| { |
| "epoch": 0.7934678194044188, |
| "grad_norm": 0.23182038962841034, |
| "learning_rate": 3.365638410092819e-05, |
| "loss": 3.289868927001953, |
| "step": 4130 |
| }, |
| { |
| "epoch": 0.7944284341978867, |
| "grad_norm": 0.2726062834262848, |
| "learning_rate": 3.3356159035621746e-05, |
| "loss": 3.287432098388672, |
| "step": 4135 |
| }, |
| { |
| "epoch": 0.7953890489913544, |
| "grad_norm": 0.21947945654392242, |
| "learning_rate": 3.3057111425202614e-05, |
| "loss": 3.286570358276367, |
| "step": 4140 |
| }, |
| { |
| "epoch": 0.7963496637848223, |
| "grad_norm": 0.28224751353263855, |
| "learning_rate": 3.275924428839043e-05, |
| "loss": 3.2863037109375, |
| "step": 4145 |
| }, |
| { |
| "epoch": 0.7973102785782901, |
| "grad_norm": 0.22240881621837616, |
| "learning_rate": 3.246256063198895e-05, |
| "loss": 3.2870025634765625, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.7982708933717579, |
| "grad_norm": 0.2880660891532898, |
| "learning_rate": 3.216706345085499e-05, |
| "loss": 3.2907535552978517, |
| "step": 4155 |
| }, |
| { |
| "epoch": 0.7992315081652257, |
| "grad_norm": 0.2571323812007904, |
| "learning_rate": 3.187275572786878e-05, |
| "loss": 3.283245849609375, |
| "step": 4160 |
| }, |
| { |
| "epoch": 0.8001921229586936, |
| "grad_norm": 0.3757617771625519, |
| "learning_rate": 3.15796404339036e-05, |
| "loss": 3.286081314086914, |
| "step": 4165 |
| }, |
| { |
| "epoch": 0.8011527377521613, |
| "grad_norm": 0.34448450803756714, |
| "learning_rate": 3.128772052779569e-05, |
| "loss": 3.2928192138671877, |
| "step": 4170 |
| }, |
| { |
| "epoch": 0.8021133525456292, |
| "grad_norm": 0.29771631956100464, |
| "learning_rate": 3.099699895631474e-05, |
| "loss": 3.2850051879882813, |
| "step": 4175 |
| }, |
| { |
| "epoch": 0.803073967339097, |
| "grad_norm": 0.19579406082630157, |
| "learning_rate": 3.0707478654133706e-05, |
| "loss": 3.2885662078857423, |
| "step": 4180 |
| }, |
| { |
| "epoch": 0.8040345821325648, |
| "grad_norm": 0.21206071972846985, |
| "learning_rate": 3.041916254379949e-05, |
| "loss": 3.289264678955078, |
| "step": 4185 |
| }, |
| { |
| "epoch": 0.8049951969260326, |
| "grad_norm": 0.2306758463382721, |
| "learning_rate": 3.0132053535703342e-05, |
| "loss": 3.289895248413086, |
| "step": 4190 |
| }, |
| { |
| "epoch": 0.8059558117195005, |
| "grad_norm": 0.24324500560760498, |
| "learning_rate": 2.984615452805147e-05, |
| "loss": 3.289009857177734, |
| "step": 4195 |
| }, |
| { |
| "epoch": 0.8069164265129684, |
| "grad_norm": 0.24182389676570892, |
| "learning_rate": 2.9561468406835865e-05, |
| "loss": 3.2901374816894533, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.8078770413064361, |
| "grad_norm": 0.19510437548160553, |
| "learning_rate": 2.927799804580495e-05, |
| "loss": 3.288174057006836, |
| "step": 4205 |
| }, |
| { |
| "epoch": 0.808837656099904, |
| "grad_norm": 0.20101343095302582, |
| "learning_rate": 2.8995746306434853e-05, |
| "loss": 3.2845272064208983, |
| "step": 4210 |
| }, |
| { |
| "epoch": 0.8097982708933718, |
| "grad_norm": 0.219615176320076, |
| "learning_rate": 2.871471603790035e-05, |
| "loss": 3.288011932373047, |
| "step": 4215 |
| }, |
| { |
| "epoch": 0.8107588856868396, |
| "grad_norm": 0.2699833810329437, |
| "learning_rate": 2.8434910077046163e-05, |
| "loss": 3.2884559631347656, |
| "step": 4220 |
| }, |
| { |
| "epoch": 0.8117195004803074, |
| "grad_norm": 0.2704545855522156, |
| "learning_rate": 2.8156331248358295e-05, |
| "loss": 3.283679962158203, |
| "step": 4225 |
| }, |
| { |
| "epoch": 0.8126801152737753, |
| "grad_norm": 0.22715161740779877, |
| "learning_rate": 2.787898236393556e-05, |
| "loss": 3.285501480102539, |
| "step": 4230 |
| }, |
| { |
| "epoch": 0.813640730067243, |
| "grad_norm": 0.2033311128616333, |
| "learning_rate": 2.7602866223461044e-05, |
| "loss": 3.284902572631836, |
| "step": 4235 |
| }, |
| { |
| "epoch": 0.8146013448607109, |
| "grad_norm": 0.2283061444759369, |
| "learning_rate": 2.7327985614174143e-05, |
| "loss": 3.2900623321533202, |
| "step": 4240 |
| }, |
| { |
| "epoch": 0.8155619596541787, |
| "grad_norm": 0.21356871724128723, |
| "learning_rate": 2.7054343310842115e-05, |
| "loss": 3.283802032470703, |
| "step": 4245 |
| }, |
| { |
| "epoch": 0.8165225744476465, |
| "grad_norm": 0.21679487824440002, |
| "learning_rate": 2.6781942075732294e-05, |
| "loss": 3.284503936767578, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.8174831892411143, |
| "grad_norm": 0.2814090847969055, |
| "learning_rate": 2.65107846585841e-05, |
| "loss": 3.2889778137207033, |
| "step": 4255 |
| }, |
| { |
| "epoch": 0.8184438040345822, |
| "grad_norm": 0.20454630255699158, |
| "learning_rate": 2.624087379658123e-05, |
| "loss": 3.2832550048828124, |
| "step": 4260 |
| }, |
| { |
| "epoch": 0.8194044188280499, |
| "grad_norm": 0.25826799869537354, |
| "learning_rate": 2.5972212214324162e-05, |
| "loss": 3.2887802124023438, |
| "step": 4265 |
| }, |
| { |
| "epoch": 0.8203650336215178, |
| "grad_norm": 0.24519048631191254, |
| "learning_rate": 2.5704802623802595e-05, |
| "loss": 3.2866302490234376, |
| "step": 4270 |
| }, |
| { |
| "epoch": 0.8213256484149856, |
| "grad_norm": 0.1912376880645752, |
| "learning_rate": 2.5438647724368054e-05, |
| "loss": 3.2893089294433593, |
| "step": 4275 |
| }, |
| { |
| "epoch": 0.8222862632084534, |
| "grad_norm": 0.18562142550945282, |
| "learning_rate": 2.5173750202706666e-05, |
| "loss": 3.2848739624023438, |
| "step": 4280 |
| }, |
| { |
| "epoch": 0.8232468780019212, |
| "grad_norm": 0.21541374921798706, |
| "learning_rate": 2.491011273281189e-05, |
| "loss": 3.285577392578125, |
| "step": 4285 |
| }, |
| { |
| "epoch": 0.8242074927953891, |
| "grad_norm": 0.2195151448249817, |
| "learning_rate": 2.4647737975957954e-05, |
| "loss": 3.2902549743652343, |
| "step": 4290 |
| }, |
| { |
| "epoch": 0.8251681075888568, |
| "grad_norm": 0.2502996623516083, |
| "learning_rate": 2.4386628580672396e-05, |
| "loss": 3.2855270385742186, |
| "step": 4295 |
| }, |
| { |
| "epoch": 0.8261287223823247, |
| "grad_norm": 0.168674036860466, |
| "learning_rate": 2.4126787182709796e-05, |
| "loss": 3.2874530792236327, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.8270893371757925, |
| "grad_norm": 0.22533266246318817, |
| "learning_rate": 2.3868216405025002e-05, |
| "loss": 3.286944580078125, |
| "step": 4305 |
| }, |
| { |
| "epoch": 0.8280499519692603, |
| "grad_norm": 0.21998296678066254, |
| "learning_rate": 2.361091885774652e-05, |
| "loss": 3.286793518066406, |
| "step": 4310 |
| }, |
| { |
| "epoch": 0.8290105667627281, |
| "grad_norm": 0.2057517021894455, |
| "learning_rate": 2.3354897138150536e-05, |
| "loss": 3.2883201599121095, |
| "step": 4315 |
| }, |
| { |
| "epoch": 0.829971181556196, |
| "grad_norm": 0.2263990193605423, |
| "learning_rate": 2.3100153830634218e-05, |
| "loss": 3.2877071380615233, |
| "step": 4320 |
| }, |
| { |
| "epoch": 0.8309317963496637, |
| "grad_norm": 0.17990660667419434, |
| "learning_rate": 2.284669150669001e-05, |
| "loss": 3.2851654052734376, |
| "step": 4325 |
| }, |
| { |
| "epoch": 0.8318924111431316, |
| "grad_norm": 0.19730253517627716, |
| "learning_rate": 2.259451272487957e-05, |
| "loss": 3.2849620819091796, |
| "step": 4330 |
| }, |
| { |
| "epoch": 0.8328530259365994, |
| "grad_norm": 0.1865607649087906, |
| "learning_rate": 2.234362003080772e-05, |
| "loss": 3.2859230041503906, |
| "step": 4335 |
| }, |
| { |
| "epoch": 0.8338136407300673, |
| "grad_norm": 0.21946489810943604, |
| "learning_rate": 2.2094015957097215e-05, |
| "loss": 3.2830989837646483, |
| "step": 4340 |
| }, |
| { |
| "epoch": 0.834774255523535, |
| "grad_norm": 0.22765038907527924, |
| "learning_rate": 2.1845703023362647e-05, |
| "loss": 3.2854949951171877, |
| "step": 4345 |
| }, |
| { |
| "epoch": 0.8357348703170029, |
| "grad_norm": 0.1913105696439743, |
| "learning_rate": 2.159868373618544e-05, |
| "loss": 3.2875335693359373, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.8366954851104706, |
| "grad_norm": 0.21100343763828278, |
| "learning_rate": 2.13529605890883e-05, |
| "loss": 3.2848331451416017, |
| "step": 4355 |
| }, |
| { |
| "epoch": 0.8376560999039385, |
| "grad_norm": 0.20388315618038177, |
| "learning_rate": 2.110853606251004e-05, |
| "loss": 3.283430480957031, |
| "step": 4360 |
| }, |
| { |
| "epoch": 0.8386167146974063, |
| "grad_norm": 0.18785250186920166, |
| "learning_rate": 2.0865412623780858e-05, |
| "loss": 3.2853065490722657, |
| "step": 4365 |
| }, |
| { |
| "epoch": 0.8395773294908742, |
| "grad_norm": 0.19692327082157135, |
| "learning_rate": 2.0623592727096916e-05, |
| "loss": 3.2828216552734375, |
| "step": 4370 |
| }, |
| { |
| "epoch": 0.840537944284342, |
| "grad_norm": 0.2322104126214981, |
| "learning_rate": 2.0383078813496e-05, |
| "loss": 3.2829490661621095, |
| "step": 4375 |
| }, |
| { |
| "epoch": 0.8414985590778098, |
| "grad_norm": 0.2523021697998047, |
| "learning_rate": 2.014387331083268e-05, |
| "loss": 3.2841728210449217, |
| "step": 4380 |
| }, |
| { |
| "epoch": 0.8424591738712777, |
| "grad_norm": 0.2998299300670624, |
| "learning_rate": 1.990597863375389e-05, |
| "loss": 3.2823081970214845, |
| "step": 4385 |
| }, |
| { |
| "epoch": 0.8434197886647454, |
| "grad_norm": 0.22866974771022797, |
| "learning_rate": 1.966939718367444e-05, |
| "loss": 3.2830490112304687, |
| "step": 4390 |
| }, |
| { |
| "epoch": 0.8443804034582133, |
| "grad_norm": 0.2374458909034729, |
| "learning_rate": 1.9434131348752842e-05, |
| "loss": 3.2861083984375, |
| "step": 4395 |
| }, |
| { |
| "epoch": 0.8453410182516811, |
| "grad_norm": 0.271164208650589, |
| "learning_rate": 1.920018350386725e-05, |
| "loss": 3.2853363037109373, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.8463016330451489, |
| "grad_norm": 0.26674726605415344, |
| "learning_rate": 1.8967556010591423e-05, |
| "loss": 3.2835784912109376, |
| "step": 4405 |
| }, |
| { |
| "epoch": 0.8472622478386167, |
| "grad_norm": 0.2277594804763794, |
| "learning_rate": 1.873625121717089e-05, |
| "loss": 3.287038040161133, |
| "step": 4410 |
| }, |
| { |
| "epoch": 0.8482228626320846, |
| "grad_norm": 0.21174593269824982, |
| "learning_rate": 1.850627145849926e-05, |
| "loss": 3.281999206542969, |
| "step": 4415 |
| }, |
| { |
| "epoch": 0.8491834774255523, |
| "grad_norm": 0.19385212659835815, |
| "learning_rate": 1.8277619056094684e-05, |
| "loss": 3.283008575439453, |
| "step": 4420 |
| }, |
| { |
| "epoch": 0.8501440922190202, |
| "grad_norm": 0.22816501557826996, |
| "learning_rate": 1.805029631807632e-05, |
| "loss": 3.282820129394531, |
| "step": 4425 |
| }, |
| { |
| "epoch": 0.851104707012488, |
| "grad_norm": 0.20349720120429993, |
| "learning_rate": 1.7824305539141165e-05, |
| "loss": 3.2845672607421874, |
| "step": 4430 |
| }, |
| { |
| "epoch": 0.8520653218059558, |
| "grad_norm": 0.17361263930797577, |
| "learning_rate": 1.7599649000540828e-05, |
| "loss": 3.284697341918945, |
| "step": 4435 |
| }, |
| { |
| "epoch": 0.8530259365994236, |
| "grad_norm": 0.19036982953548431, |
| "learning_rate": 1.7376328970058382e-05, |
| "loss": 3.2811737060546875, |
| "step": 4440 |
| }, |
| { |
| "epoch": 0.8539865513928915, |
| "grad_norm": 0.18866802752017975, |
| "learning_rate": 1.715434770198582e-05, |
| "loss": 3.2872802734375, |
| "step": 4445 |
| }, |
| { |
| "epoch": 0.8549471661863592, |
| "grad_norm": 0.17417368292808533, |
| "learning_rate": 1.6933707437100852e-05, |
| "loss": 3.2829563140869142, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.8559077809798271, |
| "grad_norm": 0.1611868292093277, |
| "learning_rate": 1.67144104026446e-05, |
| "loss": 3.281053161621094, |
| "step": 4455 |
| }, |
| { |
| "epoch": 0.8568683957732949, |
| "grad_norm": 0.18905870616436005, |
| "learning_rate": 1.6496458812299073e-05, |
| "loss": 3.281689453125, |
| "step": 4460 |
| }, |
| { |
| "epoch": 0.8578290105667628, |
| "grad_norm": 0.1868743747472763, |
| "learning_rate": 1.6279854866164586e-05, |
| "loss": 3.2842777252197264, |
| "step": 4465 |
| }, |
| { |
| "epoch": 0.8587896253602305, |
| "grad_norm": 0.16863790154457092, |
| "learning_rate": 1.6064600750737995e-05, |
| "loss": 3.2836139678955076, |
| "step": 4470 |
| }, |
| { |
| "epoch": 0.8597502401536984, |
| "grad_norm": 0.15220773220062256, |
| "learning_rate": 1.5850698638890093e-05, |
| "loss": 3.28204460144043, |
| "step": 4475 |
| }, |
| { |
| "epoch": 0.8607108549471661, |
| "grad_norm": 0.16620075702667236, |
| "learning_rate": 1.563815068984418e-05, |
| "loss": 3.2853729248046877, |
| "step": 4480 |
| }, |
| { |
| "epoch": 0.861671469740634, |
| "grad_norm": 0.14935848116874695, |
| "learning_rate": 1.54269590491539e-05, |
| "loss": 3.2866542816162108, |
| "step": 4485 |
| }, |
| { |
| "epoch": 0.8626320845341018, |
| "grad_norm": 0.1583867073059082, |
| "learning_rate": 1.521712584868166e-05, |
| "loss": 3.283392333984375, |
| "step": 4490 |
| }, |
| { |
| "epoch": 0.8635926993275697, |
| "grad_norm": 0.1681807041168213, |
| "learning_rate": 1.5008653206577376e-05, |
| "loss": 3.281229782104492, |
| "step": 4495 |
| }, |
| { |
| "epoch": 0.8645533141210374, |
| "grad_norm": 0.1857963651418686, |
| "learning_rate": 1.4801543227256685e-05, |
| "loss": 3.2861068725585936, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.8655139289145053, |
| "grad_norm": 0.16560649871826172, |
| "learning_rate": 1.4595798001379965e-05, |
| "loss": 3.2863983154296874, |
| "step": 4505 |
| }, |
| { |
| "epoch": 0.866474543707973, |
| "grad_norm": 0.17330175638198853, |
| "learning_rate": 1.4391419605831218e-05, |
| "loss": 3.284455490112305, |
| "step": 4510 |
| }, |
| { |
| "epoch": 0.8674351585014409, |
| "grad_norm": 0.15583863854408264, |
| "learning_rate": 1.4188410103696957e-05, |
| "loss": 3.283338165283203, |
| "step": 4515 |
| }, |
| { |
| "epoch": 0.8683957732949087, |
| "grad_norm": 0.1486099660396576, |
| "learning_rate": 1.398677154424559e-05, |
| "loss": 3.282981109619141, |
| "step": 4520 |
| }, |
| { |
| "epoch": 0.8693563880883766, |
| "grad_norm": 0.16416342556476593, |
| "learning_rate": 1.3786505962906475e-05, |
| "loss": 3.281314468383789, |
| "step": 4525 |
| }, |
| { |
| "epoch": 0.8703170028818443, |
| "grad_norm": 0.17155486345291138, |
| "learning_rate": 1.3587615381249622e-05, |
| "loss": 3.2852855682373048, |
| "step": 4530 |
| }, |
| { |
| "epoch": 0.8712776176753122, |
| "grad_norm": 0.17230698466300964, |
| "learning_rate": 1.3390101806965165e-05, |
| "loss": 3.2855682373046875, |
| "step": 4535 |
| }, |
| { |
| "epoch": 0.8722382324687801, |
| "grad_norm": 0.16426202654838562, |
| "learning_rate": 1.3193967233843083e-05, |
| "loss": 3.284606170654297, |
| "step": 4540 |
| }, |
| { |
| "epoch": 0.8731988472622478, |
| "grad_norm": 0.18724007904529572, |
| "learning_rate": 1.2999213641753164e-05, |
| "loss": 3.281108093261719, |
| "step": 4545 |
| }, |
| { |
| "epoch": 0.8741594620557157, |
| "grad_norm": 0.171888530254364, |
| "learning_rate": 1.280584299662486e-05, |
| "loss": 3.2837615966796876, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.8751200768491835, |
| "grad_norm": 0.15725582838058472, |
| "learning_rate": 1.2613857250427622e-05, |
| "loss": 3.281300354003906, |
| "step": 4555 |
| }, |
| { |
| "epoch": 0.8760806916426513, |
| "grad_norm": 0.1605101078748703, |
| "learning_rate": 1.2423258341151088e-05, |
| "loss": 3.2830268859863283, |
| "step": 4560 |
| }, |
| { |
| "epoch": 0.8770413064361191, |
| "grad_norm": 0.16305667161941528, |
| "learning_rate": 1.2234048192785539e-05, |
| "loss": 3.28362922668457, |
| "step": 4565 |
| }, |
| { |
| "epoch": 0.878001921229587, |
| "grad_norm": 0.1558937281370163, |
| "learning_rate": 1.2046228715302509e-05, |
| "loss": 3.28128662109375, |
| "step": 4570 |
| }, |
| { |
| "epoch": 0.8789625360230547, |
| "grad_norm": 0.14114581048488617, |
| "learning_rate": 1.1859801804635471e-05, |
| "loss": 3.2840221405029295, |
| "step": 4575 |
| }, |
| { |
| "epoch": 0.8799231508165226, |
| "grad_norm": 0.14987899363040924, |
| "learning_rate": 1.167476934266065e-05, |
| "loss": 3.2838886260986326, |
| "step": 4580 |
| }, |
| { |
| "epoch": 0.8808837656099904, |
| "grad_norm": 0.15901191532611847, |
| "learning_rate": 1.1491133197178177e-05, |
| "loss": 3.2861114501953126, |
| "step": 4585 |
| }, |
| { |
| "epoch": 0.8818443804034583, |
| "grad_norm": 0.1510273516178131, |
| "learning_rate": 1.1308895221893088e-05, |
| "loss": 3.2844474792480467, |
| "step": 4590 |
| }, |
| { |
| "epoch": 0.882804995196926, |
| "grad_norm": 0.14691267907619476, |
| "learning_rate": 1.1128057256396684e-05, |
| "loss": 3.2848670959472654, |
| "step": 4595 |
| }, |
| { |
| "epoch": 0.8837656099903939, |
| "grad_norm": 0.16120979189872742, |
| "learning_rate": 1.0948621126147978e-05, |
| "loss": 3.2807106018066405, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.8847262247838616, |
| "grad_norm": 0.161734938621521, |
| "learning_rate": 1.0770588642455092e-05, |
| "loss": 3.279821014404297, |
| "step": 4605 |
| }, |
| { |
| "epoch": 0.8856868395773295, |
| "grad_norm": 0.17960962653160095, |
| "learning_rate": 1.0593961602457346e-05, |
| "loss": 3.2869110107421875, |
| "step": 4610 |
| }, |
| { |
| "epoch": 0.8866474543707973, |
| "grad_norm": 0.13539910316467285, |
| "learning_rate": 1.041874178910666e-05, |
| "loss": 3.283498001098633, |
| "step": 4615 |
| }, |
| { |
| "epoch": 0.8876080691642652, |
| "grad_norm": 0.147013857960701, |
| "learning_rate": 1.0244930971149918e-05, |
| "loss": 3.28480224609375, |
| "step": 4620 |
| }, |
| { |
| "epoch": 0.8885686839577329, |
| "grad_norm": 0.16810062527656555, |
| "learning_rate": 1.0072530903110942e-05, |
| "loss": 3.284687805175781, |
| "step": 4625 |
| }, |
| { |
| "epoch": 0.8895292987512008, |
| "grad_norm": 0.16976973414421082, |
| "learning_rate": 9.901543325272753e-06, |
| "loss": 3.282553863525391, |
| "step": 4630 |
| }, |
| { |
| "epoch": 0.8904899135446686, |
| "grad_norm": 0.16181102395057678, |
| "learning_rate": 9.731969963660224e-06, |
| "loss": 3.2854312896728515, |
| "step": 4635 |
| }, |
| { |
| "epoch": 0.8914505283381364, |
| "grad_norm": 0.16076242923736572, |
| "learning_rate": 9.56381253002233e-06, |
| "loss": 3.2812950134277346, |
| "step": 4640 |
| }, |
| { |
| "epoch": 0.8924111431316042, |
| "grad_norm": 0.14188607037067413, |
| "learning_rate": 9.397072721815113e-06, |
| "loss": 3.284113311767578, |
| "step": 4645 |
| }, |
| { |
| "epoch": 0.8933717579250721, |
| "grad_norm": 0.16452723741531372, |
| "learning_rate": 9.231752222184496e-06, |
| "loss": 3.2799339294433594, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.8943323727185398, |
| "grad_norm": 0.13960616290569305, |
| "learning_rate": 9.067852699949197e-06, |
| "loss": 3.2871246337890625, |
| "step": 4655 |
| }, |
| { |
| "epoch": 0.8952929875120077, |
| "grad_norm": 0.1369549185037613, |
| "learning_rate": 8.905375809584053e-06, |
| "loss": 3.283050537109375, |
| "step": 4660 |
| }, |
| { |
| "epoch": 0.8962536023054755, |
| "grad_norm": 0.13010026514530182, |
| "learning_rate": 8.744323191203129e-06, |
| "loss": 3.2822250366210937, |
| "step": 4665 |
| }, |
| { |
| "epoch": 0.8972142170989433, |
| "grad_norm": 0.14469455182552338, |
| "learning_rate": 8.5846964705433e-06, |
| "loss": 3.2816314697265625, |
| "step": 4670 |
| }, |
| { |
| "epoch": 0.8981748318924111, |
| "grad_norm": 0.15112441778182983, |
| "learning_rate": 8.426497258947813e-06, |
| "loss": 3.279371643066406, |
| "step": 4675 |
| }, |
| { |
| "epoch": 0.899135446685879, |
| "grad_norm": 0.14177513122558594, |
| "learning_rate": 8.269727153349915e-06, |
| "loss": 3.2836181640625, |
| "step": 4680 |
| }, |
| { |
| "epoch": 0.9000960614793467, |
| "grad_norm": 0.1383771449327469, |
| "learning_rate": 8.114387736256966e-06, |
| "loss": 3.2833229064941407, |
| "step": 4685 |
| }, |
| { |
| "epoch": 0.9010566762728146, |
| "grad_norm": 0.1385938823223114, |
| "learning_rate": 7.960480575734162e-06, |
| "loss": 3.283488464355469, |
| "step": 4690 |
| }, |
| { |
| "epoch": 0.9020172910662824, |
| "grad_norm": 0.1253698468208313, |
| "learning_rate": 7.808007225388963e-06, |
| "loss": 3.282206726074219, |
| "step": 4695 |
| }, |
| { |
| "epoch": 0.9029779058597502, |
| "grad_norm": 0.14205971360206604, |
| "learning_rate": 7.656969224355285e-06, |
| "loss": 3.2821189880371096, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.9039385206532181, |
| "grad_norm": 0.13180512189865112, |
| "learning_rate": 7.507368097277994e-06, |
| "loss": 3.278270721435547, |
| "step": 4705 |
| }, |
| { |
| "epoch": 0.9048991354466859, |
| "grad_norm": 0.1405053585767746, |
| "learning_rate": 7.359205354297499e-06, |
| "loss": 3.279631423950195, |
| "step": 4710 |
| }, |
| { |
| "epoch": 0.9058597502401537, |
| "grad_norm": 0.1261477768421173, |
| "learning_rate": 7.21248249103451e-06, |
| "loss": 3.2792625427246094, |
| "step": 4715 |
| }, |
| { |
| "epoch": 0.9068203650336215, |
| "grad_norm": 0.1316026747226715, |
| "learning_rate": 7.067200988574983e-06, |
| "loss": 3.2806537628173826, |
| "step": 4720 |
| }, |
| { |
| "epoch": 0.9077809798270894, |
| "grad_norm": 0.1313539296388626, |
| "learning_rate": 6.923362313455094e-06, |
| "loss": 3.2771621704101563, |
| "step": 4725 |
| }, |
| { |
| "epoch": 0.9087415946205571, |
| "grad_norm": 0.1397256702184677, |
| "learning_rate": 6.780967917646518e-06, |
| "loss": 3.279193115234375, |
| "step": 4730 |
| }, |
| { |
| "epoch": 0.909702209414025, |
| "grad_norm": 0.13661499321460724, |
| "learning_rate": 6.640019238541727e-06, |
| "loss": 3.2864913940429688, |
| "step": 4735 |
| }, |
| { |
| "epoch": 0.9106628242074928, |
| "grad_norm": 0.13143934309482574, |
| "learning_rate": 6.5005176989394335e-06, |
| "loss": 3.2848949432373047, |
| "step": 4740 |
| }, |
| { |
| "epoch": 0.9116234390009607, |
| "grad_norm": 0.12046822905540466, |
| "learning_rate": 6.362464707030334e-06, |
| "loss": 3.283034896850586, |
| "step": 4745 |
| }, |
| { |
| "epoch": 0.9125840537944284, |
| "grad_norm": 0.12852244079113007, |
| "learning_rate": 6.225861656382825e-06, |
| "loss": 3.2837677001953125, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.9135446685878963, |
| "grad_norm": 0.13605345785617828, |
| "learning_rate": 6.090709925928938e-06, |
| "loss": 3.2784072875976564, |
| "step": 4755 |
| }, |
| { |
| "epoch": 0.914505283381364, |
| "grad_norm": 0.1370062530040741, |
| "learning_rate": 5.957010879950446e-06, |
| "loss": 3.279316711425781, |
| "step": 4760 |
| }, |
| { |
| "epoch": 0.9154658981748319, |
| "grad_norm": 0.14122441411018372, |
| "learning_rate": 5.824765868065101e-06, |
| "loss": 3.2786293029785156, |
| "step": 4765 |
| }, |
| { |
| "epoch": 0.9164265129682997, |
| "grad_norm": 0.1247406154870987, |
| "learning_rate": 5.69397622521291e-06, |
| "loss": 3.2822948455810548, |
| "step": 4770 |
| }, |
| { |
| "epoch": 0.9173871277617676, |
| "grad_norm": 0.13668857514858246, |
| "learning_rate": 5.564643271642799e-06, |
| "loss": 3.282684326171875, |
| "step": 4775 |
| }, |
| { |
| "epoch": 0.9183477425552353, |
| "grad_norm": 0.14488086104393005, |
| "learning_rate": 5.436768312899226e-06, |
| "loss": 3.2865325927734377, |
| "step": 4780 |
| }, |
| { |
| "epoch": 0.9193083573487032, |
| "grad_norm": 0.13214252889156342, |
| "learning_rate": 5.310352639808968e-06, |
| "loss": 3.2808204650878907, |
| "step": 4785 |
| }, |
| { |
| "epoch": 0.920268972142171, |
| "grad_norm": 0.12839952111244202, |
| "learning_rate": 5.185397528468155e-06, |
| "loss": 3.2781848907470703, |
| "step": 4790 |
| }, |
| { |
| "epoch": 0.9212295869356388, |
| "grad_norm": 0.12066492438316345, |
| "learning_rate": 5.061904240229309e-06, |
| "loss": 3.285063934326172, |
| "step": 4795 |
| }, |
| { |
| "epoch": 0.9221902017291066, |
| "grad_norm": 0.1341644525527954, |
| "learning_rate": 4.939874021688739e-06, |
| "loss": 3.2813148498535156, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.9231508165225745, |
| "grad_norm": 0.11596546322107315, |
| "learning_rate": 4.819308104673769e-06, |
| "loss": 3.281114196777344, |
| "step": 4805 |
| }, |
| { |
| "epoch": 0.9241114313160422, |
| "grad_norm": 0.1205928698182106, |
| "learning_rate": 4.700207706230513e-06, |
| "loss": 3.2828353881835937, |
| "step": 4810 |
| }, |
| { |
| "epoch": 0.9250720461095101, |
| "grad_norm": 0.1389501690864563, |
| "learning_rate": 4.582574028611435e-06, |
| "loss": 3.2821121215820312, |
| "step": 4815 |
| }, |
| { |
| "epoch": 0.9260326609029779, |
| "grad_norm": 0.11629052460193634, |
| "learning_rate": 4.466408259263243e-06, |
| "loss": 3.2819660186767576, |
| "step": 4820 |
| }, |
| { |
| "epoch": 0.9269932756964457, |
| "grad_norm": 0.13381557166576385, |
| "learning_rate": 4.351711570815014e-06, |
| "loss": 3.278516387939453, |
| "step": 4825 |
| }, |
| { |
| "epoch": 0.9279538904899135, |
| "grad_norm": 0.12584474682807922, |
| "learning_rate": 4.238485121066154e-06, |
| "loss": 3.281688690185547, |
| "step": 4830 |
| }, |
| { |
| "epoch": 0.9289145052833814, |
| "grad_norm": 0.12782976031303406, |
| "learning_rate": 4.126730052974908e-06, |
| "loss": 3.279840087890625, |
| "step": 4835 |
| }, |
| { |
| "epoch": 0.9298751200768491, |
| "grad_norm": 0.1391313374042511, |
| "learning_rate": 4.016447494646718e-06, |
| "loss": 3.2828041076660157, |
| "step": 4840 |
| }, |
| { |
| "epoch": 0.930835734870317, |
| "grad_norm": 0.11799421906471252, |
| "learning_rate": 3.907638559322817e-06, |
| "loss": 3.284708786010742, |
| "step": 4845 |
| }, |
| { |
| "epoch": 0.9317963496637848, |
| "grad_norm": 0.14186686277389526, |
| "learning_rate": 3.8003043453691207e-06, |
| "loss": 3.2787437438964844, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.9327569644572526, |
| "grad_norm": 0.11602967977523804, |
| "learning_rate": 3.69444593626495e-06, |
| "loss": 3.28221435546875, |
| "step": 4855 |
| }, |
| { |
| "epoch": 0.9337175792507204, |
| "grad_norm": 0.13371752202510834, |
| "learning_rate": 3.5900644005922465e-06, |
| "loss": 3.2848297119140626, |
| "step": 4860 |
| }, |
| { |
| "epoch": 0.9346781940441883, |
| "grad_norm": 0.12414630502462387, |
| "learning_rate": 3.48716079202474e-06, |
| "loss": 3.2800872802734373, |
| "step": 4865 |
| }, |
| { |
| "epoch": 0.9356388088376562, |
| "grad_norm": 0.11889371275901794, |
| "learning_rate": 3.385736149317264e-06, |
| "loss": 3.2823089599609374, |
| "step": 4870 |
| }, |
| { |
| "epoch": 0.9365994236311239, |
| "grad_norm": 0.1263454407453537, |
| "learning_rate": 3.2857914962953926e-06, |
| "loss": 3.2813446044921877, |
| "step": 4875 |
| }, |
| { |
| "epoch": 0.9375600384245918, |
| "grad_norm": 0.12067841738462448, |
| "learning_rate": 3.1873278418449e-06, |
| "loss": 3.2796573638916016, |
| "step": 4880 |
| }, |
| { |
| "epoch": 0.9385206532180596, |
| "grad_norm": 0.137226864695549, |
| "learning_rate": 3.090346179901837e-06, |
| "loss": 3.2818023681640627, |
| "step": 4885 |
| }, |
| { |
| "epoch": 0.9394812680115274, |
| "grad_norm": 0.11869537830352783, |
| "learning_rate": 2.99484748944227e-06, |
| "loss": 3.2809097290039064, |
| "step": 4890 |
| }, |
| { |
| "epoch": 0.9404418828049952, |
| "grad_norm": 0.11975245177745819, |
| "learning_rate": 2.9008327344724913e-06, |
| "loss": 3.282526397705078, |
| "step": 4895 |
| }, |
| { |
| "epoch": 0.9414024975984631, |
| "grad_norm": 0.11280205845832825, |
| "learning_rate": 2.808302864019374e-06, |
| "loss": 3.279458999633789, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.9423631123919308, |
| "grad_norm": 0.13222162425518036, |
| "learning_rate": 2.7172588121205983e-06, |
| "loss": 3.2817481994628905, |
| "step": 4905 |
| }, |
| { |
| "epoch": 0.9433237271853987, |
| "grad_norm": 0.09990071505308151, |
| "learning_rate": 2.62770149781541e-06, |
| "loss": 3.281599426269531, |
| "step": 4910 |
| }, |
| { |
| "epoch": 0.9442843419788665, |
| "grad_norm": 0.1150190606713295, |
| "learning_rate": 2.539631825135191e-06, |
| "loss": 3.2885650634765624, |
| "step": 4915 |
| }, |
| { |
| "epoch": 0.9452449567723343, |
| "grad_norm": 0.13486024737358093, |
| "learning_rate": 2.4530506830944876e-06, |
| "loss": 3.2844482421875, |
| "step": 4920 |
| }, |
| { |
| "epoch": 0.9462055715658021, |
| "grad_norm": 0.11893726140260696, |
| "learning_rate": 2.367958945681897e-06, |
| "loss": 3.28011474609375, |
| "step": 4925 |
| }, |
| { |
| "epoch": 0.94716618635927, |
| "grad_norm": 0.1119316965341568, |
| "learning_rate": 2.2843574718513114e-06, |
| "loss": 3.2823932647705076, |
| "step": 4930 |
| }, |
| { |
| "epoch": 0.9481268011527377, |
| "grad_norm": 0.09861624985933304, |
| "learning_rate": 2.2022471055132553e-06, |
| "loss": 3.281215286254883, |
| "step": 4935 |
| }, |
| { |
| "epoch": 0.9490874159462056, |
| "grad_norm": 0.12267674505710602, |
| "learning_rate": 2.1216286755263434e-06, |
| "loss": 3.281680679321289, |
| "step": 4940 |
| }, |
| { |
| "epoch": 0.9500480307396734, |
| "grad_norm": 0.11797958612442017, |
| "learning_rate": 2.042502995688905e-06, |
| "loss": 3.2819839477539063, |
| "step": 4945 |
| }, |
| { |
| "epoch": 0.9510086455331412, |
| "grad_norm": 0.10683488100767136, |
| "learning_rate": 1.964870864730789e-06, |
| "loss": 3.2817310333251952, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.951969260326609, |
| "grad_norm": 0.12260305136442184, |
| "learning_rate": 1.8887330663053536e-06, |
| "loss": 3.2814239501953124, |
| "step": 4955 |
| }, |
| { |
| "epoch": 0.9529298751200769, |
| "grad_norm": 0.12215402722358704, |
| "learning_rate": 1.8140903689814079e-06, |
| "loss": 3.283163070678711, |
| "step": 4960 |
| }, |
| { |
| "epoch": 0.9538904899135446, |
| "grad_norm": 0.1066419780254364, |
| "learning_rate": 1.740943526235583e-06, |
| "loss": 3.2810131072998048, |
| "step": 4965 |
| }, |
| { |
| "epoch": 0.9548511047070125, |
| "grad_norm": 0.10814165323972702, |
| "learning_rate": 1.6692932764447054e-06, |
| "loss": 3.281442642211914, |
| "step": 4970 |
| }, |
| { |
| "epoch": 0.9558117195004803, |
| "grad_norm": 0.10270337015390396, |
| "learning_rate": 1.5991403428783188e-06, |
| "loss": 3.283908462524414, |
| "step": 4975 |
| }, |
| { |
| "epoch": 0.9567723342939481, |
| "grad_norm": 0.10882619768381119, |
| "learning_rate": 1.5304854336913752e-06, |
| "loss": 3.2817230224609375, |
| "step": 4980 |
| }, |
| { |
| "epoch": 0.9577329490874159, |
| "grad_norm": 0.10903850197792053, |
| "learning_rate": 1.463329241917105e-06, |
| "loss": 3.282952880859375, |
| "step": 4985 |
| }, |
| { |
| "epoch": 0.9586935638808838, |
| "grad_norm": 0.11845772713422775, |
| "learning_rate": 1.397672445460024e-06, |
| "loss": 3.2793148040771483, |
| "step": 4990 |
| }, |
| { |
| "epoch": 0.9596541786743515, |
| "grad_norm": 0.10239589214324951, |
| "learning_rate": 1.333515707089089e-06, |
| "loss": 3.28063850402832, |
| "step": 4995 |
| }, |
| { |
| "epoch": 0.9606147934678194, |
| "grad_norm": 0.10683233290910721, |
| "learning_rate": 1.2708596744309685e-06, |
| "loss": 3.2818748474121096, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.9615754082612872, |
| "grad_norm": 0.1100321039557457, |
| "learning_rate": 1.209704979963616e-06, |
| "loss": 3.287214660644531, |
| "step": 5005 |
| }, |
| { |
| "epoch": 0.962536023054755, |
| "grad_norm": 0.10722316801548004, |
| "learning_rate": 1.1500522410096912e-06, |
| "loss": 3.2820858001708983, |
| "step": 5010 |
| }, |
| { |
| "epoch": 0.9634966378482228, |
| "grad_norm": 0.09480059891939163, |
| "learning_rate": 1.0919020597305649e-06, |
| "loss": 3.281991195678711, |
| "step": 5015 |
| }, |
| { |
| "epoch": 0.9644572526416907, |
| "grad_norm": 0.09626810252666473, |
| "learning_rate": 1.0352550231200407e-06, |
| "loss": 3.282924270629883, |
| "step": 5020 |
| }, |
| { |
| "epoch": 0.9654178674351584, |
| "grad_norm": 0.09960578382015228, |
| "learning_rate": 9.801117029985596e-07, |
| "loss": 3.280469512939453, |
| "step": 5025 |
| }, |
| { |
| "epoch": 0.9663784822286263, |
| "grad_norm": 0.10461635142564774, |
| "learning_rate": 9.264726560073377e-07, |
| "loss": 3.282293701171875, |
| "step": 5030 |
| }, |
| { |
| "epoch": 0.9673390970220941, |
| "grad_norm": 0.09554687887430191, |
| "learning_rate": 8.743384236028051e-07, |
| "loss": 3.2805267333984376, |
| "step": 5035 |
| }, |
| { |
| "epoch": 0.968299711815562, |
| "grad_norm": 0.09053874015808105, |
| "learning_rate": 8.237095320511589e-07, |
| "loss": 3.282093048095703, |
| "step": 5040 |
| }, |
| { |
| "epoch": 0.9692603266090298, |
| "grad_norm": 0.09056618809700012, |
| "learning_rate": 7.745864924229351e-07, |
| "loss": 3.2803783416748047, |
| "step": 5045 |
| }, |
| { |
| "epoch": 0.9702209414024976, |
| "grad_norm": 0.09100056439638138, |
| "learning_rate": 7.26969800588012e-07, |
| "loss": 3.279576873779297, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.9711815561959655, |
| "grad_norm": 0.10496609658002853, |
| "learning_rate": 6.808599372104817e-07, |
| "loss": 3.2778125762939454, |
| "step": 5055 |
| }, |
| { |
| "epoch": 0.9721421709894332, |
| "grad_norm": 0.09851839393377304, |
| "learning_rate": 6.362573677438199e-07, |
| "loss": 3.2856407165527344, |
| "step": 5060 |
| }, |
| { |
| "epoch": 0.9731027857829011, |
| "grad_norm": 0.1031249463558197, |
| "learning_rate": 5.931625424262731e-07, |
| "loss": 3.281835174560547, |
| "step": 5065 |
| }, |
| { |
| "epoch": 0.9740634005763689, |
| "grad_norm": 0.0869988277554512, |
| "learning_rate": 5.51575896276163e-07, |
| "loss": 3.2821426391601562, |
| "step": 5070 |
| }, |
| { |
| "epoch": 0.9750240153698367, |
| "grad_norm": 0.10024792701005936, |
| "learning_rate": 5.11497849087622e-07, |
| "loss": 3.278700256347656, |
| "step": 5075 |
| }, |
| { |
| "epoch": 0.9759846301633045, |
| "grad_norm": 0.08857379108667374, |
| "learning_rate": 4.7292880542634805e-07, |
| "loss": 3.2780426025390623, |
| "step": 5080 |
| }, |
| { |
| "epoch": 0.9769452449567724, |
| "grad_norm": 0.08493303507566452, |
| "learning_rate": 4.358691546254067e-07, |
| "loss": 3.279518890380859, |
| "step": 5085 |
| }, |
| { |
| "epoch": 0.9779058597502401, |
| "grad_norm": 0.09684525430202484, |
| "learning_rate": 4.0031927078145176e-07, |
| "loss": 3.281083679199219, |
| "step": 5090 |
| }, |
| { |
| "epoch": 0.978866474543708, |
| "grad_norm": 0.09520915895700455, |
| "learning_rate": 3.662795127508111e-07, |
| "loss": 3.2831382751464844, |
| "step": 5095 |
| }, |
| { |
| "epoch": 0.9798270893371758, |
| "grad_norm": 0.09378495812416077, |
| "learning_rate": 3.3375022414598994e-07, |
| "loss": 3.280741882324219, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.9807877041306436, |
| "grad_norm": 0.0972544252872467, |
| "learning_rate": 3.027317333321233e-07, |
| "loss": 3.282391357421875, |
| "step": 5105 |
| }, |
| { |
| "epoch": 0.9817483189241114, |
| "grad_norm": 0.09050152450799942, |
| "learning_rate": 2.7322435342364556e-07, |
| "loss": 3.2794769287109373, |
| "step": 5110 |
| }, |
| { |
| "epoch": 0.9827089337175793, |
| "grad_norm": 0.098331019282341, |
| "learning_rate": 2.452283822812262e-07, |
| "loss": 3.283687210083008, |
| "step": 5115 |
| }, |
| { |
| "epoch": 0.983669548511047, |
| "grad_norm": 0.09614613652229309, |
| "learning_rate": 2.1874410250863893e-07, |
| "loss": 3.2843692779541014, |
| "step": 5120 |
| }, |
| { |
| "epoch": 0.9846301633045149, |
| "grad_norm": 0.08628969639539719, |
| "learning_rate": 1.9377178145003059e-07, |
| "loss": 3.2782211303710938, |
| "step": 5125 |
| }, |
| { |
| "epoch": 0.9855907780979827, |
| "grad_norm": 0.09115748107433319, |
| "learning_rate": 1.7031167118708998e-07, |
| "loss": 3.282467269897461, |
| "step": 5130 |
| }, |
| { |
| "epoch": 0.9865513928914506, |
| "grad_norm": 0.08609933406114578, |
| "learning_rate": 1.4836400853666662e-07, |
| "loss": 3.280436706542969, |
| "step": 5135 |
| }, |
| { |
| "epoch": 0.9875120076849183, |
| "grad_norm": 0.08899597823619843, |
| "learning_rate": 1.2792901504820595e-07, |
| "loss": 3.2786022186279298, |
| "step": 5140 |
| }, |
| { |
| "epoch": 0.9884726224783862, |
| "grad_norm": 0.0838359072804451, |
| "learning_rate": 1.0900689700166776e-07, |
| "loss": 3.283920669555664, |
| "step": 5145 |
| }, |
| { |
| "epoch": 0.989433237271854, |
| "grad_norm": 0.08638288825750351, |
| "learning_rate": 9.159784540531124e-08, |
| "loss": 3.2787384033203124, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.9903938520653218, |
| "grad_norm": 0.08724058419466019, |
| "learning_rate": 7.57020359938798e-08, |
| "loss": 3.2797054290771483, |
| "step": 5155 |
| }, |
| { |
| "epoch": 0.9913544668587896, |
| "grad_norm": 0.08608481287956238, |
| "learning_rate": 6.131962922673595e-08, |
| "loss": 3.2824764251708984, |
| "step": 5160 |
| }, |
| { |
| "epoch": 0.9923150816522575, |
| "grad_norm": 0.08680781722068787, |
| "learning_rate": 4.845077028631239e-08, |
| "loss": 3.2829864501953123, |
| "step": 5165 |
| }, |
| { |
| "epoch": 0.9932756964457252, |
| "grad_norm": 0.08766987174749374, |
| "learning_rate": 3.709558907659671e-08, |
| "loss": 3.279438018798828, |
| "step": 5170 |
| }, |
| { |
| "epoch": 0.9942363112391931, |
| "grad_norm": 0.08375875651836395, |
| "learning_rate": 2.7254200221848988e-08, |
| "loss": 3.28199462890625, |
| "step": 5175 |
| }, |
| { |
| "epoch": 0.9951969260326609, |
| "grad_norm": 0.08243842422962189, |
| "learning_rate": 1.8926703065436087e-08, |
| "loss": 3.281136322021484, |
| "step": 5180 |
| }, |
| { |
| "epoch": 0.9961575408261287, |
| "grad_norm": 0.08867152035236359, |
| "learning_rate": 1.2113181668815808e-08, |
| "loss": 3.2808319091796876, |
| "step": 5185 |
| }, |
| { |
| "epoch": 0.9971181556195965, |
| "grad_norm": 0.08233336359262466, |
| "learning_rate": 6.813704810704201e-09, |
| "loss": 3.28259162902832, |
| "step": 5190 |
| }, |
| { |
| "epoch": 0.9980787704130644, |
| "grad_norm": 0.08597059547901154, |
| "learning_rate": 3.028325986392799e-09, |
| "loss": 3.281261444091797, |
| "step": 5195 |
| }, |
| { |
| "epoch": 0.9990393852065321, |
| "grad_norm": 0.08740798383951187, |
| "learning_rate": 7.570834071823905e-10, |
| "loss": 3.2809608459472654, |
| "step": 5200 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.08548780530691147, |
| "learning_rate": 0.0, |
| "loss": 3.2842376708984373, |
| "step": 5205 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 5205, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 9.770678188442963e+19, |
| "train_batch_size": 64, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|