| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.9997049277072882, |
| "eval_steps": 500, |
| "global_step": 10166, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.009835743090390479, |
| "grad_norm": 0.05815335735678673, |
| "learning_rate": 0.00019999032134813635, |
| "loss": 0.9008, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.019671486180780958, |
| "grad_norm": 0.06804105639457703, |
| "learning_rate": 0.00019995686668528316, |
| "loss": 0.7784, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.029507229271171435, |
| "grad_norm": 0.07179820537567139, |
| "learning_rate": 0.0001998995245149411, |
| "loss": 0.765, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.039342972361561916, |
| "grad_norm": 0.07887151092290878, |
| "learning_rate": 0.00019981830854063617, |
| "loss": 0.7539, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.0491787154519524, |
| "grad_norm": 0.06816300749778748, |
| "learning_rate": 0.00019971323817121194, |
| "loss": 0.7355, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.05901445854234287, |
| "grad_norm": 0.0781150683760643, |
| "learning_rate": 0.00019958433851619142, |
| "loss": 0.7295, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.06885020163273335, |
| "grad_norm": 0.07763133943080902, |
| "learning_rate": 0.00019943164037977625, |
| "loss": 0.7326, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.07868594472312383, |
| "grad_norm": 0.07704794406890869, |
| "learning_rate": 0.00019925518025348511, |
| "loss": 0.7267, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.08852168781351431, |
| "grad_norm": 0.06781066209077835, |
| "learning_rate": 0.00019905500030743332, |
| "loss": 0.7203, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.0983574309039048, |
| "grad_norm": 0.06790652126073837, |
| "learning_rate": 0.00019883114838025484, |
| "loss": 0.7243, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.10819317399429527, |
| "grad_norm": 0.07213232666254044, |
| "learning_rate": 0.00019858367796767002, |
| "loss": 0.7159, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.11802891708468574, |
| "grad_norm": 0.06770846992731094, |
| "learning_rate": 0.00019831264820970127, |
| "loss": 0.7115, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.12786466017507622, |
| "grad_norm": 0.07667895406484604, |
| "learning_rate": 0.00019801812387653983, |
| "loss": 0.7135, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.1377004032654667, |
| "grad_norm": 0.06398730725049973, |
| "learning_rate": 0.00019770017535306717, |
| "loss": 0.707, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.14753614635585718, |
| "grad_norm": 0.0711282268166542, |
| "learning_rate": 0.00019735887862203457, |
| "loss": 0.7045, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.15737188944624766, |
| "grad_norm": 0.06350903958082199, |
| "learning_rate": 0.00019699431524590477, |
| "loss": 0.6977, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.16720763253663815, |
| "grad_norm": 0.06291402131319046, |
| "learning_rate": 0.00019660657234736045, |
| "loss": 0.7017, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.17704337562702863, |
| "grad_norm": 0.07249592989683151, |
| "learning_rate": 0.00019619574258848376, |
| "loss": 0.6997, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.1868791187174191, |
| "grad_norm": 0.07148096710443497, |
| "learning_rate": 0.00019576192414861215, |
| "loss": 0.6931, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.1967148618078096, |
| "grad_norm": 0.080138199031353, |
| "learning_rate": 0.00019530522070087554, |
| "loss": 0.6956, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.20655060489820007, |
| "grad_norm": 0.07587867230176926, |
| "learning_rate": 0.00019482574138742086, |
| "loss": 0.6972, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.21638634798859055, |
| "grad_norm": 0.0822625681757927, |
| "learning_rate": 0.0001943236007933294, |
| "loss": 0.6887, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.22622209107898103, |
| "grad_norm": 0.06909151375293732, |
| "learning_rate": 0.0001937989189192334, |
| "loss": 0.6872, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.23605783416937148, |
| "grad_norm": 0.06804945319890976, |
| "learning_rate": 0.00019325182115263859, |
| "loss": 0.6875, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.24589357725976196, |
| "grad_norm": 0.07097342610359192, |
| "learning_rate": 0.00019268243823795936, |
| "loss": 0.6794, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.25572932035015244, |
| "grad_norm": 0.06648524850606918, |
| "learning_rate": 0.0001920909062452736, |
| "loss": 0.6844, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.2655650634405429, |
| "grad_norm": 0.06843659281730652, |
| "learning_rate": 0.00019147736653780508, |
| "loss": 0.684, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.2754008065309334, |
| "grad_norm": 0.06251444667577744, |
| "learning_rate": 0.00019084196573814047, |
| "loss": 0.6856, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.2852365496213239, |
| "grad_norm": 0.07899219542741776, |
| "learning_rate": 0.00019018485569319003, |
| "loss": 0.6753, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.29507229271171437, |
| "grad_norm": 0.07931499183177948, |
| "learning_rate": 0.00018950619343789934, |
| "loss": 0.6808, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.30490803580210485, |
| "grad_norm": 0.061664845794439316, |
| "learning_rate": 0.00018880614115772138, |
| "loss": 0.6797, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.31474377889249533, |
| "grad_norm": 0.0717281773686409, |
| "learning_rate": 0.00018808486614985782, |
| "loss": 0.6774, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.3245795219828858, |
| "grad_norm": 0.07137037068605423, |
| "learning_rate": 0.00018734254078327868, |
| "loss": 0.6787, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.3344152650732763, |
| "grad_norm": 0.07462574541568756, |
| "learning_rate": 0.0001865793424575298, |
| "loss": 0.6836, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.34425100816366677, |
| "grad_norm": 0.06719885766506195, |
| "learning_rate": 0.00018579545356033823, |
| "loss": 0.6737, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.35408675125405725, |
| "grad_norm": 0.06752898544073105, |
| "learning_rate": 0.00018499106142402563, |
| "loss": 0.6767, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.36392249434444773, |
| "grad_norm": 0.072923943400383, |
| "learning_rate": 0.00018416635828073994, |
| "loss": 0.6756, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.3737582374348382, |
| "grad_norm": 0.07471544295549393, |
| "learning_rate": 0.00018332154121651598, |
| "loss": 0.675, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.3835939805252287, |
| "grad_norm": 0.0654042437672615, |
| "learning_rate": 0.00018245681212417631, |
| "loss": 0.6662, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.3934297236156192, |
| "grad_norm": 0.07551202178001404, |
| "learning_rate": 0.00018157237765508325, |
| "loss": 0.6635, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.40326546670600966, |
| "grad_norm": 0.07835708558559418, |
| "learning_rate": 0.00018066844916975354, |
| "loss": 0.6701, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.41310120979640014, |
| "grad_norm": 0.06583021581172943, |
| "learning_rate": 0.0001797452426873481, |
| "loss": 0.6644, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.4229369528867906, |
| "grad_norm": 0.07056088745594025, |
| "learning_rate": 0.0001788029788340479, |
| "loss": 0.6696, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.4327726959771811, |
| "grad_norm": 0.061976175755262375, |
| "learning_rate": 0.00017784188279032932, |
| "loss": 0.6641, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.4426084390675716, |
| "grad_norm": 0.06996449083089828, |
| "learning_rate": 0.00017686218423715072, |
| "loss": 0.6666, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.45244418215796206, |
| "grad_norm": 0.07750783860683441, |
| "learning_rate": 0.00017586411730106399, |
| "loss": 0.6608, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.4622799252483525, |
| "grad_norm": 0.06929858028888702, |
| "learning_rate": 0.00017484792049826306, |
| "loss": 0.6693, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.47211566833874297, |
| "grad_norm": 0.07935669273138046, |
| "learning_rate": 0.00017381383667758416, |
| "loss": 0.6709, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.48195141142913345, |
| "grad_norm": 0.06911829113960266, |
| "learning_rate": 0.00017276211296246986, |
| "loss": 0.6592, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.49178715451952393, |
| "grad_norm": 0.06784563511610031, |
| "learning_rate": 0.00017169300069191224, |
| "loss": 0.6631, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.5016228976099144, |
| "grad_norm": 0.06838862597942352, |
| "learning_rate": 0.00017060675536038818, |
| "loss": 0.6685, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.5114586407003049, |
| "grad_norm": 0.07473283261060715, |
| "learning_rate": 0.000169503636556802, |
| "loss": 0.6579, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.5212943837906954, |
| "grad_norm": 0.07981903851032257, |
| "learning_rate": 0.000168383907902449, |
| "loss": 0.6531, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.5311301268810859, |
| "grad_norm": 0.07383566349744797, |
| "learning_rate": 0.00016724783698801615, |
| "loss": 0.6552, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.5409658699714763, |
| "grad_norm": 0.075548455119133, |
| "learning_rate": 0.00016609569530963352, |
| "loss": 0.665, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.5508016130618668, |
| "grad_norm": 0.0663590133190155, |
| "learning_rate": 0.00016492775820399258, |
| "loss": 0.663, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.5606373561522573, |
| "grad_norm": 0.07384879887104034, |
| "learning_rate": 0.00016374430478254697, |
| "loss": 0.656, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.5704730992426478, |
| "grad_norm": 0.06946936994791031, |
| "learning_rate": 0.00016254561786481077, |
| "loss": 0.6585, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.5803088423330383, |
| "grad_norm": 0.06723761558532715, |
| "learning_rate": 0.00016133198391077096, |
| "loss": 0.6563, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.5901445854234287, |
| "grad_norm": 0.06926653534173965, |
| "learning_rate": 0.00016010369295242955, |
| "loss": 0.6542, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.5999803285138192, |
| "grad_norm": 0.07461749017238617, |
| "learning_rate": 0.00015886103852449254, |
| "loss": 0.6618, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.6098160716042097, |
| "grad_norm": 0.07314767688512802, |
| "learning_rate": 0.00015760431759422118, |
| "loss": 0.6509, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.6196518146946002, |
| "grad_norm": 0.07121975719928741, |
| "learning_rate": 0.00015633383049046365, |
| "loss": 0.6574, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.6294875577849907, |
| "grad_norm": 0.08515851944684982, |
| "learning_rate": 0.00015504988083188281, |
| "loss": 0.6576, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.6393233008753811, |
| "grad_norm": 0.07423117756843567, |
| "learning_rate": 0.000153752775454398, |
| "loss": 0.6512, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.6491590439657716, |
| "grad_norm": 0.06631764024496078, |
| "learning_rate": 0.00015244282433785796, |
| "loss": 0.6552, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.6589947870561621, |
| "grad_norm": 0.06597639620304108, |
| "learning_rate": 0.00015112034053196247, |
| "loss": 0.6572, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.6688305301465526, |
| "grad_norm": 0.07060743868350983, |
| "learning_rate": 0.00014978564008145032, |
| "loss": 0.6468, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.6786662732369431, |
| "grad_norm": 0.07211815565824509, |
| "learning_rate": 0.00014843904195057137, |
| "loss": 0.6476, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.6885020163273335, |
| "grad_norm": 0.0679902508854866, |
| "learning_rate": 0.00014708086794686128, |
| "loss": 0.6516, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.698337759417724, |
| "grad_norm": 0.06677327305078506, |
| "learning_rate": 0.00014571144264423642, |
| "loss": 0.654, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.7081735025081145, |
| "grad_norm": 0.07432771474123001, |
| "learning_rate": 0.00014433109330542769, |
| "loss": 0.6484, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.718009245598505, |
| "grad_norm": 0.07363554835319519, |
| "learning_rate": 0.00014294014980377213, |
| "loss": 0.6477, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.7278449886888955, |
| "grad_norm": 0.07398252189159393, |
| "learning_rate": 0.00014153894454438018, |
| "loss": 0.6494, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.737680731779286, |
| "grad_norm": 0.06380564719438553, |
| "learning_rate": 0.00014012781238469823, |
| "loss": 0.6511, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.7475164748696764, |
| "grad_norm": 0.06673400104045868, |
| "learning_rate": 0.00013870709055448504, |
| "loss": 0.649, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.7573522179600669, |
| "grad_norm": 0.07433997839689255, |
| "learning_rate": 0.00013727711857522095, |
| "loss": 0.642, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.7671879610504574, |
| "grad_norm": 0.07073818892240524, |
| "learning_rate": 0.00013583823817897006, |
| "loss": 0.6371, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.7770237041408479, |
| "grad_norm": 0.06574368476867676, |
| "learning_rate": 0.0001343907932267134, |
| "loss": 0.6493, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.7868594472312384, |
| "grad_norm": 0.067158043384552, |
| "learning_rate": 0.00013293512962617377, |
| "loss": 0.6433, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.7966951903216288, |
| "grad_norm": 0.07653222978115082, |
| "learning_rate": 0.0001314715952491514, |
| "loss": 0.6457, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.8065309334120193, |
| "grad_norm": 0.07320449501276016, |
| "learning_rate": 0.0001300005398483902, |
| "loss": 0.642, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.8163666765024098, |
| "grad_norm": 0.07599062472581863, |
| "learning_rate": 0.0001285223149739944, |
| "loss": 0.6444, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.8262024195928003, |
| "grad_norm": 0.07315944135189056, |
| "learning_rate": 0.00012703727388941577, |
| "loss": 0.646, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.8360381626831908, |
| "grad_norm": 0.0766439437866211, |
| "learning_rate": 0.00012554577148703148, |
| "loss": 0.6391, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.8458739057735812, |
| "grad_norm": 0.07688874751329422, |
| "learning_rate": 0.00012404816420333247, |
| "loss": 0.6423, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.8557096488639717, |
| "grad_norm": 0.07058276981115341, |
| "learning_rate": 0.0001225448099337429, |
| "loss": 0.6473, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.8655453919543622, |
| "grad_norm": 0.07005015760660172, |
| "learning_rate": 0.00012103606794709112, |
| "loss": 0.638, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.8753811350447527, |
| "grad_norm": 0.06887346506118774, |
| "learning_rate": 0.00011952229879975207, |
| "loss": 0.6428, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.8852168781351432, |
| "grad_norm": 0.06948423385620117, |
| "learning_rate": 0.00011800386424948227, |
| "loss": 0.6413, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.8950526212255336, |
| "grad_norm": 0.08220444619655609, |
| "learning_rate": 0.00011648112716896771, |
| "loss": 0.6413, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.9048883643159241, |
| "grad_norm": 0.07942084223031998, |
| "learning_rate": 0.000114954451459105, |
| "loss": 0.642, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.9147241074063146, |
| "grad_norm": 0.074773870408535, |
| "learning_rate": 0.00011342420196203719, |
| "loss": 0.64, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.924559850496705, |
| "grad_norm": 0.06527584791183472, |
| "learning_rate": 0.00011189074437396438, |
| "loss": 0.6448, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.9343955935870955, |
| "grad_norm": 0.0640060305595398, |
| "learning_rate": 0.00011035444515775035, |
| "loss": 0.6381, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.9442313366774859, |
| "grad_norm": 0.06932298839092255, |
| "learning_rate": 0.00010881567145534591, |
| "loss": 0.6424, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.9540670797678764, |
| "grad_norm": 0.0729447677731514, |
| "learning_rate": 0.00010727479100005005, |
| "loss": 0.6398, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.9639028228582669, |
| "grad_norm": 0.06698109209537506, |
| "learning_rate": 0.00010573217202862959, |
| "loss": 0.64, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.9737385659486574, |
| "grad_norm": 0.0725114643573761, |
| "learning_rate": 0.0001041881831933188, |
| "loss": 0.6389, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.9835743090390479, |
| "grad_norm": 0.06850885599851608, |
| "learning_rate": 0.00010264319347371951, |
| "loss": 0.6341, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.9934100521294383, |
| "grad_norm": 0.07343582063913345, |
| "learning_rate": 0.00010109757208862299, |
| "loss": 0.6315, |
| "step": 5050 |
| }, |
| { |
| "epoch": 1.003147437788925, |
| "grad_norm": 0.07858431339263916, |
| "learning_rate": 9.955168840777474e-05, |
| "loss": 0.6336, |
| "step": 5100 |
| }, |
| { |
| "epoch": 1.0129831808793155, |
| "grad_norm": 0.07285405695438385, |
| "learning_rate": 9.800591186360323e-05, |
| "loss": 0.6202, |
| "step": 5150 |
| }, |
| { |
| "epoch": 1.0228189239697059, |
| "grad_norm": 0.07421938329935074, |
| "learning_rate": 9.646061186293367e-05, |
| "loss": 0.6256, |
| "step": 5200 |
| }, |
| { |
| "epoch": 1.0326546670600965, |
| "grad_norm": 0.06922327727079391, |
| "learning_rate": 9.491615769870769e-05, |
| "loss": 0.6214, |
| "step": 5250 |
| }, |
| { |
| "epoch": 1.0424904101504868, |
| "grad_norm": 0.08220986276865005, |
| "learning_rate": 9.337291846173059e-05, |
| "loss": 0.6272, |
| "step": 5300 |
| }, |
| { |
| "epoch": 1.0523261532408774, |
| "grad_norm": 0.06896129250526428, |
| "learning_rate": 9.183126295246645e-05, |
| "loss": 0.6219, |
| "step": 5350 |
| }, |
| { |
| "epoch": 1.0621618963312678, |
| "grad_norm": 0.07372142374515533, |
| "learning_rate": 9.029155959290319e-05, |
| "loss": 0.6171, |
| "step": 5400 |
| }, |
| { |
| "epoch": 1.0719976394216584, |
| "grad_norm": 0.07114165276288986, |
| "learning_rate": 8.875417633850746e-05, |
| "loss": 0.6189, |
| "step": 5450 |
| }, |
| { |
| "epoch": 1.0818333825120487, |
| "grad_norm": 0.0711623951792717, |
| "learning_rate": 8.721948059029161e-05, |
| "loss": 0.6167, |
| "step": 5500 |
| }, |
| { |
| "epoch": 1.0916691256024393, |
| "grad_norm": 0.07456561177968979, |
| "learning_rate": 8.568783910701252e-05, |
| "loss": 0.6141, |
| "step": 5550 |
| }, |
| { |
| "epoch": 1.1015048686928297, |
| "grad_norm": 0.0760912150144577, |
| "learning_rate": 8.415961791752472e-05, |
| "loss": 0.6167, |
| "step": 5600 |
| }, |
| { |
| "epoch": 1.1113406117832203, |
| "grad_norm": 0.07034651935100555, |
| "learning_rate": 8.263518223330697e-05, |
| "loss": 0.6125, |
| "step": 5650 |
| }, |
| { |
| "epoch": 1.1211763548736107, |
| "grad_norm": 0.07252858579158783, |
| "learning_rate": 8.111489636118522e-05, |
| "loss": 0.6181, |
| "step": 5700 |
| }, |
| { |
| "epoch": 1.1310120979640013, |
| "grad_norm": 0.07909699529409409, |
| "learning_rate": 7.959912361627082e-05, |
| "loss": 0.6214, |
| "step": 5750 |
| }, |
| { |
| "epoch": 1.1408478410543916, |
| "grad_norm": 0.07374490797519684, |
| "learning_rate": 7.808822623513643e-05, |
| "loss": 0.625, |
| "step": 5800 |
| }, |
| { |
| "epoch": 1.1506835841447822, |
| "grad_norm": 0.07597927004098892, |
| "learning_rate": 7.658256528924909e-05, |
| "loss": 0.6199, |
| "step": 5850 |
| }, |
| { |
| "epoch": 1.1605193272351726, |
| "grad_norm": 0.07946628332138062, |
| "learning_rate": 7.508250059868249e-05, |
| "loss": 0.615, |
| "step": 5900 |
| }, |
| { |
| "epoch": 1.1703550703255632, |
| "grad_norm": 0.07706974446773529, |
| "learning_rate": 7.358839064612726e-05, |
| "loss": 0.6126, |
| "step": 5950 |
| }, |
| { |
| "epoch": 1.1801908134159536, |
| "grad_norm": 0.0789932906627655, |
| "learning_rate": 7.210059249122193e-05, |
| "loss": 0.6192, |
| "step": 6000 |
| }, |
| { |
| "epoch": 1.1900265565063441, |
| "grad_norm": 0.08002398908138275, |
| "learning_rate": 7.061946168522318e-05, |
| "loss": 0.6167, |
| "step": 6050 |
| }, |
| { |
| "epoch": 1.1998622995967345, |
| "grad_norm": 0.07844787836074829, |
| "learning_rate": 6.914535218603708e-05, |
| "loss": 0.6208, |
| "step": 6100 |
| }, |
| { |
| "epoch": 1.209698042687125, |
| "grad_norm": 0.08679769188165665, |
| "learning_rate": 6.767861627363054e-05, |
| "loss": 0.6157, |
| "step": 6150 |
| }, |
| { |
| "epoch": 1.2195337857775155, |
| "grad_norm": 0.07716654241085052, |
| "learning_rate": 6.621960446584452e-05, |
| "loss": 0.61, |
| "step": 6200 |
| }, |
| { |
| "epoch": 1.229369528867906, |
| "grad_norm": 0.08282492309808731, |
| "learning_rate": 6.476866543462761e-05, |
| "loss": 0.6145, |
| "step": 6250 |
| }, |
| { |
| "epoch": 1.2392052719582964, |
| "grad_norm": 0.09066256135702133, |
| "learning_rate": 6.332614592271122e-05, |
| "loss": 0.6237, |
| "step": 6300 |
| }, |
| { |
| "epoch": 1.2490410150486868, |
| "grad_norm": 0.07888253778219223, |
| "learning_rate": 6.18923906607455e-05, |
| "loss": 0.6149, |
| "step": 6350 |
| }, |
| { |
| "epoch": 1.2588767581390774, |
| "grad_norm": 0.08149804919958115, |
| "learning_rate": 6.046774228491643e-05, |
| "loss": 0.6166, |
| "step": 6400 |
| }, |
| { |
| "epoch": 1.268712501229468, |
| "grad_norm": 0.07619079202413559, |
| "learning_rate": 5.905254125506301e-05, |
| "loss": 0.6111, |
| "step": 6450 |
| }, |
| { |
| "epoch": 1.2785482443198584, |
| "grad_norm": 0.08567807078361511, |
| "learning_rate": 5.76471257733151e-05, |
| "loss": 0.6153, |
| "step": 6500 |
| }, |
| { |
| "epoch": 1.2883839874102487, |
| "grad_norm": 0.0711495652794838, |
| "learning_rate": 5.625183170327017e-05, |
| "loss": 0.6126, |
| "step": 6550 |
| }, |
| { |
| "epoch": 1.2982197305006393, |
| "grad_norm": 0.08142837882041931, |
| "learning_rate": 5.4866992489729554e-05, |
| "loss": 0.6212, |
| "step": 6600 |
| }, |
| { |
| "epoch": 1.30805547359103, |
| "grad_norm": 0.07909776270389557, |
| "learning_rate": 5.3492939079012206e-05, |
| "loss": 0.6139, |
| "step": 6650 |
| }, |
| { |
| "epoch": 1.3178912166814203, |
| "grad_norm": 0.08278420567512512, |
| "learning_rate": 5.2129999839865796e-05, |
| "loss": 0.6115, |
| "step": 6700 |
| }, |
| { |
| "epoch": 1.3277269597718107, |
| "grad_norm": 0.08086064457893372, |
| "learning_rate": 5.077850048499388e-05, |
| "loss": 0.6111, |
| "step": 6750 |
| }, |
| { |
| "epoch": 1.3375627028622012, |
| "grad_norm": 0.07711977511644363, |
| "learning_rate": 4.9438763993217495e-05, |
| "loss": 0.6101, |
| "step": 6800 |
| }, |
| { |
| "epoch": 1.3473984459525918, |
| "grad_norm": 0.08161487430334091, |
| "learning_rate": 4.811111053229043e-05, |
| "loss": 0.6167, |
| "step": 6850 |
| }, |
| { |
| "epoch": 1.3572341890429822, |
| "grad_norm": 0.0841764435172081, |
| "learning_rate": 4.6795857382386044e-05, |
| "loss": 0.6121, |
| "step": 6900 |
| }, |
| { |
| "epoch": 1.3670699321333726, |
| "grad_norm": 0.0816132128238678, |
| "learning_rate": 4.549331886027429e-05, |
| "loss": 0.6078, |
| "step": 6950 |
| }, |
| { |
| "epoch": 1.3769056752237632, |
| "grad_norm": 0.08330381661653519, |
| "learning_rate": 4.4203806244206756e-05, |
| "loss": 0.6195, |
| "step": 7000 |
| }, |
| { |
| "epoch": 1.3867414183141538, |
| "grad_norm": 0.0845024362206459, |
| "learning_rate": 4.292762769952816e-05, |
| "loss": 0.6133, |
| "step": 7050 |
| }, |
| { |
| "epoch": 1.3965771614045441, |
| "grad_norm": 0.07662446796894073, |
| "learning_rate": 4.1665088205031334e-05, |
| "loss": 0.6129, |
| "step": 7100 |
| }, |
| { |
| "epoch": 1.4064129044949345, |
| "grad_norm": 0.08240839838981628, |
| "learning_rate": 4.041648948007416e-05, |
| "loss": 0.6156, |
| "step": 7150 |
| }, |
| { |
| "epoch": 1.416248647585325, |
| "grad_norm": 0.07303918898105621, |
| "learning_rate": 3.918212991247514e-05, |
| "loss": 0.6156, |
| "step": 7200 |
| }, |
| { |
| "epoch": 1.4260843906757157, |
| "grad_norm": 0.07239814847707748, |
| "learning_rate": 3.796230448720526e-05, |
| "loss": 0.6131, |
| "step": 7250 |
| }, |
| { |
| "epoch": 1.435920133766106, |
| "grad_norm": 0.07538265734910965, |
| "learning_rate": 3.675730471589286e-05, |
| "loss": 0.6116, |
| "step": 7300 |
| }, |
| { |
| "epoch": 1.4457558768564964, |
| "grad_norm": 0.08947084844112396, |
| "learning_rate": 3.556741856715907e-05, |
| "loss": 0.6147, |
| "step": 7350 |
| }, |
| { |
| "epoch": 1.455591619946887, |
| "grad_norm": 0.08240395039319992, |
| "learning_rate": 3.4392930397799194e-05, |
| "loss": 0.6125, |
| "step": 7400 |
| }, |
| { |
| "epoch": 1.4654273630372774, |
| "grad_norm": 0.07328338176012039, |
| "learning_rate": 3.3234120884828e-05, |
| "loss": 0.6041, |
| "step": 7450 |
| }, |
| { |
| "epoch": 1.475263106127668, |
| "grad_norm": 0.07761271297931671, |
| "learning_rate": 3.209126695840382e-05, |
| "loss": 0.6116, |
| "step": 7500 |
| }, |
| { |
| "epoch": 1.4850988492180583, |
| "grad_norm": 0.084846630692482, |
| "learning_rate": 3.0964641735648423e-05, |
| "loss": 0.6141, |
| "step": 7550 |
| }, |
| { |
| "epoch": 1.494934592308449, |
| "grad_norm": 0.08141667395830154, |
| "learning_rate": 2.9854514455377837e-05, |
| "loss": 0.6073, |
| "step": 7600 |
| }, |
| { |
| "epoch": 1.5047703353988395, |
| "grad_norm": 0.07505550980567932, |
| "learning_rate": 2.876115041376034e-05, |
| "loss": 0.6091, |
| "step": 7650 |
| }, |
| { |
| "epoch": 1.51460607848923, |
| "grad_norm": 0.07819739729166031, |
| "learning_rate": 2.7684810900916315e-05, |
| "loss": 0.6141, |
| "step": 7700 |
| }, |
| { |
| "epoch": 1.5244418215796203, |
| "grad_norm": 0.07576938718557358, |
| "learning_rate": 2.6625753138475718e-05, |
| "loss": 0.6046, |
| "step": 7750 |
| }, |
| { |
| "epoch": 1.5342775646700109, |
| "grad_norm": 0.07528570294380188, |
| "learning_rate": 2.5584230218107718e-05, |
| "loss": 0.6114, |
| "step": 7800 |
| }, |
| { |
| "epoch": 1.5441133077604015, |
| "grad_norm": 0.08330899477005005, |
| "learning_rate": 2.456049104103726e-05, |
| "loss": 0.609, |
| "step": 7850 |
| }, |
| { |
| "epoch": 1.5539490508507918, |
| "grad_norm": 0.0729718878865242, |
| "learning_rate": 2.3554780258563125e-05, |
| "loss": 0.6138, |
| "step": 7900 |
| }, |
| { |
| "epoch": 1.5637847939411822, |
| "grad_norm": 0.07709172368049622, |
| "learning_rate": 2.256733821359168e-05, |
| "loss": 0.6038, |
| "step": 7950 |
| }, |
| { |
| "epoch": 1.5736205370315728, |
| "grad_norm": 0.07105692476034164, |
| "learning_rate": 2.1598400883200065e-05, |
| "loss": 0.6078, |
| "step": 8000 |
| }, |
| { |
| "epoch": 1.5834562801219634, |
| "grad_norm": 0.0719430074095726, |
| "learning_rate": 2.0648199822242953e-05, |
| "loss": 0.6121, |
| "step": 8050 |
| }, |
| { |
| "epoch": 1.5932920232123537, |
| "grad_norm": 0.08640766143798828, |
| "learning_rate": 1.971696210801589e-05, |
| "loss": 0.5973, |
| "step": 8100 |
| }, |
| { |
| "epoch": 1.6031277663027441, |
| "grad_norm": 0.07854746282100677, |
| "learning_rate": 1.8804910285988885e-05, |
| "loss": 0.6176, |
| "step": 8150 |
| }, |
| { |
| "epoch": 1.6129635093931345, |
| "grad_norm": 0.09165063500404358, |
| "learning_rate": 1.7912262316622753e-05, |
| "loss": 0.6112, |
| "step": 8200 |
| }, |
| { |
| "epoch": 1.622799252483525, |
| "grad_norm": 0.07630006968975067, |
| "learning_rate": 1.703923152328145e-05, |
| "loss": 0.6053, |
| "step": 8250 |
| }, |
| { |
| "epoch": 1.6326349955739157, |
| "grad_norm": 0.0875934585928917, |
| "learning_rate": 1.6186026541252452e-05, |
| "loss": 0.6082, |
| "step": 8300 |
| }, |
| { |
| "epoch": 1.642470738664306, |
| "grad_norm": 0.08449984341859818, |
| "learning_rate": 1.5352851267887423e-05, |
| "loss": 0.6139, |
| "step": 8350 |
| }, |
| { |
| "epoch": 1.6523064817546964, |
| "grad_norm": 0.07857895642518997, |
| "learning_rate": 1.453990481387526e-05, |
| "loss": 0.6134, |
| "step": 8400 |
| }, |
| { |
| "epoch": 1.662142224845087, |
| "grad_norm": 0.0947275385260582, |
| "learning_rate": 1.3747381455658848e-05, |
| "loss": 0.6018, |
| "step": 8450 |
| }, |
| { |
| "epoch": 1.6719779679354776, |
| "grad_norm": 0.08134233206510544, |
| "learning_rate": 1.2975470589007454e-05, |
| "loss": 0.6127, |
| "step": 8500 |
| }, |
| { |
| "epoch": 1.681813711025868, |
| "grad_norm": 0.07631494104862213, |
| "learning_rate": 1.2224356683755089e-05, |
| "loss": 0.6105, |
| "step": 8550 |
| }, |
| { |
| "epoch": 1.6916494541162583, |
| "grad_norm": 0.07870069146156311, |
| "learning_rate": 1.1494219239716353e-05, |
| "loss": 0.6076, |
| "step": 8600 |
| }, |
| { |
| "epoch": 1.701485197206649, |
| "grad_norm": 0.08626607805490494, |
| "learning_rate": 1.0785232743789808e-05, |
| "loss": 0.6102, |
| "step": 8650 |
| }, |
| { |
| "epoch": 1.7113209402970395, |
| "grad_norm": 0.08474079519510269, |
| "learning_rate": 1.0097566628259614e-05, |
| "loss": 0.6018, |
| "step": 8700 |
| }, |
| { |
| "epoch": 1.7211566833874299, |
| "grad_norm": 0.08202967047691345, |
| "learning_rate": 9.431385230304613e-06, |
| "loss": 0.6016, |
| "step": 8750 |
| }, |
| { |
| "epoch": 1.7309924264778203, |
| "grad_norm": 0.07875273376703262, |
| "learning_rate": 8.786847752725614e-06, |
| "loss": 0.616, |
| "step": 8800 |
| }, |
| { |
| "epoch": 1.7408281695682108, |
| "grad_norm": 0.09360364079475403, |
| "learning_rate": 8.164108225899214e-06, |
| "loss": 0.6046, |
| "step": 8850 |
| }, |
| { |
| "epoch": 1.7506639126586014, |
| "grad_norm": 0.07391002029180527, |
| "learning_rate": 7.5633154709680575e-06, |
| "loss": 0.6117, |
| "step": 8900 |
| }, |
| { |
| "epoch": 1.7604996557489918, |
| "grad_norm": 0.06737073510885239, |
| "learning_rate": 6.984613064275658e-06, |
| "loss": 0.6034, |
| "step": 8950 |
| }, |
| { |
| "epoch": 1.7703353988393822, |
| "grad_norm": 0.0800662711262703, |
| "learning_rate": 6.428139303054981e-06, |
| "loss": 0.6054, |
| "step": 9000 |
| }, |
| { |
| "epoch": 1.7801711419297728, |
| "grad_norm": 0.07634767144918442, |
| "learning_rate": 5.894027172378247e-06, |
| "loss": 0.6041, |
| "step": 9050 |
| }, |
| { |
| "epoch": 1.7900068850201634, |
| "grad_norm": 0.0848456546664238, |
| "learning_rate": 5.382404313376555e-06, |
| "loss": 0.6013, |
| "step": 9100 |
| }, |
| { |
| "epoch": 1.7998426281105537, |
| "grad_norm": 0.08109795302152634, |
| "learning_rate": 4.893392992736301e-06, |
| "loss": 0.6037, |
| "step": 9150 |
| }, |
| { |
| "epoch": 1.809678371200944, |
| "grad_norm": 0.08365663886070251, |
| "learning_rate": 4.427110073480245e-06, |
| "loss": 0.604, |
| "step": 9200 |
| }, |
| { |
| "epoch": 1.8195141142913347, |
| "grad_norm": 0.07841967046260834, |
| "learning_rate": 3.983666987039691e-06, |
| "loss": 0.602, |
| "step": 9250 |
| }, |
| { |
| "epoch": 1.8293498573817253, |
| "grad_norm": 0.07770159840583801, |
| "learning_rate": 3.5631697066249024e-06, |
| "loss": 0.6051, |
| "step": 9300 |
| }, |
| { |
| "epoch": 1.8391856004721157, |
| "grad_norm": 0.07518593221902847, |
| "learning_rate": 3.1657187218997464e-06, |
| "loss": 0.6065, |
| "step": 9350 |
| }, |
| { |
| "epoch": 1.849021343562506, |
| "grad_norm": 0.08144789189100266, |
| "learning_rate": 2.791409014966895e-06, |
| "loss": 0.6096, |
| "step": 9400 |
| }, |
| { |
| "epoch": 1.8588570866528966, |
| "grad_norm": 0.08269818127155304, |
| "learning_rate": 2.4403300376691473e-06, |
| "loss": 0.61, |
| "step": 9450 |
| }, |
| { |
| "epoch": 1.8686928297432872, |
| "grad_norm": 0.07822602987289429, |
| "learning_rate": 2.112565690212465e-06, |
| "loss": 0.6063, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.8785285728336776, |
| "grad_norm": 0.09586796164512634, |
| "learning_rate": 1.8081943011155623e-06, |
| "loss": 0.6025, |
| "step": 9550 |
| }, |
| { |
| "epoch": 1.888364315924068, |
| "grad_norm": 0.0846213847398758, |
| "learning_rate": 1.527288608491173e-06, |
| "loss": 0.6098, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.8982000590144585, |
| "grad_norm": 0.08927745372056961, |
| "learning_rate": 1.2699157426631858e-06, |
| "loss": 0.6077, |
| "step": 9650 |
| }, |
| { |
| "epoch": 1.9080358021048491, |
| "grad_norm": 0.07482849806547165, |
| "learning_rate": 1.0361372101239997e-06, |
| "loss": 0.6045, |
| "step": 9700 |
| }, |
| { |
| "epoch": 1.9178715451952395, |
| "grad_norm": 0.07779071480035782, |
| "learning_rate": 8.260088788357711e-07, |
| "loss": 0.6068, |
| "step": 9750 |
| }, |
| { |
| "epoch": 1.9277072882856299, |
| "grad_norm": 0.079287588596344, |
| "learning_rate": 6.395809648792384e-07, |
| "loss": 0.6053, |
| "step": 9800 |
| }, |
| { |
| "epoch": 1.9375430313760205, |
| "grad_norm": 0.08390713483095169, |
| "learning_rate": 4.768980204531426e-07, |
| "loss": 0.6083, |
| "step": 9850 |
| }, |
| { |
| "epoch": 1.947378774466411, |
| "grad_norm": 0.07710904628038406, |
| "learning_rate": 3.3799892322726735e-07, |
| "loss": 0.6058, |
| "step": 9900 |
| }, |
| { |
| "epoch": 1.9572145175568014, |
| "grad_norm": 0.08125531673431396, |
| "learning_rate": 2.229168670514481e-07, |
| "loss": 0.6087, |
| "step": 9950 |
| }, |
| { |
| "epoch": 1.9670502606471918, |
| "grad_norm": 0.07178232818841934, |
| "learning_rate": 1.3167935402305098e-07, |
| "loss": 0.6064, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.9768860037375824, |
| "grad_norm": 0.08764079213142395, |
| "learning_rate": 6.430818791451909e-08, |
| "loss": 0.6119, |
| "step": 10050 |
| }, |
| { |
| "epoch": 1.986721746827973, |
| "grad_norm": 0.07766247540712357, |
| "learning_rate": 2.0819468962773868e-08, |
| "loss": 0.6038, |
| "step": 10100 |
| }, |
| { |
| "epoch": 1.9965574899183633, |
| "grad_norm": 0.08261710405349731, |
| "learning_rate": 1.2235900215817885e-09, |
| "loss": 0.6115, |
| "step": 10150 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 10166, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 50, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 5.464591551750786e+19, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|