| { |
| "best_metric": 0.65735477, |
| "best_model_checkpoint": "/cpfs04/shared/rlproject/zhangqi/model_garden/0709_intern2b_v7-1-part15-19-resize-decay/v0-20250710-072707/checkpoint-3000", |
| "epoch": 0.9599616015359386, |
| "eval_steps": 250, |
| "global_step": 3000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0003199872005119795, |
| "grad_norm": 0.06346331978668461, |
| "learning_rate": 2.0000000000000002e-07, |
| "loss": 0.4898327589035034, |
| "memory(GiB)": 37.39, |
| "step": 1, |
| "token_acc": 0.8589147286821706, |
| "train_speed(iter/s)": 0.017141 |
| }, |
| { |
| "epoch": 0.0015999360025598975, |
| "grad_norm": 0.06701772816278752, |
| "learning_rate": 1.0000000000000002e-06, |
| "loss": 0.479672372341156, |
| "memory(GiB)": 57.22, |
| "step": 5, |
| "token_acc": 0.8759901666211418, |
| "train_speed(iter/s)": 0.029212 |
| }, |
| { |
| "epoch": 0.003199872005119795, |
| "grad_norm": 0.0577242460197625, |
| "learning_rate": 2.0000000000000003e-06, |
| "loss": 0.4767899990081787, |
| "memory(GiB)": 57.22, |
| "step": 10, |
| "token_acc": 0.8920853213584058, |
| "train_speed(iter/s)": 0.030985 |
| }, |
| { |
| "epoch": 0.004799808007679693, |
| "grad_norm": 0.05378464715409511, |
| "learning_rate": 3e-06, |
| "loss": 0.46750531196594236, |
| "memory(GiB)": 57.22, |
| "step": 15, |
| "token_acc": 0.846137292877125, |
| "train_speed(iter/s)": 0.030693 |
| }, |
| { |
| "epoch": 0.00639974401023959, |
| "grad_norm": 0.05482863789544427, |
| "learning_rate": 4.000000000000001e-06, |
| "loss": 0.47017059326171873, |
| "memory(GiB)": 57.22, |
| "step": 20, |
| "token_acc": 0.8701638201463925, |
| "train_speed(iter/s)": 0.031839 |
| }, |
| { |
| "epoch": 0.007999680012799487, |
| "grad_norm": 0.05575025715512655, |
| "learning_rate": 5e-06, |
| "loss": 0.46974716186523435, |
| "memory(GiB)": 57.22, |
| "step": 25, |
| "token_acc": 0.8939419941650936, |
| "train_speed(iter/s)": 0.031203 |
| }, |
| { |
| "epoch": 0.009599616015359386, |
| "grad_norm": 0.05007200892621057, |
| "learning_rate": 6e-06, |
| "loss": 0.4700496196746826, |
| "memory(GiB)": 57.22, |
| "step": 30, |
| "token_acc": 0.872761844398669, |
| "train_speed(iter/s)": 0.030738 |
| }, |
| { |
| "epoch": 0.011199552017919284, |
| "grad_norm": 0.06284973201247342, |
| "learning_rate": 7e-06, |
| "loss": 0.477018404006958, |
| "memory(GiB)": 57.22, |
| "step": 35, |
| "token_acc": 0.8738273921200751, |
| "train_speed(iter/s)": 0.031491 |
| }, |
| { |
| "epoch": 0.01279948802047918, |
| "grad_norm": 0.051858307829802386, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 0.46584124565124513, |
| "memory(GiB)": 57.22, |
| "step": 40, |
| "token_acc": 0.8791312559017942, |
| "train_speed(iter/s)": 0.031035 |
| }, |
| { |
| "epoch": 0.014399424023039079, |
| "grad_norm": 0.05024484287076301, |
| "learning_rate": 9e-06, |
| "loss": 0.4685808658599854, |
| "memory(GiB)": 57.22, |
| "step": 45, |
| "token_acc": 0.8846325167037862, |
| "train_speed(iter/s)": 0.031503 |
| }, |
| { |
| "epoch": 0.015999360025598975, |
| "grad_norm": 0.06025612278216295, |
| "learning_rate": 1e-05, |
| "loss": 0.4697974681854248, |
| "memory(GiB)": 57.22, |
| "step": 50, |
| "token_acc": 0.854153041203401, |
| "train_speed(iter/s)": 0.031549 |
| }, |
| { |
| "epoch": 0.017599296028158875, |
| "grad_norm": 0.05171252494611451, |
| "learning_rate": 1.1000000000000001e-05, |
| "loss": 0.4622661113739014, |
| "memory(GiB)": 73.06, |
| "step": 55, |
| "token_acc": 0.8281821878812525, |
| "train_speed(iter/s)": 0.031181 |
| }, |
| { |
| "epoch": 0.01919923203071877, |
| "grad_norm": 0.06238294030934267, |
| "learning_rate": 1.2e-05, |
| "loss": 0.47000856399536134, |
| "memory(GiB)": 73.06, |
| "step": 60, |
| "token_acc": 0.8703465982028241, |
| "train_speed(iter/s)": 0.031617 |
| }, |
| { |
| "epoch": 0.020799168033278668, |
| "grad_norm": 0.05576663204628902, |
| "learning_rate": 1.3000000000000001e-05, |
| "loss": 0.4765446186065674, |
| "memory(GiB)": 73.06, |
| "step": 65, |
| "token_acc": 0.867019517036057, |
| "train_speed(iter/s)": 0.031138 |
| }, |
| { |
| "epoch": 0.022399104035838568, |
| "grad_norm": 0.054406694286476175, |
| "learning_rate": 1.4e-05, |
| "loss": 0.47959036827087403, |
| "memory(GiB)": 73.06, |
| "step": 70, |
| "token_acc": 0.8927940657011657, |
| "train_speed(iter/s)": 0.030807 |
| }, |
| { |
| "epoch": 0.023999040038398464, |
| "grad_norm": 0.05458413282309297, |
| "learning_rate": 1.5000000000000002e-05, |
| "loss": 0.45742173194885255, |
| "memory(GiB)": 73.06, |
| "step": 75, |
| "token_acc": 0.8916037316748112, |
| "train_speed(iter/s)": 0.031204 |
| }, |
| { |
| "epoch": 0.02559897604095836, |
| "grad_norm": 0.05986601718141533, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 0.46221466064453126, |
| "memory(GiB)": 73.06, |
| "step": 80, |
| "token_acc": 0.8513287970214405, |
| "train_speed(iter/s)": 0.030907 |
| }, |
| { |
| "epoch": 0.02719891204351826, |
| "grad_norm": 0.0610597448330617, |
| "learning_rate": 1.7e-05, |
| "loss": 0.4636178493499756, |
| "memory(GiB)": 73.06, |
| "step": 85, |
| "token_acc": 0.8755391250770179, |
| "train_speed(iter/s)": 0.030852 |
| }, |
| { |
| "epoch": 0.028798848046078157, |
| "grad_norm": 0.0537272421077024, |
| "learning_rate": 1.8e-05, |
| "loss": 0.4674212455749512, |
| "memory(GiB)": 73.06, |
| "step": 90, |
| "token_acc": 0.8330635370295427, |
| "train_speed(iter/s)": 0.03096 |
| }, |
| { |
| "epoch": 0.030398784048638054, |
| "grad_norm": 0.05874793146275866, |
| "learning_rate": 1.9e-05, |
| "loss": 0.459043550491333, |
| "memory(GiB)": 73.06, |
| "step": 95, |
| "token_acc": 0.8686557483379277, |
| "train_speed(iter/s)": 0.030744 |
| }, |
| { |
| "epoch": 0.03199872005119795, |
| "grad_norm": 0.0635697375742028, |
| "learning_rate": 2e-05, |
| "loss": 0.4696988582611084, |
| "memory(GiB)": 73.06, |
| "step": 100, |
| "token_acc": 0.8658624414836155, |
| "train_speed(iter/s)": 0.030967 |
| }, |
| { |
| "epoch": 0.03359865605375785, |
| "grad_norm": 0.05212754941929234, |
| "learning_rate": 1.9999865178850847e-05, |
| "loss": 0.47245235443115235, |
| "memory(GiB)": 73.06, |
| "step": 105, |
| "token_acc": 0.8773064185459707, |
| "train_speed(iter/s)": 0.030857 |
| }, |
| { |
| "epoch": 0.03519859205631775, |
| "grad_norm": 0.0584305183084298, |
| "learning_rate": 1.999946071903873e-05, |
| "loss": 0.4617309093475342, |
| "memory(GiB)": 73.06, |
| "step": 110, |
| "token_acc": 0.8634434872058634, |
| "train_speed(iter/s)": 0.030617 |
| }, |
| { |
| "epoch": 0.03679852805887764, |
| "grad_norm": 0.0594956422290189, |
| "learning_rate": 1.9998786631469602e-05, |
| "loss": 0.4737922191619873, |
| "memory(GiB)": 73.06, |
| "step": 115, |
| "token_acc": 0.8586429725363489, |
| "train_speed(iter/s)": 0.030917 |
| }, |
| { |
| "epoch": 0.03839846406143754, |
| "grad_norm": 0.05542732612540538, |
| "learning_rate": 1.999784293431971e-05, |
| "loss": 0.46903514862060547, |
| "memory(GiB)": 73.06, |
| "step": 120, |
| "token_acc": 0.8490970309152127, |
| "train_speed(iter/s)": 0.030707 |
| }, |
| { |
| "epoch": 0.03999840006399744, |
| "grad_norm": 0.06279644130355853, |
| "learning_rate": 1.9996629653035128e-05, |
| "loss": 0.47660508155822756, |
| "memory(GiB)": 73.06, |
| "step": 125, |
| "token_acc": 0.8797399783315276, |
| "train_speed(iter/s)": 0.03059 |
| }, |
| { |
| "epoch": 0.041598336066557336, |
| "grad_norm": 0.0589281224878424, |
| "learning_rate": 1.999514682033104e-05, |
| "loss": 0.47182955741882326, |
| "memory(GiB)": 73.06, |
| "step": 130, |
| "token_acc": 0.8641215106732348, |
| "train_speed(iter/s)": 0.030792 |
| }, |
| { |
| "epoch": 0.043198272069117236, |
| "grad_norm": 0.06201767062726238, |
| "learning_rate": 1.99933944761909e-05, |
| "loss": 0.4760914325714111, |
| "memory(GiB)": 73.06, |
| "step": 135, |
| "token_acc": 0.856929955290611, |
| "train_speed(iter/s)": 0.030572 |
| }, |
| { |
| "epoch": 0.044798208071677136, |
| "grad_norm": 0.06842512509591067, |
| "learning_rate": 1.999137266786531e-05, |
| "loss": 0.4673017501831055, |
| "memory(GiB)": 73.06, |
| "step": 140, |
| "token_acc": 0.8725868725868726, |
| "train_speed(iter/s)": 0.030583 |
| }, |
| { |
| "epoch": 0.04639814407423703, |
| "grad_norm": 0.06206860242026748, |
| "learning_rate": 1.998908144987078e-05, |
| "loss": 0.4650784969329834, |
| "memory(GiB)": 73.06, |
| "step": 145, |
| "token_acc": 0.8592991206195039, |
| "train_speed(iter/s)": 0.030589 |
| }, |
| { |
| "epoch": 0.04799808007679693, |
| "grad_norm": 0.07443005443822684, |
| "learning_rate": 1.9986520883988233e-05, |
| "loss": 0.46671695709228517, |
| "memory(GiB)": 73.06, |
| "step": 150, |
| "token_acc": 0.8422288642186165, |
| "train_speed(iter/s)": 0.030331 |
| }, |
| { |
| "epoch": 0.04959801607935683, |
| "grad_norm": 0.06105984389773331, |
| "learning_rate": 1.9983691039261358e-05, |
| "loss": 0.47356271743774414, |
| "memory(GiB)": 73.06, |
| "step": 155, |
| "token_acc": 0.8814565604591332, |
| "train_speed(iter/s)": 0.030476 |
| }, |
| { |
| "epoch": 0.05119795208191672, |
| "grad_norm": 0.08001789963695773, |
| "learning_rate": 1.998059199199474e-05, |
| "loss": 0.4710524559020996, |
| "memory(GiB)": 73.06, |
| "step": 160, |
| "token_acc": 0.8710629921259843, |
| "train_speed(iter/s)": 0.03042 |
| }, |
| { |
| "epoch": 0.05279788808447662, |
| "grad_norm": 0.05874066660842649, |
| "learning_rate": 1.9977223825751802e-05, |
| "loss": 0.46933708190917967, |
| "memory(GiB)": 73.06, |
| "step": 165, |
| "token_acc": 0.894688221709007, |
| "train_speed(iter/s)": 0.030265 |
| }, |
| { |
| "epoch": 0.05439782408703652, |
| "grad_norm": 0.06544476253513323, |
| "learning_rate": 1.997358663135255e-05, |
| "loss": 0.46097607612609864, |
| "memory(GiB)": 73.06, |
| "step": 170, |
| "token_acc": 0.8832285384319261, |
| "train_speed(iter/s)": 0.030391 |
| }, |
| { |
| "epoch": 0.055997760089596414, |
| "grad_norm": 0.052612734205809274, |
| "learning_rate": 1.9969680506871138e-05, |
| "loss": 0.4674376010894775, |
| "memory(GiB)": 73.06, |
| "step": 175, |
| "token_acc": 0.8822400558269365, |
| "train_speed(iter/s)": 0.030298 |
| }, |
| { |
| "epoch": 0.057597696092156314, |
| "grad_norm": 0.05831524002340728, |
| "learning_rate": 1.9965505557633188e-05, |
| "loss": 0.47021942138671874, |
| "memory(GiB)": 73.06, |
| "step": 180, |
| "token_acc": 0.8729036501150937, |
| "train_speed(iter/s)": 0.030222 |
| }, |
| { |
| "epoch": 0.059197632094716214, |
| "grad_norm": 0.05558463298745032, |
| "learning_rate": 1.9961061896213006e-05, |
| "loss": 0.4707474708557129, |
| "memory(GiB)": 73.06, |
| "step": 185, |
| "token_acc": 0.8507278220269157, |
| "train_speed(iter/s)": 0.030322 |
| }, |
| { |
| "epoch": 0.06079756809727611, |
| "grad_norm": 0.05835065786169716, |
| "learning_rate": 1.9956349642430494e-05, |
| "loss": 0.4792951583862305, |
| "memory(GiB)": 73.06, |
| "step": 190, |
| "token_acc": 0.8596896665566194, |
| "train_speed(iter/s)": 0.030221 |
| }, |
| { |
| "epoch": 0.06239750409983601, |
| "grad_norm": 0.055138234334700054, |
| "learning_rate": 1.9951368923347945e-05, |
| "loss": 0.4755210876464844, |
| "memory(GiB)": 73.06, |
| "step": 195, |
| "token_acc": 0.8501170960187353, |
| "train_speed(iter/s)": 0.030259 |
| }, |
| { |
| "epoch": 0.0639974401023959, |
| "grad_norm": 0.06763946666899583, |
| "learning_rate": 1.9946119873266615e-05, |
| "loss": 0.4560092926025391, |
| "memory(GiB)": 73.06, |
| "step": 200, |
| "token_acc": 0.8891928864569083, |
| "train_speed(iter/s)": 0.030234 |
| }, |
| { |
| "epoch": 0.0655973761049558, |
| "grad_norm": 0.056758126149685124, |
| "learning_rate": 1.9940602633723097e-05, |
| "loss": 0.470977258682251, |
| "memory(GiB)": 73.06, |
| "step": 205, |
| "token_acc": 0.8635224424698109, |
| "train_speed(iter/s)": 0.030086 |
| }, |
| { |
| "epoch": 0.0671973121075157, |
| "grad_norm": 0.06089496841996932, |
| "learning_rate": 1.99348173534855e-05, |
| "loss": 0.4699739456176758, |
| "memory(GiB)": 73.06, |
| "step": 210, |
| "token_acc": 0.8621679827709978, |
| "train_speed(iter/s)": 0.03019 |
| }, |
| { |
| "epoch": 0.06879724811007559, |
| "grad_norm": 0.05802383504947049, |
| "learning_rate": 1.9928764188549462e-05, |
| "loss": 0.46386079788208007, |
| "memory(GiB)": 73.06, |
| "step": 215, |
| "token_acc": 0.8754250939681403, |
| "train_speed(iter/s)": 0.030114 |
| }, |
| { |
| "epoch": 0.0703971841126355, |
| "grad_norm": 0.05515703750577632, |
| "learning_rate": 1.9922443302133906e-05, |
| "loss": 0.4679898262023926, |
| "memory(GiB)": 73.06, |
| "step": 220, |
| "token_acc": 0.88738807102747, |
| "train_speed(iter/s)": 0.030007 |
| }, |
| { |
| "epoch": 0.07199712011519539, |
| "grad_norm": 0.05967474480415036, |
| "learning_rate": 1.9915854864676665e-05, |
| "loss": 0.47310919761657716, |
| "memory(GiB)": 73.06, |
| "step": 225, |
| "token_acc": 0.8383795309168444, |
| "train_speed(iter/s)": 0.030071 |
| }, |
| { |
| "epoch": 0.07359705611775529, |
| "grad_norm": 0.06076387420670948, |
| "learning_rate": 1.990899905382988e-05, |
| "loss": 0.4678232192993164, |
| "memory(GiB)": 73.06, |
| "step": 230, |
| "token_acc": 0.8767123287671232, |
| "train_speed(iter/s)": 0.029948 |
| }, |
| { |
| "epoch": 0.07519699212031519, |
| "grad_norm": 0.05635803811030448, |
| "learning_rate": 1.9901876054455217e-05, |
| "loss": 0.4821170330047607, |
| "memory(GiB)": 73.06, |
| "step": 235, |
| "token_acc": 0.8841222879684418, |
| "train_speed(iter/s)": 0.029913 |
| }, |
| { |
| "epoch": 0.07679692812287509, |
| "grad_norm": 0.053886114557468945, |
| "learning_rate": 1.9894486058618863e-05, |
| "loss": 0.46213107109069823, |
| "memory(GiB)": 73.06, |
| "step": 240, |
| "token_acc": 0.8886558627264061, |
| "train_speed(iter/s)": 0.02992 |
| }, |
| { |
| "epoch": 0.07839686412543498, |
| "grad_norm": 0.06048992108753748, |
| "learning_rate": 1.9886829265586368e-05, |
| "loss": 0.4749046802520752, |
| "memory(GiB)": 73.06, |
| "step": 245, |
| "token_acc": 0.8758281279575999, |
| "train_speed(iter/s)": 0.0298 |
| }, |
| { |
| "epoch": 0.07999680012799489, |
| "grad_norm": 0.06662896613700448, |
| "learning_rate": 1.9878905881817254e-05, |
| "loss": 0.47487664222717285, |
| "memory(GiB)": 73.06, |
| "step": 250, |
| "token_acc": 0.8239743295897318, |
| "train_speed(iter/s)": 0.029855 |
| }, |
| { |
| "epoch": 0.07999680012799489, |
| "eval_loss": 0.6802101731300354, |
| "eval_runtime": 108.8605, |
| "eval_samples_per_second": 184.53, |
| "eval_steps_per_second": 0.928, |
| "eval_token_acc": 0.8656411339267154, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.08159673613055478, |
| "grad_norm": 0.058967589577093804, |
| "learning_rate": 1.9870716120959462e-05, |
| "loss": 0.4691306591033936, |
| "memory(GiB)": 73.24, |
| "step": 255, |
| "token_acc": 0.8612697569398327, |
| "train_speed(iter/s)": 0.029606 |
| }, |
| { |
| "epoch": 0.08319667213311467, |
| "grad_norm": 0.057353651690814994, |
| "learning_rate": 1.986226020384359e-05, |
| "loss": 0.46143622398376466, |
| "memory(GiB)": 73.24, |
| "step": 260, |
| "token_acc": 0.8685547371094742, |
| "train_speed(iter/s)": 0.029681 |
| }, |
| { |
| "epoch": 0.08479660813567458, |
| "grad_norm": 0.05409688809510523, |
| "learning_rate": 1.9853538358476933e-05, |
| "loss": 0.4704445838928223, |
| "memory(GiB)": 73.24, |
| "step": 265, |
| "token_acc": 0.8804637020144431, |
| "train_speed(iter/s)": 0.02979 |
| }, |
| { |
| "epoch": 0.08639654413823447, |
| "grad_norm": 0.06968473514476099, |
| "learning_rate": 1.9844550820037326e-05, |
| "loss": 0.4717890739440918, |
| "memory(GiB)": 73.24, |
| "step": 270, |
| "token_acc": 0.8638003174145145, |
| "train_speed(iter/s)": 0.029834 |
| }, |
| { |
| "epoch": 0.08799648014079436, |
| "grad_norm": 0.06009720175343309, |
| "learning_rate": 1.9835297830866827e-05, |
| "loss": 0.4709662437438965, |
| "memory(GiB)": 73.24, |
| "step": 275, |
| "token_acc": 0.8634590377113134, |
| "train_speed(iter/s)": 0.029835 |
| }, |
| { |
| "epoch": 0.08959641614335427, |
| "grad_norm": 0.058778539356308675, |
| "learning_rate": 1.9825779640465157e-05, |
| "loss": 0.47084336280822753, |
| "memory(GiB)": 73.24, |
| "step": 280, |
| "token_acc": 0.9203691779351793, |
| "train_speed(iter/s)": 0.029942 |
| }, |
| { |
| "epoch": 0.09119635214591416, |
| "grad_norm": 0.054325246749067864, |
| "learning_rate": 1.9815996505483e-05, |
| "loss": 0.4666774749755859, |
| "memory(GiB)": 73.24, |
| "step": 285, |
| "token_acc": 0.8521723454119344, |
| "train_speed(iter/s)": 0.029906 |
| }, |
| { |
| "epoch": 0.09279628814847406, |
| "grad_norm": 0.058698263071843435, |
| "learning_rate": 1.9805948689715043e-05, |
| "loss": 0.45826416015625, |
| "memory(GiB)": 73.24, |
| "step": 290, |
| "token_acc": 0.8421138211382114, |
| "train_speed(iter/s)": 0.029895 |
| }, |
| { |
| "epoch": 0.09439622415103396, |
| "grad_norm": 0.05517972536097747, |
| "learning_rate": 1.979563646409291e-05, |
| "loss": 0.47627692222595214, |
| "memory(GiB)": 73.24, |
| "step": 295, |
| "token_acc": 0.8784122999686226, |
| "train_speed(iter/s)": 0.029966 |
| }, |
| { |
| "epoch": 0.09599616015359386, |
| "grad_norm": 0.06223926082468345, |
| "learning_rate": 1.9785060106677818e-05, |
| "loss": 0.4711057186126709, |
| "memory(GiB)": 73.24, |
| "step": 300, |
| "token_acc": 0.876372039283651, |
| "train_speed(iter/s)": 0.029911 |
| }, |
| { |
| "epoch": 0.09759609615615375, |
| "grad_norm": 0.060678733702642235, |
| "learning_rate": 1.97742199026531e-05, |
| "loss": 0.46833024024963377, |
| "memory(GiB)": 73.24, |
| "step": 305, |
| "token_acc": 0.8586995355484102, |
| "train_speed(iter/s)": 0.029998 |
| }, |
| { |
| "epoch": 0.09919603215871366, |
| "grad_norm": 0.06117494885421727, |
| "learning_rate": 1.9763116144316506e-05, |
| "loss": 0.4692807197570801, |
| "memory(GiB)": 73.24, |
| "step": 310, |
| "token_acc": 0.8383072793304911, |
| "train_speed(iter/s)": 0.029983 |
| }, |
| { |
| "epoch": 0.10079596816127355, |
| "grad_norm": 0.059512004342169564, |
| "learning_rate": 1.9751749131072335e-05, |
| "loss": 0.462421178817749, |
| "memory(GiB)": 73.24, |
| "step": 315, |
| "token_acc": 0.865073787772312, |
| "train_speed(iter/s)": 0.029959 |
| }, |
| { |
| "epoch": 0.10239590416383344, |
| "grad_norm": 0.05759903892800583, |
| "learning_rate": 1.9740119169423337e-05, |
| "loss": 0.4749638080596924, |
| "memory(GiB)": 73.24, |
| "step": 320, |
| "token_acc": 0.8657438292194797, |
| "train_speed(iter/s)": 0.030064 |
| }, |
| { |
| "epoch": 0.10399584016639335, |
| "grad_norm": 0.05512670495542287, |
| "learning_rate": 1.9728226572962474e-05, |
| "loss": 0.48053979873657227, |
| "memory(GiB)": 73.24, |
| "step": 325, |
| "token_acc": 0.9068181818181819, |
| "train_speed(iter/s)": 0.03004 |
| }, |
| { |
| "epoch": 0.10559577616895324, |
| "grad_norm": 0.05723038100011267, |
| "learning_rate": 1.9716071662364454e-05, |
| "loss": 0.47551665306091306, |
| "memory(GiB)": 73.24, |
| "step": 330, |
| "token_acc": 0.8362432269717038, |
| "train_speed(iter/s)": 0.030003 |
| }, |
| { |
| "epoch": 0.10719571217151314, |
| "grad_norm": 0.057638605082885846, |
| "learning_rate": 1.970365476537707e-05, |
| "loss": 0.4652701854705811, |
| "memory(GiB)": 73.24, |
| "step": 335, |
| "token_acc": 0.8735049401976079, |
| "train_speed(iter/s)": 0.030082 |
| }, |
| { |
| "epoch": 0.10879564817407304, |
| "grad_norm": 0.05903871731521889, |
| "learning_rate": 1.9690976216812397e-05, |
| "loss": 0.4698742389678955, |
| "memory(GiB)": 73.24, |
| "step": 340, |
| "token_acc": 0.8620361560418649, |
| "train_speed(iter/s)": 0.030027 |
| }, |
| { |
| "epoch": 0.11039558417663294, |
| "grad_norm": 0.053856521964694516, |
| "learning_rate": 1.9678036358537726e-05, |
| "loss": 0.4701416015625, |
| "memory(GiB)": 73.24, |
| "step": 345, |
| "token_acc": 0.8708435421771089, |
| "train_speed(iter/s)": 0.03002 |
| }, |
| { |
| "epoch": 0.11199552017919283, |
| "grad_norm": 0.05586893539038131, |
| "learning_rate": 1.966483553946637e-05, |
| "loss": 0.47447028160095217, |
| "memory(GiB)": 73.24, |
| "step": 350, |
| "token_acc": 0.8617533718689788, |
| "train_speed(iter/s)": 0.030041 |
| }, |
| { |
| "epoch": 0.11359545618175274, |
| "grad_norm": 0.052599438001953325, |
| "learning_rate": 1.9651374115548255e-05, |
| "loss": 0.4637298583984375, |
| "memory(GiB)": 73.24, |
| "step": 355, |
| "token_acc": 0.8874341610233258, |
| "train_speed(iter/s)": 0.029967 |
| }, |
| { |
| "epoch": 0.11519539218431263, |
| "grad_norm": 0.05804143123407663, |
| "learning_rate": 1.9637652449760297e-05, |
| "loss": 0.4660144329071045, |
| "memory(GiB)": 73.24, |
| "step": 360, |
| "token_acc": 0.8349885408708938, |
| "train_speed(iter/s)": 0.030034 |
| }, |
| { |
| "epoch": 0.11679532818687252, |
| "grad_norm": 0.06055547849970778, |
| "learning_rate": 1.9623670912096656e-05, |
| "loss": 0.4716383934020996, |
| "memory(GiB)": 73.24, |
| "step": 365, |
| "token_acc": 0.8751012473675684, |
| "train_speed(iter/s)": 0.02998 |
| }, |
| { |
| "epoch": 0.11839526418943243, |
| "grad_norm": 0.058520598293842735, |
| "learning_rate": 1.9609429879558726e-05, |
| "loss": 0.46298699378967284, |
| "memory(GiB)": 73.24, |
| "step": 370, |
| "token_acc": 0.8553921568627451, |
| "train_speed(iter/s)": 0.029931 |
| }, |
| { |
| "epoch": 0.11999520019199232, |
| "grad_norm": 0.058584318589478955, |
| "learning_rate": 1.9594929736144978e-05, |
| "loss": 0.4756875514984131, |
| "memory(GiB)": 73.24, |
| "step": 375, |
| "token_acc": 0.8618346545866364, |
| "train_speed(iter/s)": 0.030006 |
| }, |
| { |
| "epoch": 0.12159513619455221, |
| "grad_norm": 0.05966533070217228, |
| "learning_rate": 1.958017087284061e-05, |
| "loss": 0.4596414089202881, |
| "memory(GiB)": 73.24, |
| "step": 380, |
| "token_acc": 0.8836156297165856, |
| "train_speed(iter/s)": 0.029968 |
| }, |
| { |
| "epoch": 0.12319507219711212, |
| "grad_norm": 0.06510894340277039, |
| "learning_rate": 1.9565153687607006e-05, |
| "loss": 0.4687026023864746, |
| "memory(GiB)": 73.24, |
| "step": 385, |
| "token_acc": 0.8694005270092227, |
| "train_speed(iter/s)": 0.029956 |
| }, |
| { |
| "epoch": 0.12479500819967201, |
| "grad_norm": 0.05180588304383506, |
| "learning_rate": 1.9549878585371006e-05, |
| "loss": 0.4649878978729248, |
| "memory(GiB)": 73.24, |
| "step": 390, |
| "token_acc": 0.8677233429394813, |
| "train_speed(iter/s)": 0.029999 |
| }, |
| { |
| "epoch": 0.1263949442022319, |
| "grad_norm": 0.060875424512666344, |
| "learning_rate": 1.9534345978013972e-05, |
| "loss": 0.47073874473571775, |
| "memory(GiB)": 73.24, |
| "step": 395, |
| "token_acc": 0.8484663512894858, |
| "train_speed(iter/s)": 0.029947 |
| }, |
| { |
| "epoch": 0.1279948802047918, |
| "grad_norm": 0.0571374353277554, |
| "learning_rate": 1.9518556284360696e-05, |
| "loss": 0.4666412353515625, |
| "memory(GiB)": 73.24, |
| "step": 400, |
| "token_acc": 0.8975701436434421, |
| "train_speed(iter/s)": 0.029998 |
| }, |
| { |
| "epoch": 0.1295948162073517, |
| "grad_norm": 0.05215050598306155, |
| "learning_rate": 1.9502509930168113e-05, |
| "loss": 0.4628121376037598, |
| "memory(GiB)": 73.24, |
| "step": 405, |
| "token_acc": 0.8816677696889477, |
| "train_speed(iter/s)": 0.029966 |
| }, |
| { |
| "epoch": 0.1311947522099116, |
| "grad_norm": 0.07947570193916972, |
| "learning_rate": 1.9486207348113803e-05, |
| "loss": 0.4593012809753418, |
| "memory(GiB)": 73.24, |
| "step": 410, |
| "token_acc": 0.8692473832862602, |
| "train_speed(iter/s)": 0.029911 |
| }, |
| { |
| "epoch": 0.1327946882124715, |
| "grad_norm": 0.07262611466641217, |
| "learning_rate": 1.946964897778433e-05, |
| "loss": 0.47004990577697753, |
| "memory(GiB)": 73.24, |
| "step": 415, |
| "token_acc": 0.8736337958983176, |
| "train_speed(iter/s)": 0.029969 |
| }, |
| { |
| "epoch": 0.1343946242150314, |
| "grad_norm": 0.053754461298334506, |
| "learning_rate": 1.9452835265663404e-05, |
| "loss": 0.4695271015167236, |
| "memory(GiB)": 73.24, |
| "step": 420, |
| "token_acc": 0.8747993579454254, |
| "train_speed(iter/s)": 0.029901 |
| }, |
| { |
| "epoch": 0.1359945602175913, |
| "grad_norm": 0.0742051800083311, |
| "learning_rate": 1.9435766665119823e-05, |
| "loss": 0.47011446952819824, |
| "memory(GiB)": 73.24, |
| "step": 425, |
| "token_acc": 0.8356736242884251, |
| "train_speed(iter/s)": 0.029856 |
| }, |
| { |
| "epoch": 0.13759449622015119, |
| "grad_norm": 0.06429200177825628, |
| "learning_rate": 1.941844363639525e-05, |
| "loss": 0.476796817779541, |
| "memory(GiB)": 73.24, |
| "step": 430, |
| "token_acc": 0.869019972131909, |
| "train_speed(iter/s)": 0.0299 |
| }, |
| { |
| "epoch": 0.13919443222271108, |
| "grad_norm": 0.06544854557851852, |
| "learning_rate": 1.9400866646591816e-05, |
| "loss": 0.4666853904724121, |
| "memory(GiB)": 73.24, |
| "step": 435, |
| "token_acc": 0.8204667863554758, |
| "train_speed(iter/s)": 0.029847 |
| }, |
| { |
| "epoch": 0.140794368225271, |
| "grad_norm": 0.0546565929911768, |
| "learning_rate": 1.9383036169659513e-05, |
| "loss": 0.4738778591156006, |
| "memory(GiB)": 73.24, |
| "step": 440, |
| "token_acc": 0.8605809128630706, |
| "train_speed(iter/s)": 0.029855 |
| }, |
| { |
| "epoch": 0.1423943042278309, |
| "grad_norm": 0.06789336848906298, |
| "learning_rate": 1.936495268638342e-05, |
| "loss": 0.47726120948791506, |
| "memory(GiB)": 73.24, |
| "step": 445, |
| "token_acc": 0.8404369243949454, |
| "train_speed(iter/s)": 0.029875 |
| }, |
| { |
| "epoch": 0.14399424023039079, |
| "grad_norm": 0.049909982274150465, |
| "learning_rate": 1.934661668437073e-05, |
| "loss": 0.47165632247924805, |
| "memory(GiB)": 73.24, |
| "step": 450, |
| "token_acc": 0.848471615720524, |
| "train_speed(iter/s)": 0.029826 |
| }, |
| { |
| "epoch": 0.14559417623295068, |
| "grad_norm": 0.057441474731933166, |
| "learning_rate": 1.932802865803763e-05, |
| "loss": 0.4703391075134277, |
| "memory(GiB)": 73.24, |
| "step": 455, |
| "token_acc": 0.8466442358774571, |
| "train_speed(iter/s)": 0.029877 |
| }, |
| { |
| "epoch": 0.14719411223551057, |
| "grad_norm": 0.07263904251491092, |
| "learning_rate": 1.930918910859592e-05, |
| "loss": 0.467697811126709, |
| "memory(GiB)": 73.24, |
| "step": 460, |
| "token_acc": 0.8491142333536957, |
| "train_speed(iter/s)": 0.02984 |
| }, |
| { |
| "epoch": 0.14879404823807046, |
| "grad_norm": 0.06769237623086669, |
| "learning_rate": 1.9290098544039546e-05, |
| "loss": 0.46541628837585447, |
| "memory(GiB)": 73.24, |
| "step": 465, |
| "token_acc": 0.8555353301340394, |
| "train_speed(iter/s)": 0.02978 |
| }, |
| { |
| "epoch": 0.15039398424063039, |
| "grad_norm": 0.06751583633556477, |
| "learning_rate": 1.927075747913088e-05, |
| "loss": 0.47134056091308596, |
| "memory(GiB)": 73.24, |
| "step": 470, |
| "token_acc": 0.8708000507163687, |
| "train_speed(iter/s)": 0.029832 |
| }, |
| { |
| "epoch": 0.15199392024319028, |
| "grad_norm": 0.0539492567012165, |
| "learning_rate": 1.9251166435386837e-05, |
| "loss": 0.4645866394042969, |
| "memory(GiB)": 73.24, |
| "step": 475, |
| "token_acc": 0.849832526981764, |
| "train_speed(iter/s)": 0.029779 |
| }, |
| { |
| "epoch": 0.15359385624575017, |
| "grad_norm": 0.06038706866556876, |
| "learning_rate": 1.923132594106483e-05, |
| "loss": 0.46890692710876464, |
| "memory(GiB)": 73.24, |
| "step": 480, |
| "token_acc": 0.8665925514174542, |
| "train_speed(iter/s)": 0.02976 |
| }, |
| { |
| "epoch": 0.15519379224831006, |
| "grad_norm": 0.05215840717634863, |
| "learning_rate": 1.92112365311485e-05, |
| "loss": 0.46829919815063475, |
| "memory(GiB)": 73.24, |
| "step": 485, |
| "token_acc": 0.861963565228023, |
| "train_speed(iter/s)": 0.029794 |
| }, |
| { |
| "epoch": 0.15679372825086996, |
| "grad_norm": 0.06554142579397569, |
| "learning_rate": 1.919089874733332e-05, |
| "loss": 0.4702622413635254, |
| "memory(GiB)": 73.24, |
| "step": 490, |
| "token_acc": 0.8809186723297153, |
| "train_speed(iter/s)": 0.029747 |
| }, |
| { |
| "epoch": 0.15839366425342985, |
| "grad_norm": 0.0601172563145885, |
| "learning_rate": 1.9170313138011964e-05, |
| "loss": 0.46490135192871096, |
| "memory(GiB)": 73.24, |
| "step": 495, |
| "token_acc": 0.8890911637025627, |
| "train_speed(iter/s)": 0.02977 |
| }, |
| { |
| "epoch": 0.15999360025598977, |
| "grad_norm": 0.05924399402367875, |
| "learning_rate": 1.9149480258259535e-05, |
| "loss": 0.46698894500732424, |
| "memory(GiB)": 73.24, |
| "step": 500, |
| "token_acc": 0.8781434114096853, |
| "train_speed(iter/s)": 0.029766 |
| }, |
| { |
| "epoch": 0.15999360025598977, |
| "eval_loss": 0.677643895149231, |
| "eval_runtime": 109.3458, |
| "eval_samples_per_second": 183.711, |
| "eval_steps_per_second": 0.924, |
| "eval_token_acc": 0.8661408286670019, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.16159353625854966, |
| "grad_norm": 0.05118070522939682, |
| "learning_rate": 1.9128400669818586e-05, |
| "loss": 0.4606743812561035, |
| "memory(GiB)": 73.24, |
| "step": 505, |
| "token_acc": 0.8727327237295758, |
| "train_speed(iter/s)": 0.029628 |
| }, |
| { |
| "epoch": 0.16319347226110956, |
| "grad_norm": 0.05904937387674259, |
| "learning_rate": 1.9107074941083987e-05, |
| "loss": 0.47115492820739746, |
| "memory(GiB)": 73.24, |
| "step": 510, |
| "token_acc": 0.8801781737193763, |
| "train_speed(iter/s)": 0.029663 |
| }, |
| { |
| "epoch": 0.16479340826366945, |
| "grad_norm": 0.061211680590962145, |
| "learning_rate": 1.9085503647087588e-05, |
| "loss": 0.46154184341430665, |
| "memory(GiB)": 73.24, |
| "step": 515, |
| "token_acc": 0.8573438874230431, |
| "train_speed(iter/s)": 0.029714 |
| }, |
| { |
| "epoch": 0.16639334426622934, |
| "grad_norm": 0.05461804298242196, |
| "learning_rate": 1.906368736948272e-05, |
| "loss": 0.46891465187072756, |
| "memory(GiB)": 73.24, |
| "step": 520, |
| "token_acc": 0.8665508756694167, |
| "train_speed(iter/s)": 0.029721 |
| }, |
| { |
| "epoch": 0.16799328026878924, |
| "grad_norm": 0.059072521440841075, |
| "learning_rate": 1.9041626696528503e-05, |
| "loss": 0.4666083812713623, |
| "memory(GiB)": 73.24, |
| "step": 525, |
| "token_acc": 0.8742783835792175, |
| "train_speed(iter/s)": 0.029735 |
| }, |
| { |
| "epoch": 0.16959321627134916, |
| "grad_norm": 0.06762878495647719, |
| "learning_rate": 1.9019322223073997e-05, |
| "loss": 0.4684437274932861, |
| "memory(GiB)": 73.24, |
| "step": 530, |
| "token_acc": 0.8906074591493077, |
| "train_speed(iter/s)": 0.029782 |
| }, |
| { |
| "epoch": 0.17119315227390905, |
| "grad_norm": 0.05741557316745661, |
| "learning_rate": 1.899677455054215e-05, |
| "loss": 0.4690097332000732, |
| "memory(GiB)": 73.24, |
| "step": 535, |
| "token_acc": 0.8231878958479943, |
| "train_speed(iter/s)": 0.029785 |
| }, |
| { |
| "epoch": 0.17279308827646894, |
| "grad_norm": 0.049026865135578496, |
| "learning_rate": 1.8973984286913584e-05, |
| "loss": 0.469140625, |
| "memory(GiB)": 73.24, |
| "step": 540, |
| "token_acc": 0.8849415539766216, |
| "train_speed(iter/s)": 0.029789 |
| }, |
| { |
| "epoch": 0.17439302427902884, |
| "grad_norm": 0.059746465018255104, |
| "learning_rate": 1.895095204671021e-05, |
| "loss": 0.4646149158477783, |
| "memory(GiB)": 73.24, |
| "step": 545, |
| "token_acc": 0.8944385405596883, |
| "train_speed(iter/s)": 0.029813 |
| }, |
| { |
| "epoch": 0.17599296028158873, |
| "grad_norm": 0.049833714934798115, |
| "learning_rate": 1.892767845097864e-05, |
| "loss": 0.47077240943908694, |
| "memory(GiB)": 73.24, |
| "step": 550, |
| "token_acc": 0.8640860961638605, |
| "train_speed(iter/s)": 0.029794 |
| }, |
| { |
| "epoch": 0.17759289628414862, |
| "grad_norm": 0.06593845007149325, |
| "learning_rate": 1.890416412727346e-05, |
| "loss": 0.46265759468078616, |
| "memory(GiB)": 73.24, |
| "step": 555, |
| "token_acc": 0.8249895412076419, |
| "train_speed(iter/s)": 0.02984 |
| }, |
| { |
| "epoch": 0.17919283228670854, |
| "grad_norm": 0.058254003445636866, |
| "learning_rate": 1.88804097096403e-05, |
| "loss": 0.459829044342041, |
| "memory(GiB)": 73.24, |
| "step": 560, |
| "token_acc": 0.8835873095178616, |
| "train_speed(iter/s)": 0.029842 |
| }, |
| { |
| "epoch": 0.18079276828926844, |
| "grad_norm": 0.07335953644753283, |
| "learning_rate": 1.8856415838598738e-05, |
| "loss": 0.45765042304992676, |
| "memory(GiB)": 73.24, |
| "step": 565, |
| "token_acc": 0.8755007210382951, |
| "train_speed(iter/s)": 0.029818 |
| }, |
| { |
| "epoch": 0.18239270429182833, |
| "grad_norm": 0.06659181547700674, |
| "learning_rate": 1.8832183161125026e-05, |
| "loss": 0.4609128475189209, |
| "memory(GiB)": 73.24, |
| "step": 570, |
| "token_acc": 0.8344311377245509, |
| "train_speed(iter/s)": 0.029871 |
| }, |
| { |
| "epoch": 0.18399264029438822, |
| "grad_norm": 0.05836437871791382, |
| "learning_rate": 1.8807712330634645e-05, |
| "loss": 0.4691438674926758, |
| "memory(GiB)": 73.24, |
| "step": 575, |
| "token_acc": 0.8848027659908848, |
| "train_speed(iter/s)": 0.029828 |
| }, |
| { |
| "epoch": 0.18559257629694811, |
| "grad_norm": 0.05735059462858394, |
| "learning_rate": 1.87830040069647e-05, |
| "loss": 0.4602513790130615, |
| "memory(GiB)": 73.24, |
| "step": 580, |
| "token_acc": 0.8959147903465012, |
| "train_speed(iter/s)": 0.029816 |
| }, |
| { |
| "epoch": 0.187192512299508, |
| "grad_norm": 0.05337219773586585, |
| "learning_rate": 1.87580588563561e-05, |
| "loss": 0.46318631172180175, |
| "memory(GiB)": 73.24, |
| "step": 585, |
| "token_acc": 0.8725881039706586, |
| "train_speed(iter/s)": 0.029851 |
| }, |
| { |
| "epoch": 0.18879244830206793, |
| "grad_norm": 0.05886716832883729, |
| "learning_rate": 1.873287755143563e-05, |
| "loss": 0.4604507923126221, |
| "memory(GiB)": 73.24, |
| "step": 590, |
| "token_acc": 0.9041755130927105, |
| "train_speed(iter/s)": 0.029822 |
| }, |
| { |
| "epoch": 0.19039238430462782, |
| "grad_norm": 0.053483810048332456, |
| "learning_rate": 1.8707460771197773e-05, |
| "loss": 0.46618080139160156, |
| "memory(GiB)": 73.24, |
| "step": 595, |
| "token_acc": 0.8785046728971962, |
| "train_speed(iter/s)": 0.029819 |
| }, |
| { |
| "epoch": 0.1919923203071877, |
| "grad_norm": 0.0518592001281956, |
| "learning_rate": 1.868180920098644e-05, |
| "loss": 0.4680916786193848, |
| "memory(GiB)": 73.24, |
| "step": 600, |
| "token_acc": 0.8467063770147162, |
| "train_speed(iter/s)": 0.029843 |
| }, |
| { |
| "epoch": 0.1935922563097476, |
| "grad_norm": 0.07018232236413237, |
| "learning_rate": 1.8655923532476463e-05, |
| "loss": 0.46170759201049805, |
| "memory(GiB)": 73.24, |
| "step": 605, |
| "token_acc": 0.889030612244898, |
| "train_speed(iter/s)": 0.02981 |
| }, |
| { |
| "epoch": 0.1951921923123075, |
| "grad_norm": 0.06030421269833889, |
| "learning_rate": 1.8629804463654956e-05, |
| "loss": 0.46511187553405764, |
| "memory(GiB)": 73.24, |
| "step": 610, |
| "token_acc": 0.8554680664916885, |
| "train_speed(iter/s)": 0.029852 |
| }, |
| { |
| "epoch": 0.1967921283148674, |
| "grad_norm": 0.056137765321266526, |
| "learning_rate": 1.8603452698802498e-05, |
| "loss": 0.47327299118041993, |
| "memory(GiB)": 76.61, |
| "step": 615, |
| "token_acc": 0.8645191852202747, |
| "train_speed(iter/s)": 0.029831 |
| }, |
| { |
| "epoch": 0.1983920643174273, |
| "grad_norm": 0.05458475201274465, |
| "learning_rate": 1.857686894847413e-05, |
| "loss": 0.45963249206542967, |
| "memory(GiB)": 76.61, |
| "step": 620, |
| "token_acc": 0.8517509197438343, |
| "train_speed(iter/s)": 0.029791 |
| }, |
| { |
| "epoch": 0.1999920003199872, |
| "grad_norm": 0.059902578480064236, |
| "learning_rate": 1.8550053929480202e-05, |
| "loss": 0.4687147617340088, |
| "memory(GiB)": 76.61, |
| "step": 625, |
| "token_acc": 0.8958185683912119, |
| "train_speed(iter/s)": 0.029833 |
| }, |
| { |
| "epoch": 0.2015919363225471, |
| "grad_norm": 0.0539478773118384, |
| "learning_rate": 1.8523008364867056e-05, |
| "loss": 0.4696544647216797, |
| "memory(GiB)": 76.61, |
| "step": 630, |
| "token_acc": 0.8439355385920272, |
| "train_speed(iter/s)": 0.029796 |
| }, |
| { |
| "epoch": 0.203191872325107, |
| "grad_norm": 0.05688926646164217, |
| "learning_rate": 1.8495732983897504e-05, |
| "loss": 0.4628334045410156, |
| "memory(GiB)": 76.61, |
| "step": 635, |
| "token_acc": 0.8406656465187452, |
| "train_speed(iter/s)": 0.029775 |
| }, |
| { |
| "epoch": 0.20479180832766689, |
| "grad_norm": 0.055104479428209605, |
| "learning_rate": 1.8468228522031197e-05, |
| "loss": 0.4559271812438965, |
| "memory(GiB)": 76.61, |
| "step": 640, |
| "token_acc": 0.8823529411764706, |
| "train_speed(iter/s)": 0.029794 |
| }, |
| { |
| "epoch": 0.20639174433022678, |
| "grad_norm": 0.058080447436547736, |
| "learning_rate": 1.8440495720904758e-05, |
| "loss": 0.4649765968322754, |
| "memory(GiB)": 76.61, |
| "step": 645, |
| "token_acc": 0.8708735027753433, |
| "train_speed(iter/s)": 0.029752 |
| }, |
| { |
| "epoch": 0.2079916803327867, |
| "grad_norm": 0.06300003986546152, |
| "learning_rate": 1.8412535328311813e-05, |
| "loss": 0.47095327377319335, |
| "memory(GiB)": 76.61, |
| "step": 650, |
| "token_acc": 0.8504976200778883, |
| "train_speed(iter/s)": 0.029755 |
| }, |
| { |
| "epoch": 0.2095916163353466, |
| "grad_norm": 0.06584526718748161, |
| "learning_rate": 1.8384348098182815e-05, |
| "loss": 0.46697392463684084, |
| "memory(GiB)": 76.61, |
| "step": 655, |
| "token_acc": 0.8224407171775593, |
| "train_speed(iter/s)": 0.029762 |
| }, |
| { |
| "epoch": 0.21119155233790649, |
| "grad_norm": 0.07147957728971413, |
| "learning_rate": 1.8355934790564718e-05, |
| "loss": 0.4684570789337158, |
| "memory(GiB)": 76.61, |
| "step": 660, |
| "token_acc": 0.8842165898617511, |
| "train_speed(iter/s)": 0.029723 |
| }, |
| { |
| "epoch": 0.21279148834046638, |
| "grad_norm": 0.06592046292925295, |
| "learning_rate": 1.832729617160047e-05, |
| "loss": 0.461454439163208, |
| "memory(GiB)": 76.61, |
| "step": 665, |
| "token_acc": 0.9114801444043321, |
| "train_speed(iter/s)": 0.02976 |
| }, |
| { |
| "epoch": 0.21439142434302627, |
| "grad_norm": 0.0656829490109071, |
| "learning_rate": 1.8298433013508384e-05, |
| "loss": 0.46404447555541994, |
| "memory(GiB)": 76.61, |
| "step": 670, |
| "token_acc": 0.8516549891278087, |
| "train_speed(iter/s)": 0.029736 |
| }, |
| { |
| "epoch": 0.21599136034558616, |
| "grad_norm": 0.05417998837874903, |
| "learning_rate": 1.826934609456129e-05, |
| "loss": 0.47208566665649415, |
| "memory(GiB)": 76.61, |
| "step": 675, |
| "token_acc": 0.8798815733822078, |
| "train_speed(iter/s)": 0.029718 |
| }, |
| { |
| "epoch": 0.21759129634814608, |
| "grad_norm": 0.06917195844649823, |
| "learning_rate": 1.8240036199065546e-05, |
| "loss": 0.4724391460418701, |
| "memory(GiB)": 76.61, |
| "step": 680, |
| "token_acc": 0.875845675626257, |
| "train_speed(iter/s)": 0.029745 |
| }, |
| { |
| "epoch": 0.21919123235070598, |
| "grad_norm": 0.055849189404917746, |
| "learning_rate": 1.8210504117339917e-05, |
| "loss": 0.463816499710083, |
| "memory(GiB)": 76.61, |
| "step": 685, |
| "token_acc": 0.8841904379268782, |
| "train_speed(iter/s)": 0.029711 |
| }, |
| { |
| "epoch": 0.22079116835326587, |
| "grad_norm": 0.059563786969142496, |
| "learning_rate": 1.8180750645694236e-05, |
| "loss": 0.4678086757659912, |
| "memory(GiB)": 76.61, |
| "step": 690, |
| "token_acc": 0.8675231977159172, |
| "train_speed(iter/s)": 0.029714 |
| }, |
| { |
| "epoch": 0.22239110435582576, |
| "grad_norm": 0.05908606421708839, |
| "learning_rate": 1.8150776586407957e-05, |
| "loss": 0.46315860748291016, |
| "memory(GiB)": 76.61, |
| "step": 695, |
| "token_acc": 0.8914956011730205, |
| "train_speed(iter/s)": 0.029731 |
| }, |
| { |
| "epoch": 0.22399104035838566, |
| "grad_norm": 0.05617530731492468, |
| "learning_rate": 1.8120582747708503e-05, |
| "loss": 0.46682062149047854, |
| "memory(GiB)": 76.61, |
| "step": 700, |
| "token_acc": 0.8805088596092685, |
| "train_speed(iter/s)": 0.029689 |
| }, |
| { |
| "epoch": 0.22559097636094555, |
| "grad_norm": 0.06138477303861948, |
| "learning_rate": 1.8090169943749477e-05, |
| "loss": 0.47155141830444336, |
| "memory(GiB)": 76.61, |
| "step": 705, |
| "token_acc": 0.8753766681015928, |
| "train_speed(iter/s)": 0.029703 |
| }, |
| { |
| "epoch": 0.22719091236350547, |
| "grad_norm": 0.07073141016351848, |
| "learning_rate": 1.8059538994588715e-05, |
| "loss": 0.45953845977783203, |
| "memory(GiB)": 76.61, |
| "step": 710, |
| "token_acc": 0.8449233877757198, |
| "train_speed(iter/s)": 0.0297 |
| }, |
| { |
| "epoch": 0.22879084836606536, |
| "grad_norm": 0.06266619359839708, |
| "learning_rate": 1.8028690726166172e-05, |
| "loss": 0.4604049205780029, |
| "memory(GiB)": 76.61, |
| "step": 715, |
| "token_acc": 0.8688032048072108, |
| "train_speed(iter/s)": 0.02966 |
| }, |
| { |
| "epoch": 0.23039078436862526, |
| "grad_norm": 0.0563660774004587, |
| "learning_rate": 1.7997625970281652e-05, |
| "loss": 0.4622708797454834, |
| "memory(GiB)": 76.61, |
| "step": 720, |
| "token_acc": 0.8698216735253772, |
| "train_speed(iter/s)": 0.029685 |
| }, |
| { |
| "epoch": 0.23199072037118515, |
| "grad_norm": 0.06596213612143108, |
| "learning_rate": 1.796634556457236e-05, |
| "loss": 0.4681892395019531, |
| "memory(GiB)": 76.61, |
| "step": 725, |
| "token_acc": 0.8842619184376795, |
| "train_speed(iter/s)": 0.029661 |
| }, |
| { |
| "epoch": 0.23359065637374504, |
| "grad_norm": 0.05364579438678848, |
| "learning_rate": 1.793485035249036e-05, |
| "loss": 0.46258745193481443, |
| "memory(GiB)": 76.61, |
| "step": 730, |
| "token_acc": 0.8599531615925059, |
| "train_speed(iter/s)": 0.02965 |
| }, |
| { |
| "epoch": 0.23519059237630494, |
| "grad_norm": 0.07509450433159735, |
| "learning_rate": 1.7903141183279776e-05, |
| "loss": 0.47242441177368166, |
| "memory(GiB)": 76.61, |
| "step": 735, |
| "token_acc": 0.8404958677685951, |
| "train_speed(iter/s)": 0.029665 |
| }, |
| { |
| "epoch": 0.23679052837886486, |
| "grad_norm": 0.06478313540282635, |
| "learning_rate": 1.7871218911953942e-05, |
| "loss": 0.4565444469451904, |
| "memory(GiB)": 76.61, |
| "step": 740, |
| "token_acc": 0.8338650865998177, |
| "train_speed(iter/s)": 0.029634 |
| }, |
| { |
| "epoch": 0.23839046438142475, |
| "grad_norm": 0.06348939893307848, |
| "learning_rate": 1.7839084399272317e-05, |
| "loss": 0.4670473575592041, |
| "memory(GiB)": 76.61, |
| "step": 745, |
| "token_acc": 0.8652410477034038, |
| "train_speed(iter/s)": 0.029638 |
| }, |
| { |
| "epoch": 0.23999040038398464, |
| "grad_norm": 0.07434587030241245, |
| "learning_rate": 1.780673851171728e-05, |
| "loss": 0.47047910690307615, |
| "memory(GiB)": 76.61, |
| "step": 750, |
| "token_acc": 0.88801504530689, |
| "train_speed(iter/s)": 0.029638 |
| }, |
| { |
| "epoch": 0.23999040038398464, |
| "eval_loss": 0.6746003031730652, |
| "eval_runtime": 113.2223, |
| "eval_samples_per_second": 177.421, |
| "eval_steps_per_second": 0.892, |
| "eval_token_acc": 0.8668385651547512, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.24159033638654454, |
| "grad_norm": 0.06732795706859432, |
| "learning_rate": 1.777418212147079e-05, |
| "loss": 0.46190509796142576, |
| "memory(GiB)": 76.61, |
| "step": 755, |
| "token_acc": 0.8881346728210697, |
| "train_speed(iter/s)": 0.029543 |
| }, |
| { |
| "epoch": 0.24319027238910443, |
| "grad_norm": 0.06836940989947664, |
| "learning_rate": 1.7741416106390828e-05, |
| "loss": 0.46283302307128904, |
| "memory(GiB)": 76.61, |
| "step": 760, |
| "token_acc": 0.8831443688586545, |
| "train_speed(iter/s)": 0.029566 |
| }, |
| { |
| "epoch": 0.24479020839166432, |
| "grad_norm": 0.07072489516219096, |
| "learning_rate": 1.7708441349987753e-05, |
| "loss": 0.4619740962982178, |
| "memory(GiB)": 76.61, |
| "step": 765, |
| "token_acc": 0.8610668789808917, |
| "train_speed(iter/s)": 0.0296 |
| }, |
| { |
| "epoch": 0.24639014439422424, |
| "grad_norm": 0.07152232857362027, |
| "learning_rate": 1.767525874140048e-05, |
| "loss": 0.46694121360778806, |
| "memory(GiB)": 76.61, |
| "step": 770, |
| "token_acc": 0.8397869022869023, |
| "train_speed(iter/s)": 0.029606 |
| }, |
| { |
| "epoch": 0.24799008039678413, |
| "grad_norm": 0.059354056163304685, |
| "learning_rate": 1.7641869175372493e-05, |
| "loss": 0.4596868991851807, |
| "memory(GiB)": 76.61, |
| "step": 775, |
| "token_acc": 0.8582827406764961, |
| "train_speed(iter/s)": 0.029599 |
| }, |
| { |
| "epoch": 0.24959001639934403, |
| "grad_norm": 0.0629690289705531, |
| "learning_rate": 1.7608273552227723e-05, |
| "loss": 0.4583168029785156, |
| "memory(GiB)": 76.61, |
| "step": 780, |
| "token_acc": 0.8841532106646639, |
| "train_speed(iter/s)": 0.029639 |
| }, |
| { |
| "epoch": 0.25118995240190395, |
| "grad_norm": 0.05810355160479093, |
| "learning_rate": 1.7574472777846276e-05, |
| "loss": 0.47337069511413576, |
| "memory(GiB)": 76.61, |
| "step": 785, |
| "token_acc": 0.8676557863501484, |
| "train_speed(iter/s)": 0.029632 |
| }, |
| { |
| "epoch": 0.2527898884044638, |
| "grad_norm": 0.05365185572887828, |
| "learning_rate": 1.7540467763639994e-05, |
| "loss": 0.46567063331604003, |
| "memory(GiB)": 76.61, |
| "step": 790, |
| "token_acc": 0.8745288099084545, |
| "train_speed(iter/s)": 0.029629 |
| }, |
| { |
| "epoch": 0.25438982440702373, |
| "grad_norm": 0.054672322658953366, |
| "learning_rate": 1.7506259426527903e-05, |
| "loss": 0.47023472785949705, |
| "memory(GiB)": 76.61, |
| "step": 795, |
| "token_acc": 0.874407844001322, |
| "train_speed(iter/s)": 0.02965 |
| }, |
| { |
| "epoch": 0.2559897604095836, |
| "grad_norm": 0.057060955079149434, |
| "learning_rate": 1.7471848688911465e-05, |
| "loss": 0.4684537410736084, |
| "memory(GiB)": 76.61, |
| "step": 800, |
| "token_acc": 0.8839382448537378, |
| "train_speed(iter/s)": 0.029634 |
| }, |
| { |
| "epoch": 0.2575896964121435, |
| "grad_norm": 0.06051290772323595, |
| "learning_rate": 1.7437236478649718e-05, |
| "loss": 0.46199979782104494, |
| "memory(GiB)": 76.61, |
| "step": 805, |
| "token_acc": 0.8673650919153983, |
| "train_speed(iter/s)": 0.02966 |
| }, |
| { |
| "epoch": 0.2591896324147034, |
| "grad_norm": 0.0643397562387603, |
| "learning_rate": 1.7402423729034252e-05, |
| "loss": 0.4548381805419922, |
| "memory(GiB)": 76.61, |
| "step": 810, |
| "token_acc": 0.83125, |
| "train_speed(iter/s)": 0.029652 |
| }, |
| { |
| "epoch": 0.2607895684172633, |
| "grad_norm": 0.065624934571794, |
| "learning_rate": 1.736741137876405e-05, |
| "loss": 0.46353764533996583, |
| "memory(GiB)": 76.61, |
| "step": 815, |
| "token_acc": 0.8907202528787537, |
| "train_speed(iter/s)": 0.029628 |
| }, |
| { |
| "epoch": 0.2623895044198232, |
| "grad_norm": 0.053961693017135055, |
| "learning_rate": 1.7332200371920173e-05, |
| "loss": 0.46685361862182617, |
| "memory(GiB)": 76.61, |
| "step": 820, |
| "token_acc": 0.8522188711762172, |
| "train_speed(iter/s)": 0.029672 |
| }, |
| { |
| "epoch": 0.2639894404223831, |
| "grad_norm": 0.054388550053431586, |
| "learning_rate": 1.72967916579403e-05, |
| "loss": 0.46024084091186523, |
| "memory(GiB)": 76.61, |
| "step": 825, |
| "token_acc": 0.8684630384683567, |
| "train_speed(iter/s)": 0.02966 |
| }, |
| { |
| "epoch": 0.265589376424943, |
| "grad_norm": 0.0583019332597641, |
| "learning_rate": 1.7261186191593135e-05, |
| "loss": 0.47214059829711913, |
| "memory(GiB)": 76.61, |
| "step": 830, |
| "token_acc": 0.8717123935666982, |
| "train_speed(iter/s)": 0.029645 |
| }, |
| { |
| "epoch": 0.2671893124275029, |
| "grad_norm": 0.06004272220759217, |
| "learning_rate": 1.7225384932952655e-05, |
| "loss": 0.4626835823059082, |
| "memory(GiB)": 76.61, |
| "step": 835, |
| "token_acc": 0.8737211788059246, |
| "train_speed(iter/s)": 0.02967 |
| }, |
| { |
| "epoch": 0.2687892484300628, |
| "grad_norm": 0.05611993161069816, |
| "learning_rate": 1.7189388847372227e-05, |
| "loss": 0.46799750328063966, |
| "memory(GiB)": 76.61, |
| "step": 840, |
| "token_acc": 0.8781684382665577, |
| "train_speed(iter/s)": 0.029642 |
| }, |
| { |
| "epoch": 0.2703891844326227, |
| "grad_norm": 0.06345947319153013, |
| "learning_rate": 1.715319890545857e-05, |
| "loss": 0.4568619728088379, |
| "memory(GiB)": 76.61, |
| "step": 845, |
| "token_acc": 0.860916969527537, |
| "train_speed(iter/s)": 0.029655 |
| }, |
| { |
| "epoch": 0.2719891204351826, |
| "grad_norm": 0.0592531603954309, |
| "learning_rate": 1.7116816083045603e-05, |
| "loss": 0.46942729949951173, |
| "memory(GiB)": 76.61, |
| "step": 850, |
| "token_acc": 0.8726317245194303, |
| "train_speed(iter/s)": 0.029655 |
| }, |
| { |
| "epoch": 0.2735890564377425, |
| "grad_norm": 0.05711267065318382, |
| "learning_rate": 1.7080241361168108e-05, |
| "loss": 0.45801239013671874, |
| "memory(GiB)": 76.61, |
| "step": 855, |
| "token_acc": 0.8834167608590344, |
| "train_speed(iter/s)": 0.02963 |
| }, |
| { |
| "epoch": 0.27518899244030237, |
| "grad_norm": 0.05715792257951623, |
| "learning_rate": 1.704347572603529e-05, |
| "loss": 0.4675910472869873, |
| "memory(GiB)": 76.61, |
| "step": 860, |
| "token_acc": 0.8361073624231519, |
| "train_speed(iter/s)": 0.029659 |
| }, |
| { |
| "epoch": 0.2767889284428623, |
| "grad_norm": 0.056617536923221766, |
| "learning_rate": 1.700652016900419e-05, |
| "loss": 0.467483377456665, |
| "memory(GiB)": 76.61, |
| "step": 865, |
| "token_acc": 0.8753590807532716, |
| "train_speed(iter/s)": 0.029639 |
| }, |
| { |
| "epoch": 0.27838886444542216, |
| "grad_norm": 0.060433939578350394, |
| "learning_rate": 1.696937568655294e-05, |
| "loss": 0.46129570007324217, |
| "memory(GiB)": 76.61, |
| "step": 870, |
| "token_acc": 0.8700755748512623, |
| "train_speed(iter/s)": 0.029622 |
| }, |
| { |
| "epoch": 0.2799888004479821, |
| "grad_norm": 0.06826391103956585, |
| "learning_rate": 1.6932043280253892e-05, |
| "loss": 0.47449960708618166, |
| "memory(GiB)": 76.61, |
| "step": 875, |
| "token_acc": 0.8767408356010885, |
| "train_speed(iter/s)": 0.02965 |
| }, |
| { |
| "epoch": 0.281588736450542, |
| "grad_norm": 0.060978189753072065, |
| "learning_rate": 1.689452395674664e-05, |
| "loss": 0.464243745803833, |
| "memory(GiB)": 76.61, |
| "step": 880, |
| "token_acc": 0.8622170179547228, |
| "train_speed(iter/s)": 0.029624 |
| }, |
| { |
| "epoch": 0.28318867245310186, |
| "grad_norm": 0.0760276206328267, |
| "learning_rate": 1.6856818727710847e-05, |
| "loss": 0.4566212177276611, |
| "memory(GiB)": 76.61, |
| "step": 885, |
| "token_acc": 0.8465499485066942, |
| "train_speed(iter/s)": 0.029618 |
| }, |
| { |
| "epoch": 0.2847886084556618, |
| "grad_norm": 0.05693121191664627, |
| "learning_rate": 1.6818928609838967e-05, |
| "loss": 0.46042599678039553, |
| "memory(GiB)": 76.61, |
| "step": 890, |
| "token_acc": 0.8798391728891441, |
| "train_speed(iter/s)": 0.029627 |
| }, |
| { |
| "epoch": 0.28638854445822165, |
| "grad_norm": 0.05744826995499506, |
| "learning_rate": 1.678085462480885e-05, |
| "loss": 0.4604465961456299, |
| "memory(GiB)": 76.61, |
| "step": 895, |
| "token_acc": 0.8780676542118063, |
| "train_speed(iter/s)": 0.029599 |
| }, |
| { |
| "epoch": 0.28798848046078157, |
| "grad_norm": 0.06271464886952488, |
| "learning_rate": 1.6742597799256182e-05, |
| "loss": 0.46231966018676757, |
| "memory(GiB)": 76.61, |
| "step": 900, |
| "token_acc": 0.8866765515780555, |
| "train_speed(iter/s)": 0.029611 |
| }, |
| { |
| "epoch": 0.2895884164633415, |
| "grad_norm": 0.06044356676681803, |
| "learning_rate": 1.6704159164746797e-05, |
| "loss": 0.47655544281005857, |
| "memory(GiB)": 76.61, |
| "step": 905, |
| "token_acc": 0.8872944211544663, |
| "train_speed(iter/s)": 0.029601 |
| }, |
| { |
| "epoch": 0.29118835246590136, |
| "grad_norm": 0.05103569816400521, |
| "learning_rate": 1.6665539757748866e-05, |
| "loss": 0.4603917121887207, |
| "memory(GiB)": 76.61, |
| "step": 910, |
| "token_acc": 0.8611705475141599, |
| "train_speed(iter/s)": 0.029574 |
| }, |
| { |
| "epoch": 0.2927882884684613, |
| "grad_norm": 0.055811472748585486, |
| "learning_rate": 1.6626740619604967e-05, |
| "loss": 0.46213679313659667, |
| "memory(GiB)": 76.61, |
| "step": 915, |
| "token_acc": 0.8148507643775783, |
| "train_speed(iter/s)": 0.029594 |
| }, |
| { |
| "epoch": 0.29438822447102114, |
| "grad_norm": 0.05463929857953068, |
| "learning_rate": 1.658776279650397e-05, |
| "loss": 0.4658839702606201, |
| "memory(GiB)": 76.61, |
| "step": 920, |
| "token_acc": 0.8766637089618456, |
| "train_speed(iter/s)": 0.029577 |
| }, |
| { |
| "epoch": 0.29598816047358106, |
| "grad_norm": 0.06343067949686905, |
| "learning_rate": 1.6548607339452853e-05, |
| "loss": 0.46423888206481934, |
| "memory(GiB)": 76.61, |
| "step": 925, |
| "token_acc": 0.8785782119115453, |
| "train_speed(iter/s)": 0.029564 |
| }, |
| { |
| "epoch": 0.29758809647614093, |
| "grad_norm": 0.052431934937864355, |
| "learning_rate": 1.6509275304248366e-05, |
| "loss": 0.46324734687805175, |
| "memory(GiB)": 76.61, |
| "step": 930, |
| "token_acc": 0.8571011956838729, |
| "train_speed(iter/s)": 0.02958 |
| }, |
| { |
| "epoch": 0.29918803247870085, |
| "grad_norm": 0.059009943510604755, |
| "learning_rate": 1.6469767751448538e-05, |
| "loss": 0.46290836334228513, |
| "memory(GiB)": 76.61, |
| "step": 935, |
| "token_acc": 0.8388616290480864, |
| "train_speed(iter/s)": 0.029556 |
| }, |
| { |
| "epoch": 0.30078796848126077, |
| "grad_norm": 0.05160057372757322, |
| "learning_rate": 1.6430085746344107e-05, |
| "loss": 0.45898871421813964, |
| "memory(GiB)": 76.61, |
| "step": 940, |
| "token_acc": 0.8690580344123651, |
| "train_speed(iter/s)": 0.029556 |
| }, |
| { |
| "epoch": 0.30238790448382064, |
| "grad_norm": 0.05612231994140208, |
| "learning_rate": 1.639023035892978e-05, |
| "loss": 0.4546724796295166, |
| "memory(GiB)": 76.61, |
| "step": 945, |
| "token_acc": 0.876509544215037, |
| "train_speed(iter/s)": 0.02956 |
| }, |
| { |
| "epoch": 0.30398784048638056, |
| "grad_norm": 0.06733149115024578, |
| "learning_rate": 1.6350202663875385e-05, |
| "loss": 0.4598522663116455, |
| "memory(GiB)": 76.61, |
| "step": 950, |
| "token_acc": 0.8623452294246177, |
| "train_speed(iter/s)": 0.029531 |
| }, |
| { |
| "epoch": 0.3055877764889404, |
| "grad_norm": 0.05450569676621943, |
| "learning_rate": 1.6310003740496887e-05, |
| "loss": 0.4602477550506592, |
| "memory(GiB)": 76.61, |
| "step": 955, |
| "token_acc": 0.8647700701480904, |
| "train_speed(iter/s)": 0.029548 |
| }, |
| { |
| "epoch": 0.30718771249150034, |
| "grad_norm": 0.06736921151917717, |
| "learning_rate": 1.6269634672727296e-05, |
| "loss": 0.4589672565460205, |
| "memory(GiB)": 76.61, |
| "step": 960, |
| "token_acc": 0.877502001601281, |
| "train_speed(iter/s)": 0.029536 |
| }, |
| { |
| "epoch": 0.30878764849406026, |
| "grad_norm": 0.06166660436042404, |
| "learning_rate": 1.6229096549087434e-05, |
| "loss": 0.4601268291473389, |
| "memory(GiB)": 76.61, |
| "step": 965, |
| "token_acc": 0.8723534201954397, |
| "train_speed(iter/s)": 0.029518 |
| }, |
| { |
| "epoch": 0.31038758449662013, |
| "grad_norm": 0.055128746386822226, |
| "learning_rate": 1.618839046265658e-05, |
| "loss": 0.4666788101196289, |
| "memory(GiB)": 76.61, |
| "step": 970, |
| "token_acc": 0.8550563360689943, |
| "train_speed(iter/s)": 0.029541 |
| }, |
| { |
| "epoch": 0.31198752049918005, |
| "grad_norm": 0.056867326711030626, |
| "learning_rate": 1.614751751104301e-05, |
| "loss": 0.4646125793457031, |
| "memory(GiB)": 76.61, |
| "step": 975, |
| "token_acc": 0.8651571964234208, |
| "train_speed(iter/s)": 0.029524 |
| }, |
| { |
| "epoch": 0.3135874565017399, |
| "grad_norm": 0.05501107287069041, |
| "learning_rate": 1.6106478796354382e-05, |
| "loss": 0.4588280200958252, |
| "memory(GiB)": 76.61, |
| "step": 980, |
| "token_acc": 0.8767766331985918, |
| "train_speed(iter/s)": 0.029517 |
| }, |
| { |
| "epoch": 0.31518739250429983, |
| "grad_norm": 0.08099201898186387, |
| "learning_rate": 1.6065275425168034e-05, |
| "loss": 0.4589373111724854, |
| "memory(GiB)": 76.61, |
| "step": 985, |
| "token_acc": 0.8917890157694399, |
| "train_speed(iter/s)": 0.029526 |
| }, |
| { |
| "epoch": 0.3167873285068597, |
| "grad_norm": 0.0522899382710734, |
| "learning_rate": 1.602390850850113e-05, |
| "loss": 0.46761279106140136, |
| "memory(GiB)": 76.61, |
| "step": 990, |
| "token_acc": 0.8461229409401366, |
| "train_speed(iter/s)": 0.029505 |
| }, |
| { |
| "epoch": 0.3183872645094196, |
| "grad_norm": 0.05838858698011934, |
| "learning_rate": 1.5982379161780722e-05, |
| "loss": 0.44941887855529783, |
| "memory(GiB)": 76.61, |
| "step": 995, |
| "token_acc": 0.8547228871294421, |
| "train_speed(iter/s)": 0.029511 |
| }, |
| { |
| "epoch": 0.31998720051197954, |
| "grad_norm": 0.054930484370324516, |
| "learning_rate": 1.5940688504813664e-05, |
| "loss": 0.4591392517089844, |
| "memory(GiB)": 76.61, |
| "step": 1000, |
| "token_acc": 0.8995555555555556, |
| "train_speed(iter/s)": 0.029505 |
| }, |
| { |
| "epoch": 0.31998720051197954, |
| "eval_loss": 0.671963095664978, |
| "eval_runtime": 110.8694, |
| "eval_samples_per_second": 181.186, |
| "eval_steps_per_second": 0.911, |
| "eval_token_acc": 0.8676077802864524, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.3215871365145394, |
| "grad_norm": 0.0578985798516978, |
| "learning_rate": 1.5898837661756405e-05, |
| "loss": 0.46222972869873047, |
| "memory(GiB)": 76.61, |
| "step": 1005, |
| "token_acc": 0.8840002569208042, |
| "train_speed(iter/s)": 0.029425 |
| }, |
| { |
| "epoch": 0.3231870725170993, |
| "grad_norm": 0.05872050053297838, |
| "learning_rate": 1.5856827761084698e-05, |
| "loss": 0.45543718338012695, |
| "memory(GiB)": 76.61, |
| "step": 1010, |
| "token_acc": 0.8753668220265838, |
| "train_speed(iter/s)": 0.02945 |
| }, |
| { |
| "epoch": 0.3247870085196592, |
| "grad_norm": 0.05268695066428434, |
| "learning_rate": 1.5814659935563165e-05, |
| "loss": 0.46614727973937986, |
| "memory(GiB)": 76.61, |
| "step": 1015, |
| "token_acc": 0.8792250035355678, |
| "train_speed(iter/s)": 0.029474 |
| }, |
| { |
| "epoch": 0.3263869445222191, |
| "grad_norm": 0.059454673806441594, |
| "learning_rate": 1.577233532221474e-05, |
| "loss": 0.45902605056762696, |
| "memory(GiB)": 76.61, |
| "step": 1020, |
| "token_acc": 0.86709886547812, |
| "train_speed(iter/s)": 0.029475 |
| }, |
| { |
| "epoch": 0.32798688052477903, |
| "grad_norm": 0.053728974295076275, |
| "learning_rate": 1.5729855062290024e-05, |
| "loss": 0.46491541862487795, |
| "memory(GiB)": 76.61, |
| "step": 1025, |
| "token_acc": 0.8708870261478794, |
| "train_speed(iter/s)": 0.029469 |
| }, |
| { |
| "epoch": 0.3295868165273389, |
| "grad_norm": 0.07030309576814114, |
| "learning_rate": 1.568722030123651e-05, |
| "loss": 0.453840970993042, |
| "memory(GiB)": 76.61, |
| "step": 1030, |
| "token_acc": 0.8568111455108359, |
| "train_speed(iter/s)": 0.029496 |
| }, |
| { |
| "epoch": 0.3311867525298988, |
| "grad_norm": 0.07385415365022158, |
| "learning_rate": 1.5644432188667695e-05, |
| "loss": 0.45582828521728513, |
| "memory(GiB)": 76.61, |
| "step": 1035, |
| "token_acc": 0.8800162140251317, |
| "train_speed(iter/s)": 0.029488 |
| }, |
| { |
| "epoch": 0.3327866885324587, |
| "grad_norm": 0.05407863995123405, |
| "learning_rate": 1.5601491878332077e-05, |
| "loss": 0.4665637969970703, |
| "memory(GiB)": 76.61, |
| "step": 1040, |
| "token_acc": 0.8628481345244351, |
| "train_speed(iter/s)": 0.029487 |
| }, |
| { |
| "epoch": 0.3343866245350186, |
| "grad_norm": 0.05879461372080454, |
| "learning_rate": 1.5558400528082057e-05, |
| "loss": 0.4657593250274658, |
| "memory(GiB)": 76.61, |
| "step": 1045, |
| "token_acc": 0.879185119574845, |
| "train_speed(iter/s)": 0.02951 |
| }, |
| { |
| "epoch": 0.33598656053757847, |
| "grad_norm": 0.06618244368029796, |
| "learning_rate": 1.551515929984271e-05, |
| "loss": 0.45760574340820315, |
| "memory(GiB)": 76.61, |
| "step": 1050, |
| "token_acc": 0.8899380348185305, |
| "train_speed(iter/s)": 0.029502 |
| }, |
| { |
| "epoch": 0.3375864965401384, |
| "grad_norm": 0.06388796415692906, |
| "learning_rate": 1.547176935958044e-05, |
| "loss": 0.46065597534179686, |
| "memory(GiB)": 76.61, |
| "step": 1055, |
| "token_acc": 0.8536853685368537, |
| "train_speed(iter/s)": 0.029524 |
| }, |
| { |
| "epoch": 0.3391864325426983, |
| "grad_norm": 0.05811152365312673, |
| "learning_rate": 1.5428231877271584e-05, |
| "loss": 0.46312780380249025, |
| "memory(GiB)": 76.61, |
| "step": 1060, |
| "token_acc": 0.8520375161707633, |
| "train_speed(iter/s)": 0.029515 |
| }, |
| { |
| "epoch": 0.3407863685452582, |
| "grad_norm": 0.05545936328508829, |
| "learning_rate": 1.538454802687081e-05, |
| "loss": 0.4615220546722412, |
| "memory(GiB)": 76.61, |
| "step": 1065, |
| "token_acc": 0.8744265080713679, |
| "train_speed(iter/s)": 0.029504 |
| }, |
| { |
| "epoch": 0.3423863045478181, |
| "grad_norm": 0.05964362984731802, |
| "learning_rate": 1.5340718986279505e-05, |
| "loss": 0.46706466674804686, |
| "memory(GiB)": 76.61, |
| "step": 1070, |
| "token_acc": 0.8592233009708737, |
| "train_speed(iter/s)": 0.029536 |
| }, |
| { |
| "epoch": 0.34398624055037796, |
| "grad_norm": 0.05356886450328198, |
| "learning_rate": 1.529674593731399e-05, |
| "loss": 0.45301499366760256, |
| "memory(GiB)": 76.61, |
| "step": 1075, |
| "token_acc": 0.8575192096597146, |
| "train_speed(iter/s)": 0.029526 |
| }, |
| { |
| "epoch": 0.3455861765529379, |
| "grad_norm": 0.05995962073425321, |
| "learning_rate": 1.5252630065673662e-05, |
| "loss": 0.46819314956665037, |
| "memory(GiB)": 76.61, |
| "step": 1080, |
| "token_acc": 0.8875031814711123, |
| "train_speed(iter/s)": 0.029518 |
| }, |
| { |
| "epoch": 0.3471861125554978, |
| "grad_norm": 0.05389432634852101, |
| "learning_rate": 1.5208372560909031e-05, |
| "loss": 0.46298394203186033, |
| "memory(GiB)": 76.61, |
| "step": 1085, |
| "token_acc": 0.8872426699937617, |
| "train_speed(iter/s)": 0.029543 |
| }, |
| { |
| "epoch": 0.34878604855805767, |
| "grad_norm": 0.06642390255342462, |
| "learning_rate": 1.5163974616389621e-05, |
| "loss": 0.45978522300720215, |
| "memory(GiB)": 76.61, |
| "step": 1090, |
| "token_acc": 0.8246258860593332, |
| "train_speed(iter/s)": 0.029525 |
| }, |
| { |
| "epoch": 0.3503859845606176, |
| "grad_norm": 0.06115184110491886, |
| "learning_rate": 1.5119437429271813e-05, |
| "loss": 0.4637304782867432, |
| "memory(GiB)": 76.61, |
| "step": 1095, |
| "token_acc": 0.8666082895504962, |
| "train_speed(iter/s)": 0.029534 |
| }, |
| { |
| "epoch": 0.35198592056317746, |
| "grad_norm": 0.060865150660591956, |
| "learning_rate": 1.5074762200466557e-05, |
| "loss": 0.4542848587036133, |
| "memory(GiB)": 76.61, |
| "step": 1100, |
| "token_acc": 0.8913602663035255, |
| "train_speed(iter/s)": 0.029544 |
| }, |
| { |
| "epoch": 0.3535858565657374, |
| "grad_norm": 0.057666943430007674, |
| "learning_rate": 1.5029950134606991e-05, |
| "loss": 0.4574248790740967, |
| "memory(GiB)": 76.61, |
| "step": 1105, |
| "token_acc": 0.8634470336597996, |
| "train_speed(iter/s)": 0.029524 |
| }, |
| { |
| "epoch": 0.35518579256829724, |
| "grad_norm": 0.054034554153381265, |
| "learning_rate": 1.4985002440015959e-05, |
| "loss": 0.4520272254943848, |
| "memory(GiB)": 76.61, |
| "step": 1110, |
| "token_acc": 0.8674898358680921, |
| "train_speed(iter/s)": 0.029551 |
| }, |
| { |
| "epoch": 0.35678572857085716, |
| "grad_norm": 0.06416854479766453, |
| "learning_rate": 1.4939920328673422e-05, |
| "loss": 0.4668846130371094, |
| "memory(GiB)": 76.61, |
| "step": 1115, |
| "token_acc": 0.9170854271356784, |
| "train_speed(iter/s)": 0.029541 |
| }, |
| { |
| "epoch": 0.3583856645734171, |
| "grad_norm": 0.05775941336987237, |
| "learning_rate": 1.4894705016183803e-05, |
| "loss": 0.4518620491027832, |
| "memory(GiB)": 76.61, |
| "step": 1120, |
| "token_acc": 0.8672782874617737, |
| "train_speed(iter/s)": 0.029531 |
| }, |
| { |
| "epoch": 0.35998560057597695, |
| "grad_norm": 0.0625175589581686, |
| "learning_rate": 1.4849357721743169e-05, |
| "loss": 0.4566941738128662, |
| "memory(GiB)": 76.61, |
| "step": 1125, |
| "token_acc": 0.8505491793163026, |
| "train_speed(iter/s)": 0.029557 |
| }, |
| { |
| "epoch": 0.36158553657853687, |
| "grad_norm": 0.05911529293553411, |
| "learning_rate": 1.4803879668106393e-05, |
| "loss": 0.4640664577484131, |
| "memory(GiB)": 76.61, |
| "step": 1130, |
| "token_acc": 0.8772325625117503, |
| "train_speed(iter/s)": 0.029544 |
| }, |
| { |
| "epoch": 0.36318547258109674, |
| "grad_norm": 0.06483783687935218, |
| "learning_rate": 1.4758272081554168e-05, |
| "loss": 0.45419878959655763, |
| "memory(GiB)": 76.61, |
| "step": 1135, |
| "token_acc": 0.8594914930223667, |
| "train_speed(iter/s)": 0.029539 |
| }, |
| { |
| "epoch": 0.36478540858365666, |
| "grad_norm": 0.06032730304497941, |
| "learning_rate": 1.4712536191859934e-05, |
| "loss": 0.45779004096984866, |
| "memory(GiB)": 76.61, |
| "step": 1140, |
| "token_acc": 0.8938053097345132, |
| "train_speed(iter/s)": 0.029564 |
| }, |
| { |
| "epoch": 0.3663853445862166, |
| "grad_norm": 0.0637380940226065, |
| "learning_rate": 1.4666673232256738e-05, |
| "loss": 0.46385722160339354, |
| "memory(GiB)": 76.61, |
| "step": 1145, |
| "token_acc": 0.8621830209481808, |
| "train_speed(iter/s)": 0.029544 |
| }, |
| { |
| "epoch": 0.36798528058877644, |
| "grad_norm": 0.057006770373085346, |
| "learning_rate": 1.4620684439403962e-05, |
| "loss": 0.4613553524017334, |
| "memory(GiB)": 76.61, |
| "step": 1150, |
| "token_acc": 0.8831837819873712, |
| "train_speed(iter/s)": 0.029558 |
| }, |
| { |
| "epoch": 0.36958521659133636, |
| "grad_norm": 0.057569299635009126, |
| "learning_rate": 1.4574571053353987e-05, |
| "loss": 0.4598341464996338, |
| "memory(GiB)": 76.61, |
| "step": 1155, |
| "token_acc": 0.8825154371140721, |
| "train_speed(iter/s)": 0.029557 |
| }, |
| { |
| "epoch": 0.37118515259389623, |
| "grad_norm": 0.06747695219063263, |
| "learning_rate": 1.452833431751875e-05, |
| "loss": 0.4570640563964844, |
| "memory(GiB)": 76.61, |
| "step": 1160, |
| "token_acc": 0.8726823238566132, |
| "train_speed(iter/s)": 0.029543 |
| }, |
| { |
| "epoch": 0.37278508859645615, |
| "grad_norm": 0.05405367649749466, |
| "learning_rate": 1.448197547863622e-05, |
| "loss": 0.4516812801361084, |
| "memory(GiB)": 76.61, |
| "step": 1165, |
| "token_acc": 0.8704696273608984, |
| "train_speed(iter/s)": 0.029568 |
| }, |
| { |
| "epoch": 0.374385024599016, |
| "grad_norm": 0.06041157710672601, |
| "learning_rate": 1.4435495786736796e-05, |
| "loss": 0.465837287902832, |
| "memory(GiB)": 76.61, |
| "step": 1170, |
| "token_acc": 0.8673412029229904, |
| "train_speed(iter/s)": 0.029554 |
| }, |
| { |
| "epoch": 0.37598496060157593, |
| "grad_norm": 0.05229585247228306, |
| "learning_rate": 1.438889649510956e-05, |
| "loss": 0.4427653789520264, |
| "memory(GiB)": 76.61, |
| "step": 1175, |
| "token_acc": 0.8558139534883721, |
| "train_speed(iter/s)": 0.02954 |
| }, |
| { |
| "epoch": 0.37758489660413586, |
| "grad_norm": 0.0547875272797444, |
| "learning_rate": 1.4342178860268523e-05, |
| "loss": 0.45673260688781736, |
| "memory(GiB)": 76.61, |
| "step": 1180, |
| "token_acc": 0.880563238622077, |
| "train_speed(iter/s)": 0.029563 |
| }, |
| { |
| "epoch": 0.3791848326066957, |
| "grad_norm": 0.0565328006493161, |
| "learning_rate": 1.4295344141918734e-05, |
| "loss": 0.46208748817443845, |
| "memory(GiB)": 76.61, |
| "step": 1185, |
| "token_acc": 0.8671328671328671, |
| "train_speed(iter/s)": 0.029544 |
| }, |
| { |
| "epoch": 0.38078476860925564, |
| "grad_norm": 0.062473905403265834, |
| "learning_rate": 1.4248393602922299e-05, |
| "loss": 0.46883163452148435, |
| "memory(GiB)": 76.61, |
| "step": 1190, |
| "token_acc": 0.8412252145605209, |
| "train_speed(iter/s)": 0.029548 |
| }, |
| { |
| "epoch": 0.3823847046118155, |
| "grad_norm": 0.05646151042315891, |
| "learning_rate": 1.420132850926434e-05, |
| "loss": 0.45732822418212893, |
| "memory(GiB)": 76.61, |
| "step": 1195, |
| "token_acc": 0.8820655966503839, |
| "train_speed(iter/s)": 0.02956 |
| }, |
| { |
| "epoch": 0.3839846406143754, |
| "grad_norm": 0.052981558367052706, |
| "learning_rate": 1.4154150130018867e-05, |
| "loss": 0.45579113960266116, |
| "memory(GiB)": 76.61, |
| "step": 1200, |
| "token_acc": 0.8677085226240233, |
| "train_speed(iter/s)": 0.029546 |
| }, |
| { |
| "epoch": 0.38558457661693535, |
| "grad_norm": 0.052315204322432474, |
| "learning_rate": 1.4106859737314532e-05, |
| "loss": 0.45348801612854006, |
| "memory(GiB)": 76.61, |
| "step": 1205, |
| "token_acc": 0.8616187989556136, |
| "train_speed(iter/s)": 0.029561 |
| }, |
| { |
| "epoch": 0.3871845126194952, |
| "grad_norm": 0.05319888084520812, |
| "learning_rate": 1.4059458606300358e-05, |
| "loss": 0.45279593467712403, |
| "memory(GiB)": 76.61, |
| "step": 1210, |
| "token_acc": 0.86090645233311, |
| "train_speed(iter/s)": 0.029565 |
| }, |
| { |
| "epoch": 0.38878444862205513, |
| "grad_norm": 0.054475973938428034, |
| "learning_rate": 1.4011948015111334e-05, |
| "loss": 0.4616706848144531, |
| "memory(GiB)": 76.61, |
| "step": 1215, |
| "token_acc": 0.8390133684805121, |
| "train_speed(iter/s)": 0.029549 |
| }, |
| { |
| "epoch": 0.390384384624615, |
| "grad_norm": 0.054891067059900926, |
| "learning_rate": 1.396432924483396e-05, |
| "loss": 0.4553243637084961, |
| "memory(GiB)": 76.61, |
| "step": 1220, |
| "token_acc": 0.8715350793347353, |
| "train_speed(iter/s)": 0.029571 |
| }, |
| { |
| "epoch": 0.3919843206271749, |
| "grad_norm": 0.06058246643434403, |
| "learning_rate": 1.3916603579471705e-05, |
| "loss": 0.47067904472351074, |
| "memory(GiB)": 76.61, |
| "step": 1225, |
| "token_acc": 0.8662144337667232, |
| "train_speed(iter/s)": 0.029556 |
| }, |
| { |
| "epoch": 0.3935842566297348, |
| "grad_norm": 0.05715510214651738, |
| "learning_rate": 1.3868772305910376e-05, |
| "loss": 0.46147928237915037, |
| "memory(GiB)": 76.61, |
| "step": 1230, |
| "token_acc": 0.868918375552875, |
| "train_speed(iter/s)": 0.029548 |
| }, |
| { |
| "epoch": 0.3951841926322947, |
| "grad_norm": 0.06593047910666934, |
| "learning_rate": 1.3820836713883424e-05, |
| "loss": 0.45935769081115724, |
| "memory(GiB)": 76.61, |
| "step": 1235, |
| "token_acc": 0.8596291476903057, |
| "train_speed(iter/s)": 0.02957 |
| }, |
| { |
| "epoch": 0.3967841286348546, |
| "grad_norm": 0.056071042953882384, |
| "learning_rate": 1.3772798095937172e-05, |
| "loss": 0.4495890140533447, |
| "memory(GiB)": 76.61, |
| "step": 1240, |
| "token_acc": 0.8471917163476623, |
| "train_speed(iter/s)": 0.029553 |
| }, |
| { |
| "epoch": 0.3983840646374145, |
| "grad_norm": 0.05810589720196263, |
| "learning_rate": 1.3724657747395957e-05, |
| "loss": 0.4619898319244385, |
| "memory(GiB)": 76.61, |
| "step": 1245, |
| "token_acc": 0.8691186216037111, |
| "train_speed(iter/s)": 0.029561 |
| }, |
| { |
| "epoch": 0.3999840006399744, |
| "grad_norm": 0.055604926632171425, |
| "learning_rate": 1.3676416966327201e-05, |
| "loss": 0.4587514400482178, |
| "memory(GiB)": 76.61, |
| "step": 1250, |
| "token_acc": 0.8369355461211887, |
| "train_speed(iter/s)": 0.029564 |
| }, |
| { |
| "epoch": 0.3999840006399744, |
| "eval_loss": 0.6690404415130615, |
| "eval_runtime": 106.3444, |
| "eval_samples_per_second": 188.896, |
| "eval_steps_per_second": 0.95, |
| "eval_token_acc": 0.8683678146748934, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.4015839366425343, |
| "grad_norm": 0.04782987834900457, |
| "learning_rate": 1.362807705350641e-05, |
| "loss": 0.46315851211547854, |
| "memory(GiB)": 76.61, |
| "step": 1255, |
| "token_acc": 0.8767961498796838, |
| "train_speed(iter/s)": 0.029512 |
| }, |
| { |
| "epoch": 0.4031838726450942, |
| "grad_norm": 0.05995996443795485, |
| "learning_rate": 1.3579639312382105e-05, |
| "loss": 0.46349530220031737, |
| "memory(GiB)": 76.61, |
| "step": 1260, |
| "token_acc": 0.8588617886178862, |
| "train_speed(iter/s)": 0.029524 |
| }, |
| { |
| "epoch": 0.4047838086476541, |
| "grad_norm": 0.06488882353036057, |
| "learning_rate": 1.3531105049040667e-05, |
| "loss": 0.45726447105407714, |
| "memory(GiB)": 76.61, |
| "step": 1265, |
| "token_acc": 0.8802249582003344, |
| "train_speed(iter/s)": 0.029543 |
| }, |
| { |
| "epoch": 0.406383744650214, |
| "grad_norm": 0.05350128050935312, |
| "learning_rate": 1.3482475572171132e-05, |
| "loss": 0.4516806125640869, |
| "memory(GiB)": 76.61, |
| "step": 1270, |
| "token_acc": 0.8560765550239234, |
| "train_speed(iter/s)": 0.029549 |
| }, |
| { |
| "epoch": 0.4079836806527739, |
| "grad_norm": 0.05672697687392494, |
| "learning_rate": 1.3433752193029888e-05, |
| "loss": 0.46581568717956545, |
| "memory(GiB)": 76.61, |
| "step": 1275, |
| "token_acc": 0.8881742738589211, |
| "train_speed(iter/s)": 0.029547 |
| }, |
| { |
| "epoch": 0.40958361665533377, |
| "grad_norm": 0.0598115330947421, |
| "learning_rate": 1.3384936225405326e-05, |
| "loss": 0.46333680152893064, |
| "memory(GiB)": 76.61, |
| "step": 1280, |
| "token_acc": 0.8608710985716804, |
| "train_speed(iter/s)": 0.029573 |
| }, |
| { |
| "epoch": 0.4111835526578937, |
| "grad_norm": 0.05384417907735887, |
| "learning_rate": 1.333602898558242e-05, |
| "loss": 0.4611030578613281, |
| "memory(GiB)": 76.61, |
| "step": 1285, |
| "token_acc": 0.8845689770746749, |
| "train_speed(iter/s)": 0.029567 |
| }, |
| { |
| "epoch": 0.41278348866045356, |
| "grad_norm": 0.06043637267465684, |
| "learning_rate": 1.3287031792307226e-05, |
| "loss": 0.46013875007629396, |
| "memory(GiB)": 76.61, |
| "step": 1290, |
| "token_acc": 0.870195210303884, |
| "train_speed(iter/s)": 0.029565 |
| }, |
| { |
| "epoch": 0.4143834246630135, |
| "grad_norm": 0.06140603532631629, |
| "learning_rate": 1.323794596675132e-05, |
| "loss": 0.45681238174438477, |
| "memory(GiB)": 76.61, |
| "step": 1295, |
| "token_acc": 0.8450012281994596, |
| "train_speed(iter/s)": 0.029583 |
| }, |
| { |
| "epoch": 0.4159833606655734, |
| "grad_norm": 0.062077229851937275, |
| "learning_rate": 1.318877283247619e-05, |
| "loss": 0.4490199565887451, |
| "memory(GiB)": 76.61, |
| "step": 1300, |
| "token_acc": 0.89259877573734, |
| "train_speed(iter/s)": 0.029573 |
| }, |
| { |
| "epoch": 0.41758329666813326, |
| "grad_norm": 0.05216177276902916, |
| "learning_rate": 1.3139513715397521e-05, |
| "loss": 0.45108351707458494, |
| "memory(GiB)": 76.61, |
| "step": 1305, |
| "token_acc": 0.8547701815372731, |
| "train_speed(iter/s)": 0.029594 |
| }, |
| { |
| "epoch": 0.4191832326706932, |
| "grad_norm": 0.05738628087610287, |
| "learning_rate": 1.3090169943749475e-05, |
| "loss": 0.4558550834655762, |
| "memory(GiB)": 76.61, |
| "step": 1310, |
| "token_acc": 0.840696686491079, |
| "train_speed(iter/s)": 0.029587 |
| }, |
| { |
| "epoch": 0.42078316867325305, |
| "grad_norm": 0.05518036740697275, |
| "learning_rate": 1.304074284804885e-05, |
| "loss": 0.4631648063659668, |
| "memory(GiB)": 76.61, |
| "step": 1315, |
| "token_acc": 0.8788111708941839, |
| "train_speed(iter/s)": 0.029578 |
| }, |
| { |
| "epoch": 0.42238310467581297, |
| "grad_norm": 0.05902492258138098, |
| "learning_rate": 1.2991233761059214e-05, |
| "loss": 0.45921921730041504, |
| "memory(GiB)": 76.61, |
| "step": 1320, |
| "token_acc": 0.866059646344682, |
| "train_speed(iter/s)": 0.029604 |
| }, |
| { |
| "epoch": 0.4239830406783729, |
| "grad_norm": 0.059535437419073044, |
| "learning_rate": 1.2941644017754964e-05, |
| "loss": 0.46445517539978026, |
| "memory(GiB)": 76.61, |
| "step": 1325, |
| "token_acc": 0.8831345826235094, |
| "train_speed(iter/s)": 0.029591 |
| }, |
| { |
| "epoch": 0.42558297668093276, |
| "grad_norm": 0.04863893443696892, |
| "learning_rate": 1.289197495528534e-05, |
| "loss": 0.45836362838745115, |
| "memory(GiB)": 76.61, |
| "step": 1330, |
| "token_acc": 0.8986429177268872, |
| "train_speed(iter/s)": 0.029582 |
| }, |
| { |
| "epoch": 0.4271829126834927, |
| "grad_norm": 0.05945822860509985, |
| "learning_rate": 1.284222791293836e-05, |
| "loss": 0.45783252716064454, |
| "memory(GiB)": 76.61, |
| "step": 1335, |
| "token_acc": 0.8500874125874126, |
| "train_speed(iter/s)": 0.029596 |
| }, |
| { |
| "epoch": 0.42878284868605254, |
| "grad_norm": 0.05989616737178823, |
| "learning_rate": 1.2792404232104699e-05, |
| "loss": 0.45293269157409666, |
| "memory(GiB)": 76.61, |
| "step": 1340, |
| "token_acc": 0.8637480798771121, |
| "train_speed(iter/s)": 0.029584 |
| }, |
| { |
| "epoch": 0.43038278468861246, |
| "grad_norm": 0.0586629819404024, |
| "learning_rate": 1.2742505256241543e-05, |
| "loss": 0.45876450538635255, |
| "memory(GiB)": 76.61, |
| "step": 1345, |
| "token_acc": 0.8296499119890475, |
| "train_speed(iter/s)": 0.029588 |
| }, |
| { |
| "epoch": 0.43198272069117233, |
| "grad_norm": 0.052924904785980484, |
| "learning_rate": 1.2692532330836346e-05, |
| "loss": 0.45821080207824705, |
| "memory(GiB)": 76.61, |
| "step": 1350, |
| "token_acc": 0.8636084374360025, |
| "train_speed(iter/s)": 0.029594 |
| }, |
| { |
| "epoch": 0.43358265669373225, |
| "grad_norm": 0.059304249814977644, |
| "learning_rate": 1.2642486803370553e-05, |
| "loss": 0.45485148429870603, |
| "memory(GiB)": 76.61, |
| "step": 1355, |
| "token_acc": 0.8686690223792697, |
| "train_speed(iter/s)": 0.029579 |
| }, |
| { |
| "epoch": 0.43518259269629217, |
| "grad_norm": 0.06253442360689314, |
| "learning_rate": 1.2592370023283268e-05, |
| "loss": 0.45198469161987304, |
| "memory(GiB)": 76.61, |
| "step": 1360, |
| "token_acc": 0.8737075332348597, |
| "train_speed(iter/s)": 0.0296 |
| }, |
| { |
| "epoch": 0.43678252869885204, |
| "grad_norm": 0.05314091037792793, |
| "learning_rate": 1.2542183341934873e-05, |
| "loss": 0.4516898155212402, |
| "memory(GiB)": 76.61, |
| "step": 1365, |
| "token_acc": 0.8714476021314387, |
| "train_speed(iter/s)": 0.029596 |
| }, |
| { |
| "epoch": 0.43838246470141196, |
| "grad_norm": 0.06014404788689081, |
| "learning_rate": 1.2491928112570568e-05, |
| "loss": 0.45399184226989747, |
| "memory(GiB)": 76.61, |
| "step": 1370, |
| "token_acc": 0.8657097288676237, |
| "train_speed(iter/s)": 0.029583 |
| }, |
| { |
| "epoch": 0.4399824007039718, |
| "grad_norm": 0.05910144328100835, |
| "learning_rate": 1.2441605690283915e-05, |
| "loss": 0.4607128143310547, |
| "memory(GiB)": 76.61, |
| "step": 1375, |
| "token_acc": 0.8990952307928232, |
| "train_speed(iter/s)": 0.029603 |
| }, |
| { |
| "epoch": 0.44158233670653174, |
| "grad_norm": 0.059073628736854025, |
| "learning_rate": 1.2391217431980273e-05, |
| "loss": 0.4515543937683105, |
| "memory(GiB)": 76.61, |
| "step": 1380, |
| "token_acc": 0.9016349860428021, |
| "train_speed(iter/s)": 0.029591 |
| }, |
| { |
| "epoch": 0.44318227270909166, |
| "grad_norm": 0.058358968679540275, |
| "learning_rate": 1.234076469634022e-05, |
| "loss": 0.45762925148010253, |
| "memory(GiB)": 76.61, |
| "step": 1385, |
| "token_acc": 0.8919261822376009, |
| "train_speed(iter/s)": 0.029584 |
| }, |
| { |
| "epoch": 0.4447822087116515, |
| "grad_norm": 0.0672513399669503, |
| "learning_rate": 1.2290248843782915e-05, |
| "loss": 0.44803729057312014, |
| "memory(GiB)": 76.61, |
| "step": 1390, |
| "token_acc": 0.8975998070196599, |
| "train_speed(iter/s)": 0.029597 |
| }, |
| { |
| "epoch": 0.44638214471421145, |
| "grad_norm": 0.05793114375836921, |
| "learning_rate": 1.2239671236429413e-05, |
| "loss": 0.4537235736846924, |
| "memory(GiB)": 76.61, |
| "step": 1395, |
| "token_acc": 0.8839514422541486, |
| "train_speed(iter/s)": 0.02958 |
| }, |
| { |
| "epoch": 0.4479820807167713, |
| "grad_norm": 0.05955306099185102, |
| "learning_rate": 1.218903323806595e-05, |
| "loss": 0.4573692798614502, |
| "memory(GiB)": 76.61, |
| "step": 1400, |
| "token_acc": 0.8418099547511312, |
| "train_speed(iter/s)": 0.029594 |
| }, |
| { |
| "epoch": 0.44958201671933123, |
| "grad_norm": 0.058484796569864064, |
| "learning_rate": 1.2138336214107148e-05, |
| "loss": 0.44894704818725584, |
| "memory(GiB)": 76.61, |
| "step": 1405, |
| "token_acc": 0.8525200458190149, |
| "train_speed(iter/s)": 0.029594 |
| }, |
| { |
| "epoch": 0.4511819527218911, |
| "grad_norm": 0.05092836798588581, |
| "learning_rate": 1.2087581531559208e-05, |
| "loss": 0.45393967628479004, |
| "memory(GiB)": 76.61, |
| "step": 1410, |
| "token_acc": 0.8791390728476821, |
| "train_speed(iter/s)": 0.02958 |
| }, |
| { |
| "epoch": 0.452781888724451, |
| "grad_norm": 0.07033477253264378, |
| "learning_rate": 1.2036770558983067e-05, |
| "loss": 0.45307221412658694, |
| "memory(GiB)": 76.61, |
| "step": 1415, |
| "token_acc": 0.8387482900136799, |
| "train_speed(iter/s)": 0.029599 |
| }, |
| { |
| "epoch": 0.45438182472701094, |
| "grad_norm": 0.05966547548288182, |
| "learning_rate": 1.1985904666457455e-05, |
| "loss": 0.455959415435791, |
| "memory(GiB)": 76.61, |
| "step": 1420, |
| "token_acc": 0.9042096902303416, |
| "train_speed(iter/s)": 0.029583 |
| }, |
| { |
| "epoch": 0.4559817607295708, |
| "grad_norm": 0.08159145764722696, |
| "learning_rate": 1.1934985225541998e-05, |
| "loss": 0.462065601348877, |
| "memory(GiB)": 76.61, |
| "step": 1425, |
| "token_acc": 0.885252444621832, |
| "train_speed(iter/s)": 0.029573 |
| }, |
| { |
| "epoch": 0.4575816967321307, |
| "grad_norm": 0.05540814227664117, |
| "learning_rate": 1.18840136092402e-05, |
| "loss": 0.4551572322845459, |
| "memory(GiB)": 76.61, |
| "step": 1430, |
| "token_acc": 0.8559651934966797, |
| "train_speed(iter/s)": 0.029592 |
| }, |
| { |
| "epoch": 0.4591816327346906, |
| "grad_norm": 0.05534004007067895, |
| "learning_rate": 1.1832991191962435e-05, |
| "loss": 0.4455368995666504, |
| "memory(GiB)": 76.61, |
| "step": 1435, |
| "token_acc": 0.875560538116592, |
| "train_speed(iter/s)": 0.029576 |
| }, |
| { |
| "epoch": 0.4607815687372505, |
| "grad_norm": 0.058276771895487044, |
| "learning_rate": 1.1781919349488894e-05, |
| "loss": 0.4590908527374268, |
| "memory(GiB)": 76.61, |
| "step": 1440, |
| "token_acc": 0.8510418460478733, |
| "train_speed(iter/s)": 0.029576 |
| }, |
| { |
| "epoch": 0.46238150473981043, |
| "grad_norm": 0.05839975543902795, |
| "learning_rate": 1.1730799458932473e-05, |
| "loss": 0.462816858291626, |
| "memory(GiB)": 76.61, |
| "step": 1445, |
| "token_acc": 0.9052378085490669, |
| "train_speed(iter/s)": 0.029586 |
| }, |
| { |
| "epoch": 0.4639814407423703, |
| "grad_norm": 0.07084434546926481, |
| "learning_rate": 1.1679632898701649e-05, |
| "loss": 0.4550295829772949, |
| "memory(GiB)": 76.61, |
| "step": 1450, |
| "token_acc": 0.8805626598465474, |
| "train_speed(iter/s)": 0.029572 |
| }, |
| { |
| "epoch": 0.4655813767449302, |
| "grad_norm": 0.06519996046237972, |
| "learning_rate": 1.1628421048463315e-05, |
| "loss": 0.46291208267211914, |
| "memory(GiB)": 76.61, |
| "step": 1455, |
| "token_acc": 0.8565744600227359, |
| "train_speed(iter/s)": 0.029581 |
| }, |
| { |
| "epoch": 0.4671813127474901, |
| "grad_norm": 0.05799269979733804, |
| "learning_rate": 1.1577165289105565e-05, |
| "loss": 0.4474311351776123, |
| "memory(GiB)": 76.61, |
| "step": 1460, |
| "token_acc": 0.8579789309403043, |
| "train_speed(iter/s)": 0.029568 |
| }, |
| { |
| "epoch": 0.46878124875005, |
| "grad_norm": 0.057120675003187855, |
| "learning_rate": 1.1525867002700484e-05, |
| "loss": 0.46109714508056643, |
| "memory(GiB)": 76.61, |
| "step": 1465, |
| "token_acc": 0.8752182516587126, |
| "train_speed(iter/s)": 0.029548 |
| }, |
| { |
| "epoch": 0.47038118475260987, |
| "grad_norm": 0.05696370798749074, |
| "learning_rate": 1.1474527572466847e-05, |
| "loss": 0.4501948833465576, |
| "memory(GiB)": 76.61, |
| "step": 1470, |
| "token_acc": 0.8529032258064516, |
| "train_speed(iter/s)": 0.029562 |
| }, |
| { |
| "epoch": 0.4719811207551698, |
| "grad_norm": 0.05518112754329221, |
| "learning_rate": 1.1423148382732854e-05, |
| "loss": 0.45941987037658694, |
| "memory(GiB)": 76.61, |
| "step": 1475, |
| "token_acc": 0.9009282399143164, |
| "train_speed(iter/s)": 0.029545 |
| }, |
| { |
| "epoch": 0.4735810567577297, |
| "grad_norm": 0.051496444525703684, |
| "learning_rate": 1.1371730818898785e-05, |
| "loss": 0.45296878814697267, |
| "memory(GiB)": 76.61, |
| "step": 1480, |
| "token_acc": 0.8814303638644918, |
| "train_speed(iter/s)": 0.029538 |
| }, |
| { |
| "epoch": 0.4751809927602896, |
| "grad_norm": 0.0677105428949175, |
| "learning_rate": 1.132027626739965e-05, |
| "loss": 0.45635080337524414, |
| "memory(GiB)": 76.61, |
| "step": 1485, |
| "token_acc": 0.880248833592535, |
| "train_speed(iter/s)": 0.029546 |
| }, |
| { |
| "epoch": 0.4767809287628495, |
| "grad_norm": 0.0673509631098402, |
| "learning_rate": 1.1268786115667798e-05, |
| "loss": 0.4614115715026855, |
| "memory(GiB)": 76.61, |
| "step": 1490, |
| "token_acc": 0.8609592251210748, |
| "train_speed(iter/s)": 0.029525 |
| }, |
| { |
| "epoch": 0.47838086476540936, |
| "grad_norm": 0.053337771378298794, |
| "learning_rate": 1.1217261752095518e-05, |
| "loss": 0.45500664710998534, |
| "memory(GiB)": 76.61, |
| "step": 1495, |
| "token_acc": 0.8794466403162056, |
| "train_speed(iter/s)": 0.029522 |
| }, |
| { |
| "epoch": 0.4799808007679693, |
| "grad_norm": 0.05429302474155136, |
| "learning_rate": 1.1165704565997593e-05, |
| "loss": 0.44763407707214353, |
| "memory(GiB)": 76.61, |
| "step": 1500, |
| "token_acc": 0.8700440528634361, |
| "train_speed(iter/s)": 0.02952 |
| }, |
| { |
| "epoch": 0.4799808007679693, |
| "eval_loss": 0.6668144464492798, |
| "eval_runtime": 124.2589, |
| "eval_samples_per_second": 161.662, |
| "eval_steps_per_second": 0.813, |
| "eval_token_acc": 0.8694583558206896, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.4815807367705292, |
| "grad_norm": 0.06146640524587408, |
| "learning_rate": 1.1114115947573834e-05, |
| "loss": 0.45711498260498046, |
| "memory(GiB)": 76.61, |
| "step": 1505, |
| "token_acc": 0.8695166967121641, |
| "train_speed(iter/s)": 0.029461 |
| }, |
| { |
| "epoch": 0.48318067277308907, |
| "grad_norm": 0.06183782693437151, |
| "learning_rate": 1.1062497287871606e-05, |
| "loss": 0.4499336242675781, |
| "memory(GiB)": 76.61, |
| "step": 1510, |
| "token_acc": 0.8487557381009906, |
| "train_speed(iter/s)": 0.029468 |
| }, |
| { |
| "epoch": 0.484780608775649, |
| "grad_norm": 0.056753269624682155, |
| "learning_rate": 1.1010849978748314e-05, |
| "loss": 0.4551094055175781, |
| "memory(GiB)": 76.61, |
| "step": 1515, |
| "token_acc": 0.8579035448045033, |
| "train_speed(iter/s)": 0.02948 |
| }, |
| { |
| "epoch": 0.48638054477820886, |
| "grad_norm": 0.05188962595699218, |
| "learning_rate": 1.0959175412833869e-05, |
| "loss": 0.4483503818511963, |
| "memory(GiB)": 76.61, |
| "step": 1520, |
| "token_acc": 0.8649127992905705, |
| "train_speed(iter/s)": 0.029477 |
| }, |
| { |
| "epoch": 0.4879804807807688, |
| "grad_norm": 0.0526414480661873, |
| "learning_rate": 1.0907474983493144e-05, |
| "loss": 0.45140752792358396, |
| "memory(GiB)": 76.61, |
| "step": 1525, |
| "token_acc": 0.8700204290091931, |
| "train_speed(iter/s)": 0.029464 |
| }, |
| { |
| "epoch": 0.48958041678332864, |
| "grad_norm": 0.06682159988119828, |
| "learning_rate": 1.08557500847884e-05, |
| "loss": 0.4480952262878418, |
| "memory(GiB)": 76.61, |
| "step": 1530, |
| "token_acc": 0.8906385187748745, |
| "train_speed(iter/s)": 0.029478 |
| }, |
| { |
| "epoch": 0.49118035278588856, |
| "grad_norm": 0.06117568492897364, |
| "learning_rate": 1.080400211144169e-05, |
| "loss": 0.453688907623291, |
| "memory(GiB)": 76.61, |
| "step": 1535, |
| "token_acc": 0.8130096719135217, |
| "train_speed(iter/s)": 0.029469 |
| }, |
| { |
| "epoch": 0.4927802887884485, |
| "grad_norm": 0.05143211947513191, |
| "learning_rate": 1.0752232458797262e-05, |
| "loss": 0.44568753242492676, |
| "memory(GiB)": 76.61, |
| "step": 1540, |
| "token_acc": 0.847257743677181, |
| "train_speed(iter/s)": 0.029464 |
| }, |
| { |
| "epoch": 0.49438022479100835, |
| "grad_norm": 0.05201971134010435, |
| "learning_rate": 1.070044252278393e-05, |
| "loss": 0.46500363349914553, |
| "memory(GiB)": 76.61, |
| "step": 1545, |
| "token_acc": 0.84, |
| "train_speed(iter/s)": 0.029474 |
| }, |
| { |
| "epoch": 0.49598016079356827, |
| "grad_norm": 0.05304880581645989, |
| "learning_rate": 1.064863369987743e-05, |
| "loss": 0.4501206398010254, |
| "memory(GiB)": 76.61, |
| "step": 1550, |
| "token_acc": 0.8888641920426762, |
| "train_speed(iter/s)": 0.029465 |
| }, |
| { |
| "epoch": 0.49758009679612814, |
| "grad_norm": 0.050584443072610216, |
| "learning_rate": 1.0596807387062772e-05, |
| "loss": 0.456621789932251, |
| "memory(GiB)": 76.61, |
| "step": 1555, |
| "token_acc": 0.8793768317137128, |
| "train_speed(iter/s)": 0.02948 |
| }, |
| { |
| "epoch": 0.49918003279868806, |
| "grad_norm": 0.05907676168100355, |
| "learning_rate": 1.0544964981796563e-05, |
| "loss": 0.4567122936248779, |
| "memory(GiB)": 76.61, |
| "step": 1560, |
| "token_acc": 0.8505747126436781, |
| "train_speed(iter/s)": 0.029477 |
| }, |
| { |
| "epoch": 0.500779968801248, |
| "grad_norm": 0.055037989511506104, |
| "learning_rate": 1.0493107881969335e-05, |
| "loss": 0.44720020294189455, |
| "memory(GiB)": 76.61, |
| "step": 1565, |
| "token_acc": 0.8853304383227032, |
| "train_speed(iter/s)": 0.029466 |
| }, |
| { |
| "epoch": 0.5023799048038079, |
| "grad_norm": 0.0597376748229471, |
| "learning_rate": 1.0441237485867845e-05, |
| "loss": 0.4492997169494629, |
| "memory(GiB)": 76.61, |
| "step": 1570, |
| "token_acc": 0.8809347181008902, |
| "train_speed(iter/s)": 0.029489 |
| }, |
| { |
| "epoch": 0.5039798408063677, |
| "grad_norm": 0.060265741182571844, |
| "learning_rate": 1.0389355192137379e-05, |
| "loss": 0.4525942325592041, |
| "memory(GiB)": 76.61, |
| "step": 1575, |
| "token_acc": 0.8839541547277937, |
| "train_speed(iter/s)": 0.029481 |
| }, |
| { |
| "epoch": 0.5055797768089276, |
| "grad_norm": 0.06015007204584338, |
| "learning_rate": 1.0337462399744025e-05, |
| "loss": 0.4606604099273682, |
| "memory(GiB)": 76.61, |
| "step": 1580, |
| "token_acc": 0.8439696373348328, |
| "train_speed(iter/s)": 0.029471 |
| }, |
| { |
| "epoch": 0.5071797128114875, |
| "grad_norm": 0.0539606724017438, |
| "learning_rate": 1.0285560507936962e-05, |
| "loss": 0.46471481323242186, |
| "memory(GiB)": 76.61, |
| "step": 1585, |
| "token_acc": 0.8212732305258995, |
| "train_speed(iter/s)": 0.029486 |
| }, |
| { |
| "epoch": 0.5087796488140475, |
| "grad_norm": 0.0588254805138369, |
| "learning_rate": 1.0233650916210736e-05, |
| "loss": 0.45154604911804197, |
| "memory(GiB)": 76.61, |
| "step": 1590, |
| "token_acc": 0.883357041251778, |
| "train_speed(iter/s)": 0.029474 |
| }, |
| { |
| "epoch": 0.5103795848166074, |
| "grad_norm": 0.06409304780777438, |
| "learning_rate": 1.0181735024267504e-05, |
| "loss": 0.45000271797180175, |
| "memory(GiB)": 76.61, |
| "step": 1595, |
| "token_acc": 0.8340197693574959, |
| "train_speed(iter/s)": 0.029485 |
| }, |
| { |
| "epoch": 0.5119795208191672, |
| "grad_norm": 0.058200459243944895, |
| "learning_rate": 1.012981423197931e-05, |
| "loss": 0.4608008861541748, |
| "memory(GiB)": 76.61, |
| "step": 1600, |
| "token_acc": 0.8793913904007917, |
| "train_speed(iter/s)": 0.029484 |
| }, |
| { |
| "epoch": 0.5135794568217271, |
| "grad_norm": 0.052541653818392466, |
| "learning_rate": 1.007788993935033e-05, |
| "loss": 0.45448942184448243, |
| "memory(GiB)": 76.61, |
| "step": 1605, |
| "token_acc": 0.8615229110512129, |
| "train_speed(iter/s)": 0.029472 |
| }, |
| { |
| "epoch": 0.515179392824287, |
| "grad_norm": 0.06526426191856917, |
| "learning_rate": 1.002596354647912e-05, |
| "loss": 0.45614914894104003, |
| "memory(GiB)": 76.61, |
| "step": 1610, |
| "token_acc": 0.871312462372065, |
| "train_speed(iter/s)": 0.029489 |
| }, |
| { |
| "epoch": 0.516779328826847, |
| "grad_norm": 0.05743481751682084, |
| "learning_rate": 9.974036453520881e-06, |
| "loss": 0.447450590133667, |
| "memory(GiB)": 76.61, |
| "step": 1615, |
| "token_acc": 0.8760352658295485, |
| "train_speed(iter/s)": 0.02948 |
| }, |
| { |
| "epoch": 0.5183792648294068, |
| "grad_norm": 0.06577378911981274, |
| "learning_rate": 9.922110060649672e-06, |
| "loss": 0.45809640884399416, |
| "memory(GiB)": 76.61, |
| "step": 1620, |
| "token_acc": 0.9048299514146899, |
| "train_speed(iter/s)": 0.029468 |
| }, |
| { |
| "epoch": 0.5199792008319667, |
| "grad_norm": 0.05180626448607971, |
| "learning_rate": 9.870185768020694e-06, |
| "loss": 0.4360641002655029, |
| "memory(GiB)": 76.61, |
| "step": 1625, |
| "token_acc": 0.890797148412184, |
| "train_speed(iter/s)": 0.029481 |
| }, |
| { |
| "epoch": 0.5215791368345266, |
| "grad_norm": 0.048795347699578454, |
| "learning_rate": 9.818264975732497e-06, |
| "loss": 0.4505919933319092, |
| "memory(GiB)": 76.61, |
| "step": 1630, |
| "token_acc": 0.8830155979202773, |
| "train_speed(iter/s)": 0.029462 |
| }, |
| { |
| "epoch": 0.5231790728370865, |
| "grad_norm": 0.054325754690003863, |
| "learning_rate": 9.766349083789266e-06, |
| "loss": 0.4518167495727539, |
| "memory(GiB)": 76.61, |
| "step": 1635, |
| "token_acc": 0.8740636704119851, |
| "train_speed(iter/s)": 0.029458 |
| }, |
| { |
| "epoch": 0.5247790088396465, |
| "grad_norm": 0.05473270131153146, |
| "learning_rate": 9.71443949206304e-06, |
| "loss": 0.4629377841949463, |
| "memory(GiB)": 76.61, |
| "step": 1640, |
| "token_acc": 0.8927648578811369, |
| "train_speed(iter/s)": 0.029469 |
| }, |
| { |
| "epoch": 0.5263789448422063, |
| "grad_norm": 0.05962553624327487, |
| "learning_rate": 9.662537600255979e-06, |
| "loss": 0.4535552501678467, |
| "memory(GiB)": 76.61, |
| "step": 1645, |
| "token_acc": 0.8980960623106881, |
| "train_speed(iter/s)": 0.029457 |
| }, |
| { |
| "epoch": 0.5279788808447662, |
| "grad_norm": 0.06367541172972058, |
| "learning_rate": 9.610644807862625e-06, |
| "loss": 0.44418978691101074, |
| "memory(GiB)": 76.61, |
| "step": 1650, |
| "token_acc": 0.8769371011850501, |
| "train_speed(iter/s)": 0.029468 |
| }, |
| { |
| "epoch": 0.5295788168473261, |
| "grad_norm": 0.05288367644088033, |
| "learning_rate": 9.558762514132157e-06, |
| "loss": 0.4513704299926758, |
| "memory(GiB)": 76.61, |
| "step": 1655, |
| "token_acc": 0.8576478906434126, |
| "train_speed(iter/s)": 0.029464 |
| }, |
| { |
| "epoch": 0.531178752849886, |
| "grad_norm": 0.054919719691940275, |
| "learning_rate": 9.506892118030668e-06, |
| "loss": 0.4454075336456299, |
| "memory(GiB)": 76.61, |
| "step": 1660, |
| "token_acc": 0.8535078688042359, |
| "train_speed(iter/s)": 0.029456 |
| }, |
| { |
| "epoch": 0.532778688852446, |
| "grad_norm": 0.0561391497524031, |
| "learning_rate": 9.455035018203439e-06, |
| "loss": 0.4459484100341797, |
| "memory(GiB)": 76.61, |
| "step": 1665, |
| "token_acc": 0.8793124922157181, |
| "train_speed(iter/s)": 0.029471 |
| }, |
| { |
| "epoch": 0.5343786248550058, |
| "grad_norm": 0.051526152917333715, |
| "learning_rate": 9.40319261293723e-06, |
| "loss": 0.4593966484069824, |
| "memory(GiB)": 76.61, |
| "step": 1670, |
| "token_acc": 0.8957568638966378, |
| "train_speed(iter/s)": 0.029466 |
| }, |
| { |
| "epoch": 0.5359785608575657, |
| "grad_norm": 0.05336577516465571, |
| "learning_rate": 9.351366300122569e-06, |
| "loss": 0.45195541381835935, |
| "memory(GiB)": 76.61, |
| "step": 1675, |
| "token_acc": 0.8254120659305488, |
| "train_speed(iter/s)": 0.029459 |
| }, |
| { |
| "epoch": 0.5375784968601256, |
| "grad_norm": 0.05605931671991975, |
| "learning_rate": 9.299557477216073e-06, |
| "loss": 0.4473400115966797, |
| "memory(GiB)": 76.61, |
| "step": 1680, |
| "token_acc": 0.8684433164128595, |
| "train_speed(iter/s)": 0.029474 |
| }, |
| { |
| "epoch": 0.5391784328626855, |
| "grad_norm": 0.05284398220938548, |
| "learning_rate": 9.247767541202738e-06, |
| "loss": 0.4539934158325195, |
| "memory(GiB)": 76.61, |
| "step": 1685, |
| "token_acc": 0.8787515006002401, |
| "train_speed(iter/s)": 0.029458 |
| }, |
| { |
| "epoch": 0.5407783688652454, |
| "grad_norm": 0.06074778175058899, |
| "learning_rate": 9.195997888558312e-06, |
| "loss": 0.4540121078491211, |
| "memory(GiB)": 76.61, |
| "step": 1690, |
| "token_acc": 0.882076702321941, |
| "train_speed(iter/s)": 0.029458 |
| }, |
| { |
| "epoch": 0.5423783048678052, |
| "grad_norm": 0.06072504661929311, |
| "learning_rate": 9.144249915211605e-06, |
| "loss": 0.45176243782043457, |
| "memory(GiB)": 76.61, |
| "step": 1695, |
| "token_acc": 0.8652057386094908, |
| "train_speed(iter/s)": 0.029465 |
| }, |
| { |
| "epoch": 0.5439782408703652, |
| "grad_norm": 0.058552695609385315, |
| "learning_rate": 9.092525016506858e-06, |
| "loss": 0.4491862773895264, |
| "memory(GiB)": 76.61, |
| "step": 1700, |
| "token_acc": 0.8822588020118884, |
| "train_speed(iter/s)": 0.02945 |
| }, |
| { |
| "epoch": 0.5455781768729251, |
| "grad_norm": 0.056892490634495974, |
| "learning_rate": 9.040824587166136e-06, |
| "loss": 0.45043745040893557, |
| "memory(GiB)": 76.61, |
| "step": 1705, |
| "token_acc": 0.8825789923142613, |
| "train_speed(iter/s)": 0.029461 |
| }, |
| { |
| "epoch": 0.547178112875485, |
| "grad_norm": 0.05885692671807609, |
| "learning_rate": 8.98915002125169e-06, |
| "loss": 0.4475353240966797, |
| "memory(GiB)": 76.61, |
| "step": 1710, |
| "token_acc": 0.8721031538595574, |
| "train_speed(iter/s)": 0.029454 |
| }, |
| { |
| "epoch": 0.5487780488780449, |
| "grad_norm": 0.060276094585736115, |
| "learning_rate": 8.9375027121284e-06, |
| "loss": 0.4502556800842285, |
| "memory(GiB)": 76.61, |
| "step": 1715, |
| "token_acc": 0.8562842259917189, |
| "train_speed(iter/s)": 0.029449 |
| }, |
| { |
| "epoch": 0.5503779848806047, |
| "grad_norm": 0.06782068962590707, |
| "learning_rate": 8.885884052426168e-06, |
| "loss": 0.4532322883605957, |
| "memory(GiB)": 76.61, |
| "step": 1720, |
| "token_acc": 0.8593545573484518, |
| "train_speed(iter/s)": 0.029466 |
| }, |
| { |
| "epoch": 0.5519779208831647, |
| "grad_norm": 0.06070839045848377, |
| "learning_rate": 8.83429543400241e-06, |
| "loss": 0.45258092880249023, |
| "memory(GiB)": 76.61, |
| "step": 1725, |
| "token_acc": 0.8751242791807516, |
| "train_speed(iter/s)": 0.029452 |
| }, |
| { |
| "epoch": 0.5535778568857246, |
| "grad_norm": 0.049979952181739715, |
| "learning_rate": 8.78273824790448e-06, |
| "loss": 0.4340657234191895, |
| "memory(GiB)": 76.61, |
| "step": 1730, |
| "token_acc": 0.8650843222985634, |
| "train_speed(iter/s)": 0.029451 |
| }, |
| { |
| "epoch": 0.5551777928882845, |
| "grad_norm": 0.059124658124222323, |
| "learning_rate": 8.731213884332205e-06, |
| "loss": 0.43556828498840333, |
| "memory(GiB)": 76.61, |
| "step": 1735, |
| "token_acc": 0.8524390243902439, |
| "train_speed(iter/s)": 0.029459 |
| }, |
| { |
| "epoch": 0.5567777288908443, |
| "grad_norm": 0.05228309031135195, |
| "learning_rate": 8.679723732600355e-06, |
| "loss": 0.4483633041381836, |
| "memory(GiB)": 76.61, |
| "step": 1740, |
| "token_acc": 0.9039166284928997, |
| "train_speed(iter/s)": 0.029445 |
| }, |
| { |
| "epoch": 0.5583776648934042, |
| "grad_norm": 0.05659321921396489, |
| "learning_rate": 8.628269181101216e-06, |
| "loss": 0.45377864837646487, |
| "memory(GiB)": 76.61, |
| "step": 1745, |
| "token_acc": 0.8812897628687102, |
| "train_speed(iter/s)": 0.029449 |
| }, |
| { |
| "epoch": 0.5599776008959642, |
| "grad_norm": 0.0610469666222746, |
| "learning_rate": 8.576851617267151e-06, |
| "loss": 0.4495216369628906, |
| "memory(GiB)": 76.61, |
| "step": 1750, |
| "token_acc": 0.8734145104008117, |
| "train_speed(iter/s)": 0.029452 |
| }, |
| { |
| "epoch": 0.5599776008959642, |
| "eval_loss": 0.6640093922615051, |
| "eval_runtime": 114.9985, |
| "eval_samples_per_second": 174.681, |
| "eval_steps_per_second": 0.878, |
| "eval_token_acc": 0.8701128116616424, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.5615775368985241, |
| "grad_norm": 0.061306770256929585, |
| "learning_rate": 8.525472427533156e-06, |
| "loss": 0.44908857345581055, |
| "memory(GiB)": 77.63, |
| "step": 1755, |
| "token_acc": 0.8715457946180765, |
| "train_speed(iter/s)": 0.029409 |
| }, |
| { |
| "epoch": 0.563177472901084, |
| "grad_norm": 0.05235639827008535, |
| "learning_rate": 8.474132997299521e-06, |
| "loss": 0.4579316139221191, |
| "memory(GiB)": 77.63, |
| "step": 1760, |
| "token_acc": 0.8922923256201098, |
| "train_speed(iter/s)": 0.029422 |
| }, |
| { |
| "epoch": 0.5647774089036438, |
| "grad_norm": 0.051281007426132216, |
| "learning_rate": 8.422834710894434e-06, |
| "loss": 0.45467004776000974, |
| "memory(GiB)": 77.63, |
| "step": 1765, |
| "token_acc": 0.903878366189924, |
| "train_speed(iter/s)": 0.029438 |
| }, |
| { |
| "epoch": 0.5663773449062037, |
| "grad_norm": 0.05049109520782513, |
| "learning_rate": 8.371578951536689e-06, |
| "loss": 0.45294957160949706, |
| "memory(GiB)": 77.63, |
| "step": 1770, |
| "token_acc": 0.8928110202324581, |
| "train_speed(iter/s)": 0.029439 |
| }, |
| { |
| "epoch": 0.5679772809087636, |
| "grad_norm": 0.04946427707777728, |
| "learning_rate": 8.320367101298351e-06, |
| "loss": 0.4473431587219238, |
| "memory(GiB)": 77.63, |
| "step": 1775, |
| "token_acc": 0.8723599632690542, |
| "train_speed(iter/s)": 0.029439 |
| }, |
| { |
| "epoch": 0.5695772169113236, |
| "grad_norm": 0.053606352244487274, |
| "learning_rate": 8.26920054106753e-06, |
| "loss": 0.4495864391326904, |
| "memory(GiB)": 77.63, |
| "step": 1780, |
| "token_acc": 0.8844315111203492, |
| "train_speed(iter/s)": 0.029459 |
| }, |
| { |
| "epoch": 0.5711771529138835, |
| "grad_norm": 0.05525614374940963, |
| "learning_rate": 8.218080650511107e-06, |
| "loss": 0.44890499114990234, |
| "memory(GiB)": 77.63, |
| "step": 1785, |
| "token_acc": 0.8749736453721273, |
| "train_speed(iter/s)": 0.02946 |
| }, |
| { |
| "epoch": 0.5727770889164433, |
| "grad_norm": 0.05882148265537131, |
| "learning_rate": 8.167008808037568e-06, |
| "loss": 0.44676194190979, |
| "memory(GiB)": 77.63, |
| "step": 1790, |
| "token_acc": 0.8807511737089202, |
| "train_speed(iter/s)": 0.029457 |
| }, |
| { |
| "epoch": 0.5743770249190032, |
| "grad_norm": 0.048821121641334515, |
| "learning_rate": 8.115986390759805e-06, |
| "loss": 0.4417415142059326, |
| "memory(GiB)": 77.63, |
| "step": 1795, |
| "token_acc": 0.8531673379714391, |
| "train_speed(iter/s)": 0.029469 |
| }, |
| { |
| "epoch": 0.5759769609215631, |
| "grad_norm": 0.054949264031140505, |
| "learning_rate": 8.065014774458004e-06, |
| "loss": 0.46439437866210936, |
| "memory(GiB)": 77.63, |
| "step": 1800, |
| "token_acc": 0.8333022213925705, |
| "train_speed(iter/s)": 0.029467 |
| }, |
| { |
| "epoch": 0.5775768969241231, |
| "grad_norm": 0.059507220518762304, |
| "learning_rate": 8.014095333542548e-06, |
| "loss": 0.4539642333984375, |
| "memory(GiB)": 77.63, |
| "step": 1805, |
| "token_acc": 0.8577178858942948, |
| "train_speed(iter/s)": 0.029483 |
| }, |
| { |
| "epoch": 0.579176832926683, |
| "grad_norm": 0.05302143027350534, |
| "learning_rate": 7.963229441016938e-06, |
| "loss": 0.4606470108032227, |
| "memory(GiB)": 77.63, |
| "step": 1810, |
| "token_acc": 0.8760775862068966, |
| "train_speed(iter/s)": 0.029484 |
| }, |
| { |
| "epoch": 0.5807767689292428, |
| "grad_norm": 0.0699581228289572, |
| "learning_rate": 7.912418468440794e-06, |
| "loss": 0.4488551139831543, |
| "memory(GiB)": 77.63, |
| "step": 1815, |
| "token_acc": 0.8892276422764228, |
| "train_speed(iter/s)": 0.029481 |
| }, |
| { |
| "epoch": 0.5823767049318027, |
| "grad_norm": 0.053456667148895104, |
| "learning_rate": 7.861663785892857e-06, |
| "loss": 0.45035881996154786, |
| "memory(GiB)": 77.63, |
| "step": 1820, |
| "token_acc": 0.8806643202815662, |
| "train_speed(iter/s)": 0.029498 |
| }, |
| { |
| "epoch": 0.5839766409343626, |
| "grad_norm": 0.05451209338463787, |
| "learning_rate": 7.810966761934053e-06, |
| "loss": 0.44800753593444825, |
| "memory(GiB)": 77.63, |
| "step": 1825, |
| "token_acc": 0.8771571298819255, |
| "train_speed(iter/s)": 0.029496 |
| }, |
| { |
| "epoch": 0.5855765769369226, |
| "grad_norm": 0.05912934985203241, |
| "learning_rate": 7.760328763570589e-06, |
| "loss": 0.4499057769775391, |
| "memory(GiB)": 77.63, |
| "step": 1830, |
| "token_acc": 0.868710326675956, |
| "train_speed(iter/s)": 0.029487 |
| }, |
| { |
| "epoch": 0.5871765129394825, |
| "grad_norm": 0.052841905445767515, |
| "learning_rate": 7.709751156217088e-06, |
| "loss": 0.4497323989868164, |
| "memory(GiB)": 77.63, |
| "step": 1835, |
| "token_acc": 0.8117094325984822, |
| "train_speed(iter/s)": 0.029501 |
| }, |
| { |
| "epoch": 0.5887764489420423, |
| "grad_norm": 0.060076953422732254, |
| "learning_rate": 7.659235303659784e-06, |
| "loss": 0.4582187652587891, |
| "memory(GiB)": 77.63, |
| "step": 1840, |
| "token_acc": 0.8795674258561363, |
| "train_speed(iter/s)": 0.029492 |
| }, |
| { |
| "epoch": 0.5903763849446022, |
| "grad_norm": 0.06307528499562465, |
| "learning_rate": 7.608782568019729e-06, |
| "loss": 0.4430552005767822, |
| "memory(GiB)": 77.63, |
| "step": 1845, |
| "token_acc": 0.8452444922084901, |
| "train_speed(iter/s)": 0.029498 |
| }, |
| { |
| "epoch": 0.5919763209471621, |
| "grad_norm": 0.05378691938628143, |
| "learning_rate": 7.558394309716088e-06, |
| "loss": 0.459810209274292, |
| "memory(GiB)": 77.63, |
| "step": 1850, |
| "token_acc": 0.8506092736192435, |
| "train_speed(iter/s)": 0.029503 |
| }, |
| { |
| "epoch": 0.593576256949722, |
| "grad_norm": 0.0586506530143339, |
| "learning_rate": 7.508071887429433e-06, |
| "loss": 0.46239190101623534, |
| "memory(GiB)": 77.63, |
| "step": 1855, |
| "token_acc": 0.9115304709141274, |
| "train_speed(iter/s)": 0.029495 |
| }, |
| { |
| "epoch": 0.5951761929522819, |
| "grad_norm": 0.053290473896441634, |
| "learning_rate": 7.4578166580651335e-06, |
| "loss": 0.4524221897125244, |
| "memory(GiB)": 77.63, |
| "step": 1860, |
| "token_acc": 0.8817879571481345, |
| "train_speed(iter/s)": 0.029508 |
| }, |
| { |
| "epoch": 0.5967761289548418, |
| "grad_norm": 0.051901913358510056, |
| "learning_rate": 7.4076299767167325e-06, |
| "loss": 0.4579325675964355, |
| "memory(GiB)": 77.63, |
| "step": 1865, |
| "token_acc": 0.8617401668653158, |
| "train_speed(iter/s)": 0.029506 |
| }, |
| { |
| "epoch": 0.5983760649574017, |
| "grad_norm": 0.05256077511072294, |
| "learning_rate": 7.35751319662945e-06, |
| "loss": 0.45406513214111327, |
| "memory(GiB)": 77.63, |
| "step": 1870, |
| "token_acc": 0.8924402944873406, |
| "train_speed(iter/s)": 0.029497 |
| }, |
| { |
| "epoch": 0.5999760009599616, |
| "grad_norm": 0.056121622843709036, |
| "learning_rate": 7.307467669163655e-06, |
| "loss": 0.450104284286499, |
| "memory(GiB)": 77.63, |
| "step": 1875, |
| "token_acc": 0.8646184340931615, |
| "train_speed(iter/s)": 0.02951 |
| }, |
| { |
| "epoch": 0.6015759369625215, |
| "grad_norm": 0.051068951060234354, |
| "learning_rate": 7.25749474375846e-06, |
| "loss": 0.45695791244506834, |
| "memory(GiB)": 77.63, |
| "step": 1880, |
| "token_acc": 0.9112655568126717, |
| "train_speed(iter/s)": 0.029503 |
| }, |
| { |
| "epoch": 0.6031758729650813, |
| "grad_norm": 0.05120698584703106, |
| "learning_rate": 7.207595767895303e-06, |
| "loss": 0.4460740089416504, |
| "memory(GiB)": 77.63, |
| "step": 1885, |
| "token_acc": 0.8637192342752963, |
| "train_speed(iter/s)": 0.029499 |
| }, |
| { |
| "epoch": 0.6047758089676413, |
| "grad_norm": 0.05826366701259215, |
| "learning_rate": 7.157772087061645e-06, |
| "loss": 0.4498391628265381, |
| "memory(GiB)": 77.63, |
| "step": 1890, |
| "token_acc": 0.8602477214302408, |
| "train_speed(iter/s)": 0.029509 |
| }, |
| { |
| "epoch": 0.6063757449702012, |
| "grad_norm": 0.05454678875604061, |
| "learning_rate": 7.108025044714661e-06, |
| "loss": 0.44768247604370115, |
| "memory(GiB)": 77.63, |
| "step": 1895, |
| "token_acc": 0.8998014357721094, |
| "train_speed(iter/s)": 0.029496 |
| }, |
| { |
| "epoch": 0.6079756809727611, |
| "grad_norm": 0.04862560763785379, |
| "learning_rate": 7.058355982245038e-06, |
| "loss": 0.44283151626586914, |
| "memory(GiB)": 77.63, |
| "step": 1900, |
| "token_acc": 0.8749580958766343, |
| "train_speed(iter/s)": 0.029501 |
| }, |
| { |
| "epoch": 0.609575616975321, |
| "grad_norm": 0.05390239428952395, |
| "learning_rate": 7.00876623894079e-06, |
| "loss": 0.4445077419281006, |
| "memory(GiB)": 77.63, |
| "step": 1905, |
| "token_acc": 0.8588156123822341, |
| "train_speed(iter/s)": 0.029504 |
| }, |
| { |
| "epoch": 0.6111755529778808, |
| "grad_norm": 0.052917745372876655, |
| "learning_rate": 6.959257151951153e-06, |
| "loss": 0.45001955032348634, |
| "memory(GiB)": 77.63, |
| "step": 1910, |
| "token_acc": 0.8768155911013054, |
| "train_speed(iter/s)": 0.029494 |
| }, |
| { |
| "epoch": 0.6127754889804408, |
| "grad_norm": 0.05432256049056495, |
| "learning_rate": 6.909830056250527e-06, |
| "loss": 0.44944238662719727, |
| "memory(GiB)": 77.63, |
| "step": 1915, |
| "token_acc": 0.8941244909831297, |
| "train_speed(iter/s)": 0.029507 |
| }, |
| { |
| "epoch": 0.6143754249830007, |
| "grad_norm": 0.05852297407331436, |
| "learning_rate": 6.860486284602479e-06, |
| "loss": 0.4477729797363281, |
| "memory(GiB)": 77.63, |
| "step": 1920, |
| "token_acc": 0.8854845719252499, |
| "train_speed(iter/s)": 0.029501 |
| }, |
| { |
| "epoch": 0.6159753609855606, |
| "grad_norm": 0.05474007805899836, |
| "learning_rate": 6.8112271675238154e-06, |
| "loss": 0.4501204013824463, |
| "memory(GiB)": 77.63, |
| "step": 1925, |
| "token_acc": 0.8803290949887809, |
| "train_speed(iter/s)": 0.029496 |
| }, |
| { |
| "epoch": 0.6175752969881205, |
| "grad_norm": 0.05545012433641634, |
| "learning_rate": 6.762054033248681e-06, |
| "loss": 0.44565958976745607, |
| "memory(GiB)": 77.63, |
| "step": 1930, |
| "token_acc": 0.8480542195015304, |
| "train_speed(iter/s)": 0.029507 |
| }, |
| { |
| "epoch": 0.6191752329906803, |
| "grad_norm": 0.05495247298953925, |
| "learning_rate": 6.712968207692778e-06, |
| "loss": 0.44170804023742677, |
| "memory(GiB)": 77.63, |
| "step": 1935, |
| "token_acc": 0.8709073900841908, |
| "train_speed(iter/s)": 0.029498 |
| }, |
| { |
| "epoch": 0.6207751689932403, |
| "grad_norm": 0.05792014047889592, |
| "learning_rate": 6.663971014417585e-06, |
| "loss": 0.4454016208648682, |
| "memory(GiB)": 77.63, |
| "step": 1940, |
| "token_acc": 0.8606651376146789, |
| "train_speed(iter/s)": 0.0295 |
| }, |
| { |
| "epoch": 0.6223751049958002, |
| "grad_norm": 0.04853659131630362, |
| "learning_rate": 6.615063774594677e-06, |
| "loss": 0.4387532711029053, |
| "memory(GiB)": 77.63, |
| "step": 1945, |
| "token_acc": 0.8920454545454546, |
| "train_speed(iter/s)": 0.029507 |
| }, |
| { |
| "epoch": 0.6239750409983601, |
| "grad_norm": 0.05266495974136303, |
| "learning_rate": 6.566247806970119e-06, |
| "loss": 0.4472493171691895, |
| "memory(GiB)": 77.63, |
| "step": 1950, |
| "token_acc": 0.848505251817937, |
| "train_speed(iter/s)": 0.029497 |
| }, |
| { |
| "epoch": 0.62557497700092, |
| "grad_norm": 0.054994759694813, |
| "learning_rate": 6.5175244278288705e-06, |
| "loss": 0.44487895965576174, |
| "memory(GiB)": 77.63, |
| "step": 1955, |
| "token_acc": 0.8689320388349514, |
| "train_speed(iter/s)": 0.029506 |
| }, |
| { |
| "epoch": 0.6271749130034798, |
| "grad_norm": 0.057067387083368365, |
| "learning_rate": 6.468894950959336e-06, |
| "loss": 0.4466127395629883, |
| "memory(GiB)": 77.63, |
| "step": 1960, |
| "token_acc": 0.846737755286463, |
| "train_speed(iter/s)": 0.0295 |
| }, |
| { |
| "epoch": 0.6287748490060397, |
| "grad_norm": 0.05417940634204734, |
| "learning_rate": 6.420360687617897e-06, |
| "loss": 0.44883151054382325, |
| "memory(GiB)": 77.63, |
| "step": 1965, |
| "token_acc": 0.8795967892477132, |
| "train_speed(iter/s)": 0.02949 |
| }, |
| { |
| "epoch": 0.6303747850085997, |
| "grad_norm": 0.05176488752108695, |
| "learning_rate": 6.3719229464935915e-06, |
| "loss": 0.4542849063873291, |
| "memory(GiB)": 77.63, |
| "step": 1970, |
| "token_acc": 0.8568893191352049, |
| "train_speed(iter/s)": 0.029503 |
| }, |
| { |
| "epoch": 0.6319747210111596, |
| "grad_norm": 0.04867135924369273, |
| "learning_rate": 6.323583033672799e-06, |
| "loss": 0.44331774711608884, |
| "memory(GiB)": 77.63, |
| "step": 1975, |
| "token_acc": 0.8647865559204172, |
| "train_speed(iter/s)": 0.029491 |
| }, |
| { |
| "epoch": 0.6335746570137194, |
| "grad_norm": 0.06076783884358601, |
| "learning_rate": 6.275342252604044e-06, |
| "loss": 0.44751858711242676, |
| "memory(GiB)": 77.63, |
| "step": 1980, |
| "token_acc": 0.871765773944621, |
| "train_speed(iter/s)": 0.029486 |
| }, |
| { |
| "epoch": 0.6351745930162793, |
| "grad_norm": 0.0520886449098567, |
| "learning_rate": 6.22720190406283e-06, |
| "loss": 0.46150927543640136, |
| "memory(GiB)": 77.63, |
| "step": 1985, |
| "token_acc": 0.8921661480178595, |
| "train_speed(iter/s)": 0.029497 |
| }, |
| { |
| "epoch": 0.6367745290188392, |
| "grad_norm": 0.058090405193780774, |
| "learning_rate": 6.179163286116581e-06, |
| "loss": 0.44019436836242676, |
| "memory(GiB)": 77.63, |
| "step": 1990, |
| "token_acc": 0.9157033805888768, |
| "train_speed(iter/s)": 0.029488 |
| }, |
| { |
| "epoch": 0.6383744650213992, |
| "grad_norm": 0.057472120727550105, |
| "learning_rate": 6.13122769408963e-06, |
| "loss": 0.4466409683227539, |
| "memory(GiB)": 77.63, |
| "step": 1995, |
| "token_acc": 0.8608313968499871, |
| "train_speed(iter/s)": 0.029492 |
| }, |
| { |
| "epoch": 0.6399744010239591, |
| "grad_norm": 0.05665485079826101, |
| "learning_rate": 6.083396420528298e-06, |
| "loss": 0.451153039932251, |
| "memory(GiB)": 77.63, |
| "step": 2000, |
| "token_acc": 0.8910367046369808, |
| "train_speed(iter/s)": 0.029496 |
| }, |
| { |
| "epoch": 0.6399744010239591, |
| "eval_loss": 0.6622327566146851, |
| "eval_runtime": 115.9166, |
| "eval_samples_per_second": 173.297, |
| "eval_steps_per_second": 0.871, |
| "eval_token_acc": 0.8709580958089251, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.6415743370265189, |
| "grad_norm": 0.05247711719910443, |
| "learning_rate": 6.0356707551660434e-06, |
| "loss": 0.45055346488952636, |
| "memory(GiB)": 77.63, |
| "step": 2005, |
| "token_acc": 0.8961562482257424, |
| "train_speed(iter/s)": 0.029453 |
| }, |
| { |
| "epoch": 0.6431742730290788, |
| "grad_norm": 0.0533769803289562, |
| "learning_rate": 5.988051984888668e-06, |
| "loss": 0.4436792373657227, |
| "memory(GiB)": 77.63, |
| "step": 2010, |
| "token_acc": 0.8894836272040302, |
| "train_speed(iter/s)": 0.029463 |
| }, |
| { |
| "epoch": 0.6447742090316387, |
| "grad_norm": 0.05428299581856707, |
| "learning_rate": 5.940541393699646e-06, |
| "loss": 0.44562363624572754, |
| "memory(GiB)": 77.63, |
| "step": 2015, |
| "token_acc": 0.8804031789106416, |
| "train_speed(iter/s)": 0.029477 |
| }, |
| { |
| "epoch": 0.6463741450341987, |
| "grad_norm": 0.055930584071511934, |
| "learning_rate": 5.893140262685469e-06, |
| "loss": 0.4412201404571533, |
| "memory(GiB)": 77.63, |
| "step": 2020, |
| "token_acc": 0.8791348600508906, |
| "train_speed(iter/s)": 0.029482 |
| }, |
| { |
| "epoch": 0.6479740810367586, |
| "grad_norm": 0.06077731970466293, |
| "learning_rate": 5.845849869981137e-06, |
| "loss": 0.44964237213134767, |
| "memory(GiB)": 77.63, |
| "step": 2025, |
| "token_acc": 0.8710450018908358, |
| "train_speed(iter/s)": 0.029484 |
| }, |
| { |
| "epoch": 0.6495740170393184, |
| "grad_norm": 0.05824848510516177, |
| "learning_rate": 5.7986714907356614e-06, |
| "loss": 0.4586543083190918, |
| "memory(GiB)": 77.63, |
| "step": 2030, |
| "token_acc": 0.8852591792656588, |
| "train_speed(iter/s)": 0.029498 |
| }, |
| { |
| "epoch": 0.6511739530418783, |
| "grad_norm": 0.06066761562869553, |
| "learning_rate": 5.751606397077703e-06, |
| "loss": 0.44632205963134763, |
| "memory(GiB)": 77.63, |
| "step": 2035, |
| "token_acc": 0.8871352785145888, |
| "train_speed(iter/s)": 0.029494 |
| }, |
| { |
| "epoch": 0.6527738890444382, |
| "grad_norm": 0.055201144436432543, |
| "learning_rate": 5.704655858081268e-06, |
| "loss": 0.43164916038513185, |
| "memory(GiB)": 77.63, |
| "step": 2040, |
| "token_acc": 0.8937977909940527, |
| "train_speed(iter/s)": 0.029496 |
| }, |
| { |
| "epoch": 0.6543738250469981, |
| "grad_norm": 0.05987076771844116, |
| "learning_rate": 5.6578211397314765e-06, |
| "loss": 0.4560856819152832, |
| "memory(GiB)": 77.63, |
| "step": 2045, |
| "token_acc": 0.8462914545204349, |
| "train_speed(iter/s)": 0.029507 |
| }, |
| { |
| "epoch": 0.6559737610495581, |
| "grad_norm": 0.05549604363093839, |
| "learning_rate": 5.611103504890444e-06, |
| "loss": 0.44809746742248535, |
| "memory(GiB)": 77.63, |
| "step": 2050, |
| "token_acc": 0.8783254279232832, |
| "train_speed(iter/s)": 0.029503 |
| }, |
| { |
| "epoch": 0.6575736970521179, |
| "grad_norm": 0.060605872447174955, |
| "learning_rate": 5.564504213263205e-06, |
| "loss": 0.43492536544799804, |
| "memory(GiB)": 77.63, |
| "step": 2055, |
| "token_acc": 0.8383036405886909, |
| "train_speed(iter/s)": 0.029515 |
| }, |
| { |
| "epoch": 0.6591736330546778, |
| "grad_norm": 0.05003885513998493, |
| "learning_rate": 5.5180245213637785e-06, |
| "loss": 0.44741315841674806, |
| "memory(GiB)": 77.63, |
| "step": 2060, |
| "token_acc": 0.8784857874174862, |
| "train_speed(iter/s)": 0.029515 |
| }, |
| { |
| "epoch": 0.6607735690572377, |
| "grad_norm": 0.053248832099036005, |
| "learning_rate": 5.4716656824812505e-06, |
| "loss": 0.4469279766082764, |
| "memory(GiB)": 77.63, |
| "step": 2065, |
| "token_acc": 0.8853107344632768, |
| "train_speed(iter/s)": 0.029507 |
| }, |
| { |
| "epoch": 0.6623735050597976, |
| "grad_norm": 0.050937098304756526, |
| "learning_rate": 5.425428946646016e-06, |
| "loss": 0.44948582649230956, |
| "memory(GiB)": 77.63, |
| "step": 2070, |
| "token_acc": 0.8934210526315789, |
| "train_speed(iter/s)": 0.029526 |
| }, |
| { |
| "epoch": 0.6639734410623576, |
| "grad_norm": 0.050018563981163396, |
| "learning_rate": 5.379315560596038e-06, |
| "loss": 0.4475410461425781, |
| "memory(GiB)": 77.63, |
| "step": 2075, |
| "token_acc": 0.8478816857555876, |
| "train_speed(iter/s)": 0.029518 |
| }, |
| { |
| "epoch": 0.6655733770649174, |
| "grad_norm": 0.05745892430696422, |
| "learning_rate": 5.333326767743263e-06, |
| "loss": 0.45008225440979005, |
| "memory(GiB)": 77.63, |
| "step": 2080, |
| "token_acc": 0.8264099454214675, |
| "train_speed(iter/s)": 0.029515 |
| }, |
| { |
| "epoch": 0.6671733130674773, |
| "grad_norm": 0.05536507134607956, |
| "learning_rate": 5.287463808140069e-06, |
| "loss": 0.4393789291381836, |
| "memory(GiB)": 77.63, |
| "step": 2085, |
| "token_acc": 0.8450008816787162, |
| "train_speed(iter/s)": 0.029527 |
| }, |
| { |
| "epoch": 0.6687732490700372, |
| "grad_norm": 0.06142641017026178, |
| "learning_rate": 5.241727918445836e-06, |
| "loss": 0.4437687873840332, |
| "memory(GiB)": 77.63, |
| "step": 2090, |
| "token_acc": 0.8837277242185217, |
| "train_speed(iter/s)": 0.02952 |
| }, |
| { |
| "epoch": 0.6703731850725971, |
| "grad_norm": 0.05426196603270913, |
| "learning_rate": 5.1961203318936116e-06, |
| "loss": 0.4427367687225342, |
| "memory(GiB)": 77.63, |
| "step": 2095, |
| "token_acc": 0.856048805815161, |
| "train_speed(iter/s)": 0.029524 |
| }, |
| { |
| "epoch": 0.6719731210751569, |
| "grad_norm": 0.054169398190345976, |
| "learning_rate": 5.1506422782568345e-06, |
| "loss": 0.4520686626434326, |
| "memory(GiB)": 77.63, |
| "step": 2100, |
| "token_acc": 0.8747133027522935, |
| "train_speed(iter/s)": 0.029529 |
| }, |
| { |
| "epoch": 0.6735730570777169, |
| "grad_norm": 0.054436537230257924, |
| "learning_rate": 5.105294983816203e-06, |
| "loss": 0.44482645988464353, |
| "memory(GiB)": 77.63, |
| "step": 2105, |
| "token_acc": 0.8637377049180328, |
| "train_speed(iter/s)": 0.029517 |
| }, |
| { |
| "epoch": 0.6751729930802768, |
| "grad_norm": 0.05860088154390529, |
| "learning_rate": 5.060079671326577e-06, |
| "loss": 0.44719686508178713, |
| "memory(GiB)": 77.63, |
| "step": 2110, |
| "token_acc": 0.8593150866058442, |
| "train_speed(iter/s)": 0.029529 |
| }, |
| { |
| "epoch": 0.6767729290828367, |
| "grad_norm": 0.05264024149284518, |
| "learning_rate": 5.014997559984045e-06, |
| "loss": 0.43972039222717285, |
| "memory(GiB)": 77.63, |
| "step": 2115, |
| "token_acc": 0.8533221194280909, |
| "train_speed(iter/s)": 0.029526 |
| }, |
| { |
| "epoch": 0.6783728650853966, |
| "grad_norm": 0.0534652970629265, |
| "learning_rate": 4.970049865393009e-06, |
| "loss": 0.4468375205993652, |
| "memory(GiB)": 77.63, |
| "step": 2120, |
| "token_acc": 0.8628782287822878, |
| "train_speed(iter/s)": 0.029518 |
| }, |
| { |
| "epoch": 0.6799728010879564, |
| "grad_norm": 0.05246927821047006, |
| "learning_rate": 4.925237799533445e-06, |
| "loss": 0.4498266696929932, |
| "memory(GiB)": 77.63, |
| "step": 2125, |
| "token_acc": 0.9048205760049284, |
| "train_speed(iter/s)": 0.029533 |
| }, |
| { |
| "epoch": 0.6815727370905164, |
| "grad_norm": 0.05104237083350841, |
| "learning_rate": 4.880562570728188e-06, |
| "loss": 0.4389338970184326, |
| "memory(GiB)": 77.63, |
| "step": 2130, |
| "token_acc": 0.8844444444444445, |
| "train_speed(iter/s)": 0.029525 |
| }, |
| { |
| "epoch": 0.6831726730930763, |
| "grad_norm": 0.05297787940328326, |
| "learning_rate": 4.836025383610382e-06, |
| "loss": 0.4495584487915039, |
| "memory(GiB)": 77.63, |
| "step": 2135, |
| "token_acc": 0.8647426233038984, |
| "train_speed(iter/s)": 0.029524 |
| }, |
| { |
| "epoch": 0.6847726090956362, |
| "grad_norm": 0.05092547333787791, |
| "learning_rate": 4.791627439090975e-06, |
| "loss": 0.4421692848205566, |
| "memory(GiB)": 77.63, |
| "step": 2140, |
| "token_acc": 0.8828041384231181, |
| "train_speed(iter/s)": 0.029534 |
| }, |
| { |
| "epoch": 0.6863725450981961, |
| "grad_norm": 0.053418572817851825, |
| "learning_rate": 4.74736993432634e-06, |
| "loss": 0.44208922386169436, |
| "memory(GiB)": 77.63, |
| "step": 2145, |
| "token_acc": 0.8888520238885202, |
| "train_speed(iter/s)": 0.029522 |
| }, |
| { |
| "epoch": 0.6879724811007559, |
| "grad_norm": 0.053760421496406786, |
| "learning_rate": 4.703254062686017e-06, |
| "loss": 0.4469425201416016, |
| "memory(GiB)": 77.63, |
| "step": 2150, |
| "token_acc": 0.8764145324597975, |
| "train_speed(iter/s)": 0.02953 |
| }, |
| { |
| "epoch": 0.6895724171033158, |
| "grad_norm": 0.05280613332203591, |
| "learning_rate": 4.6592810137205e-06, |
| "loss": 0.45023741722106936, |
| "memory(GiB)": 77.63, |
| "step": 2155, |
| "token_acc": 0.8968010517090271, |
| "train_speed(iter/s)": 0.029531 |
| }, |
| { |
| "epoch": 0.6911723531058758, |
| "grad_norm": 0.05438679052798784, |
| "learning_rate": 4.615451973129196e-06, |
| "loss": 0.4470167636871338, |
| "memory(GiB)": 77.63, |
| "step": 2160, |
| "token_acc": 0.8761696818465378, |
| "train_speed(iter/s)": 0.02952 |
| }, |
| { |
| "epoch": 0.6927722891084357, |
| "grad_norm": 0.056426544102266905, |
| "learning_rate": 4.571768122728421e-06, |
| "loss": 0.4486443042755127, |
| "memory(GiB)": 77.63, |
| "step": 2165, |
| "token_acc": 0.8781996587030717, |
| "train_speed(iter/s)": 0.029533 |
| }, |
| { |
| "epoch": 0.6943722251109956, |
| "grad_norm": 0.05337656902490804, |
| "learning_rate": 4.528230640419562e-06, |
| "loss": 0.4497722625732422, |
| "memory(GiB)": 77.63, |
| "step": 2170, |
| "token_acc": 0.8571428571428571, |
| "train_speed(iter/s)": 0.029522 |
| }, |
| { |
| "epoch": 0.6959721611135554, |
| "grad_norm": 0.054129658100736014, |
| "learning_rate": 4.4848407001572945e-06, |
| "loss": 0.44121665954589845, |
| "memory(GiB)": 77.63, |
| "step": 2175, |
| "token_acc": 0.8674278464954012, |
| "train_speed(iter/s)": 0.029518 |
| }, |
| { |
| "epoch": 0.6975720971161153, |
| "grad_norm": 0.05332136363084243, |
| "learning_rate": 4.441599471917946e-06, |
| "loss": 0.43872866630554197, |
| "memory(GiB)": 77.63, |
| "step": 2180, |
| "token_acc": 0.8575067664384652, |
| "train_speed(iter/s)": 0.029529 |
| }, |
| { |
| "epoch": 0.6991720331186753, |
| "grad_norm": 0.06093731322456081, |
| "learning_rate": 4.398508121667925e-06, |
| "loss": 0.42902402877807616, |
| "memory(GiB)": 77.63, |
| "step": 2185, |
| "token_acc": 0.8526187576126675, |
| "train_speed(iter/s)": 0.029519 |
| }, |
| { |
| "epoch": 0.7007719691212352, |
| "grad_norm": 0.05959591977220614, |
| "learning_rate": 4.355567811332311e-06, |
| "loss": 0.44504075050354003, |
| "memory(GiB)": 77.63, |
| "step": 2190, |
| "token_acc": 0.8604511878618487, |
| "train_speed(iter/s)": 0.029519 |
| }, |
| { |
| "epoch": 0.702371905123795, |
| "grad_norm": 0.052994813328955795, |
| "learning_rate": 4.312779698763493e-06, |
| "loss": 0.4408130168914795, |
| "memory(GiB)": 77.63, |
| "step": 2195, |
| "token_acc": 0.8787728847105394, |
| "train_speed(iter/s)": 0.029526 |
| }, |
| { |
| "epoch": 0.7039718411263549, |
| "grad_norm": 0.05402320661185779, |
| "learning_rate": 4.270144937709981e-06, |
| "loss": 0.4396658897399902, |
| "memory(GiB)": 77.63, |
| "step": 2200, |
| "token_acc": 0.8677248677248677, |
| "train_speed(iter/s)": 0.029515 |
| }, |
| { |
| "epoch": 0.7055717771289148, |
| "grad_norm": 0.053270864469091045, |
| "learning_rate": 4.227664677785264e-06, |
| "loss": 0.4493250846862793, |
| "memory(GiB)": 77.63, |
| "step": 2205, |
| "token_acc": 0.8801270542742715, |
| "train_speed(iter/s)": 0.029523 |
| }, |
| { |
| "epoch": 0.7071717131314748, |
| "grad_norm": 0.06248819511930574, |
| "learning_rate": 4.1853400644368395e-06, |
| "loss": 0.44740095138549807, |
| "memory(GiB)": 77.63, |
| "step": 2210, |
| "token_acc": 0.8988747408942849, |
| "train_speed(iter/s)": 0.029514 |
| }, |
| { |
| "epoch": 0.7087716491340347, |
| "grad_norm": 0.05807018681919018, |
| "learning_rate": 4.143172238915302e-06, |
| "loss": 0.4508991241455078, |
| "memory(GiB)": 77.63, |
| "step": 2215, |
| "token_acc": 0.8739803562510404, |
| "train_speed(iter/s)": 0.029506 |
| }, |
| { |
| "epoch": 0.7103715851365945, |
| "grad_norm": 0.05376168037907147, |
| "learning_rate": 4.101162338243595e-06, |
| "loss": 0.4486696243286133, |
| "memory(GiB)": 77.63, |
| "step": 2220, |
| "token_acc": 0.8750778169744761, |
| "train_speed(iter/s)": 0.029514 |
| }, |
| { |
| "epoch": 0.7119715211391544, |
| "grad_norm": 0.05460944158847028, |
| "learning_rate": 4.059311495186338e-06, |
| "loss": 0.4484865188598633, |
| "memory(GiB)": 77.63, |
| "step": 2225, |
| "token_acc": 0.8524350054924936, |
| "train_speed(iter/s)": 0.029506 |
| }, |
| { |
| "epoch": 0.7135714571417143, |
| "grad_norm": 0.05716955035585288, |
| "learning_rate": 4.017620838219276e-06, |
| "loss": 0.44258599281311034, |
| "memory(GiB)": 77.63, |
| "step": 2230, |
| "token_acc": 0.8597191629955947, |
| "train_speed(iter/s)": 0.029504 |
| }, |
| { |
| "epoch": 0.7151713931442742, |
| "grad_norm": 0.05984913995816041, |
| "learning_rate": 3.9760914914988716e-06, |
| "loss": 0.4547589778900146, |
| "memory(GiB)": 77.63, |
| "step": 2235, |
| "token_acc": 0.8679617117117117, |
| "train_speed(iter/s)": 0.029511 |
| }, |
| { |
| "epoch": 0.7167713291468342, |
| "grad_norm": 0.05686589162715874, |
| "learning_rate": 3.93472457483197e-06, |
| "loss": 0.4416301727294922, |
| "memory(GiB)": 77.63, |
| "step": 2240, |
| "token_acc": 0.826577064816822, |
| "train_speed(iter/s)": 0.029498 |
| }, |
| { |
| "epoch": 0.718371265149394, |
| "grad_norm": 0.05780707586931182, |
| "learning_rate": 3.893521203645618e-06, |
| "loss": 0.45052361488342285, |
| "memory(GiB)": 77.63, |
| "step": 2245, |
| "token_acc": 0.8836182062608028, |
| "train_speed(iter/s)": 0.0295 |
| }, |
| { |
| "epoch": 0.7199712011519539, |
| "grad_norm": 0.049110615928360885, |
| "learning_rate": 3.852482488956992e-06, |
| "loss": 0.4427218437194824, |
| "memory(GiB)": 77.63, |
| "step": 2250, |
| "token_acc": 0.8621255642183012, |
| "train_speed(iter/s)": 0.029501 |
| }, |
| { |
| "epoch": 0.7199712011519539, |
| "eval_loss": 0.659950852394104, |
| "eval_runtime": 108.3142, |
| "eval_samples_per_second": 185.461, |
| "eval_steps_per_second": 0.932, |
| "eval_token_acc": 0.8716289458342705, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.7215711371545138, |
| "grad_norm": 0.04899882607919235, |
| "learning_rate": 3.8116095373434204e-06, |
| "loss": 0.4487879753112793, |
| "memory(GiB)": 77.63, |
| "step": 2255, |
| "token_acc": 0.8912671818368324, |
| "train_speed(iter/s)": 0.029473 |
| }, |
| { |
| "epoch": 0.7231710731570737, |
| "grad_norm": 0.05080548488435112, |
| "learning_rate": 3.7709034509125706e-06, |
| "loss": 0.44452829360961915, |
| "memory(GiB)": 77.63, |
| "step": 2260, |
| "token_acc": 0.8442477876106195, |
| "train_speed(iter/s)": 0.029482 |
| }, |
| { |
| "epoch": 0.7247710091596337, |
| "grad_norm": 0.048986009146357284, |
| "learning_rate": 3.7303653272727057e-06, |
| "loss": 0.4472095012664795, |
| "memory(GiB)": 77.63, |
| "step": 2265, |
| "token_acc": 0.870567815521944, |
| "train_speed(iter/s)": 0.029495 |
| }, |
| { |
| "epoch": 0.7263709451621935, |
| "grad_norm": 0.05152412916361422, |
| "learning_rate": 3.689996259503116e-06, |
| "loss": 0.440493106842041, |
| "memory(GiB)": 77.63, |
| "step": 2270, |
| "token_acc": 0.8795436455293181, |
| "train_speed(iter/s)": 0.029499 |
| }, |
| { |
| "epoch": 0.7279708811647534, |
| "grad_norm": 0.055480142184644934, |
| "learning_rate": 3.6497973361246153e-06, |
| "loss": 0.4417555809020996, |
| "memory(GiB)": 77.63, |
| "step": 2275, |
| "token_acc": 0.8660460713158725, |
| "train_speed(iter/s)": 0.029499 |
| }, |
| { |
| "epoch": 0.7295708171673133, |
| "grad_norm": 0.05625540509082736, |
| "learning_rate": 3.609769641070221e-06, |
| "loss": 0.4407214164733887, |
| "memory(GiB)": 77.63, |
| "step": 2280, |
| "token_acc": 0.8890608875128999, |
| "train_speed(iter/s)": 0.029513 |
| }, |
| { |
| "epoch": 0.7311707531698732, |
| "grad_norm": 0.05002588428206622, |
| "learning_rate": 3.569914253655896e-06, |
| "loss": 0.4413386344909668, |
| "memory(GiB)": 77.63, |
| "step": 2285, |
| "token_acc": 0.8921049390319005, |
| "train_speed(iter/s)": 0.029513 |
| }, |
| { |
| "epoch": 0.7327706891724332, |
| "grad_norm": 0.05088814815973685, |
| "learning_rate": 3.530232248551466e-06, |
| "loss": 0.4507819652557373, |
| "memory(GiB)": 77.63, |
| "step": 2290, |
| "token_acc": 0.8278411830895355, |
| "train_speed(iter/s)": 0.029513 |
| }, |
| { |
| "epoch": 0.734370625174993, |
| "grad_norm": 0.05399937134620822, |
| "learning_rate": 3.4907246957516416e-06, |
| "loss": 0.4447961330413818, |
| "memory(GiB)": 77.63, |
| "step": 2295, |
| "token_acc": 0.8888263967004124, |
| "train_speed(iter/s)": 0.029522 |
| }, |
| { |
| "epoch": 0.7359705611775529, |
| "grad_norm": 0.06200035405309708, |
| "learning_rate": 3.4513926605471504e-06, |
| "loss": 0.45868444442749023, |
| "memory(GiB)": 77.63, |
| "step": 2300, |
| "token_acc": 0.8513141426783479, |
| "train_speed(iter/s)": 0.02952 |
| }, |
| { |
| "epoch": 0.7375704971801128, |
| "grad_norm": 0.055876255912378235, |
| "learning_rate": 3.412237203496036e-06, |
| "loss": 0.4431456089019775, |
| "memory(GiB)": 77.63, |
| "step": 2305, |
| "token_acc": 0.8651997041420119, |
| "train_speed(iter/s)": 0.029532 |
| }, |
| { |
| "epoch": 0.7391704331826727, |
| "grad_norm": 0.06032844036632358, |
| "learning_rate": 3.3732593803950354e-06, |
| "loss": 0.4452229976654053, |
| "memory(GiB)": 77.63, |
| "step": 2310, |
| "token_acc": 0.8915232899706252, |
| "train_speed(iter/s)": 0.02953 |
| }, |
| { |
| "epoch": 0.7407703691852325, |
| "grad_norm": 0.05255216039270682, |
| "learning_rate": 3.3344602422511343e-06, |
| "loss": 0.4414207458496094, |
| "memory(GiB)": 77.63, |
| "step": 2315, |
| "token_acc": 0.8901802257032171, |
| "train_speed(iter/s)": 0.029522 |
| }, |
| { |
| "epoch": 0.7423703051877925, |
| "grad_norm": 0.05535966142690852, |
| "learning_rate": 3.2958408352532055e-06, |
| "loss": 0.43938393592834474, |
| "memory(GiB)": 77.63, |
| "step": 2320, |
| "token_acc": 0.8354077253218885, |
| "train_speed(iter/s)": 0.029536 |
| }, |
| { |
| "epoch": 0.7439702411903524, |
| "grad_norm": 0.0505418855319798, |
| "learning_rate": 3.257402200743821e-06, |
| "loss": 0.44445362091064455, |
| "memory(GiB)": 77.63, |
| "step": 2325, |
| "token_acc": 0.864262790258637, |
| "train_speed(iter/s)": 0.02953 |
| }, |
| { |
| "epoch": 0.7455701771929123, |
| "grad_norm": 0.0563107101835597, |
| "learning_rate": 3.2191453751911505e-06, |
| "loss": 0.45569453239440916, |
| "memory(GiB)": 77.63, |
| "step": 2330, |
| "token_acc": 0.8784313725490196, |
| "train_speed(iter/s)": 0.029523 |
| }, |
| { |
| "epoch": 0.7471701131954722, |
| "grad_norm": 0.05000152007266613, |
| "learning_rate": 3.1810713901610367e-06, |
| "loss": 0.4395348072052002, |
| "memory(GiB)": 77.63, |
| "step": 2335, |
| "token_acc": 0.8867111781175964, |
| "train_speed(iter/s)": 0.029536 |
| }, |
| { |
| "epoch": 0.748770049198032, |
| "grad_norm": 0.057169590375126145, |
| "learning_rate": 3.1431812722891598e-06, |
| "loss": 0.4397278785705566, |
| "memory(GiB)": 77.63, |
| "step": 2340, |
| "token_acc": 0.8577532891037895, |
| "train_speed(iter/s)": 0.029529 |
| }, |
| { |
| "epoch": 0.750369985200592, |
| "grad_norm": 0.05730745865195846, |
| "learning_rate": 3.1054760432533626e-06, |
| "loss": 0.45998029708862304, |
| "memory(GiB)": 77.63, |
| "step": 2345, |
| "token_acc": 0.8845755097339016, |
| "train_speed(iter/s)": 0.029534 |
| }, |
| { |
| "epoch": 0.7519699212031519, |
| "grad_norm": 0.05180470840824953, |
| "learning_rate": 3.0679567197461135e-06, |
| "loss": 0.45008273124694825, |
| "memory(GiB)": 77.63, |
| "step": 2350, |
| "token_acc": 0.8394425931535898, |
| "train_speed(iter/s)": 0.029537 |
| }, |
| { |
| "epoch": 0.7535698572057118, |
| "grad_norm": 0.06025883780673481, |
| "learning_rate": 3.0306243134470668e-06, |
| "loss": 0.4444745540618896, |
| "memory(GiB)": 77.63, |
| "step": 2355, |
| "token_acc": 0.889631386074585, |
| "train_speed(iter/s)": 0.02953 |
| }, |
| { |
| "epoch": 0.7551697932082717, |
| "grad_norm": 0.05199872680450009, |
| "learning_rate": 2.993479830995815e-06, |
| "loss": 0.451768159866333, |
| "memory(GiB)": 77.63, |
| "step": 2360, |
| "token_acc": 0.8736520199581522, |
| "train_speed(iter/s)": 0.029541 |
| }, |
| { |
| "epoch": 0.7567697292108315, |
| "grad_norm": 0.05489027404588469, |
| "learning_rate": 2.9565242739647115e-06, |
| "loss": 0.4442115306854248, |
| "memory(GiB)": 77.63, |
| "step": 2365, |
| "token_acc": 0.8865552903739061, |
| "train_speed(iter/s)": 0.029538 |
| }, |
| { |
| "epoch": 0.7583696652133914, |
| "grad_norm": 0.06334021131539457, |
| "learning_rate": 2.919758638831893e-06, |
| "loss": 0.4570741653442383, |
| "memory(GiB)": 77.63, |
| "step": 2370, |
| "token_acc": 0.8652606912712361, |
| "train_speed(iter/s)": 0.029531 |
| }, |
| { |
| "epoch": 0.7599696012159514, |
| "grad_norm": 0.053831666314624105, |
| "learning_rate": 2.8831839169543998e-06, |
| "loss": 0.44495415687561035, |
| "memory(GiB)": 77.63, |
| "step": 2375, |
| "token_acc": 0.8756476683937824, |
| "train_speed(iter/s)": 0.029541 |
| }, |
| { |
| "epoch": 0.7615695372185113, |
| "grad_norm": 0.0527583973457582, |
| "learning_rate": 2.84680109454143e-06, |
| "loss": 0.4472104549407959, |
| "memory(GiB)": 77.63, |
| "step": 2380, |
| "token_acc": 0.8725328947368421, |
| "train_speed(iter/s)": 0.029536 |
| }, |
| { |
| "epoch": 0.7631694732210712, |
| "grad_norm": 0.058941021305098804, |
| "learning_rate": 2.810611152627777e-06, |
| "loss": 0.4499720573425293, |
| "memory(GiB)": 77.63, |
| "step": 2385, |
| "token_acc": 0.8632213889794588, |
| "train_speed(iter/s)": 0.029535 |
| }, |
| { |
| "epoch": 0.764769409223631, |
| "grad_norm": 0.05393008170855123, |
| "learning_rate": 2.774615067047346e-06, |
| "loss": 0.43872222900390623, |
| "memory(GiB)": 77.63, |
| "step": 2390, |
| "token_acc": 0.8742202234150588, |
| "train_speed(iter/s)": 0.02954 |
| }, |
| { |
| "epoch": 0.7663693452261909, |
| "grad_norm": 0.0556417645809335, |
| "learning_rate": 2.738813808406866e-06, |
| "loss": 0.4399220943450928, |
| "memory(GiB)": 77.63, |
| "step": 2395, |
| "token_acc": 0.8997599039615847, |
| "train_speed(iter/s)": 0.029531 |
| }, |
| { |
| "epoch": 0.7679692812287509, |
| "grad_norm": 0.05379432234404155, |
| "learning_rate": 2.7032083420597e-06, |
| "loss": 0.4382453441619873, |
| "memory(GiB)": 77.63, |
| "step": 2400, |
| "token_acc": 0.8875784668061633, |
| "train_speed(iter/s)": 0.029541 |
| }, |
| { |
| "epoch": 0.7695692172313108, |
| "grad_norm": 0.05806842163630925, |
| "learning_rate": 2.667799628079829e-06, |
| "loss": 0.44454326629638674, |
| "memory(GiB)": 77.63, |
| "step": 2405, |
| "token_acc": 0.8880662020905923, |
| "train_speed(iter/s)": 0.029535 |
| }, |
| { |
| "epoch": 0.7711691532338707, |
| "grad_norm": 0.06148704112133217, |
| "learning_rate": 2.6325886212359496e-06, |
| "loss": 0.43945813179016113, |
| "memory(GiB)": 77.63, |
| "step": 2410, |
| "token_acc": 0.8767772511848341, |
| "train_speed(iter/s)": 0.029527 |
| }, |
| { |
| "epoch": 0.7727690892364305, |
| "grad_norm": 0.056530065759685846, |
| "learning_rate": 2.5975762709657506e-06, |
| "loss": 0.4438450813293457, |
| "memory(GiB)": 77.63, |
| "step": 2415, |
| "token_acc": 0.8570975416336241, |
| "train_speed(iter/s)": 0.029538 |
| }, |
| { |
| "epoch": 0.7743690252389904, |
| "grad_norm": 0.056107845444701834, |
| "learning_rate": 2.5627635213502832e-06, |
| "loss": 0.43836054801940916, |
| "memory(GiB)": 77.63, |
| "step": 2420, |
| "token_acc": 0.8966822253059165, |
| "train_speed(iter/s)": 0.029532 |
| }, |
| { |
| "epoch": 0.7759689612415503, |
| "grad_norm": 0.05796065696017405, |
| "learning_rate": 2.528151311088537e-06, |
| "loss": 0.4400279998779297, |
| "memory(GiB)": 77.63, |
| "step": 2425, |
| "token_acc": 0.8552805280528053, |
| "train_speed(iter/s)": 0.029526 |
| }, |
| { |
| "epoch": 0.7775688972441103, |
| "grad_norm": 0.05418546630146028, |
| "learning_rate": 2.4937405734720964e-06, |
| "loss": 0.44541444778442385, |
| "memory(GiB)": 77.63, |
| "step": 2430, |
| "token_acc": 0.8620764552562988, |
| "train_speed(iter/s)": 0.029532 |
| }, |
| { |
| "epoch": 0.7791688332466701, |
| "grad_norm": 0.05654250831277805, |
| "learning_rate": 2.459532236360007e-06, |
| "loss": 0.43491110801696775, |
| "memory(GiB)": 77.63, |
| "step": 2435, |
| "token_acc": 0.8570184983677911, |
| "train_speed(iter/s)": 0.029522 |
| }, |
| { |
| "epoch": 0.78076876924923, |
| "grad_norm": 0.05526372242621089, |
| "learning_rate": 2.4255272221537295e-06, |
| "loss": 0.4378859043121338, |
| "memory(GiB)": 77.63, |
| "step": 2440, |
| "token_acc": 0.8631236857197476, |
| "train_speed(iter/s)": 0.029526 |
| }, |
| { |
| "epoch": 0.7823687052517899, |
| "grad_norm": 0.05404315424969483, |
| "learning_rate": 2.391726447772279e-06, |
| "loss": 0.45857391357421873, |
| "memory(GiB)": 77.63, |
| "step": 2445, |
| "token_acc": 0.8634816932081122, |
| "train_speed(iter/s)": 0.029527 |
| }, |
| { |
| "epoch": 0.7839686412543498, |
| "grad_norm": 0.05765113554061621, |
| "learning_rate": 2.3581308246275103e-06, |
| "loss": 0.4473139762878418, |
| "memory(GiB)": 77.63, |
| "step": 2450, |
| "token_acc": 0.8979846898922044, |
| "train_speed(iter/s)": 0.029518 |
| }, |
| { |
| "epoch": 0.7855685772569098, |
| "grad_norm": 0.058353221842389495, |
| "learning_rate": 2.324741258599521e-06, |
| "loss": 0.44444866180419923, |
| "memory(GiB)": 77.63, |
| "step": 2455, |
| "token_acc": 0.8648913576213038, |
| "train_speed(iter/s)": 0.029527 |
| }, |
| { |
| "epoch": 0.7871685132594696, |
| "grad_norm": 0.05309289512529438, |
| "learning_rate": 2.29155865001225e-06, |
| "loss": 0.43857607841491697, |
| "memory(GiB)": 77.63, |
| "step": 2460, |
| "token_acc": 0.894580549368968, |
| "train_speed(iter/s)": 0.029521 |
| }, |
| { |
| "epoch": 0.7887684492620295, |
| "grad_norm": 0.0540412312893473, |
| "learning_rate": 2.2585838936091753e-06, |
| "loss": 0.43953213691711424, |
| "memory(GiB)": 77.63, |
| "step": 2465, |
| "token_acc": 0.8868672731513879, |
| "train_speed(iter/s)": 0.029515 |
| }, |
| { |
| "epoch": 0.7903683852645894, |
| "grad_norm": 0.05880520800742855, |
| "learning_rate": 2.225817878529214e-06, |
| "loss": 0.4457580089569092, |
| "memory(GiB)": 77.63, |
| "step": 2470, |
| "token_acc": 0.8630282437884901, |
| "train_speed(iter/s)": 0.029525 |
| }, |
| { |
| "epoch": 0.7919683212671493, |
| "grad_norm": 0.058328885138964066, |
| "learning_rate": 2.1932614882827196e-06, |
| "loss": 0.4424918174743652, |
| "memory(GiB)": 77.63, |
| "step": 2475, |
| "token_acc": 0.8814697747925722, |
| "train_speed(iter/s)": 0.029517 |
| }, |
| { |
| "epoch": 0.7935682572697093, |
| "grad_norm": 0.05685263085809571, |
| "learning_rate": 2.160915600727688e-06, |
| "loss": 0.43921732902526855, |
| "memory(GiB)": 77.63, |
| "step": 2480, |
| "token_acc": 0.913681738109219, |
| "train_speed(iter/s)": 0.029516 |
| }, |
| { |
| "epoch": 0.7951681932722691, |
| "grad_norm": 0.056639845561812056, |
| "learning_rate": 2.1287810880460636e-06, |
| "loss": 0.44060502052307127, |
| "memory(GiB)": 77.63, |
| "step": 2485, |
| "token_acc": 0.8829075425790754, |
| "train_speed(iter/s)": 0.02952 |
| }, |
| { |
| "epoch": 0.796768129274829, |
| "grad_norm": 0.05230960490349676, |
| "learning_rate": 2.0968588167202265e-06, |
| "loss": 0.43935480117797854, |
| "memory(GiB)": 77.63, |
| "step": 2490, |
| "token_acc": 0.8856997455470738, |
| "train_speed(iter/s)": 0.029511 |
| }, |
| { |
| "epoch": 0.7983680652773889, |
| "grad_norm": 0.05305183045142263, |
| "learning_rate": 2.0651496475096455e-06, |
| "loss": 0.4360368728637695, |
| "memory(GiB)": 77.63, |
| "step": 2495, |
| "token_acc": 0.8394655704008221, |
| "train_speed(iter/s)": 0.029517 |
| }, |
| { |
| "epoch": 0.7999680012799488, |
| "grad_norm": 0.05620012484228566, |
| "learning_rate": 2.03365443542764e-06, |
| "loss": 0.44507203102111814, |
| "memory(GiB)": 77.63, |
| "step": 2500, |
| "token_acc": 0.8857098429482195, |
| "train_speed(iter/s)": 0.029515 |
| }, |
| { |
| "epoch": 0.7999680012799488, |
| "eval_loss": 0.6586322784423828, |
| "eval_runtime": 105.1966, |
| "eval_samples_per_second": 190.957, |
| "eval_steps_per_second": 0.96, |
| "eval_token_acc": 0.8721292963419328, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.8015679372825087, |
| "grad_norm": 0.05541548293022976, |
| "learning_rate": 2.0023740297183536e-06, |
| "loss": 0.44654192924499514, |
| "memory(GiB)": 77.63, |
| "step": 2505, |
| "token_acc": 0.8819252077562327, |
| "train_speed(iter/s)": 0.029489 |
| }, |
| { |
| "epoch": 0.8031678732850686, |
| "grad_norm": 0.057374199419424524, |
| "learning_rate": 1.971309273833828e-06, |
| "loss": 0.44596128463745116, |
| "memory(GiB)": 77.63, |
| "step": 2510, |
| "token_acc": 0.855553561815898, |
| "train_speed(iter/s)": 0.029494 |
| }, |
| { |
| "epoch": 0.8047678092876285, |
| "grad_norm": 0.05297595418967434, |
| "learning_rate": 1.940461005411288e-06, |
| "loss": 0.45099148750305174, |
| "memory(GiB)": 77.63, |
| "step": 2515, |
| "token_acc": 0.8958361962347121, |
| "train_speed(iter/s)": 0.029503 |
| }, |
| { |
| "epoch": 0.8063677452901884, |
| "grad_norm": 0.056714681257095015, |
| "learning_rate": 1.9098300562505266e-06, |
| "loss": 0.4423669338226318, |
| "memory(GiB)": 77.63, |
| "step": 2520, |
| "token_acc": 0.879045996592845, |
| "train_speed(iter/s)": 0.029507 |
| }, |
| { |
| "epoch": 0.8079676812927483, |
| "grad_norm": 0.05868047347236466, |
| "learning_rate": 1.8794172522915022e-06, |
| "loss": 0.4462554931640625, |
| "memory(GiB)": 77.63, |
| "step": 2525, |
| "token_acc": 0.8811320754716981, |
| "train_speed(iter/s)": 0.029507 |
| }, |
| { |
| "epoch": 0.8095676172953082, |
| "grad_norm": 0.04860202133053876, |
| "learning_rate": 1.849223413592046e-06, |
| "loss": 0.4488513946533203, |
| "memory(GiB)": 77.63, |
| "step": 2530, |
| "token_acc": 0.8654490616621984, |
| "train_speed(iter/s)": 0.029517 |
| }, |
| { |
| "epoch": 0.811167553297868, |
| "grad_norm": 0.055005201314914515, |
| "learning_rate": 1.8192493543057676e-06, |
| "loss": 0.45094904899597166, |
| "memory(GiB)": 77.63, |
| "step": 2535, |
| "token_acc": 0.8800874078120732, |
| "train_speed(iter/s)": 0.029516 |
| }, |
| { |
| "epoch": 0.812767489300428, |
| "grad_norm": 0.058350813115584405, |
| "learning_rate": 1.7894958826600884e-06, |
| "loss": 0.4489152908325195, |
| "memory(GiB)": 77.63, |
| "step": 2540, |
| "token_acc": 0.8784655623365301, |
| "train_speed(iter/s)": 0.02952 |
| }, |
| { |
| "epoch": 0.8143674253029879, |
| "grad_norm": 0.05243037926696882, |
| "learning_rate": 1.7599638009344566e-06, |
| "loss": 0.4506648063659668, |
| "memory(GiB)": 77.63, |
| "step": 2545, |
| "token_acc": 0.8711162255466053, |
| "train_speed(iter/s)": 0.029531 |
| }, |
| { |
| "epoch": 0.8159673613055478, |
| "grad_norm": 0.055874525296942985, |
| "learning_rate": 1.730653905438714e-06, |
| "loss": 0.451121187210083, |
| "memory(GiB)": 77.63, |
| "step": 2550, |
| "token_acc": 0.875531914893617, |
| "train_speed(iter/s)": 0.029527 |
| }, |
| { |
| "epoch": 0.8175672973081076, |
| "grad_norm": 0.052382476873803714, |
| "learning_rate": 1.701566986491614e-06, |
| "loss": 0.43659415245056155, |
| "memory(GiB)": 77.63, |
| "step": 2555, |
| "token_acc": 0.8824301518844928, |
| "train_speed(iter/s)": 0.029539 |
| }, |
| { |
| "epoch": 0.8191672333106675, |
| "grad_norm": 0.05679998544269888, |
| "learning_rate": 1.672703828399529e-06, |
| "loss": 0.44143290519714357, |
| "memory(GiB)": 77.63, |
| "step": 2560, |
| "token_acc": 0.9194786645241921, |
| "train_speed(iter/s)": 0.029538 |
| }, |
| { |
| "epoch": 0.8207671693132275, |
| "grad_norm": 0.05177151194489443, |
| "learning_rate": 1.6440652094352838e-06, |
| "loss": 0.44036478996276857, |
| "memory(GiB)": 77.63, |
| "step": 2565, |
| "token_acc": 0.86801315171442, |
| "train_speed(iter/s)": 0.029534 |
| }, |
| { |
| "epoch": 0.8223671053157874, |
| "grad_norm": 0.047985617439266506, |
| "learning_rate": 1.6156519018171856e-06, |
| "loss": 0.44090909957885743, |
| "memory(GiB)": 77.63, |
| "step": 2570, |
| "token_acc": 0.8926761055759482, |
| "train_speed(iter/s)": 0.029546 |
| }, |
| { |
| "epoch": 0.8239670413183473, |
| "grad_norm": 0.06022869670658613, |
| "learning_rate": 1.587464671688187e-06, |
| "loss": 0.4480876922607422, |
| "memory(GiB)": 77.63, |
| "step": 2575, |
| "token_acc": 0.8749468913751594, |
| "train_speed(iter/s)": 0.029542 |
| }, |
| { |
| "epoch": 0.8255669773209071, |
| "grad_norm": 0.05591525813204934, |
| "learning_rate": 1.5595042790952442e-06, |
| "loss": 0.4516183853149414, |
| "memory(GiB)": 77.63, |
| "step": 2580, |
| "token_acc": 0.8408594319009468, |
| "train_speed(iter/s)": 0.029539 |
| }, |
| { |
| "epoch": 0.827166913323467, |
| "grad_norm": 0.05315795994218538, |
| "learning_rate": 1.5317714779688076e-06, |
| "loss": 0.44116387367248533, |
| "memory(GiB)": 77.63, |
| "step": 2585, |
| "token_acc": 0.8697747394374089, |
| "train_speed(iter/s)": 0.029549 |
| }, |
| { |
| "epoch": 0.828766849326027, |
| "grad_norm": 0.054322305095737905, |
| "learning_rate": 1.5042670161024975e-06, |
| "loss": 0.4457075119018555, |
| "memory(GiB)": 77.63, |
| "step": 2590, |
| "token_acc": 0.8946940985381701, |
| "train_speed(iter/s)": 0.029543 |
| }, |
| { |
| "epoch": 0.8303667853285869, |
| "grad_norm": 0.05625328099370699, |
| "learning_rate": 1.4769916351329495e-06, |
| "loss": 0.4413478851318359, |
| "memory(GiB)": 77.63, |
| "step": 2595, |
| "token_acc": 0.8992218637312583, |
| "train_speed(iter/s)": 0.029548 |
| }, |
| { |
| "epoch": 0.8319667213311468, |
| "grad_norm": 0.055362561635933255, |
| "learning_rate": 1.4499460705198e-06, |
| "loss": 0.4511932373046875, |
| "memory(GiB)": 77.63, |
| "step": 2600, |
| "token_acc": 0.8438552188552189, |
| "train_speed(iter/s)": 0.029552 |
| }, |
| { |
| "epoch": 0.8335666573337066, |
| "grad_norm": 0.052438473367666195, |
| "learning_rate": 1.4231310515258745e-06, |
| "loss": 0.441973352432251, |
| "memory(GiB)": 77.63, |
| "step": 2605, |
| "token_acc": 0.8753952017853822, |
| "train_speed(iter/s)": 0.029546 |
| }, |
| { |
| "epoch": 0.8351665933362665, |
| "grad_norm": 0.05274974730016364, |
| "learning_rate": 1.396547301197504e-06, |
| "loss": 0.4393311977386475, |
| "memory(GiB)": 77.63, |
| "step": 2610, |
| "token_acc": 0.8518848700967906, |
| "train_speed(iter/s)": 0.029557 |
| }, |
| { |
| "epoch": 0.8367665293388264, |
| "grad_norm": 0.05853294004614818, |
| "learning_rate": 1.3701955363450447e-06, |
| "loss": 0.4380232810974121, |
| "memory(GiB)": 77.63, |
| "step": 2615, |
| "token_acc": 0.8570597362296354, |
| "train_speed(iter/s)": 0.029554 |
| }, |
| { |
| "epoch": 0.8383664653413864, |
| "grad_norm": 0.05410978790127522, |
| "learning_rate": 1.3440764675235384e-06, |
| "loss": 0.4373164653778076, |
| "memory(GiB)": 77.63, |
| "step": 2620, |
| "token_acc": 0.8798353909465021, |
| "train_speed(iter/s)": 0.029552 |
| }, |
| { |
| "epoch": 0.8399664013439463, |
| "grad_norm": 0.048967955063799855, |
| "learning_rate": 1.3181907990135624e-06, |
| "loss": 0.4333020210266113, |
| "memory(GiB)": 77.63, |
| "step": 2625, |
| "token_acc": 0.8836341008089608, |
| "train_speed(iter/s)": 0.029564 |
| }, |
| { |
| "epoch": 0.8415663373465061, |
| "grad_norm": 0.048274089580157754, |
| "learning_rate": 1.2925392288022299e-06, |
| "loss": 0.4414947509765625, |
| "memory(GiB)": 77.63, |
| "step": 2630, |
| "token_acc": 0.8760546404178385, |
| "train_speed(iter/s)": 0.02956 |
| }, |
| { |
| "epoch": 0.843166273349066, |
| "grad_norm": 0.053684823491607934, |
| "learning_rate": 1.267122448564374e-06, |
| "loss": 0.44922800064086915, |
| "memory(GiB)": 77.63, |
| "step": 2635, |
| "token_acc": 0.8554064052425748, |
| "train_speed(iter/s)": 0.029558 |
| }, |
| { |
| "epoch": 0.8447662093516259, |
| "grad_norm": 0.05262528572569429, |
| "learning_rate": 1.2419411436439021e-06, |
| "loss": 0.4328805923461914, |
| "memory(GiB)": 77.63, |
| "step": 2640, |
| "token_acc": 0.8400081317340923, |
| "train_speed(iter/s)": 0.029565 |
| }, |
| { |
| "epoch": 0.8463661453541859, |
| "grad_norm": 0.05549273276976771, |
| "learning_rate": 1.2169959930353049e-06, |
| "loss": 0.4460554599761963, |
| "memory(GiB)": 77.63, |
| "step": 2645, |
| "token_acc": 0.8804424157303371, |
| "train_speed(iter/s)": 0.029559 |
| }, |
| { |
| "epoch": 0.8479660813567458, |
| "grad_norm": 0.05180537064402683, |
| "learning_rate": 1.1922876693653584e-06, |
| "loss": 0.4503427505493164, |
| "memory(GiB)": 77.63, |
| "step": 2650, |
| "token_acc": 0.8934362934362934, |
| "train_speed(iter/s)": 0.029564 |
| }, |
| { |
| "epoch": 0.8495660173593056, |
| "grad_norm": 0.05011013702559624, |
| "learning_rate": 1.1678168388749788e-06, |
| "loss": 0.4415099620819092, |
| "memory(GiB)": 77.63, |
| "step": 2655, |
| "token_acc": 0.8995949690897463, |
| "train_speed(iter/s)": 0.029566 |
| }, |
| { |
| "epoch": 0.8511659533618655, |
| "grad_norm": 0.057170743989864214, |
| "learning_rate": 1.1435841614012666e-06, |
| "loss": 0.44884433746337893, |
| "memory(GiB)": 77.63, |
| "step": 2660, |
| "token_acc": 0.854816112084063, |
| "train_speed(iter/s)": 0.02956 |
| }, |
| { |
| "epoch": 0.8527658893644254, |
| "grad_norm": 0.054388272142607975, |
| "learning_rate": 1.1195902903597023e-06, |
| "loss": 0.439667797088623, |
| "memory(GiB)": 77.63, |
| "step": 2665, |
| "token_acc": 0.8846260387811634, |
| "train_speed(iter/s)": 0.02957 |
| }, |
| { |
| "epoch": 0.8543658253669854, |
| "grad_norm": 0.05039207218424233, |
| "learning_rate": 1.0958358727265438e-06, |
| "loss": 0.4384475231170654, |
| "memory(GiB)": 77.63, |
| "step": 2670, |
| "token_acc": 0.8525793222533995, |
| "train_speed(iter/s)": 0.029565 |
| }, |
| { |
| "epoch": 0.8559657613695452, |
| "grad_norm": 0.0543645938545219, |
| "learning_rate": 1.0723215490213635e-06, |
| "loss": 0.4338691711425781, |
| "memory(GiB)": 77.63, |
| "step": 2675, |
| "token_acc": 0.853824495541999, |
| "train_speed(iter/s)": 0.029558 |
| }, |
| { |
| "epoch": 0.8575656973721051, |
| "grad_norm": 0.0579168704227633, |
| "learning_rate": 1.0490479532897946e-06, |
| "loss": 0.458463716506958, |
| "memory(GiB)": 77.63, |
| "step": 2680, |
| "token_acc": 0.867092866756393, |
| "train_speed(iter/s)": 0.029566 |
| }, |
| { |
| "epoch": 0.859165633374665, |
| "grad_norm": 0.04996147776053655, |
| "learning_rate": 1.0260157130864178e-06, |
| "loss": 0.43754091262817385, |
| "memory(GiB)": 77.63, |
| "step": 2685, |
| "token_acc": 0.8611873713109128, |
| "train_speed(iter/s)": 0.029558 |
| }, |
| { |
| "epoch": 0.8607655693772249, |
| "grad_norm": 0.053537353272037014, |
| "learning_rate": 1.0032254494578519e-06, |
| "loss": 0.44204487800598147, |
| "memory(GiB)": 77.63, |
| "step": 2690, |
| "token_acc": 0.8575780654988576, |
| "train_speed(iter/s)": 0.029558 |
| }, |
| { |
| "epoch": 0.8623655053797848, |
| "grad_norm": 0.055119127073411836, |
| "learning_rate": 9.806777769260034e-07, |
| "loss": 0.4500781536102295, |
| "memory(GiB)": 77.63, |
| "step": 2695, |
| "token_acc": 0.8872294372294373, |
| "train_speed(iter/s)": 0.029562 |
| }, |
| { |
| "epoch": 0.8639654413823447, |
| "grad_norm": 0.0558713442289911, |
| "learning_rate": 9.583733034714982e-07, |
| "loss": 0.43947248458862304, |
| "memory(GiB)": 77.63, |
| "step": 2700, |
| "token_acc": 0.8926744522729466, |
| "train_speed(iter/s)": 0.029555 |
| }, |
| { |
| "epoch": 0.8655653773849046, |
| "grad_norm": 0.0552196144062876, |
| "learning_rate": 9.363126305172831e-07, |
| "loss": 0.4443229675292969, |
| "memory(GiB)": 77.63, |
| "step": 2705, |
| "token_acc": 0.9038031319910514, |
| "train_speed(iter/s)": 0.029561 |
| }, |
| { |
| "epoch": 0.8671653133874645, |
| "grad_norm": 0.055723745057826034, |
| "learning_rate": 9.144963529124163e-07, |
| "loss": 0.42942004203796386, |
| "memory(GiB)": 77.63, |
| "step": 2710, |
| "token_acc": 0.8680161943319838, |
| "train_speed(iter/s)": 0.029557 |
| }, |
| { |
| "epoch": 0.8687652493900244, |
| "grad_norm": 0.0589746686821641, |
| "learning_rate": 8.929250589160166e-07, |
| "loss": 0.4397599220275879, |
| "memory(GiB)": 77.63, |
| "step": 2715, |
| "token_acc": 0.8713878713878714, |
| "train_speed(iter/s)": 0.029552 |
| }, |
| { |
| "epoch": 0.8703651853925843, |
| "grad_norm": 0.04909314017257213, |
| "learning_rate": 8.715993301814174e-07, |
| "loss": 0.44155421257019045, |
| "memory(GiB)": 77.63, |
| "step": 2720, |
| "token_acc": 0.8710053650571495, |
| "train_speed(iter/s)": 0.029561 |
| }, |
| { |
| "epoch": 0.8719651213951441, |
| "grad_norm": 0.05047518453544575, |
| "learning_rate": 8.505197417404687e-07, |
| "loss": 0.43677616119384766, |
| "memory(GiB)": 77.63, |
| "step": 2725, |
| "token_acc": 0.8809886575249704, |
| "train_speed(iter/s)": 0.029556 |
| }, |
| { |
| "epoch": 0.8735650573977041, |
| "grad_norm": 0.05102151204327215, |
| "learning_rate": 8.296868619880372e-07, |
| "loss": 0.44188566207885743, |
| "memory(GiB)": 77.63, |
| "step": 2730, |
| "token_acc": 0.8547172833573602, |
| "train_speed(iter/s)": 0.029553 |
| }, |
| { |
| "epoch": 0.875164993400264, |
| "grad_norm": 0.04729834705444575, |
| "learning_rate": 8.091012526666797e-07, |
| "loss": 0.4441237926483154, |
| "memory(GiB)": 77.63, |
| "step": 2735, |
| "token_acc": 0.8537975972307066, |
| "train_speed(iter/s)": 0.029561 |
| }, |
| { |
| "epoch": 0.8767649294028239, |
| "grad_norm": 0.047668539210598965, |
| "learning_rate": 7.887634688515e-07, |
| "loss": 0.4462736129760742, |
| "memory(GiB)": 77.63, |
| "step": 2740, |
| "token_acc": 0.903437815975733, |
| "train_speed(iter/s)": 0.029554 |
| }, |
| { |
| "epoch": 0.8783648654053838, |
| "grad_norm": 0.052216823887528664, |
| "learning_rate": 7.686740589351704e-07, |
| "loss": 0.44857120513916016, |
| "memory(GiB)": 77.63, |
| "step": 2745, |
| "token_acc": 0.8033573141486811, |
| "train_speed(iter/s)": 0.029556 |
| }, |
| { |
| "epoch": 0.8799648014079436, |
| "grad_norm": 0.055862979558343906, |
| "learning_rate": 7.488335646131628e-07, |
| "loss": 0.44959425926208496, |
| "memory(GiB)": 77.63, |
| "step": 2750, |
| "token_acc": 0.8605054151624548, |
| "train_speed(iter/s)": 0.029558 |
| }, |
| { |
| "epoch": 0.8799648014079436, |
| "eval_loss": 0.6577034592628479, |
| "eval_runtime": 106.4875, |
| "eval_samples_per_second": 188.642, |
| "eval_steps_per_second": 0.948, |
| "eval_token_acc": 0.8722958612553617, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.8815647374105036, |
| "grad_norm": 0.051528350081992934, |
| "learning_rate": 7.292425208691212e-07, |
| "loss": 0.43878631591796874, |
| "memory(GiB)": 77.63, |
| "step": 2755, |
| "token_acc": 0.8812832745626772, |
| "train_speed(iter/s)": 0.029532 |
| }, |
| { |
| "epoch": 0.8831646734130635, |
| "grad_norm": 0.05310175482611414, |
| "learning_rate": 7.099014559604556e-07, |
| "loss": 0.45635418891906737, |
| "memory(GiB)": 77.63, |
| "step": 2760, |
| "token_acc": 0.8894999360532038, |
| "train_speed(iter/s)": 0.029537 |
| }, |
| { |
| "epoch": 0.8847646094156234, |
| "grad_norm": 0.04975738369955541, |
| "learning_rate": 6.908108914040823e-07, |
| "loss": 0.4421397686004639, |
| "memory(GiB)": 77.63, |
| "step": 2765, |
| "token_acc": 0.9070493575117089, |
| "train_speed(iter/s)": 0.029548 |
| }, |
| { |
| "epoch": 0.8863645454181833, |
| "grad_norm": 0.053564472486824076, |
| "learning_rate": 6.71971341962373e-07, |
| "loss": 0.4513510227203369, |
| "memory(GiB)": 77.63, |
| "step": 2770, |
| "token_acc": 0.8660589060308556, |
| "train_speed(iter/s)": 0.02955 |
| }, |
| { |
| "epoch": 0.8879644814207431, |
| "grad_norm": 0.06332925713320489, |
| "learning_rate": 6.53383315629268e-07, |
| "loss": 0.4404273509979248, |
| "memory(GiB)": 77.63, |
| "step": 2775, |
| "token_acc": 0.8507806501151779, |
| "train_speed(iter/s)": 0.029546 |
| }, |
| { |
| "epoch": 0.889564417423303, |
| "grad_norm": 0.063294227744794, |
| "learning_rate": 6.350473136165836e-07, |
| "loss": 0.4379493236541748, |
| "memory(GiB)": 77.63, |
| "step": 2780, |
| "token_acc": 0.8879898461050294, |
| "train_speed(iter/s)": 0.029561 |
| }, |
| { |
| "epoch": 0.891164353425863, |
| "grad_norm": 0.05151642870451994, |
| "learning_rate": 6.169638303404912e-07, |
| "loss": 0.4380655765533447, |
| "memory(GiB)": 77.63, |
| "step": 2785, |
| "token_acc": 0.8904059040590406, |
| "train_speed(iter/s)": 0.02956 |
| }, |
| { |
| "epoch": 0.8927642894284229, |
| "grad_norm": 0.05406168921762394, |
| "learning_rate": 5.991333534081878e-07, |
| "loss": 0.4479250907897949, |
| "memory(GiB)": 77.63, |
| "step": 2790, |
| "token_acc": 0.8831118813787792, |
| "train_speed(iter/s)": 0.02956 |
| }, |
| { |
| "epoch": 0.8943642254309827, |
| "grad_norm": 0.054911478200183335, |
| "learning_rate": 5.815563636047539e-07, |
| "loss": 0.43634886741638185, |
| "memory(GiB)": 77.63, |
| "step": 2795, |
| "token_acc": 0.8714865708931917, |
| "train_speed(iter/s)": 0.02957 |
| }, |
| { |
| "epoch": 0.8959641614335426, |
| "grad_norm": 0.05257509941727236, |
| "learning_rate": 5.64233334880181e-07, |
| "loss": 0.44048466682434084, |
| "memory(GiB)": 77.63, |
| "step": 2800, |
| "token_acc": 0.891296869625043, |
| "train_speed(iter/s)": 0.029568 |
| }, |
| { |
| "epoch": 0.8975640974361025, |
| "grad_norm": 0.05532206374459385, |
| "learning_rate": 5.471647343365982e-07, |
| "loss": 0.44726853370666503, |
| "memory(GiB)": 77.63, |
| "step": 2805, |
| "token_acc": 0.8828892005610098, |
| "train_speed(iter/s)": 0.029576 |
| }, |
| { |
| "epoch": 0.8991640334386625, |
| "grad_norm": 0.05107208389411162, |
| "learning_rate": 5.303510222156716e-07, |
| "loss": 0.4470540523529053, |
| "memory(GiB)": 77.63, |
| "step": 2810, |
| "token_acc": 0.8489765812281025, |
| "train_speed(iter/s)": 0.029574 |
| }, |
| { |
| "epoch": 0.9007639694412224, |
| "grad_norm": 0.055452171728558604, |
| "learning_rate": 5.137926518862013e-07, |
| "loss": 0.4417248249053955, |
| "memory(GiB)": 77.63, |
| "step": 2815, |
| "token_acc": 0.8739084132055378, |
| "train_speed(iter/s)": 0.029569 |
| }, |
| { |
| "epoch": 0.9023639054437822, |
| "grad_norm": 0.04866974543895629, |
| "learning_rate": 4.974900698318885e-07, |
| "loss": 0.4414045810699463, |
| "memory(GiB)": 77.63, |
| "step": 2820, |
| "token_acc": 0.8659420289855072, |
| "train_speed(iter/s)": 0.029581 |
| }, |
| { |
| "epoch": 0.9039638414463421, |
| "grad_norm": 0.05275403785935388, |
| "learning_rate": 4.814437156393048e-07, |
| "loss": 0.4543337821960449, |
| "memory(GiB)": 77.63, |
| "step": 2825, |
| "token_acc": 0.8396122896854425, |
| "train_speed(iter/s)": 0.029574 |
| }, |
| { |
| "epoch": 0.905563777448902, |
| "grad_norm": 0.05372217346495556, |
| "learning_rate": 4.656540219860317e-07, |
| "loss": 0.45271754264831543, |
| "memory(GiB)": 77.63, |
| "step": 2830, |
| "token_acc": 0.8707617789520036, |
| "train_speed(iter/s)": 0.029571 |
| }, |
| { |
| "epoch": 0.907163713451462, |
| "grad_norm": 0.05715427837146615, |
| "learning_rate": 4.501214146289956e-07, |
| "loss": 0.4418344497680664, |
| "memory(GiB)": 77.63, |
| "step": 2835, |
| "token_acc": 0.8788416882939489, |
| "train_speed(iter/s)": 0.02958 |
| }, |
| { |
| "epoch": 0.9087636494540219, |
| "grad_norm": 0.04986985198768239, |
| "learning_rate": 4.3484631239299356e-07, |
| "loss": 0.4437891960144043, |
| "memory(GiB)": 77.63, |
| "step": 2840, |
| "token_acc": 0.8431597023468803, |
| "train_speed(iter/s)": 0.029572 |
| }, |
| { |
| "epoch": 0.9103635854565817, |
| "grad_norm": 0.04999552116510165, |
| "learning_rate": 4.198291271593924e-07, |
| "loss": 0.44283204078674315, |
| "memory(GiB)": 77.63, |
| "step": 2845, |
| "token_acc": 0.8843727072633896, |
| "train_speed(iter/s)": 0.029575 |
| }, |
| { |
| "epoch": 0.9119635214591416, |
| "grad_norm": 0.047844581858855956, |
| "learning_rate": 4.0507026385502747e-07, |
| "loss": 0.4449836254119873, |
| "memory(GiB)": 77.63, |
| "step": 2850, |
| "token_acc": 0.9034812490661885, |
| "train_speed(iter/s)": 0.029576 |
| }, |
| { |
| "epoch": 0.9135634574617015, |
| "grad_norm": 0.05675006848739315, |
| "learning_rate": 3.9057012044127817e-07, |
| "loss": 0.44204154014587405, |
| "memory(GiB)": 77.63, |
| "step": 2855, |
| "token_acc": 0.8655569782330346, |
| "train_speed(iter/s)": 0.029569 |
| }, |
| { |
| "epoch": 0.9151633934642615, |
| "grad_norm": 0.054077703664588216, |
| "learning_rate": 3.7632908790334656e-07, |
| "loss": 0.4383398532867432, |
| "memory(GiB)": 77.63, |
| "step": 2860, |
| "token_acc": 0.8899396929824561, |
| "train_speed(iter/s)": 0.029578 |
| }, |
| { |
| "epoch": 0.9167633294668214, |
| "grad_norm": 0.055142279260510525, |
| "learning_rate": 3.6234755023970447e-07, |
| "loss": 0.4388674259185791, |
| "memory(GiB)": 77.63, |
| "step": 2865, |
| "token_acc": 0.8406133828996283, |
| "train_speed(iter/s)": 0.029574 |
| }, |
| { |
| "epoch": 0.9183632654693812, |
| "grad_norm": 0.05242909357272202, |
| "learning_rate": 3.4862588445174985e-07, |
| "loss": 0.44350008964538573, |
| "memory(GiB)": 77.63, |
| "step": 2870, |
| "token_acc": 0.8773854961832062, |
| "train_speed(iter/s)": 0.029568 |
| }, |
| { |
| "epoch": 0.9199632014719411, |
| "grad_norm": 0.051431062939426, |
| "learning_rate": 3.3516446053363015e-07, |
| "loss": 0.43948516845703123, |
| "memory(GiB)": 77.63, |
| "step": 2875, |
| "token_acc": 0.8812238692512353, |
| "train_speed(iter/s)": 0.029578 |
| }, |
| { |
| "epoch": 0.921563137474501, |
| "grad_norm": 0.04842108410960974, |
| "learning_rate": 3.219636414622751e-07, |
| "loss": 0.44395694732666013, |
| "memory(GiB)": 77.63, |
| "step": 2880, |
| "token_acc": 0.872349158571624, |
| "train_speed(iter/s)": 0.029573 |
| }, |
| { |
| "epoch": 0.923163073477061, |
| "grad_norm": 0.050854557927068264, |
| "learning_rate": 3.090237831876053e-07, |
| "loss": 0.4437469482421875, |
| "memory(GiB)": 77.63, |
| "step": 2885, |
| "token_acc": 0.8500566251415629, |
| "train_speed(iter/s)": 0.029572 |
| }, |
| { |
| "epoch": 0.9247630094796209, |
| "grad_norm": 0.05289158935619814, |
| "learning_rate": 2.9634523462293005e-07, |
| "loss": 0.439394474029541, |
| "memory(GiB)": 77.63, |
| "step": 2890, |
| "token_acc": 0.9039268013724743, |
| "train_speed(iter/s)": 0.029576 |
| }, |
| { |
| "epoch": 0.9263629454821807, |
| "grad_norm": 0.0532633695315608, |
| "learning_rate": 2.839283376355506e-07, |
| "loss": 0.4414195537567139, |
| "memory(GiB)": 77.63, |
| "step": 2895, |
| "token_acc": 0.8979206049149339, |
| "train_speed(iter/s)": 0.029567 |
| }, |
| { |
| "epoch": 0.9279628814847406, |
| "grad_norm": 0.05197494947423765, |
| "learning_rate": 2.717734270375272e-07, |
| "loss": 0.4303572177886963, |
| "memory(GiB)": 77.63, |
| "step": 2900, |
| "token_acc": 0.8932318992654774, |
| "train_speed(iter/s)": 0.029573 |
| }, |
| { |
| "epoch": 0.9295628174873005, |
| "grad_norm": 0.05715199054523632, |
| "learning_rate": 2.5988083057666534e-07, |
| "loss": 0.4488718032836914, |
| "memory(GiB)": 77.63, |
| "step": 2905, |
| "token_acc": 0.8734599589322382, |
| "train_speed(iter/s)": 0.029572 |
| }, |
| { |
| "epoch": 0.9311627534898604, |
| "grad_norm": 0.052180972248620894, |
| "learning_rate": 2.4825086892766745e-07, |
| "loss": 0.44499683380126953, |
| "memory(GiB)": 77.63, |
| "step": 2910, |
| "token_acc": 0.8798773215198501, |
| "train_speed(iter/s)": 0.029568 |
| }, |
| { |
| "epoch": 0.9327626894924202, |
| "grad_norm": 0.05195036988843101, |
| "learning_rate": 2.3688385568349515e-07, |
| "loss": 0.4348268508911133, |
| "memory(GiB)": 77.63, |
| "step": 2915, |
| "token_acc": 0.8380835380835381, |
| "train_speed(iter/s)": 0.029578 |
| }, |
| { |
| "epoch": 0.9343626254949802, |
| "grad_norm": 0.05340475761749026, |
| "learning_rate": 2.2578009734690264e-07, |
| "loss": 0.4533662796020508, |
| "memory(GiB)": 77.63, |
| "step": 2920, |
| "token_acc": 0.8854700854700854, |
| "train_speed(iter/s)": 0.029569 |
| }, |
| { |
| "epoch": 0.9359625614975401, |
| "grad_norm": 0.047434143218971526, |
| "learning_rate": 2.1493989332218468e-07, |
| "loss": 0.4382183074951172, |
| "memory(GiB)": 77.63, |
| "step": 2925, |
| "token_acc": 0.880750496121234, |
| "train_speed(iter/s)": 0.029565 |
| }, |
| { |
| "epoch": 0.9375624975001, |
| "grad_norm": 0.052509681961836426, |
| "learning_rate": 2.043635359070928e-07, |
| "loss": 0.44708704948425293, |
| "memory(GiB)": 77.63, |
| "step": 2930, |
| "token_acc": 0.8942012598933936, |
| "train_speed(iter/s)": 0.029576 |
| }, |
| { |
| "epoch": 0.9391624335026599, |
| "grad_norm": 0.0531539858574761, |
| "learning_rate": 1.9405131028495838e-07, |
| "loss": 0.45058341026306153, |
| "memory(GiB)": 77.63, |
| "step": 2935, |
| "token_acc": 0.8614694335389792, |
| "train_speed(iter/s)": 0.029569 |
| }, |
| { |
| "epoch": 0.9407623695052197, |
| "grad_norm": 0.05931701393554437, |
| "learning_rate": 1.8400349451700438e-07, |
| "loss": 0.44367156028747556, |
| "memory(GiB)": 77.63, |
| "step": 2940, |
| "token_acc": 0.881156184096266, |
| "train_speed(iter/s)": 0.029573 |
| }, |
| { |
| "epoch": 0.9423623055077797, |
| "grad_norm": 0.05380572562718686, |
| "learning_rate": 1.742203595348435e-07, |
| "loss": 0.4424111843109131, |
| "memory(GiB)": 77.63, |
| "step": 2945, |
| "token_acc": 0.8769617074701821, |
| "train_speed(iter/s)": 0.029576 |
| }, |
| { |
| "epoch": 0.9439622415103396, |
| "grad_norm": 0.06156259021000458, |
| "learning_rate": 1.6470216913317628e-07, |
| "loss": 0.4509577751159668, |
| "memory(GiB)": 77.63, |
| "step": 2950, |
| "token_acc": 0.8544532947139754, |
| "train_speed(iter/s)": 0.029567 |
| }, |
| { |
| "epoch": 0.9455621775128995, |
| "grad_norm": 0.050697077683688974, |
| "learning_rate": 1.5544917996267562e-07, |
| "loss": 0.44117283821105957, |
| "memory(GiB)": 77.63, |
| "step": 2955, |
| "token_acc": 0.8515226026101759, |
| "train_speed(iter/s)": 0.029575 |
| }, |
| { |
| "epoch": 0.9471621135154594, |
| "grad_norm": 0.05205296295346333, |
| "learning_rate": 1.464616415230702e-07, |
| "loss": 0.4488182067871094, |
| "memory(GiB)": 77.63, |
| "step": 2960, |
| "token_acc": 0.874439461883408, |
| "train_speed(iter/s)": 0.029569 |
| }, |
| { |
| "epoch": 0.9487620495180192, |
| "grad_norm": 0.047632029637579856, |
| "learning_rate": 1.3773979615640976e-07, |
| "loss": 0.4415272235870361, |
| "memory(GiB)": 77.63, |
| "step": 2965, |
| "token_acc": 0.8889883616830797, |
| "train_speed(iter/s)": 0.029564 |
| }, |
| { |
| "epoch": 0.9503619855205792, |
| "grad_norm": 0.051226023380966074, |
| "learning_rate": 1.292838790405393e-07, |
| "loss": 0.4453396797180176, |
| "memory(GiB)": 77.63, |
| "step": 2970, |
| "token_acc": 0.8701866977829639, |
| "train_speed(iter/s)": 0.029572 |
| }, |
| { |
| "epoch": 0.9519619215231391, |
| "grad_norm": 0.05296626405711913, |
| "learning_rate": 1.2109411818274851e-07, |
| "loss": 0.44417614936828614, |
| "memory(GiB)": 77.63, |
| "step": 2975, |
| "token_acc": 0.8997547959036493, |
| "train_speed(iter/s)": 0.029565 |
| }, |
| { |
| "epoch": 0.953561857525699, |
| "grad_norm": 0.0535756814263424, |
| "learning_rate": 1.1317073441363458e-07, |
| "loss": 0.444796085357666, |
| "memory(GiB)": 77.63, |
| "step": 2980, |
| "token_acc": 0.9101887677336147, |
| "train_speed(iter/s)": 0.029563 |
| }, |
| { |
| "epoch": 0.9551617935282589, |
| "grad_norm": 0.05005027675979017, |
| "learning_rate": 1.055139413811379e-07, |
| "loss": 0.45203323364257814, |
| "memory(GiB)": 77.63, |
| "step": 2985, |
| "token_acc": 0.8569892473118279, |
| "train_speed(iter/s)": 0.029568 |
| }, |
| { |
| "epoch": 0.9567617295308187, |
| "grad_norm": 0.04934807615166247, |
| "learning_rate": 9.812394554478355e-08, |
| "loss": 0.43912034034729003, |
| "memory(GiB)": 77.63, |
| "step": 2990, |
| "token_acc": 0.8557346268189642, |
| "train_speed(iter/s)": 0.029559 |
| }, |
| { |
| "epoch": 0.9583616655333786, |
| "grad_norm": 0.051618825055470385, |
| "learning_rate": 9.10009461701189e-08, |
| "loss": 0.4506105899810791, |
| "memory(GiB)": 77.63, |
| "step": 2995, |
| "token_acc": 0.7809948032665182, |
| "train_speed(iter/s)": 0.029565 |
| }, |
| { |
| "epoch": 0.9599616015359386, |
| "grad_norm": 0.054833054855342726, |
| "learning_rate": 8.41451353233369e-08, |
| "loss": 0.442844820022583, |
| "memory(GiB)": 77.63, |
| "step": 3000, |
| "token_acc": 0.8733862959285005, |
| "train_speed(iter/s)": 0.029563 |
| }, |
| { |
| "epoch": 0.9599616015359386, |
| "eval_loss": 0.6573547720909119, |
| "eval_runtime": 106.0877, |
| "eval_samples_per_second": 189.353, |
| "eval_steps_per_second": 0.952, |
| "eval_token_acc": 0.8724322608695082, |
| "step": 3000 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 3125, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.944159044825703e+20, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|