| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.985781990521327, |
| "eval_steps": 100, |
| "global_step": 52, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.018957345971563982, |
| "grad_norm": 3.021558692298453, |
| "learning_rate": 8.333333333333333e-07, |
| "logits/chosen": -0.5616407990455627, |
| "logits/rejected": -0.42516714334487915, |
| "logps/chosen": -373.83282470703125, |
| "logps/rejected": -493.9554748535156, |
| "loss": 0.6931, |
| "rewards/accuracies": 0.0, |
| "rewards/chosen": 0.0, |
| "rewards/margins": 0.0, |
| "rewards/rejected": 0.0, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.1895734597156398, |
| "grad_norm": 3.23447571449701, |
| "learning_rate": 4.907293218369499e-06, |
| "logits/chosen": -0.1685400903224945, |
| "logits/rejected": -0.19170813262462616, |
| "logps/chosen": -464.29693603515625, |
| "logps/rejected": -501.8587341308594, |
| "loss": 0.6936, |
| "rewards/accuracies": 0.4097222089767456, |
| "rewards/chosen": 0.0008683237829245627, |
| "rewards/margins": 0.0007554867188446224, |
| "rewards/rejected": 0.00011283738422207534, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.3791469194312796, |
| "grad_norm": 3.9389946840001264, |
| "learning_rate": 3.941700805287169e-06, |
| "logits/chosen": -0.27595698833465576, |
| "logits/rejected": -0.21783480048179626, |
| "logps/chosen": -438.6907653808594, |
| "logps/rejected": -475.0658264160156, |
| "loss": 0.6894, |
| "rewards/accuracies": 0.4937500059604645, |
| "rewards/chosen": -0.003008360043168068, |
| "rewards/margins": 0.0011918289819732308, |
| "rewards/rejected": -0.00420018844306469, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.5687203791469194, |
| "grad_norm": 3.2779169381175386, |
| "learning_rate": 2.3293939665883233e-06, |
| "logits/chosen": -0.012437907047569752, |
| "logits/rejected": -0.20707044005393982, |
| "logps/chosen": -476.6265563964844, |
| "logps/rejected": -474.9481506347656, |
| "loss": 0.6873, |
| "rewards/accuracies": 0.53125, |
| "rewards/chosen": -0.01446505170315504, |
| "rewards/margins": 0.007696834392845631, |
| "rewards/rejected": -0.02216188609600067, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.7582938388625592, |
| "grad_norm": 2.8592366061141954, |
| "learning_rate": 7.936171419533653e-07, |
| "logits/chosen": -0.4586310386657715, |
| "logits/rejected": -0.5929983854293823, |
| "logps/chosen": -422.44293212890625, |
| "logps/rejected": -402.6429748535156, |
| "loss": 0.6827, |
| "rewards/accuracies": 0.574999988079071, |
| "rewards/chosen": -0.018473980948328972, |
| "rewards/margins": 0.014362658374011517, |
| "rewards/rejected": -0.03283664211630821, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.9478672985781991, |
| "grad_norm": 3.296382116125289, |
| "learning_rate": 2.3285134909173113e-08, |
| "logits/chosen": 0.08067023754119873, |
| "logits/rejected": 0.19673587381839752, |
| "logps/chosen": -483.49212646484375, |
| "logps/rejected": -536.2080688476562, |
| "loss": 0.679, |
| "rewards/accuracies": 0.5874999761581421, |
| "rewards/chosen": -0.017066333442926407, |
| "rewards/margins": 0.01992700807750225, |
| "rewards/rejected": -0.03699334338307381, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.985781990521327, |
| "step": 52, |
| "total_flos": 0.0, |
| "train_loss": 0.6860739244864538, |
| "train_runtime": 316.322, |
| "train_samples_per_second": 21.339, |
| "train_steps_per_second": 0.164 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 52, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": false, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|