| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 12.991322457480042, | |
| "eval_steps": 500, | |
| "global_step": 18720, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.3471017007983339, | |
| "grad_norm": 9.175567626953125, | |
| "learning_rate": 2e-05, | |
| "loss": 10.9573, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.6942034015966678, | |
| "grad_norm": 6.246857166290283, | |
| "learning_rate": 1.9451152579582877e-05, | |
| "loss": 2.2463, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.4781490564346313, | |
| "eval_runtime": 78.7792, | |
| "eval_samples_per_second": 146.282, | |
| "eval_steps_per_second": 4.582, | |
| "step": 1441 | |
| }, | |
| { | |
| "epoch": 1.0409580006942034, | |
| "grad_norm": 0.7370556592941284, | |
| "learning_rate": 1.890230515916575e-05, | |
| "loss": 1.8168, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.3880597014925373, | |
| "grad_norm": 600.9544677734375, | |
| "learning_rate": 1.835345773874863e-05, | |
| "loss": 2.8979, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.7351614022908712, | |
| "grad_norm": 0.528658390045166, | |
| "learning_rate": 1.7804610318331505e-05, | |
| "loss": 1.701, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 1.3760780096054077, | |
| "eval_runtime": 78.8326, | |
| "eval_samples_per_second": 146.183, | |
| "eval_steps_per_second": 4.579, | |
| "step": 2882 | |
| }, | |
| { | |
| "epoch": 2.081916001388407, | |
| "grad_norm": 0.45837515592575073, | |
| "learning_rate": 1.725576289791438e-05, | |
| "loss": 1.5956, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.4290177021867407, | |
| "grad_norm": 0.41563984751701355, | |
| "learning_rate": 1.6706915477497255e-05, | |
| "loss": 1.5615, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 2.7761194029850746, | |
| "grad_norm": 0.4553076922893524, | |
| "learning_rate": 1.6158068057080134e-05, | |
| "loss": 1.5261, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 1.3131805658340454, | |
| "eval_runtime": 78.7183, | |
| "eval_samples_per_second": 146.396, | |
| "eval_steps_per_second": 4.586, | |
| "step": 4323 | |
| }, | |
| { | |
| "epoch": 3.12287400208261, | |
| "grad_norm": 0.4429566562175751, | |
| "learning_rate": 1.560922063666301e-05, | |
| "loss": 1.4933, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 3.469975702880944, | |
| "grad_norm": 0.4545910656452179, | |
| "learning_rate": 1.5060373216245885e-05, | |
| "loss": 1.4766, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 3.817077403679278, | |
| "grad_norm": 0.4629240930080414, | |
| "learning_rate": 1.451152579582876e-05, | |
| "loss": 1.4626, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 1.2775077819824219, | |
| "eval_runtime": 78.6454, | |
| "eval_samples_per_second": 146.531, | |
| "eval_steps_per_second": 4.59, | |
| "step": 5764 | |
| }, | |
| { | |
| "epoch": 4.163832002776814, | |
| "grad_norm": 0.3974866271018982, | |
| "learning_rate": 1.3962678375411635e-05, | |
| "loss": 1.448, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 4.510933703575148, | |
| "grad_norm": 0.4241460859775543, | |
| "learning_rate": 1.3413830954994514e-05, | |
| "loss": 1.4241, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 4.8580354043734815, | |
| "grad_norm": 0.4666702151298523, | |
| "learning_rate": 1.2864983534577389e-05, | |
| "loss": 1.4085, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 1.254936695098877, | |
| "eval_runtime": 78.5765, | |
| "eval_samples_per_second": 146.66, | |
| "eval_steps_per_second": 4.594, | |
| "step": 7205 | |
| }, | |
| { | |
| "epoch": 5.204790003471017, | |
| "grad_norm": 0.4225795567035675, | |
| "learning_rate": 1.2316136114160264e-05, | |
| "loss": 1.395, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 5.551891704269351, | |
| "grad_norm": 0.42448151111602783, | |
| "learning_rate": 1.176728869374314e-05, | |
| "loss": 1.3938, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 5.898993405067685, | |
| "grad_norm": 0.47224199771881104, | |
| "learning_rate": 1.1218441273326017e-05, | |
| "loss": 1.3844, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_loss": 1.2364274263381958, | |
| "eval_runtime": 78.7665, | |
| "eval_samples_per_second": 146.306, | |
| "eval_steps_per_second": 4.583, | |
| "step": 8646 | |
| }, | |
| { | |
| "epoch": 6.24574800416522, | |
| "grad_norm": 0.3676798939704895, | |
| "learning_rate": 1.0669593852908892e-05, | |
| "loss": 1.3668, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 6.592849704963554, | |
| "grad_norm": 0.37000757455825806, | |
| "learning_rate": 1.0120746432491769e-05, | |
| "loss": 1.3629, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 6.939951405761888, | |
| "grad_norm": 0.45095133781433105, | |
| "learning_rate": 9.571899012074644e-06, | |
| "loss": 1.3647, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_loss": 1.222563624382019, | |
| "eval_runtime": 78.5655, | |
| "eval_samples_per_second": 146.68, | |
| "eval_steps_per_second": 4.595, | |
| "step": 10087 | |
| }, | |
| { | |
| "epoch": 7.286706004859424, | |
| "grad_norm": 0.4932011365890503, | |
| "learning_rate": 9.02305159165752e-06, | |
| "loss": 1.358, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 7.633807705657758, | |
| "grad_norm": 0.36681583523750305, | |
| "learning_rate": 8.474204171240396e-06, | |
| "loss": 1.3422, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 7.9809094064560915, | |
| "grad_norm": 0.39713332056999207, | |
| "learning_rate": 7.925356750823272e-06, | |
| "loss": 1.3442, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_loss": 1.214061975479126, | |
| "eval_runtime": 78.596, | |
| "eval_samples_per_second": 146.623, | |
| "eval_steps_per_second": 4.593, | |
| "step": 11528 | |
| }, | |
| { | |
| "epoch": 8.327664005553627, | |
| "grad_norm": 0.4920157194137573, | |
| "learning_rate": 7.376509330406147e-06, | |
| "loss": 1.3358, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 8.67476570635196, | |
| "grad_norm": 0.4292038083076477, | |
| "learning_rate": 6.827661909989024e-06, | |
| "loss": 1.3373, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_loss": 1.2043074369430542, | |
| "eval_runtime": 78.6255, | |
| "eval_samples_per_second": 146.568, | |
| "eval_steps_per_second": 4.591, | |
| "step": 12969 | |
| }, | |
| { | |
| "epoch": 9.021520305449497, | |
| "grad_norm": 0.390511691570282, | |
| "learning_rate": 6.278814489571899e-06, | |
| "loss": 1.3305, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 9.368622006247831, | |
| "grad_norm": 0.42470699548721313, | |
| "learning_rate": 5.729967069154776e-06, | |
| "loss": 1.3294, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 9.715723707046164, | |
| "grad_norm": 0.4358290135860443, | |
| "learning_rate": 5.181119648737652e-06, | |
| "loss": 1.3182, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_loss": 1.1995124816894531, | |
| "eval_runtime": 78.7492, | |
| "eval_samples_per_second": 146.338, | |
| "eval_steps_per_second": 4.584, | |
| "step": 14410 | |
| }, | |
| { | |
| "epoch": 10.0624783061437, | |
| "grad_norm": 0.37242305278778076, | |
| "learning_rate": 4.632272228320528e-06, | |
| "loss": 1.3192, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 10.409580006942035, | |
| "grad_norm": 0.4692925810813904, | |
| "learning_rate": 4.0834248079034035e-06, | |
| "loss": 1.3177, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 10.756681707740368, | |
| "grad_norm": 0.42067646980285645, | |
| "learning_rate": 3.5345773874862793e-06, | |
| "loss": 1.3169, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_loss": 1.1953924894332886, | |
| "eval_runtime": 78.6673, | |
| "eval_samples_per_second": 146.49, | |
| "eval_steps_per_second": 4.589, | |
| "step": 15851 | |
| }, | |
| { | |
| "epoch": 11.103436306837903, | |
| "grad_norm": 0.4010460674762726, | |
| "learning_rate": 2.9857299670691548e-06, | |
| "loss": 1.3145, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 11.450538007636238, | |
| "grad_norm": 0.4176557660102844, | |
| "learning_rate": 2.436882546652031e-06, | |
| "loss": 1.3063, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 11.797639708434572, | |
| "grad_norm": 0.3621806800365448, | |
| "learning_rate": 1.888035126234907e-06, | |
| "loss": 1.3195, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_loss": 1.1941872835159302, | |
| "eval_runtime": 78.6747, | |
| "eval_samples_per_second": 146.477, | |
| "eval_steps_per_second": 4.589, | |
| "step": 17292 | |
| }, | |
| { | |
| "epoch": 12.144394307532107, | |
| "grad_norm": 0.41595444083213806, | |
| "learning_rate": 1.3391877058177828e-06, | |
| "loss": 1.3222, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 12.49149600833044, | |
| "grad_norm": 0.40309804677963257, | |
| "learning_rate": 7.903402854006587e-07, | |
| "loss": 1.3122, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 12.838597709128775, | |
| "grad_norm": 0.4489055871963501, | |
| "learning_rate": 2.414928649835346e-07, | |
| "loss": 1.3016, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 12.991322457480042, | |
| "eval_loss": 1.19288170337677, | |
| "eval_runtime": 78.6526, | |
| "eval_samples_per_second": 146.518, | |
| "eval_steps_per_second": 4.59, | |
| "step": 18720 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 18720, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 13, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 4.070590055763149e+17, | |
| "train_batch_size": 32, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |