{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9956458635703918, "eval_steps": 500, "global_step": 516, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05805515239477504, "grad_norm": 0.600770890712738, "learning_rate": 2.9999908009303982e-05, "loss": 2.8908, "step": 10 }, { "epoch": 0.11611030478955008, "grad_norm": 0.3810243308544159, "learning_rate": 2.998887064264453e-05, "loss": 2.6786, "step": 20 }, { "epoch": 0.1741654571843251, "grad_norm": 0.2603137493133545, "learning_rate": 2.9959452370912904e-05, "loss": 2.5845, "step": 30 }, { "epoch": 0.23222060957910015, "grad_norm": 0.2289140671491623, "learning_rate": 2.9911693281561532e-05, "loss": 2.4006, "step": 40 }, { "epoch": 0.29027576197387517, "grad_norm": 0.18332642316818237, "learning_rate": 2.9845658454559736e-05, "loss": 2.2503, "step": 50 }, { "epoch": 0.3483309143686502, "grad_norm": 0.21847300231456757, "learning_rate": 2.9761437873711118e-05, "loss": 2.0564, "step": 60 }, { "epoch": 0.40638606676342526, "grad_norm": 0.2699638307094574, "learning_rate": 2.9659146304035112e-05, "loss": 1.8541, "step": 70 }, { "epoch": 0.4644412191582003, "grad_norm": 0.17481808364391327, "learning_rate": 2.953892313537998e-05, "loss": 1.5994, "step": 80 }, { "epoch": 0.5224963715529753, "grad_norm": 0.12553584575653076, "learning_rate": 2.9400932192480125e-05, "loss": 1.4072, "step": 90 }, { "epoch": 0.5805515239477503, "grad_norm": 0.12803253531455994, "learning_rate": 2.9245361511716764e-05, "loss": 1.3509, "step": 100 }, { "epoch": 0.6386066763425254, "grad_norm": 0.0883030965924263, "learning_rate": 2.907242308488601e-05, "loss": 1.3609, "step": 110 }, { "epoch": 0.6966618287373004, "grad_norm": 0.08672440052032471, "learning_rate": 2.8882352570323616e-05, "loss": 1.3838, "step": 120 }, { "epoch": 0.7547169811320755, "grad_norm": 0.08096282929182053, "learning_rate": 2.8675408971779997e-05, "loss": 1.3488, "step": 130 }, { "epoch": 0.8127721335268505, "grad_norm": 0.08677352964878082, "learning_rate": 2.8451874285483073e-05, "loss": 1.3902, "step": 140 }, { "epoch": 0.8708272859216255, "grad_norm": 0.06741151958703995, "learning_rate": 2.8212053115869935e-05, "loss": 1.319, "step": 150 }, { "epoch": 0.9288824383164006, "grad_norm": 0.0820663794875145, "learning_rate": 2.7956272260510948e-05, "loss": 1.3725, "step": 160 }, { "epoch": 0.9869375907111756, "grad_norm": 0.06830213963985443, "learning_rate": 2.768488026479187e-05, "loss": 1.3809, "step": 170 }, { "epoch": 1.0449927431059507, "grad_norm": 0.08063437789678574, "learning_rate": 2.7398246946960864e-05, "loss": 1.286, "step": 180 }, { "epoch": 1.1030478955007257, "grad_norm": 0.07576021552085876, "learning_rate": 2.7096762894187533e-05, "loss": 1.3281, "step": 190 }, { "epoch": 1.1611030478955007, "grad_norm": 0.06808394193649292, "learning_rate": 2.678083893032079e-05, "loss": 1.3432, "step": 200 }, { "epoch": 1.2191582002902757, "grad_norm": 0.06429944187402725, "learning_rate": 2.6450905556070718e-05, "loss": 1.3108, "step": 210 }, { "epoch": 1.2772133526850509, "grad_norm": 0.07570013403892517, "learning_rate": 2.610741236237734e-05, "loss": 1.2846, "step": 220 }, { "epoch": 1.3352685050798259, "grad_norm": 0.0730566531419754, "learning_rate": 2.5750827417765714e-05, "loss": 1.2743, "step": 230 }, { "epoch": 1.3933236574746009, "grad_norm": 0.07746976613998413, "learning_rate": 2.5381636630522056e-05, "loss": 1.2424, "step": 240 }, { "epoch": 1.4513788098693758, "grad_norm": 0.063927061855793, "learning_rate": 2.50003430865602e-05, "loss": 1.3057, "step": 250 }, { "epoch": 1.509433962264151, "grad_norm": 0.06818998605012894, "learning_rate": 2.4607466363880555e-05, "loss": 1.2931, "step": 260 }, { "epoch": 1.567489114658926, "grad_norm": 0.07596269994974136, "learning_rate": 2.4203541824555744e-05, "loss": 1.3029, "step": 270 }, { "epoch": 1.625544267053701, "grad_norm": 0.07179366797208786, "learning_rate": 2.378911988520776e-05, "loss": 1.2844, "step": 280 }, { "epoch": 1.683599419448476, "grad_norm": 0.06046036258339882, "learning_rate": 2.33647652669707e-05, "loss": 1.3078, "step": 290 }, { "epoch": 1.741654571843251, "grad_norm": 0.07098483294248581, "learning_rate": 2.2931056225961137e-05, "loss": 1.3307, "step": 300 }, { "epoch": 1.799709724238026, "grad_norm": 0.07308895885944366, "learning_rate": 2.2488583765304748e-05, "loss": 1.2819, "step": 310 }, { "epoch": 1.8577648766328012, "grad_norm": 0.06851565092802048, "learning_rate": 2.203795082979298e-05, "loss": 1.267, "step": 320 }, { "epoch": 1.9158200290275762, "grad_norm": 0.07384185492992401, "learning_rate": 2.1579771484267097e-05, "loss": 1.2629, "step": 330 }, { "epoch": 1.9738751814223512, "grad_norm": 0.0664062350988388, "learning_rate": 2.111467007684928e-05, "loss": 1.2841, "step": 340 }, { "epoch": 2.0319303338171264, "grad_norm": 0.07113119959831238, "learning_rate": 2.0643280388160994e-05, "loss": 1.2779, "step": 350 }, { "epoch": 2.0899854862119014, "grad_norm": 0.07706930488348007, "learning_rate": 2.0166244767687922e-05, "loss": 1.2854, "step": 360 }, { "epoch": 2.1480406386066764, "grad_norm": 0.06886892020702362, "learning_rate": 1.968421325846838e-05, "loss": 1.2983, "step": 370 }, { "epoch": 2.2060957910014514, "grad_norm": 0.07846487313508987, "learning_rate": 1.919784271129792e-05, "loss": 1.286, "step": 380 }, { "epoch": 2.2641509433962264, "grad_norm": 0.09121862053871155, "learning_rate": 1.8707795889657207e-05, "loss": 1.2774, "step": 390 }, { "epoch": 2.3222060957910013, "grad_norm": 0.07675393670797348, "learning_rate": 1.8214740566582862e-05, "loss": 1.2635, "step": 400 }, { "epoch": 2.3802612481857763, "grad_norm": 0.06239638477563858, "learning_rate": 1.7719348614711877e-05, "loss": 1.2833, "step": 410 }, { "epoch": 2.4383164005805513, "grad_norm": 0.0633472129702568, "learning_rate": 1.7222295090739667e-05, "loss": 1.2657, "step": 420 }, { "epoch": 2.4963715529753268, "grad_norm": 0.0839238315820694, "learning_rate": 1.672425731553926e-05, "loss": 1.2424, "step": 430 }, { "epoch": 2.5544267053701017, "grad_norm": 0.061606671661138535, "learning_rate": 1.6225913951195186e-05, "loss": 1.2802, "step": 440 }, { "epoch": 2.6124818577648767, "grad_norm": 0.07224655896425247, "learning_rate": 1.5727944076209725e-05, "loss": 1.2556, "step": 450 }, { "epoch": 2.6705370101596517, "grad_norm": 0.07351450622081757, "learning_rate": 1.5231026260141692e-05, "loss": 1.292, "step": 460 }, { "epoch": 2.7285921625544267, "grad_norm": 0.06462671607732773, "learning_rate": 1.4735837638938767e-05, "loss": 1.2588, "step": 470 }, { "epoch": 2.7866473149492017, "grad_norm": 0.0673007071018219, "learning_rate": 1.4243052992223376e-05, "loss": 1.2889, "step": 480 }, { "epoch": 2.8447024673439767, "grad_norm": 0.06969607621431351, "learning_rate": 1.3753343823789447e-05, "loss": 1.2591, "step": 490 }, { "epoch": 2.9027576197387517, "grad_norm": 0.07535108923912048, "learning_rate": 1.3267377446563042e-05, "loss": 1.3026, "step": 500 }, { "epoch": 2.9608127721335267, "grad_norm": 0.06275378167629242, "learning_rate": 1.2785816073273818e-05, "loss": 1.2005, "step": 510 } ], "logging_steps": 10, "max_steps": 860, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.381176202297344e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }