| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9994672349493873, | |
| "eval_steps": 500, | |
| "global_step": 938, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 4.4688, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 5.7051, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 4.6101, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 5.1161, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 4.7314, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 5.4174, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 5.3351, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 267.06263341170535, | |
| "learning_rate": 2.1276595744680852e-07, | |
| "loss": 4.3761, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 256.6399066161345, | |
| "learning_rate": 4.2553191489361704e-07, | |
| "loss": 4.911, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 259.8979037908876, | |
| "learning_rate": 6.382978723404255e-07, | |
| "loss": 4.5067, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 206.34210154966675, | |
| "learning_rate": 8.510638297872341e-07, | |
| "loss": 4.1617, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 206.34210154966675, | |
| "learning_rate": 8.510638297872341e-07, | |
| "loss": 4.7024, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 104.25967551689651, | |
| "learning_rate": 1.0638297872340427e-06, | |
| "loss": 2.8783, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 104.25967551689651, | |
| "learning_rate": 1.0638297872340427e-06, | |
| "loss": 3.6247, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 301.56418557125875, | |
| "learning_rate": 1.276595744680851e-06, | |
| "loss": 3.8274, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 158.0102677208041, | |
| "learning_rate": 1.4893617021276596e-06, | |
| "loss": 2.5054, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 144.94261605863426, | |
| "learning_rate": 1.7021276595744682e-06, | |
| "loss": 2.6479, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 77.83778037631348, | |
| "learning_rate": 1.9148936170212767e-06, | |
| "loss": 1.7461, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 110.09097679168819, | |
| "learning_rate": 2.1276595744680853e-06, | |
| "loss": 2.005, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 64.67189678957659, | |
| "learning_rate": 2.340425531914894e-06, | |
| "loss": 1.5011, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 85.70086135587184, | |
| "learning_rate": 2.553191489361702e-06, | |
| "loss": 1.4231, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 51.55499709016579, | |
| "learning_rate": 2.765957446808511e-06, | |
| "loss": 1.3443, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 39.458031493922036, | |
| "learning_rate": 2.978723404255319e-06, | |
| "loss": 1.1782, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 30.718044498953375, | |
| "learning_rate": 3.191489361702128e-06, | |
| "loss": 1.0572, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 54.03307642588333, | |
| "learning_rate": 3.4042553191489363e-06, | |
| "loss": 1.0557, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 45.15061216435745, | |
| "learning_rate": 3.6170212765957453e-06, | |
| "loss": 0.9468, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 85.20725035566016, | |
| "learning_rate": 3.8297872340425535e-06, | |
| "loss": 1.1611, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 61.03528945297965, | |
| "learning_rate": 4.042553191489362e-06, | |
| "loss": 0.9764, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 32.57984313596529, | |
| "learning_rate": 4.255319148936171e-06, | |
| "loss": 0.9542, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 46.71878018713139, | |
| "learning_rate": 4.468085106382979e-06, | |
| "loss": 1.0575, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 28.228382153829948, | |
| "learning_rate": 4.680851063829788e-06, | |
| "loss": 0.8927, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 32.305084533281175, | |
| "learning_rate": 4.893617021276596e-06, | |
| "loss": 0.9158, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 38.80822759339413, | |
| "learning_rate": 5.106382978723404e-06, | |
| "loss": 1.0127, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 16.765291089343936, | |
| "learning_rate": 5.319148936170213e-06, | |
| "loss": 0.8579, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 46.21887734855352, | |
| "learning_rate": 5.531914893617022e-06, | |
| "loss": 1.0595, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 33.51542422627954, | |
| "learning_rate": 5.744680851063831e-06, | |
| "loss": 0.7873, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 20.62520698421828, | |
| "learning_rate": 5.957446808510638e-06, | |
| "loss": 0.7463, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 32.86540760673144, | |
| "learning_rate": 6.170212765957447e-06, | |
| "loss": 0.9393, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 16.78339589062748, | |
| "learning_rate": 6.382978723404256e-06, | |
| "loss": 0.7036, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 20.638888081681362, | |
| "learning_rate": 6.595744680851064e-06, | |
| "loss": 0.8178, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 24.957942318493615, | |
| "learning_rate": 6.808510638297873e-06, | |
| "loss": 0.8192, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 18.753926303171866, | |
| "learning_rate": 7.021276595744682e-06, | |
| "loss": 0.8053, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 20.380738488767815, | |
| "learning_rate": 7.234042553191491e-06, | |
| "loss": 0.7775, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 15.358291004654351, | |
| "learning_rate": 7.446808510638298e-06, | |
| "loss": 0.768, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 28.01853531392082, | |
| "learning_rate": 7.659574468085107e-06, | |
| "loss": 0.6495, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 90.12246442870764, | |
| "learning_rate": 7.872340425531916e-06, | |
| "loss": 1.1254, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 24.048650390369733, | |
| "learning_rate": 8.085106382978723e-06, | |
| "loss": 0.8769, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 79.51846666706089, | |
| "learning_rate": 8.297872340425532e-06, | |
| "loss": 1.2554, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 20.341370812622667, | |
| "learning_rate": 8.510638297872341e-06, | |
| "loss": 0.7153, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 48.50233018085147, | |
| "learning_rate": 8.72340425531915e-06, | |
| "loss": 1.0358, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 27.71805876780665, | |
| "learning_rate": 8.936170212765958e-06, | |
| "loss": 0.7539, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 44.56567439721268, | |
| "learning_rate": 9.148936170212767e-06, | |
| "loss": 1.0363, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 43.96614589042312, | |
| "learning_rate": 9.361702127659576e-06, | |
| "loss": 1.0823, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 36.336438544851525, | |
| "learning_rate": 9.574468085106385e-06, | |
| "loss": 0.9189, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 20.360154033813146, | |
| "learning_rate": 9.787234042553192e-06, | |
| "loss": 0.7138, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 16.415639113466074, | |
| "learning_rate": 1e-05, | |
| "loss": 0.8165, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 17.154150115372023, | |
| "learning_rate": 1.0212765957446808e-05, | |
| "loss": 0.8003, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 23.485978008365016, | |
| "learning_rate": 1.0425531914893619e-05, | |
| "loss": 0.8017, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 18.921319860871982, | |
| "learning_rate": 1.0638297872340426e-05, | |
| "loss": 0.6803, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 28.458932361446607, | |
| "learning_rate": 1.0851063829787233e-05, | |
| "loss": 0.9205, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 16.224593364238817, | |
| "learning_rate": 1.1063829787234044e-05, | |
| "loss": 0.7286, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 30.78498419558807, | |
| "learning_rate": 1.1276595744680851e-05, | |
| "loss": 0.9564, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 20.29389328745947, | |
| "learning_rate": 1.1489361702127662e-05, | |
| "loss": 0.8442, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 18.23986171950765, | |
| "learning_rate": 1.170212765957447e-05, | |
| "loss": 0.7971, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 13.607742168916245, | |
| "learning_rate": 1.1914893617021277e-05, | |
| "loss": 0.7479, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 15.589517783052612, | |
| "learning_rate": 1.2127659574468087e-05, | |
| "loss": 0.6971, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 11.730568689593209, | |
| "learning_rate": 1.2340425531914895e-05, | |
| "loss": 0.6136, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 17.919191925814953, | |
| "learning_rate": 1.2553191489361702e-05, | |
| "loss": 0.6786, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 46.93984753189794, | |
| "learning_rate": 1.2765957446808513e-05, | |
| "loss": 1.1316, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 32.097199934080784, | |
| "learning_rate": 1.297872340425532e-05, | |
| "loss": 0.8888, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 23.37535729938417, | |
| "learning_rate": 1.3191489361702127e-05, | |
| "loss": 0.7891, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 16.127458600501658, | |
| "learning_rate": 1.3404255319148938e-05, | |
| "loss": 0.5976, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 13.056670968499086, | |
| "learning_rate": 1.3617021276595745e-05, | |
| "loss": 0.6602, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 18.80577219024569, | |
| "learning_rate": 1.3829787234042556e-05, | |
| "loss": 0.7343, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 18.0606093200878, | |
| "learning_rate": 1.4042553191489363e-05, | |
| "loss": 0.7229, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 13.62137898397275, | |
| "learning_rate": 1.425531914893617e-05, | |
| "loss": 0.7897, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 10.783139067640777, | |
| "learning_rate": 1.4468085106382981e-05, | |
| "loss": 0.5573, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 28.888686230080594, | |
| "learning_rate": 1.4680851063829789e-05, | |
| "loss": 0.7468, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 24.696181606034965, | |
| "learning_rate": 1.4893617021276596e-05, | |
| "loss": 0.6625, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 32.68489697451064, | |
| "learning_rate": 1.5106382978723407e-05, | |
| "loss": 1.0193, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 20.801300816024295, | |
| "learning_rate": 1.5319148936170214e-05, | |
| "loss": 0.6352, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 10.55513433584252, | |
| "learning_rate": 1.5531914893617023e-05, | |
| "loss": 0.7007, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 15.716140126443527, | |
| "learning_rate": 1.5744680851063832e-05, | |
| "loss": 0.7187, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 21.85726232694579, | |
| "learning_rate": 1.595744680851064e-05, | |
| "loss": 0.7349, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 20.964587712026038, | |
| "learning_rate": 1.6170212765957446e-05, | |
| "loss": 0.7143, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 8.250030396732411, | |
| "learning_rate": 1.6382978723404255e-05, | |
| "loss": 0.6892, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 21.20006063299826, | |
| "learning_rate": 1.6595744680851064e-05, | |
| "loss": 0.8109, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 19.09608912821583, | |
| "learning_rate": 1.6808510638297873e-05, | |
| "loss": 0.8082, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 20.52502129407863, | |
| "learning_rate": 1.7021276595744682e-05, | |
| "loss": 0.6783, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 39.20021747253875, | |
| "learning_rate": 1.723404255319149e-05, | |
| "loss": 0.9711, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 31.095085040142717, | |
| "learning_rate": 1.74468085106383e-05, | |
| "loss": 0.8283, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 37.26207899740689, | |
| "learning_rate": 1.765957446808511e-05, | |
| "loss": 1.0039, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 16.273798667976465, | |
| "learning_rate": 1.7872340425531915e-05, | |
| "loss": 0.6814, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 32.374520273410155, | |
| "learning_rate": 1.8085106382978724e-05, | |
| "loss": 0.9674, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 34.748891951713134, | |
| "learning_rate": 1.8297872340425533e-05, | |
| "loss": 1.1996, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 26.013510708052237, | |
| "learning_rate": 1.8510638297872342e-05, | |
| "loss": 0.9372, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 19.99677186530767, | |
| "learning_rate": 1.872340425531915e-05, | |
| "loss": 0.8718, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 16.62199424485097, | |
| "learning_rate": 1.893617021276596e-05, | |
| "loss": 0.9384, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 20.111532243609883, | |
| "learning_rate": 1.914893617021277e-05, | |
| "loss": 0.9813, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 15.191733826484391, | |
| "learning_rate": 1.9361702127659575e-05, | |
| "loss": 0.8069, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 23.926673285223416, | |
| "learning_rate": 1.9574468085106384e-05, | |
| "loss": 0.8308, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 23.218285916806728, | |
| "learning_rate": 1.9787234042553193e-05, | |
| "loss": 0.9754, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 17.76104315857583, | |
| "learning_rate": 2e-05, | |
| "loss": 0.8006, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 20.836820707421783, | |
| "learning_rate": 1.9999930723752516e-05, | |
| "loss": 0.8543, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 17.0438780130028, | |
| "learning_rate": 1.9999722895969904e-05, | |
| "loss": 0.8443, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 22.36760800612908, | |
| "learning_rate": 1.9999376519531668e-05, | |
| "loss": 0.8841, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 11.874378681555248, | |
| "learning_rate": 1.999889159923694e-05, | |
| "loss": 0.6501, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 17.245286092714633, | |
| "learning_rate": 1.9998268141804412e-05, | |
| "loss": 0.8119, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 16.693402688517093, | |
| "learning_rate": 1.9997506155872246e-05, | |
| "loss": 0.7683, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 18.100428165378997, | |
| "learning_rate": 1.999660565199794e-05, | |
| "loss": 0.8259, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 10.547692312079072, | |
| "learning_rate": 1.9995566642658208e-05, | |
| "loss": 0.6651, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 10.19017430503209, | |
| "learning_rate": 1.9994389142248775e-05, | |
| "loss": 0.7619, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 13.012636488876366, | |
| "learning_rate": 1.999307316708421e-05, | |
| "loss": 0.6931, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 13.672739554862506, | |
| "learning_rate": 1.9991618735397674e-05, | |
| "loss": 0.7014, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 11.268690129927412, | |
| "learning_rate": 1.9990025867340683e-05, | |
| "loss": 0.7041, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 13.835706135922516, | |
| "learning_rate": 1.998829458498282e-05, | |
| "loss": 0.7568, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 12.580781013528279, | |
| "learning_rate": 1.998642491231143e-05, | |
| "loss": 0.7155, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 11.88807897059113, | |
| "learning_rate": 1.9984416875231303e-05, | |
| "loss": 0.7374, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 15.58619342416967, | |
| "learning_rate": 1.9982270501564286e-05, | |
| "loss": 0.8355, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 14.670349037930592, | |
| "learning_rate": 1.9979985821048925e-05, | |
| "loss": 0.623, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 17.596509727191997, | |
| "learning_rate": 1.997756286534004e-05, | |
| "loss": 0.7664, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 16.388736277383142, | |
| "learning_rate": 1.9975001668008282e-05, | |
| "loss": 0.6749, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 11.99481967344947, | |
| "learning_rate": 1.9972302264539686e-05, | |
| "loss": 0.6724, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 20.53435074276929, | |
| "learning_rate": 1.9969464692335153e-05, | |
| "loss": 0.7242, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 11.475488488305645, | |
| "learning_rate": 1.996648899070996e-05, | |
| "loss": 0.7357, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 25.553564042295182, | |
| "learning_rate": 1.996337520089319e-05, | |
| "loss": 0.7927, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 8.109394353226076, | |
| "learning_rate": 1.9960123366027187e-05, | |
| "loss": 0.7367, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 8.254315086779002, | |
| "learning_rate": 1.995673353116692e-05, | |
| "loss": 0.5336, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 28.75262073669196, | |
| "learning_rate": 1.995320574327941e-05, | |
| "loss": 0.9592, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 22.09118124557931, | |
| "learning_rate": 1.9949540051243032e-05, | |
| "loss": 0.6935, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 19.95406582271437, | |
| "learning_rate": 1.9945736505846866e-05, | |
| "loss": 0.8254, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 13.525596071741782, | |
| "learning_rate": 1.9941795159789983e-05, | |
| "loss": 0.6531, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 9.891408326838208, | |
| "learning_rate": 1.9937716067680712e-05, | |
| "loss": 0.7162, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 23.11321488500328, | |
| "learning_rate": 1.9933499286035896e-05, | |
| "loss": 0.7962, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 14.482317722843773, | |
| "learning_rate": 1.9929144873280092e-05, | |
| "loss": 0.7631, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 19.49882212964946, | |
| "learning_rate": 1.992465288974478e-05, | |
| "loss": 0.7617, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 14.59122368824527, | |
| "learning_rate": 1.992002339766751e-05, | |
| "loss": 0.6722, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 14.253058587354259, | |
| "learning_rate": 1.9915256461191054e-05, | |
| "loss": 0.6869, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 22.646528058916722, | |
| "learning_rate": 1.99103521463625e-05, | |
| "loss": 0.858, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 19.599804799046947, | |
| "learning_rate": 1.9905310521132353e-05, | |
| "loss": 0.8475, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 7.459796818444622, | |
| "learning_rate": 1.9900131655353597e-05, | |
| "loss": 0.6509, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 14.795548081947178, | |
| "learning_rate": 1.9894815620780705e-05, | |
| "loss": 0.781, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 19.287512528853632, | |
| "learning_rate": 1.9889362491068658e-05, | |
| "loss": 0.7946, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 11.780467440280454, | |
| "learning_rate": 1.9883772341771937e-05, | |
| "loss": 0.6937, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 17.590934140385546, | |
| "learning_rate": 1.9878045250343445e-05, | |
| "loss": 0.8808, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 16.151381586194, | |
| "learning_rate": 1.9872181296133473e-05, | |
| "loss": 0.8076, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 10.326349558310788, | |
| "learning_rate": 1.986618056038856e-05, | |
| "loss": 0.8052, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 9.391393835716272, | |
| "learning_rate": 1.9860043126250407e-05, | |
| "loss": 0.7323, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 13.654055394779414, | |
| "learning_rate": 1.9853769078754685e-05, | |
| "loss": 0.7861, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 13.20685189778904, | |
| "learning_rate": 1.9847358504829896e-05, | |
| "loss": 0.7136, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 8.948428090341855, | |
| "learning_rate": 1.9840811493296134e-05, | |
| "loss": 0.6935, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 9.441488025541567, | |
| "learning_rate": 1.9834128134863885e-05, | |
| "loss": 0.7222, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 12.322015322175998, | |
| "learning_rate": 1.982730852213274e-05, | |
| "loss": 0.7606, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 6.325468800857368, | |
| "learning_rate": 1.982035274959014e-05, | |
| "loss": 0.5735, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 6.406666981105762, | |
| "learning_rate": 1.9813260913610048e-05, | |
| "loss": 0.5348, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 17.612437547183127, | |
| "learning_rate": 1.9806033112451616e-05, | |
| "loss": 0.6477, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 18.78091352132794, | |
| "learning_rate": 1.9798669446257844e-05, | |
| "loss": 0.7878, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 15.546105512429289, | |
| "learning_rate": 1.979117001705415e-05, | |
| "loss": 0.852, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 16.923786294030265, | |
| "learning_rate": 1.9783534928747006e-05, | |
| "loss": 0.7754, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 13.076565312797653, | |
| "learning_rate": 1.9775764287122464e-05, | |
| "loss": 0.7425, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 14.51221649438122, | |
| "learning_rate": 1.9767858199844697e-05, | |
| "loss": 0.693, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 16.76943181344085, | |
| "learning_rate": 1.9759816776454527e-05, | |
| "loss": 0.8376, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 10.230125019828748, | |
| "learning_rate": 1.9751640128367872e-05, | |
| "loss": 0.7238, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 24.10748192789658, | |
| "learning_rate": 1.9743328368874237e-05, | |
| "loss": 0.8927, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 24.836153220553506, | |
| "learning_rate": 1.973488161313512e-05, | |
| "loss": 0.8796, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 15.039263981577998, | |
| "learning_rate": 1.972629997818243e-05, | |
| "loss": 0.7177, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 7.166311769976143, | |
| "learning_rate": 1.9717583582916862e-05, | |
| "loss": 0.7274, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 13.36802727443076, | |
| "learning_rate": 1.970873254810625e-05, | |
| "loss": 0.7425, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 12.26776202136434, | |
| "learning_rate": 1.969974699638388e-05, | |
| "loss": 0.7671, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 14.871789262668347, | |
| "learning_rate": 1.969062705224682e-05, | |
| "loss": 0.8251, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 13.01131507667653, | |
| "learning_rate": 1.968137284205417e-05, | |
| "loss": 0.8131, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 8.692440461337334, | |
| "learning_rate": 1.967198449402532e-05, | |
| "loss": 0.7195, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 15.307857384021617, | |
| "learning_rate": 1.966246213823818e-05, | |
| "loss": 0.7753, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 22.444147986646815, | |
| "learning_rate": 1.9652805906627356e-05, | |
| "loss": 0.9884, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 14.639613594215758, | |
| "learning_rate": 1.9643015932982355e-05, | |
| "loss": 0.7357, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 10.34217870581945, | |
| "learning_rate": 1.9633092352945698e-05, | |
| "loss": 0.7147, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 12.490262046586464, | |
| "learning_rate": 1.9623035304011062e-05, | |
| "loss": 0.7384, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 15.433175693682523, | |
| "learning_rate": 1.9612844925521375e-05, | |
| "loss": 0.7652, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 6.045080843254174, | |
| "learning_rate": 1.960252135866687e-05, | |
| "loss": 0.8282, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 8.861252768013037, | |
| "learning_rate": 1.9592064746483143e-05, | |
| "loss": 0.7317, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 7.8221459133117435, | |
| "learning_rate": 1.9581475233849165e-05, | |
| "loss": 0.5407, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 12.879262754926978, | |
| "learning_rate": 1.957075296748527e-05, | |
| "loss": 0.6644, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 5.635984955428702, | |
| "learning_rate": 1.9559898095951137e-05, | |
| "loss": 0.6606, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 7.056126169759745, | |
| "learning_rate": 1.9548910769643724e-05, | |
| "loss": 0.5217, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 5.817340227592049, | |
| "learning_rate": 1.953779114079517e-05, | |
| "loss": 0.694, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 11.240599667838348, | |
| "learning_rate": 1.9526539363470715e-05, | |
| "loss": 0.6613, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 6.377305729246939, | |
| "learning_rate": 1.9515155593566536e-05, | |
| "loss": 0.6127, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 12.70631731046457, | |
| "learning_rate": 1.9503639988807605e-05, | |
| "loss": 0.648, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 7.740556879132694, | |
| "learning_rate": 1.9491992708745502e-05, | |
| "loss": 0.6363, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 9.838244865368074, | |
| "learning_rate": 1.94802139147562e-05, | |
| "loss": 0.5207, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 12.292539349351499, | |
| "learning_rate": 1.946830377003782e-05, | |
| "loss": 0.7129, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 9.434980765550153, | |
| "learning_rate": 1.94562624396084e-05, | |
| "loss": 0.623, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 17.774682137642376, | |
| "learning_rate": 1.9444090090303567e-05, | |
| "loss": 0.8399, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 9.97338739760112, | |
| "learning_rate": 1.9431786890774265e-05, | |
| "loss": 0.5893, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 11.708748373782319, | |
| "learning_rate": 1.941935301148439e-05, | |
| "loss": 0.7569, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 10.640144063322005, | |
| "learning_rate": 1.9406788624708438e-05, | |
| "loss": 0.694, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 14.454447375580173, | |
| "learning_rate": 1.939409390452913e-05, | |
| "loss": 0.7807, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 10.35715254572469, | |
| "learning_rate": 1.9381269026834975e-05, | |
| "loss": 0.6799, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 6.919008432462478, | |
| "learning_rate": 1.9368314169317858e-05, | |
| "loss": 0.6883, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 9.137624019639443, | |
| "learning_rate": 1.935522951147056e-05, | |
| "loss": 0.5919, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 15.35731927734825, | |
| "learning_rate": 1.9342015234584277e-05, | |
| "loss": 0.5982, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 28.1752923954513, | |
| "learning_rate": 1.932867152174612e-05, | |
| "loss": 0.6482, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 28.1752923954513, | |
| "learning_rate": 1.932867152174612e-05, | |
| "loss": 1.748, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 1587.2702484305357, | |
| "learning_rate": 1.9315198557836555e-05, | |
| "loss": 1.6079, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 1587.2702484305357, | |
| "learning_rate": 1.9315198557836555e-05, | |
| "loss": 11.678, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 2434.8463463776825, | |
| "learning_rate": 1.9301596529526856e-05, | |
| "loss": 11.5182, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 2901.204727033937, | |
| "learning_rate": 1.928786562527652e-05, | |
| "loss": 9.7123, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 490.65682411342476, | |
| "learning_rate": 1.9274006035330654e-05, | |
| "loss": 3.1282, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 480.76138720661936, | |
| "learning_rate": 1.9260017951717334e-05, | |
| "loss": 2.4949, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 143.8799125203908, | |
| "learning_rate": 1.9245901568244945e-05, | |
| "loss": 2.1616, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 226.80553179811704, | |
| "learning_rate": 1.9231657080499507e-05, | |
| "loss": 1.9539, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 210.9739319214792, | |
| "learning_rate": 1.921728468584195e-05, | |
| "loss": 1.7932, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 102.69163082309667, | |
| "learning_rate": 1.9202784583405386e-05, | |
| "loss": 1.5288, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 80.79475053684271, | |
| "learning_rate": 1.9188156974092355e-05, | |
| "loss": 1.4608, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 131.37203972527402, | |
| "learning_rate": 1.9173402060572028e-05, | |
| "loss": 1.3191, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 641.6313774584162, | |
| "learning_rate": 1.915852004727742e-05, | |
| "loss": 1.8877, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 535.4790750541629, | |
| "learning_rate": 1.9143511140402532e-05, | |
| "loss": 1.4452, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 548.3152023208836, | |
| "learning_rate": 1.9128375547899518e-05, | |
| "loss": 1.9607, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 568.5307973204539, | |
| "learning_rate": 1.9113113479475784e-05, | |
| "loss": 1.8329, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 247.07029020503853, | |
| "learning_rate": 1.9097725146591103e-05, | |
| "loss": 1.2675, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 483.1919344273096, | |
| "learning_rate": 1.908221076245466e-05, | |
| "loss": 3.1382, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 435.67666262712555, | |
| "learning_rate": 1.9066570542022122e-05, | |
| "loss": 2.9579, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 106.36477945538998, | |
| "learning_rate": 1.905080470199264e-05, | |
| "loss": 1.9584, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 150.21543919990003, | |
| "learning_rate": 1.9034913460805868e-05, | |
| "loss": 1.5164, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 265.0211172938759, | |
| "learning_rate": 1.901889703863891e-05, | |
| "loss": 1.4234, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 135.22687197035, | |
| "learning_rate": 1.90027556574033e-05, | |
| "loss": 1.3242, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 69.09048458856181, | |
| "learning_rate": 1.8986489540741895e-05, | |
| "loss": 1.1573, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 317.48370008905204, | |
| "learning_rate": 1.89700989140258e-05, | |
| "loss": 1.6469, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 178.1220954711751, | |
| "learning_rate": 1.8953584004351243e-05, | |
| "loss": 1.3997, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 58.594800768267845, | |
| "learning_rate": 1.8936945040536413e-05, | |
| "loss": 1.4203, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 100.8540607851877, | |
| "learning_rate": 1.892018225311831e-05, | |
| "loss": 1.3959, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 40.196456568889715, | |
| "learning_rate": 1.8903295874349528e-05, | |
| "loss": 1.3036, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 49.160977991667785, | |
| "learning_rate": 1.8886286138195063e-05, | |
| "loss": 1.2698, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 38.84069117366192, | |
| "learning_rate": 1.8869153280329054e-05, | |
| "loss": 1.1846, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 43.81901383324617, | |
| "learning_rate": 1.885189753813152e-05, | |
| "loss": 1.2088, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 19.950274945570303, | |
| "learning_rate": 1.8834519150685074e-05, | |
| "loss": 1.0781, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 50.927777642645886, | |
| "learning_rate": 1.8817018358771612e-05, | |
| "loss": 1.207, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 45.137241355795545, | |
| "learning_rate": 1.879939540486897e-05, | |
| "loss": 1.1345, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 49.611501553831054, | |
| "learning_rate": 1.8781650533147572e-05, | |
| "loss": 1.1675, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 60.82753635007021, | |
| "learning_rate": 1.876378398946704e-05, | |
| "loss": 1.1098, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 36.20738946945509, | |
| "learning_rate": 1.87457960213728e-05, | |
| "loss": 1.0201, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 46.88562809493131, | |
| "learning_rate": 1.8727686878092638e-05, | |
| "loss": 1.0511, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 38.81307588678313, | |
| "learning_rate": 1.8709456810533248e-05, | |
| "loss": 0.9694, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 29.29930215775943, | |
| "learning_rate": 1.8691106071276763e-05, | |
| "loss": 0.9828, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 54.42807589658664, | |
| "learning_rate": 1.867263491457726e-05, | |
| "loss": 1.1109, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 29.031387361113243, | |
| "learning_rate": 1.8654043596357215e-05, | |
| "loss": 1.0194, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 35.21872276907598, | |
| "learning_rate": 1.8635332374203993e-05, | |
| "loss": 1.0552, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 22.43165286258535, | |
| "learning_rate": 1.8616501507366233e-05, | |
| "loss": 0.9364, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 29.83402427558291, | |
| "learning_rate": 1.85975512567503e-05, | |
| "loss": 1.0535, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 36.54199147143507, | |
| "learning_rate": 1.8578481884916645e-05, | |
| "loss": 0.9772, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 16.06774723588221, | |
| "learning_rate": 1.8559293656076167e-05, | |
| "loss": 0.9044, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 14.242434397235082, | |
| "learning_rate": 1.8539986836086568e-05, | |
| "loss": 0.8656, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 34.06087523793793, | |
| "learning_rate": 1.8520561692448655e-05, | |
| "loss": 0.9776, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 23.65421594444165, | |
| "learning_rate": 1.850101849430264e-05, | |
| "loss": 0.9739, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 53.1622333489428, | |
| "learning_rate": 1.848135751242441e-05, | |
| "loss": 0.9659, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 27.743193122343396, | |
| "learning_rate": 1.8461579019221775e-05, | |
| "loss": 0.9828, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 36.32985964499645, | |
| "learning_rate": 1.8441683288730686e-05, | |
| "loss": 1.0724, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 57.208457798230874, | |
| "learning_rate": 1.8421670596611463e-05, | |
| "loss": 1.1399, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 36.265365003042525, | |
| "learning_rate": 1.840154122014494e-05, | |
| "loss": 1.061, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 50.18323851957246, | |
| "learning_rate": 1.8381295438228656e-05, | |
| "loss": 0.9041, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 23.117360915951355, | |
| "learning_rate": 1.836093353137297e-05, | |
| "loss": 0.974, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 50.588104484420946, | |
| "learning_rate": 1.8340455781697175e-05, | |
| "loss": 0.9803, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 43.25313730795733, | |
| "learning_rate": 1.831986247292561e-05, | |
| "loss": 0.8178, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 25.22535296579869, | |
| "learning_rate": 1.8299153890383705e-05, | |
| "loss": 0.9212, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 24.74441945494526, | |
| "learning_rate": 1.8278330320994035e-05, | |
| "loss": 0.9241, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 24.98014366767532, | |
| "learning_rate": 1.8257392053272345e-05, | |
| "loss": 0.9996, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 30.455334182511248, | |
| "learning_rate": 1.823633937732357e-05, | |
| "loss": 0.8491, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 35.02803006376697, | |
| "learning_rate": 1.8215172584837776e-05, | |
| "loss": 0.9924, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 21.742297963510474, | |
| "learning_rate": 1.8193891969086164e-05, | |
| "loss": 0.9068, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 16.27634444074666, | |
| "learning_rate": 1.817249782491697e-05, | |
| "loss": 0.9532, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 27.206570441621835, | |
| "learning_rate": 1.8150990448751393e-05, | |
| "loss": 0.9585, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 9.683718717075713, | |
| "learning_rate": 1.8129370138579507e-05, | |
| "loss": 0.8514, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 46.207488174236765, | |
| "learning_rate": 1.8107637193956102e-05, | |
| "loss": 0.9744, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 24.195105770115998, | |
| "learning_rate": 1.8085791915996538e-05, | |
| "loss": 0.994, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 33.17832980287206, | |
| "learning_rate": 1.8063834607372603e-05, | |
| "loss": 1.1351, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 18.66629676451458, | |
| "learning_rate": 1.804176557230828e-05, | |
| "loss": 0.7503, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 35.58783412196323, | |
| "learning_rate": 1.8019585116575554e-05, | |
| "loss": 0.8252, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 98.83505787049795, | |
| "learning_rate": 1.799729354749018e-05, | |
| "loss": 0.772, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 21.601957059226525, | |
| "learning_rate": 1.7974891173907406e-05, | |
| "loss": 0.7751, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 25.682884674147235, | |
| "learning_rate": 1.7952378306217705e-05, | |
| "loss": 0.8812, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 14.543550272731224, | |
| "learning_rate": 1.792975525634248e-05, | |
| "loss": 0.8405, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 49.50597218889562, | |
| "learning_rate": 1.790702233772973e-05, | |
| "loss": 0.7983, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 28.92796112130337, | |
| "learning_rate": 1.7884179865349713e-05, | |
| "loss": 0.8288, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 31.306112233207404, | |
| "learning_rate": 1.786122815569058e-05, | |
| "loss": 1.0268, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 23.72305170543349, | |
| "learning_rate": 1.7838167526754002e-05, | |
| "loss": 0.9805, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 30.976885389109945, | |
| "learning_rate": 1.7814998298050744e-05, | |
| "loss": 0.9433, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 9.017186183397232, | |
| "learning_rate": 1.7791720790596242e-05, | |
| "loss": 0.9042, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 24.453663623228078, | |
| "learning_rate": 1.7768335326906182e-05, | |
| "loss": 0.916, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 16.661532775610848, | |
| "learning_rate": 1.774484223099199e-05, | |
| "loss": 0.8476, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 75.49348483174468, | |
| "learning_rate": 1.7721241828356377e-05, | |
| "loss": 0.9214, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 28.038342380284316, | |
| "learning_rate": 1.7697534445988804e-05, | |
| "loss": 0.5674, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 26.918490245230988, | |
| "learning_rate": 1.7673720412360973e-05, | |
| "loss": 0.7465, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 64.73808049621005, | |
| "learning_rate": 1.7649800057422256e-05, | |
| "loss": 0.7832, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 21.87842844322322, | |
| "learning_rate": 1.7625773712595147e-05, | |
| "loss": 0.5816, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 15.934437709078225, | |
| "learning_rate": 1.760164171077064e-05, | |
| "loss": 0.7923, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 50.69489297444858, | |
| "learning_rate": 1.7577404386303646e-05, | |
| "loss": 0.6258, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 21.105815660081543, | |
| "learning_rate": 1.755306207500834e-05, | |
| "loss": 0.7464, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 21.183565369421174, | |
| "learning_rate": 1.7528615114153524e-05, | |
| "loss": 0.6985, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 15.244682840971725, | |
| "learning_rate": 1.750406384245793e-05, | |
| "loss": 0.767, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 10.358767383759252, | |
| "learning_rate": 1.747940860008556e-05, | |
| "loss": 0.6928, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 39.96772160224854, | |
| "learning_rate": 1.7454649728640944e-05, | |
| "loss": 0.6887, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 16.221102491098176, | |
| "learning_rate": 1.742978757116443e-05, | |
| "loss": 0.6359, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 36.830429589303506, | |
| "learning_rate": 1.7404822472127406e-05, | |
| "loss": 0.6717, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 17.661289846973503, | |
| "learning_rate": 1.7379754777427554e-05, | |
| "loss": 0.6609, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 8.91929875914757, | |
| "learning_rate": 1.7354584834384036e-05, | |
| "loss": 0.6851, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 36.39317884774958, | |
| "learning_rate": 1.7329312991732687e-05, | |
| "loss": 0.8089, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 20.66408093673374, | |
| "learning_rate": 1.73039395996212e-05, | |
| "loss": 0.6783, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 13.03603460389346, | |
| "learning_rate": 1.727846500960425e-05, | |
| "loss": 0.7026, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 17.971533560556523, | |
| "learning_rate": 1.725288957463864e-05, | |
| "loss": 0.7989, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 9.851305189485732, | |
| "learning_rate": 1.7227213649078395e-05, | |
| "loss": 0.6135, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 13.9093394062407, | |
| "learning_rate": 1.720143758866988e-05, | |
| "loss": 0.7099, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 15.918012101752101, | |
| "learning_rate": 1.7175561750546833e-05, | |
| "loss": 0.6836, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 9.424349721892472, | |
| "learning_rate": 1.7149586493225453e-05, | |
| "loss": 0.6301, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 15.040725484530588, | |
| "learning_rate": 1.7123512176599413e-05, | |
| "loss": 0.5931, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 10.404277489273346, | |
| "learning_rate": 1.709733916193487e-05, | |
| "loss": 0.5216, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 10.351015057753036, | |
| "learning_rate": 1.7071067811865477e-05, | |
| "loss": 0.6885, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 12.500602158481879, | |
| "learning_rate": 1.704469849038734e-05, | |
| "loss": 0.66, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 9.587643268920083, | |
| "learning_rate": 1.7018231562853987e-05, | |
| "loss": 0.6073, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 13.14738192750954, | |
| "learning_rate": 1.6991667395971306e-05, | |
| "loss": 0.6194, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 11.464802706452456, | |
| "learning_rate": 1.6965006357792454e-05, | |
| "loss": 0.4891, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 6.153912186441177, | |
| "learning_rate": 1.6938248817712767e-05, | |
| "loss": 0.5314, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 13.329406945514984, | |
| "learning_rate": 1.691139514646464e-05, | |
| "loss": 0.7992, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 13.044412872566445, | |
| "learning_rate": 1.6884445716112388e-05, | |
| "loss": 0.7126, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 8.05329073835129, | |
| "learning_rate": 1.685740090004709e-05, | |
| "loss": 0.5594, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 12.513875641703635, | |
| "learning_rate": 1.6830261072981423e-05, | |
| "loss": 0.6085, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 12.059036404622644, | |
| "learning_rate": 1.680302661094446e-05, | |
| "loss": 0.5629, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 10.324831974232412, | |
| "learning_rate": 1.677569789127647e-05, | |
| "loss": 0.6268, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 9.702271050156932, | |
| "learning_rate": 1.6748275292623678e-05, | |
| "loss": 0.621, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 11.784449310834715, | |
| "learning_rate": 1.6720759194933037e-05, | |
| "loss": 0.6856, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 8.598925774309697, | |
| "learning_rate": 1.669314997944694e-05, | |
| "loss": 0.6918, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 11.857440160524549, | |
| "learning_rate": 1.666544802869796e-05, | |
| "loss": 0.6879, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 8.032666301516855, | |
| "learning_rate": 1.6637653726503533e-05, | |
| "loss": 0.5792, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 7.0937478718250295, | |
| "learning_rate": 1.660976745796065e-05, | |
| "loss": 0.6607, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 9.465209749684965, | |
| "learning_rate": 1.658178960944052e-05, | |
| "loss": 0.6535, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 15.131095507571468, | |
| "learning_rate": 1.655372056858322e-05, | |
| "loss": 0.6438, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 17.304531600026, | |
| "learning_rate": 1.6525560724292303e-05, | |
| "loss": 0.7653, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 8.695700770875131, | |
| "learning_rate": 1.6497310466729448e-05, | |
| "loss": 0.6045, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 11.95231240235872, | |
| "learning_rate": 1.646897018730902e-05, | |
| "loss": 0.6791, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 9.486589536350005, | |
| "learning_rate": 1.6440540278692656e-05, | |
| "loss": 0.5519, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 10.961929493661737, | |
| "learning_rate": 1.6412021134783835e-05, | |
| "loss": 0.7537, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 9.852616880372151, | |
| "learning_rate": 1.6383413150722413e-05, | |
| "loss": 0.5775, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 4.770203374130839, | |
| "learning_rate": 1.6354716722879152e-05, | |
| "loss": 0.5684, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 6.203239240182946, | |
| "learning_rate": 1.6325932248850206e-05, | |
| "loss": 0.5391, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 14.714310109670022, | |
| "learning_rate": 1.6297060127451656e-05, | |
| "loss": 0.6426, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 16.664110650870665, | |
| "learning_rate": 1.626810075871394e-05, | |
| "loss": 0.7493, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 10.82617102662038, | |
| "learning_rate": 1.6239054543876345e-05, | |
| "loss": 0.6214, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 13.189719768118387, | |
| "learning_rate": 1.6209921885381418e-05, | |
| "loss": 0.647, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 6.638748272318463, | |
| "learning_rate": 1.6180703186869414e-05, | |
| "loss": 0.6062, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 9.318868865822735, | |
| "learning_rate": 1.615139885317269e-05, | |
| "loss": 0.5877, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 13.07963389547572, | |
| "learning_rate": 1.6122009290310097e-05, | |
| "loss": 0.697, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 8.538286649824611, | |
| "learning_rate": 1.6092534905481367e-05, | |
| "loss": 0.605, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 14.845192579391924, | |
| "learning_rate": 1.6062976107061454e-05, | |
| "loss": 0.5283, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 11.398188334173918, | |
| "learning_rate": 1.6033333304594886e-05, | |
| "loss": 0.7799, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 11.098935636927996, | |
| "learning_rate": 1.6003606908790082e-05, | |
| "loss": 0.6517, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 9.295708766284237, | |
| "learning_rate": 1.5973797331513674e-05, | |
| "loss": 0.8253, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 15.251818246107037, | |
| "learning_rate": 1.5943904985784797e-05, | |
| "loss": 0.729, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 5.329126700593149, | |
| "learning_rate": 1.5913930285769356e-05, | |
| "loss": 0.5526, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 12.120816580430638, | |
| "learning_rate": 1.5883873646774296e-05, | |
| "loss": 0.699, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 10.153946398501363, | |
| "learning_rate": 1.5853735485241858e-05, | |
| "loss": 0.6955, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 6.094377398479003, | |
| "learning_rate": 1.582351621874378e-05, | |
| "loss": 0.6237, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 12.749739058129057, | |
| "learning_rate": 1.579321626597554e-05, | |
| "loss": 0.7169, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 8.376136137589857, | |
| "learning_rate": 1.5762836046750546e-05, | |
| "loss": 0.6543, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 6.832733254275183, | |
| "learning_rate": 1.573237598199432e-05, | |
| "loss": 0.6495, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 10.567138540731836, | |
| "learning_rate": 1.570183649373865e-05, | |
| "loss": 0.6424, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 4.242489418395466, | |
| "learning_rate": 1.5671218005115767e-05, | |
| "loss": 0.5227, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 10.186721290405504, | |
| "learning_rate": 1.5640520940352476e-05, | |
| "loss": 0.6831, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 5.2392014967068405, | |
| "learning_rate": 1.5609745724764264e-05, | |
| "loss": 0.5989, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 7.53125873380665, | |
| "learning_rate": 1.5578892784749414e-05, | |
| "loss": 0.5461, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 11.308114294485758, | |
| "learning_rate": 1.5547962547783126e-05, | |
| "loss": 0.5895, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 9.200894044889376, | |
| "learning_rate": 1.5516955442411533e-05, | |
| "loss": 0.7269, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 7.3282431817133, | |
| "learning_rate": 1.5485871898245824e-05, | |
| "loss": 0.5479, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 14.868640801063636, | |
| "learning_rate": 1.5454712345956254e-05, | |
| "loss": 0.6124, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 18.938959881407193, | |
| "learning_rate": 1.54234772172662e-05, | |
| "loss": 0.6996, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 9.41122569113249, | |
| "learning_rate": 1.539216694494616e-05, | |
| "loss": 0.7736, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 13.076645506243146, | |
| "learning_rate": 1.536078196280777e-05, | |
| "loss": 0.6184, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 14.603964846269042, | |
| "learning_rate": 1.532932270569778e-05, | |
| "loss": 0.7387, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 7.409964871939044, | |
| "learning_rate": 1.5297789609492062e-05, | |
| "loss": 0.6705, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 10.532417778256553, | |
| "learning_rate": 1.526618311108952e-05, | |
| "loss": 0.6359, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 7.918617892191258, | |
| "learning_rate": 1.5234503648406075e-05, | |
| "loss": 0.5597, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 8.984191116234184, | |
| "learning_rate": 1.5202751660368594e-05, | |
| "loss": 0.5886, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 8.875342349884049, | |
| "learning_rate": 1.5170927586908787e-05, | |
| "loss": 0.579, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 12.368851666890444, | |
| "learning_rate": 1.513903186895713e-05, | |
| "loss": 0.684, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 8.674893219974978, | |
| "learning_rate": 1.5107064948436758e-05, | |
| "loss": 0.5412, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 11.644842239089794, | |
| "learning_rate": 1.5075027268257328e-05, | |
| "loss": 0.6165, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 24.313226821351616, | |
| "learning_rate": 1.5042919272308895e-05, | |
| "loss": 1.0199, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 12.280306305433697, | |
| "learning_rate": 1.5010741405455751e-05, | |
| "loss": 0.5976, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 5.830472782170163, | |
| "learning_rate": 1.4978494113530268e-05, | |
| "loss": 0.5179, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 14.70235378916328, | |
| "learning_rate": 1.4946177843326726e-05, | |
| "loss": 0.7092, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 17.543216455121254, | |
| "learning_rate": 1.4913793042595109e-05, | |
| "loss": 0.8279, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 12.859056834147074, | |
| "learning_rate": 1.4881340160034909e-05, | |
| "loss": 0.7344, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 9.382814693311708, | |
| "learning_rate": 1.4848819645288915e-05, | |
| "loss": 0.5823, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 10.602701451692775, | |
| "learning_rate": 1.4816231948936967e-05, | |
| "loss": 0.706, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 9.699931667366167, | |
| "learning_rate": 1.4783577522489733e-05, | |
| "loss": 0.6573, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 8.280832106689648, | |
| "learning_rate": 1.475085681838244e-05, | |
| "loss": 0.6462, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 10.316666897706078, | |
| "learning_rate": 1.4718070289968602e-05, | |
| "loss": 0.6581, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 7.552089443149507, | |
| "learning_rate": 1.4685218391513752e-05, | |
| "loss": 0.6516, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 10.041760330765845, | |
| "learning_rate": 1.4652301578189141e-05, | |
| "loss": 0.7184, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 11.3264716915839, | |
| "learning_rate": 1.4619320306065432e-05, | |
| "loss": 0.6023, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 9.31395109143123, | |
| "learning_rate": 1.4586275032106373e-05, | |
| "loss": 0.6248, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 16.084796485739954, | |
| "learning_rate": 1.4553166214162486e-05, | |
| "loss": 0.5707, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 5.559676215156368, | |
| "learning_rate": 1.4519994310964697e-05, | |
| "loss": 0.545, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 7.5228362636067425, | |
| "learning_rate": 1.4486759782118012e-05, | |
| "loss": 0.634, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 10.426847253583567, | |
| "learning_rate": 1.4453463088095108e-05, | |
| "loss": 0.613, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 6.094049952534055, | |
| "learning_rate": 1.4420104690229992e-05, | |
| "loss": 0.5793, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 6.245104416676842, | |
| "learning_rate": 1.4386685050711593e-05, | |
| "loss": 0.5631, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 7.453663730544318, | |
| "learning_rate": 1.4353204632577354e-05, | |
| "loss": 0.5591, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 10.466741619757949, | |
| "learning_rate": 1.4319663899706818e-05, | |
| "loss": 0.7189, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 6.220499980519979, | |
| "learning_rate": 1.4286063316815209e-05, | |
| "loss": 0.4383, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 10.892786776066355, | |
| "learning_rate": 1.4252403349446986e-05, | |
| "loss": 0.5664, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 6.174114043802528, | |
| "learning_rate": 1.4218684463969396e-05, | |
| "loss": 0.5716, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 4.9332645630935605, | |
| "learning_rate": 1.4184907127566006e-05, | |
| "loss": 0.553, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 7.331476371683697, | |
| "learning_rate": 1.4151071808230246e-05, | |
| "loss": 0.5706, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 9.54440691864147, | |
| "learning_rate": 1.4117178974758903e-05, | |
| "loss": 0.6349, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 9.51093808442514, | |
| "learning_rate": 1.4083229096745644e-05, | |
| "loss": 0.6076, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 7.416880704206411, | |
| "learning_rate": 1.404922264457449e-05, | |
| "loss": 0.5238, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 9.30112890680624, | |
| "learning_rate": 1.4015160089413331e-05, | |
| "loss": 0.6277, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 9.964620600235138, | |
| "learning_rate": 1.3981041903207364e-05, | |
| "loss": 0.6501, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 8.049380804954936, | |
| "learning_rate": 1.3946868558672569e-05, | |
| "loss": 0.5853, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 9.941935672040696, | |
| "learning_rate": 1.3912640529289163e-05, | |
| "loss": 0.6632, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 4.961145697377557, | |
| "learning_rate": 1.3878358289295032e-05, | |
| "loss": 0.5542, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 4.788363887274814, | |
| "learning_rate": 1.3844022313679167e-05, | |
| "loss": 0.5203, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 12.11801121511047, | |
| "learning_rate": 1.380963307817507e-05, | |
| "loss": 0.6661, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 11.34675588679876, | |
| "learning_rate": 1.3775191059254185e-05, | |
| "loss": 0.6116, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 6.678425818352266, | |
| "learning_rate": 1.3740696734119279e-05, | |
| "loss": 0.6401, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 12.160178505852342, | |
| "learning_rate": 1.3706150580697826e-05, | |
| "loss": 0.6138, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 6.094569373598361, | |
| "learning_rate": 1.3671553077635404e-05, | |
| "loss": 0.5602, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 5.908165508182039, | |
| "learning_rate": 1.3636904704289053e-05, | |
| "loss": 0.5606, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 12.500164793835593, | |
| "learning_rate": 1.3602205940720628e-05, | |
| "loss": 0.5601, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 5.981711152222135, | |
| "learning_rate": 1.3567457267690152e-05, | |
| "loss": 0.5011, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 6.390064988076557, | |
| "learning_rate": 1.3532659166649165e-05, | |
| "loss": 0.515, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 11.203440074699836, | |
| "learning_rate": 1.3497812119734037e-05, | |
| "loss": 0.7016, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 13.733785334526278, | |
| "learning_rate": 1.3462916609759298e-05, | |
| "loss": 0.6623, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 15.73461768409232, | |
| "learning_rate": 1.342797312021094e-05, | |
| "loss": 0.7519, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 8.804555804165531, | |
| "learning_rate": 1.3392982135239736e-05, | |
| "loss": 0.531, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 10.63905280022859, | |
| "learning_rate": 1.3357944139654508e-05, | |
| "loss": 0.5914, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 14.738317600036572, | |
| "learning_rate": 1.3322859618915431e-05, | |
| "loss": 0.6387, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 9.315538586392796, | |
| "learning_rate": 1.3287729059127288e-05, | |
| "loss": 0.5793, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 7.330889488527192, | |
| "learning_rate": 1.3252552947032755e-05, | |
| "loss": 0.5955, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 5.492945122022015, | |
| "learning_rate": 1.3217331770005639e-05, | |
| "loss": 0.4936, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 6.447759334210931, | |
| "learning_rate": 1.3182066016044135e-05, | |
| "loss": 0.6306, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 10.58800825139178, | |
| "learning_rate": 1.3146756173764061e-05, | |
| "loss": 0.682, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 11.358838931871405, | |
| "learning_rate": 1.3111402732392098e-05, | |
| "loss": 0.7864, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 8.567212453461783, | |
| "learning_rate": 1.3076006181758989e-05, | |
| "loss": 0.5594, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 8.546126016623386, | |
| "learning_rate": 1.3040567012292779e-05, | |
| "loss": 0.6419, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 11.957735628493818, | |
| "learning_rate": 1.3005085715012003e-05, | |
| "loss": 0.632, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 10.122315964153882, | |
| "learning_rate": 1.2969562781518885e-05, | |
| "loss": 0.6758, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 10.274234862260917, | |
| "learning_rate": 1.2933998703992531e-05, | |
| "loss": 0.4632, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 9.56291976312044, | |
| "learning_rate": 1.2898393975182113e-05, | |
| "loss": 0.7572, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 11.309501354950024, | |
| "learning_rate": 1.2862749088400026e-05, | |
| "loss": 0.5429, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 5.4111570184934195, | |
| "learning_rate": 1.2827064537515075e-05, | |
| "loss": 0.5896, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 4.403745251090165, | |
| "learning_rate": 1.279134081694561e-05, | |
| "loss": 0.6544, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 5.2380739150735565, | |
| "learning_rate": 1.2755578421652696e-05, | |
| "loss": 0.4995, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 6.1969762293333135, | |
| "learning_rate": 1.2719777847133241e-05, | |
| "loss": 0.6669, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 6.009545288953141, | |
| "learning_rate": 1.2683939589413139e-05, | |
| "loss": 0.4661, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 11.862108381435066, | |
| "learning_rate": 1.2648064145040392e-05, | |
| "loss": 0.5745, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 8.371435047081937, | |
| "learning_rate": 1.2612152011078233e-05, | |
| "loss": 0.581, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 6.570729807144181, | |
| "learning_rate": 1.2576203685098233e-05, | |
| "loss": 0.5283, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 15.891490498465696, | |
| "learning_rate": 1.2540219665173423e-05, | |
| "loss": 0.6488, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 11.348604428897819, | |
| "learning_rate": 1.2504200449871378e-05, | |
| "loss": 0.6554, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 10.1223452644342, | |
| "learning_rate": 1.2468146538247308e-05, | |
| "loss": 0.6271, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 14.257981744049157, | |
| "learning_rate": 1.2432058429837153e-05, | |
| "loss": 0.694, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 14.906690269642054, | |
| "learning_rate": 1.2395936624650664e-05, | |
| "loss": 0.6699, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 7.166307639896277, | |
| "learning_rate": 1.2359781623164465e-05, | |
| "loss": 0.5572, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 13.74157282336694, | |
| "learning_rate": 1.2323593926315113e-05, | |
| "loss": 0.7576, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 9.280109663389997, | |
| "learning_rate": 1.2287374035492184e-05, | |
| "loss": 0.5046, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 10.090941451576859, | |
| "learning_rate": 1.2251122452531301e-05, | |
| "loss": 0.5604, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 5.903088605165492, | |
| "learning_rate": 1.2214839679707193e-05, | |
| "loss": 0.5624, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 10.167630601617299, | |
| "learning_rate": 1.2178526219726723e-05, | |
| "loss": 0.5295, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 8.331133592273517, | |
| "learning_rate": 1.2142182575721946e-05, | |
| "loss": 0.6144, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 9.705907066152204, | |
| "learning_rate": 1.2105809251243113e-05, | |
| "loss": 0.6389, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 11.02758523623926, | |
| "learning_rate": 1.2069406750251713e-05, | |
| "loss": 0.666, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 7.447181404677012, | |
| "learning_rate": 1.2032975577113474e-05, | |
| "loss": 0.7216, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 7.658897343141703, | |
| "learning_rate": 1.1996516236591398e-05, | |
| "loss": 0.6087, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 4.187819572120313, | |
| "learning_rate": 1.1960029233838737e-05, | |
| "loss": 0.5356, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 5.7176090622427385, | |
| "learning_rate": 1.1923515074392022e-05, | |
| "loss": 0.619, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 6.823057157296125, | |
| "learning_rate": 1.1886974264164037e-05, | |
| "loss": 0.6037, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 11.059981117618824, | |
| "learning_rate": 1.1850407309436831e-05, | |
| "loss": 0.604, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 9.504143330556396, | |
| "learning_rate": 1.181381471685468e-05, | |
| "loss": 0.647, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 6.557755314159755, | |
| "learning_rate": 1.1777196993417087e-05, | |
| "loss": 0.4595, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 6.9247206639927095, | |
| "learning_rate": 1.1740554646471742e-05, | |
| "loss": 0.5228, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 10.232093839271188, | |
| "learning_rate": 1.1703888183707513e-05, | |
| "loss": 0.6062, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 6.510487958239591, | |
| "learning_rate": 1.166719811314738e-05, | |
| "loss": 0.5219, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 8.968572470896715, | |
| "learning_rate": 1.1630484943141428e-05, | |
| "loss": 0.5209, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 5.0494617841107265, | |
| "learning_rate": 1.1593749182359788e-05, | |
| "loss": 0.5948, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 11.340103876058729, | |
| "learning_rate": 1.1556991339785595e-05, | |
| "loss": 0.6203, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 10.146196176770191, | |
| "learning_rate": 1.1520211924707919e-05, | |
| "loss": 0.7785, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 4.43271626514912, | |
| "learning_rate": 1.1483411446714744e-05, | |
| "loss": 0.4862, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 7.824102062875491, | |
| "learning_rate": 1.1446590415685863e-05, | |
| "loss": 0.6113, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 6.336647655938098, | |
| "learning_rate": 1.1409749341785859e-05, | |
| "loss": 0.5343, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 4.855528076878574, | |
| "learning_rate": 1.1372888735456988e-05, | |
| "loss": 0.5014, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 8.244191689910387, | |
| "learning_rate": 1.1336009107412162e-05, | |
| "loss": 0.64, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 9.72186897530395, | |
| "learning_rate": 1.1299110968627822e-05, | |
| "loss": 0.5919, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 6.153546625651434, | |
| "learning_rate": 1.1262194830336888e-05, | |
| "loss": 0.4917, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 8.261625036086883, | |
| "learning_rate": 1.1225261204021662e-05, | |
| "loss": 0.5592, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 7.32183514687012, | |
| "learning_rate": 1.118831060140676e-05, | |
| "loss": 0.4822, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 6.309347895522231, | |
| "learning_rate": 1.1151343534451995e-05, | |
| "loss": 0.5727, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 7.8065524764950265, | |
| "learning_rate": 1.1114360515345301e-05, | |
| "loss": 0.6533, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 9.553172955626218, | |
| "learning_rate": 1.107736205649564e-05, | |
| "loss": 0.4895, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 8.353357382799153, | |
| "learning_rate": 1.1040348670525889e-05, | |
| "loss": 0.5254, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 11.74491323882519, | |
| "learning_rate": 1.1003320870265741e-05, | |
| "loss": 0.6004, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 3.4228605283872797, | |
| "learning_rate": 1.096627916874461e-05, | |
| "loss": 0.4572, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 11.366613402895425, | |
| "learning_rate": 1.0929224079184514e-05, | |
| "loss": 0.648, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 5.915082226142763, | |
| "learning_rate": 1.0892156114992963e-05, | |
| "loss": 0.5606, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 5.481987693582871, | |
| "learning_rate": 1.0855075789755845e-05, | |
| "loss": 0.5845, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 6.226052256005356, | |
| "learning_rate": 1.0817983617230326e-05, | |
| "loss": 0.5025, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 8.531553272430466, | |
| "learning_rate": 1.0780880111337704e-05, | |
| "loss": 0.5657, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 6.998952295190077, | |
| "learning_rate": 1.0743765786156313e-05, | |
| "loss": 0.7092, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 6.513406245121362, | |
| "learning_rate": 1.0706641155914384e-05, | |
| "loss": 0.548, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 8.48581353131954, | |
| "learning_rate": 1.066950673498294e-05, | |
| "loss": 0.5551, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 5.34639942444931, | |
| "learning_rate": 1.0632363037868638e-05, | |
| "loss": 0.4334, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 11.269718113412772, | |
| "learning_rate": 1.0595210579206676e-05, | |
| "loss": 0.5803, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 11.41953037054744, | |
| "learning_rate": 1.0558049873753637e-05, | |
| "loss": 0.5186, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 11.07638179235055, | |
| "learning_rate": 1.0520881436380366e-05, | |
| "loss": 0.6146, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 11.357643871564994, | |
| "learning_rate": 1.0483705782064835e-05, | |
| "loss": 0.4825, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 8.12832951837023, | |
| "learning_rate": 1.0446523425885008e-05, | |
| "loss": 0.5972, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 4.573339318696573, | |
| "learning_rate": 1.040933488301171e-05, | |
| "loss": 0.4511, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 7.488115795627891, | |
| "learning_rate": 1.0372140668701483e-05, | |
| "loss": 0.5612, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 4.416399464656011, | |
| "learning_rate": 1.0334941298289441e-05, | |
| "loss": 0.5678, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 8.472247115461284, | |
| "learning_rate": 1.0297737287182144e-05, | |
| "loss": 0.5397, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 7.278212668194716, | |
| "learning_rate": 1.026052915085045e-05, | |
| "loss": 0.5597, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 5.04716478079048, | |
| "learning_rate": 1.022331740482237e-05, | |
| "loss": 0.5014, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 11.570053771414813, | |
| "learning_rate": 1.0186102564675928e-05, | |
| "loss": 0.661, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 8.829505233768511, | |
| "learning_rate": 1.014888514603202e-05, | |
| "loss": 0.5232, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 9.542821950950582, | |
| "learning_rate": 1.0111665664547267e-05, | |
| "loss": 0.5823, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 4.63327242048301, | |
| "learning_rate": 1.0074444635906875e-05, | |
| "loss": 0.5625, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 6.365741719759197, | |
| "learning_rate": 1.0037222575817476e-05, | |
| "loss": 0.6587, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 8.488178372143695, | |
| "learning_rate": 1e-05, | |
| "loss": 0.5501, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 9.897138663353587, | |
| "learning_rate": 9.962777424182527e-06, | |
| "loss": 0.6661, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 10.569907059227898, | |
| "learning_rate": 9.92555536409313e-06, | |
| "loss": 0.5958, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 3.736421534561604, | |
| "learning_rate": 9.888334335452734e-06, | |
| "loss": 0.5303, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 8.836116584372025, | |
| "learning_rate": 9.85111485396798e-06, | |
| "loss": 0.6466, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 7.181117148918435, | |
| "learning_rate": 9.813897435324075e-06, | |
| "loss": 0.54, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 4.785898028898109, | |
| "learning_rate": 9.776682595177633e-06, | |
| "loss": 0.4985, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 11.140585777357792, | |
| "learning_rate": 9.739470849149554e-06, | |
| "loss": 0.6496, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 3.531792374263498, | |
| "learning_rate": 9.702262712817857e-06, | |
| "loss": 0.5364, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 6.601039058908546, | |
| "learning_rate": 9.66505870171056e-06, | |
| "loss": 0.5116, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 8.927040349981295, | |
| "learning_rate": 9.627859331298522e-06, | |
| "loss": 0.5219, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 6.1651706502501264, | |
| "learning_rate": 9.590665116988293e-06, | |
| "loss": 0.5608, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 9.745244839875141, | |
| "learning_rate": 9.553476574114993e-06, | |
| "loss": 0.5524, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 7.008580981680822, | |
| "learning_rate": 9.51629421793517e-06, | |
| "loss": 0.5662, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 7.687272495275157, | |
| "learning_rate": 9.479118563619638e-06, | |
| "loss": 0.5806, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 5.593165759891498, | |
| "learning_rate": 9.441950126246365e-06, | |
| "loss": 0.5447, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 4.518154487773691, | |
| "learning_rate": 9.404789420793327e-06, | |
| "loss": 0.5662, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 5.979882247197896, | |
| "learning_rate": 9.367636962131364e-06, | |
| "loss": 0.5499, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 10.248432776308547, | |
| "learning_rate": 9.330493265017062e-06, | |
| "loss": 0.6341, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 7.4913799071398275, | |
| "learning_rate": 9.29335884408562e-06, | |
| "loss": 0.4995, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 12.97254908941589, | |
| "learning_rate": 9.25623421384369e-06, | |
| "loss": 0.5575, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 7.7639027583981095, | |
| "learning_rate": 9.2191198886623e-06, | |
| "loss": 0.5819, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 7.6543085882238895, | |
| "learning_rate": 9.182016382769678e-06, | |
| "loss": 0.6647, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 5.7656791466080355, | |
| "learning_rate": 9.144924210244155e-06, | |
| "loss": 0.6249, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 5.2481708291921505, | |
| "learning_rate": 9.107843885007042e-06, | |
| "loss": 0.5478, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 10.21885595901045, | |
| "learning_rate": 9.070775920815488e-06, | |
| "loss": 0.5803, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 7.389626516947618, | |
| "learning_rate": 9.033720831255391e-06, | |
| "loss": 0.5334, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 11.910444324111724, | |
| "learning_rate": 8.996679129734266e-06, | |
| "loss": 0.5788, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 10.70692641976889, | |
| "learning_rate": 8.959651329474115e-06, | |
| "loss": 0.5155, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 5.6372805638926, | |
| "learning_rate": 8.922637943504362e-06, | |
| "loss": 0.5881, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 9.8644472116661, | |
| "learning_rate": 8.8856394846547e-06, | |
| "loss": 0.5072, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 11.356856786195184, | |
| "learning_rate": 8.848656465548007e-06, | |
| "loss": 0.549, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 11.998300599364015, | |
| "learning_rate": 8.811689398593245e-06, | |
| "loss": 0.5355, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 5.815584273089378, | |
| "learning_rate": 8.774738795978341e-06, | |
| "loss": 0.4673, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 9.210932501903986, | |
| "learning_rate": 8.737805169663113e-06, | |
| "loss": 0.5137, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 10.604624719913076, | |
| "learning_rate": 8.700889031372183e-06, | |
| "loss": 0.6092, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 6.733937738438429, | |
| "learning_rate": 8.663990892587839e-06, | |
| "loss": 0.5617, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 6.687819879267525, | |
| "learning_rate": 8.627111264543012e-06, | |
| "loss": 0.4319, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 11.209429725910061, | |
| "learning_rate": 8.590250658214148e-06, | |
| "loss": 0.5174, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 5.337704061513362, | |
| "learning_rate": 8.553409584314138e-06, | |
| "loss": 0.4295, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 9.126368763120784, | |
| "learning_rate": 8.516588553285258e-06, | |
| "loss": 0.4436, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 10.856129898989279, | |
| "learning_rate": 8.479788075292083e-06, | |
| "loss": 0.6126, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 5.163627248000245, | |
| "learning_rate": 8.443008660214409e-06, | |
| "loss": 0.5352, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 6.372067670711978, | |
| "learning_rate": 8.406250817640212e-06, | |
| "loss": 0.5076, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 6.426831912031028, | |
| "learning_rate": 8.369515056858575e-06, | |
| "loss": 0.6041, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 6.985038214862641, | |
| "learning_rate": 8.332801886852624e-06, | |
| "loss": 0.6586, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 6.593624344866879, | |
| "learning_rate": 8.296111816292494e-06, | |
| "loss": 0.6751, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 7.554153671309798, | |
| "learning_rate": 8.259445353528261e-06, | |
| "loss": 0.5276, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 9.224509289941837, | |
| "learning_rate": 8.222803006582915e-06, | |
| "loss": 0.5993, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 16.50239364012574, | |
| "learning_rate": 8.186185283145325e-06, | |
| "loss": 0.6259, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 15.120105685997352, | |
| "learning_rate": 8.149592690563172e-06, | |
| "loss": 0.6467, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 7.220612802422961, | |
| "learning_rate": 8.113025735835965e-06, | |
| "loss": 0.5715, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 8.562863959435413, | |
| "learning_rate": 8.076484925607983e-06, | |
| "loss": 0.5466, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 12.490548007030851, | |
| "learning_rate": 8.039970766161264e-06, | |
| "loss": 0.5514, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 6.5845388402784435, | |
| "learning_rate": 8.003483763408604e-06, | |
| "loss": 0.5687, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 8.50284284365726, | |
| "learning_rate": 7.967024422886528e-06, | |
| "loss": 0.4796, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 7.038271102438535, | |
| "learning_rate": 7.930593249748289e-06, | |
| "loss": 0.5228, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 11.889834227941325, | |
| "learning_rate": 7.894190748756892e-06, | |
| "loss": 0.4577, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 6.711588374541616, | |
| "learning_rate": 7.857817424278056e-06, | |
| "loss": 0.498, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 7.966852766279544, | |
| "learning_rate": 7.821473780273279e-06, | |
| "loss": 0.5579, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 8.417280128263915, | |
| "learning_rate": 7.785160320292812e-06, | |
| "loss": 0.4942, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 4.894439384156987, | |
| "learning_rate": 7.7488775474687e-06, | |
| "loss": 0.5082, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 7.960251253611578, | |
| "learning_rate": 7.712625964507818e-06, | |
| "loss": 0.4997, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 8.640093158656311, | |
| "learning_rate": 7.676406073684892e-06, | |
| "loss": 0.5362, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 11.82405754186769, | |
| "learning_rate": 7.64021837683554e-06, | |
| "loss": 0.5292, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 5.258983750312472, | |
| "learning_rate": 7.604063375349338e-06, | |
| "loss": 0.547, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 10.013781925100409, | |
| "learning_rate": 7.567941570162849e-06, | |
| "loss": 0.538, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 7.8051606786973755, | |
| "learning_rate": 7.531853461752696e-06, | |
| "loss": 0.5122, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 7.2586272669558145, | |
| "learning_rate": 7.495799550128625e-06, | |
| "loss": 0.4468, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 5.774720107719125, | |
| "learning_rate": 7.459780334826578e-06, | |
| "loss": 0.5275, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 7.511213422739932, | |
| "learning_rate": 7.423796314901769e-06, | |
| "loss": 0.5768, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 13.06441655084729, | |
| "learning_rate": 7.387847988921772e-06, | |
| "loss": 0.6067, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 9.818511674108413, | |
| "learning_rate": 7.351935854959608e-06, | |
| "loss": 0.5669, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 12.085632705436097, | |
| "learning_rate": 7.31606041058686e-06, | |
| "loss": 0.6871, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 6.460535141416251, | |
| "learning_rate": 7.2802221528667604e-06, | |
| "loss": 0.4964, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 3.7229407170275715, | |
| "learning_rate": 7.244421578347307e-06, | |
| "loss": 0.4216, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 8.837577427746787, | |
| "learning_rate": 7.208659183054393e-06, | |
| "loss": 0.5951, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 4.340949956633562, | |
| "learning_rate": 7.172935462484931e-06, | |
| "loss": 0.3946, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 5.8235051717682405, | |
| "learning_rate": 7.137250911599978e-06, | |
| "loss": 0.4795, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 5.498783950116057, | |
| "learning_rate": 7.101606024817889e-06, | |
| "loss": 0.5192, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 3.1814150008692534, | |
| "learning_rate": 7.066001296007469e-06, | |
| "loss": 0.3841, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 5.5757391334979225, | |
| "learning_rate": 7.030437218481116e-06, | |
| "loss": 0.4353, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 13.254243576587063, | |
| "learning_rate": 6.9949142849880015e-06, | |
| "loss": 0.5673, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 4.1536903593774515, | |
| "learning_rate": 6.959432987707223e-06, | |
| "loss": 0.4192, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 3.4786455475133287, | |
| "learning_rate": 6.9239938182410126e-06, | |
| "loss": 0.4249, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 7.5677310862475915, | |
| "learning_rate": 6.888597267607906e-06, | |
| "loss": 0.4949, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 8.368860704109643, | |
| "learning_rate": 6.8532438262359404e-06, | |
| "loss": 0.4154, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 5.058418452049737, | |
| "learning_rate": 6.817933983955867e-06, | |
| "loss": 0.4951, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 6.42488252810342, | |
| "learning_rate": 6.7826682299943635e-06, | |
| "loss": 0.5948, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 7.090667184322379, | |
| "learning_rate": 6.747447052967246e-06, | |
| "loss": 0.564, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 9.915670648118702, | |
| "learning_rate": 6.712270940872713e-06, | |
| "loss": 0.4821, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 10.251562644163112, | |
| "learning_rate": 6.677140381084573e-06, | |
| "loss": 0.5258, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 5.692081829738655, | |
| "learning_rate": 6.642055860345494e-06, | |
| "loss": 0.4832, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 4.7513407845962545, | |
| "learning_rate": 6.607017864760265e-06, | |
| "loss": 0.4669, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 7.302874981115006, | |
| "learning_rate": 6.572026879789064e-06, | |
| "loss": 0.4492, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 13.730001115703725, | |
| "learning_rate": 6.537083390240706e-06, | |
| "loss": 0.5786, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 5.826147970750949, | |
| "learning_rate": 6.502187880265969e-06, | |
| "loss": 0.3817, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 8.854083264240913, | |
| "learning_rate": 6.467340833350837e-06, | |
| "loss": 0.4938, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 5.711323962217592, | |
| "learning_rate": 6.43254273230985e-06, | |
| "loss": 0.4934, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 6.916231992838763, | |
| "learning_rate": 6.3977940592793766e-06, | |
| "loss": 0.5225, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 9.605613693440635, | |
| "learning_rate": 6.36309529571095e-06, | |
| "loss": 0.527, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 7.226506391255898, | |
| "learning_rate": 6.328446922364595e-06, | |
| "loss": 0.4771, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 5.211969056714009, | |
| "learning_rate": 6.293849419302179e-06, | |
| "loss": 0.5099, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 11.660847177274526, | |
| "learning_rate": 6.259303265880725e-06, | |
| "loss": 0.5479, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 8.043560392139083, | |
| "learning_rate": 6.224808940745814e-06, | |
| "loss": 0.5182, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 8.052748988432281, | |
| "learning_rate": 6.190366921824933e-06, | |
| "loss": 0.6176, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 7.108683500222242, | |
| "learning_rate": 6.155977686320837e-06, | |
| "loss": 0.5878, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 11.823467597819771, | |
| "learning_rate": 6.121641710704972e-06, | |
| "loss": 0.4971, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 10.40651530100889, | |
| "learning_rate": 6.087359470710841e-06, | |
| "loss": 0.4427, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 11.976145453386437, | |
| "learning_rate": 6.053131441327431e-06, | |
| "loss": 0.5098, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 7.971563751964657, | |
| "learning_rate": 6.018958096792642e-06, | |
| "loss": 0.5695, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 6.129677385908451, | |
| "learning_rate": 5.98483991058667e-06, | |
| "loss": 0.4808, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 4.479658413986526, | |
| "learning_rate": 5.950777355425511e-06, | |
| "loss": 0.4775, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 4.309077539414612, | |
| "learning_rate": 5.916770903254363e-06, | |
| "loss": 0.4718, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 5.726658232065661, | |
| "learning_rate": 5.8828210252411e-06, | |
| "loss": 0.5147, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 6.432897045805533, | |
| "learning_rate": 5.848928191769753e-06, | |
| "loss": 0.5047, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 6.06082150623531, | |
| "learning_rate": 5.815092872433994e-06, | |
| "loss": 0.5296, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 7.760932921939526, | |
| "learning_rate": 5.781315536030607e-06, | |
| "loss": 0.5339, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 6.0611679377908825, | |
| "learning_rate": 5.74759665055302e-06, | |
| "loss": 0.4212, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 6.4875374667111565, | |
| "learning_rate": 5.713936683184795e-06, | |
| "loss": 0.5394, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 5.850168689232539, | |
| "learning_rate": 5.680336100293182e-06, | |
| "loss": 0.4512, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 9.444171703118819, | |
| "learning_rate": 5.6467953674226505e-06, | |
| "loss": 0.3942, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 8.582206095093797, | |
| "learning_rate": 5.613314949288409e-06, | |
| "loss": 0.4582, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 11.757390924671101, | |
| "learning_rate": 5.579895309770009e-06, | |
| "loss": 0.5435, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 5.02260927644731, | |
| "learning_rate": 5.546536911904896e-06, | |
| "loss": 0.4719, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 9.22536236771837, | |
| "learning_rate": 5.513240217881992e-06, | |
| "loss": 0.4668, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 10.157555843180285, | |
| "learning_rate": 5.4800056890353025e-06, | |
| "loss": 0.5409, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 12.323174404868091, | |
| "learning_rate": 5.44683378583752e-06, | |
| "loss": 0.5631, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 8.185618668543789, | |
| "learning_rate": 5.4137249678936265e-06, | |
| "loss": 0.4284, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 8.516855530899237, | |
| "learning_rate": 5.380679693934569e-06, | |
| "loss": 0.4221, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 11.008555199751408, | |
| "learning_rate": 5.347698421810861e-06, | |
| "loss": 0.5924, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 15.331344015670625, | |
| "learning_rate": 5.31478160848625e-06, | |
| "loss": 0.642, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 10.747880055046071, | |
| "learning_rate": 5.2819297100314e-06, | |
| "loss": 0.5417, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 15.482764407878273, | |
| "learning_rate": 5.249143181617562e-06, | |
| "loss": 0.63, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 4.174929188215974, | |
| "learning_rate": 5.216422477510267e-06, | |
| "loss": 0.4989, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 3.810136037355045, | |
| "learning_rate": 5.183768051063036e-06, | |
| "loss": 0.4893, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 6.8676100081227744, | |
| "learning_rate": 5.151180354711087e-06, | |
| "loss": 0.5768, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 10.044174931284044, | |
| "learning_rate": 5.11865983996509e-06, | |
| "loss": 0.4937, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 5.716263993029701, | |
| "learning_rate": 5.0862069574048956e-06, | |
| "loss": 0.4918, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 5.327730475446738, | |
| "learning_rate": 5.053822156673276e-06, | |
| "loss": 0.5112, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 9.70768865377853, | |
| "learning_rate": 5.021505886469733e-06, | |
| "loss": 0.5682, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 7.449826876587748, | |
| "learning_rate": 4.989258594544252e-06, | |
| "loss": 0.5914, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 6.246979693998428, | |
| "learning_rate": 4.957080727691107e-06, | |
| "loss": 0.5022, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 6.234169934241622, | |
| "learning_rate": 4.9249727317426765e-06, | |
| "loss": 0.5086, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 8.795959544866905, | |
| "learning_rate": 4.892935051563243e-06, | |
| "loss": 0.5631, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 6.585824557633543, | |
| "learning_rate": 4.86096813104287e-06, | |
| "loss": 0.5387, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 9.412528445371356, | |
| "learning_rate": 4.829072413091219e-06, | |
| "loss": 0.5903, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 4.934030658656449, | |
| "learning_rate": 4.797248339631411e-06, | |
| "loss": 0.5905, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 7.225794451726558, | |
| "learning_rate": 4.765496351593927e-06, | |
| "loss": 0.595, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 10.387771035955092, | |
| "learning_rate": 4.733816888910483e-06, | |
| "loss": 0.4901, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 5.585297554337063, | |
| "learning_rate": 4.7022103905079405e-06, | |
| "loss": 0.3997, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 5.2187286296388375, | |
| "learning_rate": 4.67067729430222e-06, | |
| "loss": 0.4856, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 3.4821810592954296, | |
| "learning_rate": 4.639218037192235e-06, | |
| "loss": 0.4361, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 3.959673908934417, | |
| "learning_rate": 4.607833055053843e-06, | |
| "loss": 0.3775, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 5.129251200576683, | |
| "learning_rate": 4.576522782733802e-06, | |
| "loss": 0.4947, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 4.8501294404892, | |
| "learning_rate": 4.545287654043751e-06, | |
| "loss": 0.3936, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 5.318966235176163, | |
| "learning_rate": 4.514128101754183e-06, | |
| "loss": 0.4649, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 7.407733161742192, | |
| "learning_rate": 4.483044557588473e-06, | |
| "loss": 0.5102, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 4.965311089759532, | |
| "learning_rate": 4.45203745221688e-06, | |
| "loss": 0.457, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 7.5517122869625135, | |
| "learning_rate": 4.4211072152505866e-06, | |
| "loss": 0.436, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 5.082245336395077, | |
| "learning_rate": 4.3902542752357415e-06, | |
| "loss": 0.4936, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 9.843843848674208, | |
| "learning_rate": 4.359479059647527e-06, | |
| "loss": 0.4336, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 5.231213013499434, | |
| "learning_rate": 4.3287819948842334e-06, | |
| "loss": 0.4247, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 6.305006534689214, | |
| "learning_rate": 4.2981635062613556e-06, | |
| "loss": 0.4826, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 5.533371229509814, | |
| "learning_rate": 4.267624018005686e-06, | |
| "loss": 0.4353, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 7.108534233713205, | |
| "learning_rate": 4.237163953249453e-06, | |
| "loss": 0.5991, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 4.76045650818351, | |
| "learning_rate": 4.206783734024463e-06, | |
| "loss": 0.4514, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 8.886091309499616, | |
| "learning_rate": 4.176483781256224e-06, | |
| "loss": 0.4023, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 7.524320382507302, | |
| "learning_rate": 4.1462645147581456e-06, | |
| "loss": 0.4514, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 7.482440196178911, | |
| "learning_rate": 4.116126353225703e-06, | |
| "loss": 0.5668, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 9.764506225206542, | |
| "learning_rate": 4.086069714230646e-06, | |
| "loss": 0.539, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 8.336719846814763, | |
| "learning_rate": 4.056095014215208e-06, | |
| "loss": 0.6426, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 4.148296296213888, | |
| "learning_rate": 4.0262026684863295e-06, | |
| "loss": 0.3376, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 9.203116559209336, | |
| "learning_rate": 3.99639309120992e-06, | |
| "loss": 0.5199, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 4.576118300820756, | |
| "learning_rate": 3.96666669540512e-06, | |
| "loss": 0.3731, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 9.971893856018951, | |
| "learning_rate": 3.937023892938549e-06, | |
| "loss": 0.591, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 5.809308796580302, | |
| "learning_rate": 3.907465094518636e-06, | |
| "loss": 0.5063, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 5.798443674221982, | |
| "learning_rate": 3.877990709689905e-06, | |
| "loss": 0.462, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 9.114741488810273, | |
| "learning_rate": 3.8486011468273145e-06, | |
| "loss": 0.5972, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 10.35491156365141, | |
| "learning_rate": 3.819296813130588e-06, | |
| "loss": 0.46, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 8.13256163586952, | |
| "learning_rate": 3.790078114618586e-06, | |
| "loss": 0.46, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 8.953602041543364, | |
| "learning_rate": 3.760945456123656e-06, | |
| "loss": 0.5296, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 6.830059498599811, | |
| "learning_rate": 3.731899241286061e-06, | |
| "loss": 0.4798, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 8.333772044542284, | |
| "learning_rate": 3.702939872548348e-06, | |
| "loss": 0.5452, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 7.7830518181331785, | |
| "learning_rate": 3.6740677511497958e-06, | |
| "loss": 0.52, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 9.48244358766429, | |
| "learning_rate": 3.6452832771208524e-06, | |
| "loss": 0.4604, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 5.325755301496669, | |
| "learning_rate": 3.616586849277587e-06, | |
| "loss": 0.4801, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 7.883051056212922, | |
| "learning_rate": 3.587978865216165e-06, | |
| "loss": 0.4396, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 7.231451053733015, | |
| "learning_rate": 3.559459721307349e-06, | |
| "loss": 0.5395, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 7.736510298846521, | |
| "learning_rate": 3.5310298126909814e-06, | |
| "loss": 0.4596, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 5.986545004035281, | |
| "learning_rate": 3.5026895332705504e-06, | |
| "loss": 0.4593, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 4.438626475367621, | |
| "learning_rate": 3.474439275707698e-06, | |
| "loss": 0.397, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 9.623477478063533, | |
| "learning_rate": 3.4462794314167846e-06, | |
| "loss": 0.509, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 9.394298451398283, | |
| "learning_rate": 3.418210390559481e-06, | |
| "loss": 0.4663, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 9.315469015310178, | |
| "learning_rate": 3.390232542039352e-06, | |
| "loss": 0.5065, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 6.092259066777256, | |
| "learning_rate": 3.3623462734964697e-06, | |
| "loss": 0.4477, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 4.319138234974373, | |
| "learning_rate": 3.3345519713020445e-06, | |
| "loss": 0.3925, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 6.358450139297925, | |
| "learning_rate": 3.3068500205530596e-06, | |
| "loss": 0.4679, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 7.701373209390687, | |
| "learning_rate": 3.2792408050669634e-06, | |
| "loss": 0.409, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 9.55351492544124, | |
| "learning_rate": 3.2517247073763237e-06, | |
| "loss": 0.4771, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 4.7780971794647025, | |
| "learning_rate": 3.2243021087235336e-06, | |
| "loss": 0.4689, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 7.114890807561552, | |
| "learning_rate": 3.196973389055542e-06, | |
| "loss": 0.4797, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 7.072041698936169, | |
| "learning_rate": 3.16973892701858e-06, | |
| "loss": 0.5768, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 5.961275677943624, | |
| "learning_rate": 3.142599099952912e-06, | |
| "loss": 0.3881, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 4.06251591074102, | |
| "learning_rate": 3.115554283887614e-06, | |
| "loss": 0.375, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 7.326058092329529, | |
| "learning_rate": 3.088604853535361e-06, | |
| "loss": 0.5388, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 5.274081659887408, | |
| "learning_rate": 3.0617511822872337e-06, | |
| "loss": 0.4742, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 4.793423500693722, | |
| "learning_rate": 3.03499364220755e-06, | |
| "loss": 0.4056, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 5.269835865291424, | |
| "learning_rate": 3.0083326040286977e-06, | |
| "loss": 0.4332, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 6.413197977206429, | |
| "learning_rate": 2.9817684371460155e-06, | |
| "loss": 0.5452, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 8.581131926197953, | |
| "learning_rate": 2.9553015096126638e-06, | |
| "loss": 0.4393, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 4.5160115188573045, | |
| "learning_rate": 2.9289321881345257e-06, | |
| "loss": 0.4465, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 5.2592392037223235, | |
| "learning_rate": 2.902660838065131e-06, | |
| "loss": 0.5082, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 7.05269405170704, | |
| "learning_rate": 2.8764878234005876e-06, | |
| "loss": 0.372, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 5.274091473047936, | |
| "learning_rate": 2.8504135067745463e-06, | |
| "loss": 0.4121, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 7.3781029800169655, | |
| "learning_rate": 2.8244382494531675e-06, | |
| "loss": 0.4067, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 8.32474239381458, | |
| "learning_rate": 2.798562411330126e-06, | |
| "loss": 0.3679, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 12.450031335391019, | |
| "learning_rate": 2.772786350921608e-06, | |
| "loss": 0.5521, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 7.249666510251534, | |
| "learning_rate": 2.7471104253613645e-06, | |
| "loss": 0.3938, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 7.2873098383670065, | |
| "learning_rate": 2.721534990395752e-06, | |
| "loss": 0.4884, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 5.438711289922824, | |
| "learning_rate": 2.6960604003788014e-06, | |
| "loss": 0.364, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 9.79198662580905, | |
| "learning_rate": 2.670687008267313e-06, | |
| "loss": 0.3911, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 11.420021502949318, | |
| "learning_rate": 2.6454151656159666e-06, | |
| "loss": 0.4454, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 8.204149460627178, | |
| "learning_rate": 2.6202452225724463e-06, | |
| "loss": 0.4276, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 10.999756173661007, | |
| "learning_rate": 2.5951775278725956e-06, | |
| "loss": 0.5623, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 8.568105116698392, | |
| "learning_rate": 2.5702124288355723e-06, | |
| "loss": 0.3292, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 5.56439861064704, | |
| "learning_rate": 2.545350271359055e-06, | |
| "loss": 0.3879, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 8.106162701050645, | |
| "learning_rate": 2.520591399914444e-06, | |
| "loss": 0.364, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 10.29859188596712, | |
| "learning_rate": 2.495936157542074e-06, | |
| "loss": 0.4791, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 7.657220905887133, | |
| "learning_rate": 2.471384885846482e-06, | |
| "loss": 0.4902, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 6.191995160106613, | |
| "learning_rate": 2.4469379249916614e-06, | |
| "loss": 0.4585, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 7.623947016373425, | |
| "learning_rate": 2.422595613696356e-06, | |
| "loss": 0.42, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 5.456113345068306, | |
| "learning_rate": 2.3983582892293642e-06, | |
| "loss": 0.3625, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 9.722994068985187, | |
| "learning_rate": 2.3742262874048583e-06, | |
| "loss": 0.5455, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 10.210377059221097, | |
| "learning_rate": 2.3501999425777433e-06, | |
| "loss": 0.4177, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 9.986453075138304, | |
| "learning_rate": 2.326279587639031e-06, | |
| "loss": 0.4519, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 7.444568050178271, | |
| "learning_rate": 2.3024655540111984e-06, | |
| "loss": 0.4707, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 6.882963377525352, | |
| "learning_rate": 2.2787581716436257e-06, | |
| "loss": 0.4142, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 5.621837753619931, | |
| "learning_rate": 2.255157769008011e-06, | |
| "loss": 0.3837, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 8.979013360736435, | |
| "learning_rate": 2.2316646730938197e-06, | |
| "loss": 0.4511, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 9.387346843179332, | |
| "learning_rate": 2.2082792094037585e-06, | |
| "loss": 0.3837, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 11.152697788037019, | |
| "learning_rate": 2.185001701949263e-06, | |
| "loss": 0.3815, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 7.316279559405673, | |
| "learning_rate": 2.1618324732459993e-06, | |
| "loss": 0.5763, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 4.60864460318581, | |
| "learning_rate": 2.138771844309423e-06, | |
| "loss": 0.3784, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 8.893733826091104, | |
| "learning_rate": 2.1158201346502927e-06, | |
| "loss": 0.3993, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 8.857140540604258, | |
| "learning_rate": 2.092977662270275e-06, | |
| "loss": 0.3334, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 9.000582629375513, | |
| "learning_rate": 2.0702447436575223e-06, | |
| "loss": 0.5377, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 11.407276883996886, | |
| "learning_rate": 2.0476216937822967e-06, | |
| "loss": 0.4657, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 9.343576283040107, | |
| "learning_rate": 2.0251088260925967e-06, | |
| "loss": 0.5038, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 5.742436296561597, | |
| "learning_rate": 2.002706452509824e-06, | |
| "loss": 0.462, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 9.651440289066235, | |
| "learning_rate": 1.9804148834244465e-06, | |
| "loss": 0.4588, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 5.795036174327087, | |
| "learning_rate": 1.9582344276917233e-06, | |
| "loss": 0.5565, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 7.475417877764932, | |
| "learning_rate": 1.9361653926274016e-06, | |
| "loss": 0.3644, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 7.2305893277419235, | |
| "learning_rate": 1.9142080840034647e-06, | |
| "loss": 0.451, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 7.9485644539877365, | |
| "learning_rate": 1.8923628060439037e-06, | |
| "loss": 0.5415, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 7.339019736495735, | |
| "learning_rate": 1.8706298614204942e-06, | |
| "loss": 0.4258, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 6.229226290130147, | |
| "learning_rate": 1.8490095512486072e-06, | |
| "loss": 0.372, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 8.318686997660855, | |
| "learning_rate": 1.8275021750830368e-06, | |
| "loss": 0.4337, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 11.383779868869713, | |
| "learning_rate": 1.8061080309138379e-06, | |
| "loss": 0.5517, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 4.319009051803344, | |
| "learning_rate": 1.7848274151622236e-06, | |
| "loss": 0.4032, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 6.434125961202376, | |
| "learning_rate": 1.7636606226764353e-06, | |
| "loss": 0.4165, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 5.041886706286037, | |
| "learning_rate": 1.742607946727658e-06, | |
| "loss": 0.4621, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 6.183213056857235, | |
| "learning_rate": 1.7216696790059718e-06, | |
| "loss": 0.423, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 6.373680744920821, | |
| "learning_rate": 1.7008461096162999e-06, | |
| "loss": 0.5464, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 11.493069464165945, | |
| "learning_rate": 1.6801375270743925e-06, | |
| "loss": 0.4518, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 9.805183898852514, | |
| "learning_rate": 1.6595442183028266e-06, | |
| "loss": 0.4836, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 8.05887231146552, | |
| "learning_rate": 1.6390664686270342e-06, | |
| "loss": 0.3711, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 6.975427069143952, | |
| "learning_rate": 1.6187045617713459e-06, | |
| "loss": 0.4149, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 7.272454723310934, | |
| "learning_rate": 1.5984587798550633e-06, | |
| "loss": 0.4357, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 6.941951052506783, | |
| "learning_rate": 1.578329403388541e-06, | |
| "loss": 0.3835, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 5.800438948527679, | |
| "learning_rate": 1.5583167112693153e-06, | |
| "loss": 0.4408, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 5.3480222782611175, | |
| "learning_rate": 1.5384209807782291e-06, | |
| "loss": 0.4553, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 5.909917855292465, | |
| "learning_rate": 1.518642487575591e-06, | |
| "loss": 0.497, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 7.683087880393552, | |
| "learning_rate": 1.49898150569736e-06, | |
| "loss": 0.4535, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 7.446327856244201, | |
| "learning_rate": 1.4794383075513453e-06, | |
| "loss": 0.3873, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 6.360170134685136, | |
| "learning_rate": 1.460013163913433e-06, | |
| "loss": 0.5037, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 5.053653216562189, | |
| "learning_rate": 1.4407063439238333e-06, | |
| "loss": 0.4954, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 4.2483525961998945, | |
| "learning_rate": 1.421518115083359e-06, | |
| "loss": 0.3751, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 6.345969600489226, | |
| "learning_rate": 1.4024487432497013e-06, | |
| "loss": 0.4143, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 5.646370329937249, | |
| "learning_rate": 1.3834984926337657e-06, | |
| "loss": 0.3376, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 10.69680196625431, | |
| "learning_rate": 1.36466762579601e-06, | |
| "loss": 0.5439, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 9.34187811581496, | |
| "learning_rate": 1.3459564036427841e-06, | |
| "loss": 0.4761, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 8.322291418258063, | |
| "learning_rate": 1.3273650854227438e-06, | |
| "loss": 0.4699, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 6.089668194777211, | |
| "learning_rate": 1.3088939287232383e-06, | |
| "loss": 0.4135, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 6.349508519106518, | |
| "learning_rate": 1.2905431894667552e-06, | |
| "loss": 0.4429, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 6.0067928128468955, | |
| "learning_rate": 1.272313121907366e-06, | |
| "loss": 0.5506, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 5.057193603204914, | |
| "learning_rate": 1.2542039786272008e-06, | |
| "loss": 0.4029, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 6.9264232438262, | |
| "learning_rate": 1.236216010532959e-06, | |
| "loss": 0.399, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 8.111117092935682, | |
| "learning_rate": 1.218349466852432e-06, | |
| "loss": 0.3969, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 12.617119783922481, | |
| "learning_rate": 1.200604595131033e-06, | |
| "loss": 0.3632, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 4.997880390654607, | |
| "learning_rate": 1.1829816412283912e-06, | |
| "loss": 0.3727, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 9.925641311100412, | |
| "learning_rate": 1.1654808493149283e-06, | |
| "loss": 0.4779, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 10.259671670493303, | |
| "learning_rate": 1.1481024618684821e-06, | |
| "loss": 0.4792, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 4.666467514007945, | |
| "learning_rate": 1.1308467196709505e-06, | |
| "loss": 0.3665, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 5.8390766285836895, | |
| "learning_rate": 1.1137138618049403e-06, | |
| "loss": 0.4426, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 4.079596026020399, | |
| "learning_rate": 1.096704125650474e-06, | |
| "loss": 0.3858, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 6.355034616703483, | |
| "learning_rate": 1.079817746881696e-06, | |
| "loss": 0.4322, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 4.964202739872203, | |
| "learning_rate": 1.0630549594635896e-06, | |
| "loss": 0.4246, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 9.981500108734732, | |
| "learning_rate": 1.0464159956487596e-06, | |
| "loss": 0.4048, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 5.314444267142196, | |
| "learning_rate": 1.0299010859742009e-06, | |
| "loss": 0.4226, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 6.711068748283136, | |
| "learning_rate": 1.013510459258108e-06, | |
| "loss": 0.4626, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 6.87746096070976, | |
| "learning_rate": 9.972443425967026e-07, | |
| "loss": 0.4762, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 5.584995566992966, | |
| "learning_rate": 9.811029613610913e-07, | |
| "loss": 0.4638, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 5.476555647757321, | |
| "learning_rate": 9.650865391941345e-07, | |
| "loss": 0.4174, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 5.334567697843729, | |
| "learning_rate": 9.491952980073604e-07, | |
| "loss": 0.3264, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 7.864733082511921, | |
| "learning_rate": 9.334294579778813e-07, | |
| "loss": 0.3613, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 8.345283693272092, | |
| "learning_rate": 9.177892375453413e-07, | |
| "loss": 0.5469, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 8.885940166363653, | |
| "learning_rate": 9.022748534088988e-07, | |
| "loss": 0.4455, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 7.55591370496932, | |
| "learning_rate": 8.86886520524216e-07, | |
| "loss": 0.4356, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 7.5174900881299465, | |
| "learning_rate": 8.716244521004846e-07, | |
| "loss": 0.4074, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 7.226452727128449, | |
| "learning_rate": 8.564888595974718e-07, | |
| "loss": 0.473, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 7.586424458860782, | |
| "learning_rate": 8.414799527225826e-07, | |
| "loss": 0.4757, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 5.51882334743589, | |
| "learning_rate": 8.265979394279732e-07, | |
| "loss": 0.4272, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 6.945131528688248, | |
| "learning_rate": 8.118430259076493e-07, | |
| "loss": 0.3003, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 7.957372619224033, | |
| "learning_rate": 7.972154165946155e-07, | |
| "loss": 0.4485, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 7.203632243202649, | |
| "learning_rate": 7.827153141580501e-07, | |
| "loss": 0.384, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 5.973567270845765, | |
| "learning_rate": 7.683429195004932e-07, | |
| "loss": 0.415, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 7.742846323788447, | |
| "learning_rate": 7.540984317550548e-07, | |
| "loss": 0.4261, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 4.971511012508995, | |
| "learning_rate": 7.399820482826692e-07, | |
| "loss": 0.3073, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 7.417662373162329, | |
| "learning_rate": 7.25993964669347e-07, | |
| "loss": 0.4293, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 8.474162845843916, | |
| "learning_rate": 7.12134374723481e-07, | |
| "loss": 0.5288, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 6.74629233498178, | |
| "learning_rate": 6.984034704731479e-07, | |
| "loss": 0.379, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 6.51287661591934, | |
| "learning_rate": 6.848014421634497e-07, | |
| "loss": 0.4254, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 8.203119397161476, | |
| "learning_rate": 6.71328478253881e-07, | |
| "loss": 0.4818, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 7.091765460672564, | |
| "learning_rate": 6.579847654157234e-07, | |
| "loss": 0.3937, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 5.848337659332669, | |
| "learning_rate": 6.447704885294437e-07, | |
| "loss": 0.3223, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 11.861790191542282, | |
| "learning_rate": 6.316858306821449e-07, | |
| "loss": 0.3955, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 5.89823677782504, | |
| "learning_rate": 6.18730973165027e-07, | |
| "loss": 0.52, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 7.297372429370729, | |
| "learning_rate": 6.05906095470874e-07, | |
| "loss": 0.5625, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 8.018266432717773, | |
| "learning_rate": 5.932113752915659e-07, | |
| "loss": 0.5943, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 5.441022930966068, | |
| "learning_rate": 5.806469885156163e-07, | |
| "loss": 0.3584, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 7.923651656498159, | |
| "learning_rate": 5.682131092257381e-07, | |
| "loss": 0.4374, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 4.401381730215806, | |
| "learning_rate": 5.55909909696436e-07, | |
| "loss": 0.4548, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 9.323436877274784, | |
| "learning_rate": 5.437375603916039e-07, | |
| "loss": 0.3949, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 7.6852951141865455, | |
| "learning_rate": 5.316962299621808e-07, | |
| "loss": 0.4357, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 5.468965570424503, | |
| "learning_rate": 5.19786085243803e-07, | |
| "loss": 0.4446, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 6.3812841292130855, | |
| "learning_rate": 5.080072912544987e-07, | |
| "loss": 0.4428, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 10.852851737536733, | |
| "learning_rate": 4.96360011192396e-07, | |
| "loss": 0.4991, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 5.7584503904522775, | |
| "learning_rate": 4.848444064334679e-07, | |
| "loss": 0.5114, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 6.376038086977472, | |
| "learning_rate": 4.734606365292871e-07, | |
| "loss": 0.3782, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 6.224991564333229, | |
| "learning_rate": 4.6220885920483014e-07, | |
| "loss": 0.3527, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 5.203231071738429, | |
| "learning_rate": 4.510892303562797e-07, | |
| "loss": 0.3518, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 5.884197350261294, | |
| "learning_rate": 4.401019040488652e-07, | |
| "loss": 0.4493, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 8.438339641746209, | |
| "learning_rate": 4.2924703251473355e-07, | |
| "loss": 0.5192, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 9.929683884140545, | |
| "learning_rate": 4.1852476615083957e-07, | |
| "loss": 0.4769, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 7.444778653156325, | |
| "learning_rate": 4.0793525351685905e-07, | |
| "loss": 0.3915, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 5.55575801178513, | |
| "learning_rate": 3.974786413331311e-07, | |
| "loss": 0.3861, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 5.154553375207641, | |
| "learning_rate": 3.8715507447862455e-07, | |
| "loss": 0.3713, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 9.250721341718682, | |
| "learning_rate": 3.7696469598893727e-07, | |
| "loss": 0.4575, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 5.654967363493775, | |
| "learning_rate": 3.669076470543054e-07, | |
| "loss": 0.4686, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 5.710919914401177, | |
| "learning_rate": 3.569840670176483e-07, | |
| "loss": 0.4339, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 6.7079973379018085, | |
| "learning_rate": 3.4719409337264473e-07, | |
| "loss": 0.4238, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 10.924750798758199, | |
| "learning_rate": 3.3753786176182303e-07, | |
| "loss": 0.4977, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 4.9608766209441315, | |
| "learning_rate": 3.2801550597468033e-07, | |
| "loss": 0.2732, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 7.448404330000091, | |
| "learning_rate": 3.186271579458333e-07, | |
| "loss": 0.4483, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 5.022846446335248, | |
| "learning_rate": 3.093729477531826e-07, | |
| "loss": 0.4363, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 7.140771043711367, | |
| "learning_rate": 3.002530036161222e-07, | |
| "loss": 0.4566, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 5.57307898414647, | |
| "learning_rate": 2.912674518937553e-07, | |
| "loss": 0.3565, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 7.503093403901029, | |
| "learning_rate": 2.824164170831389e-07, | |
| "loss": 0.3495, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 8.266577699990709, | |
| "learning_rate": 2.737000218175712e-07, | |
| "loss": 0.458, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 8.007643327554275, | |
| "learning_rate": 2.651183868648821e-07, | |
| "loss": 0.3738, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 6.8532106769801215, | |
| "learning_rate": 2.566716311257655e-07, | |
| "loss": 0.4602, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 8.226940714871725, | |
| "learning_rate": 2.483598716321289e-07, | |
| "loss": 0.3939, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 7.35412581775661, | |
| "learning_rate": 2.4018322354547483e-07, | |
| "loss": 0.4912, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 5.189789718764145, | |
| "learning_rate": 2.321418001553022e-07, | |
| "loss": 0.3516, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 5.643929108838239, | |
| "learning_rate": 2.2423571287753853e-07, | |
| "loss": 0.3811, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 5.764646780403799, | |
| "learning_rate": 2.1646507125299587e-07, | |
| "loss": 0.4442, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 5.42774429745915, | |
| "learning_rate": 2.0882998294585177e-07, | |
| "loss": 0.3746, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 6.171794557047173, | |
| "learning_rate": 2.013305537421606e-07, | |
| "loss": 0.447, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 5.130410450658874, | |
| "learning_rate": 1.9396688754838355e-07, | |
| "loss": 0.3469, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 7.39720002220979, | |
| "learning_rate": 1.867390863899543e-07, | |
| "loss": 0.3783, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 7.90865739350042, | |
| "learning_rate": 1.796472504098612e-07, | |
| "loss": 0.4422, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 5.815959678060295, | |
| "learning_rate": 1.726914778672606e-07, | |
| "loss": 0.378, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 5.081919878256831, | |
| "learning_rate": 1.6587186513611685e-07, | |
| "loss": 0.3205, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 6.567585188289477, | |
| "learning_rate": 1.5918850670386677e-07, | |
| "loss": 0.44, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 7.533656800692689, | |
| "learning_rate": 1.5264149517010718e-07, | |
| "loss": 0.4362, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 5.4628885270524865, | |
| "learning_rate": 1.4623092124531613e-07, | |
| "loss": 0.4497, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 5.858826602545033, | |
| "learning_rate": 1.399568737495971e-07, | |
| "loss": 0.43, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 8.572231557747521, | |
| "learning_rate": 1.3381943961144118e-07, | |
| "loss": 0.3925, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 5.873954819161939, | |
| "learning_rate": 1.2781870386653016e-07, | |
| "loss": 0.3683, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 4.581416916538669, | |
| "learning_rate": 1.2195474965655652e-07, | |
| "loss": 0.2704, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 6.965146521509025, | |
| "learning_rate": 1.1622765822806636e-07, | |
| "loss": 0.3669, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 5.699569364212127, | |
| "learning_rate": 1.1063750893134273e-07, | |
| "loss": 0.4125, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 7.0422218435015855, | |
| "learning_rate": 1.0518437921929747e-07, | |
| "loss": 0.3156, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 9.329828931384606, | |
| "learning_rate": 9.986834464640328e-08, | |
| "loss": 0.4834, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 5.211053129225831, | |
| "learning_rate": 9.468947886764557e-08, | |
| "loss": 0.4188, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 5.70280316894895, | |
| "learning_rate": 8.964785363750228e-08, | |
| "loss": 0.3474, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 5.824204755244337, | |
| "learning_rate": 8.474353880894904e-08, | |
| "loss": 0.2739, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 7.610004827598348, | |
| "learning_rate": 7.997660233249105e-08, | |
| "loss": 0.4407, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 6.558806283300632, | |
| "learning_rate": 7.534711025522168e-08, | |
| "loss": 0.3602, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 6.692057122360635, | |
| "learning_rate": 7.08551267199098e-08, | |
| "loss": 0.4439, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 6.4300501195085555, | |
| "learning_rate": 6.650071396410718e-08, | |
| "loss": 0.428, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 6.994677903035368, | |
| "learning_rate": 6.22839323192892e-08, | |
| "loss": 0.517, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 7.2674316715963325, | |
| "learning_rate": 5.820484021001993e-08, | |
| "loss": 0.3041, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 8.22191067382583, | |
| "learning_rate": 5.426349415313503e-08, | |
| "loss": 0.4481, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 5.065088639546076, | |
| "learning_rate": 5.0459948756967914e-08, | |
| "loss": 0.3333, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 8.073376899592944, | |
| "learning_rate": 4.679425672059035e-08, | |
| "loss": 0.4493, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 6.784518754930513, | |
| "learning_rate": 4.326646883307972e-08, | |
| "loss": 0.4454, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 8.57495143340707, | |
| "learning_rate": 3.987663397281627e-08, | |
| "loss": 0.4892, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 5.72551957891776, | |
| "learning_rate": 3.662479910681027e-08, | |
| "loss": 0.4259, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 4.571968724650214, | |
| "learning_rate": 3.3511009290042585e-08, | |
| "loss": 0.364, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 7.563468099974791, | |
| "learning_rate": 3.0535307664848466e-08, | |
| "loss": 0.4774, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 6.8428676776213555, | |
| "learning_rate": 2.7697735460316954e-08, | |
| "loss": 0.4322, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 3.9770591980462786, | |
| "learning_rate": 2.499833199171908e-08, | |
| "loss": 0.3225, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 5.666940466332362, | |
| "learning_rate": 2.2437134659962777e-08, | |
| "loss": 0.4234, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 5.153046102603702, | |
| "learning_rate": 2.0014178951076602e-08, | |
| "loss": 0.471, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 5.709388029240762, | |
| "learning_rate": 1.7729498435716808e-08, | |
| "loss": 0.4391, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 5.983125443983794, | |
| "learning_rate": 1.5583124768701052e-08, | |
| "loss": 0.483, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 6.445238846878224, | |
| "learning_rate": 1.3575087688570965e-08, | |
| "loss": 0.3558, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 6.816452865113282, | |
| "learning_rate": 1.1705415017183585e-08, | |
| "loss": 0.413, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 5.457722190574727, | |
| "learning_rate": 9.974132659319457e-09, | |
| "loss": 0.3282, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 6.969886154201284, | |
| "learning_rate": 8.381264602327354e-09, | |
| "loss": 0.438, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 938, | |
| "total_flos": 373627158528000.0, | |
| "train_loss": 0.7472518730773601, | |
| "train_runtime": 19046.2174, | |
| "train_samples_per_second": 6.306, | |
| "train_steps_per_second": 0.049 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 938, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 200, | |
| "total_flos": 373627158528000.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |