{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 64.51612903225806, "eval_steps": 500, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.3225806451612903, "grad_norm": 4.83341646194458, "learning_rate": 1.8e-07, "loss": 0.6146, "step": 10 }, { "epoch": 0.6451612903225806, "grad_norm": 4.601212978363037, "learning_rate": 3.8e-07, "loss": 0.6156, "step": 20 }, { "epoch": 0.967741935483871, "grad_norm": 4.288501739501953, "learning_rate": 5.8e-07, "loss": 0.6041, "step": 30 }, { "epoch": 1.2903225806451613, "grad_norm": 3.9819118976593018, "learning_rate": 7.8e-07, "loss": 0.5691, "step": 40 }, { "epoch": 1.6129032258064515, "grad_norm": 3.2057559490203857, "learning_rate": 9.8e-07, "loss": 0.4867, "step": 50 }, { "epoch": 1.935483870967742, "grad_norm": 1.8752663135528564, "learning_rate": 1.18e-06, "loss": 0.4109, "step": 60 }, { "epoch": 2.258064516129032, "grad_norm": 1.5188192129135132, "learning_rate": 1.3800000000000001e-06, "loss": 0.3164, "step": 70 }, { "epoch": 2.5806451612903225, "grad_norm": 0.7869158983230591, "learning_rate": 1.5800000000000003e-06, "loss": 0.2642, "step": 80 }, { "epoch": 2.903225806451613, "grad_norm": 0.499815970659256, "learning_rate": 1.7800000000000001e-06, "loss": 0.2078, "step": 90 }, { "epoch": 3.225806451612903, "grad_norm": 0.36157462000846863, "learning_rate": 1.98e-06, "loss": 0.1876, "step": 100 }, { "epoch": 3.5483870967741935, "grad_norm": 0.4228762686252594, "learning_rate": 2.1800000000000003e-06, "loss": 0.167, "step": 110 }, { "epoch": 3.870967741935484, "grad_norm": 0.22709406912326813, "learning_rate": 2.38e-06, "loss": 0.1528, "step": 120 }, { "epoch": 4.193548387096774, "grad_norm": 0.2506711184978485, "learning_rate": 2.5800000000000003e-06, "loss": 0.141, "step": 130 }, { "epoch": 4.516129032258064, "grad_norm": 0.17586937546730042, "learning_rate": 2.78e-06, "loss": 0.1301, "step": 140 }, { "epoch": 4.838709677419355, "grad_norm": 0.2080131471157074, "learning_rate": 2.9800000000000003e-06, "loss": 0.1252, "step": 150 }, { "epoch": 5.161290322580645, "grad_norm": 0.14357304573059082, "learning_rate": 3.1800000000000005e-06, "loss": 0.1174, "step": 160 }, { "epoch": 5.483870967741936, "grad_norm": 0.15768013894557953, "learning_rate": 3.38e-06, "loss": 0.1123, "step": 170 }, { "epoch": 5.806451612903226, "grad_norm": 0.1478148102760315, "learning_rate": 3.58e-06, "loss": 0.1059, "step": 180 }, { "epoch": 6.129032258064516, "grad_norm": 0.11791401356458664, "learning_rate": 3.7800000000000002e-06, "loss": 0.1028, "step": 190 }, { "epoch": 6.451612903225806, "grad_norm": 0.12674356997013092, "learning_rate": 3.98e-06, "loss": 0.1002, "step": 200 }, { "epoch": 6.774193548387097, "grad_norm": 0.10726606845855713, "learning_rate": 4.18e-06, "loss": 0.0936, "step": 210 }, { "epoch": 7.096774193548387, "grad_norm": 0.13341137766838074, "learning_rate": 4.38e-06, "loss": 0.093, "step": 220 }, { "epoch": 7.419354838709677, "grad_norm": 0.09624793380498886, "learning_rate": 4.58e-06, "loss": 0.0891, "step": 230 }, { "epoch": 7.741935483870968, "grad_norm": 0.14156094193458557, "learning_rate": 4.780000000000001e-06, "loss": 0.0861, "step": 240 }, { "epoch": 8.064516129032258, "grad_norm": 0.10326485335826874, "learning_rate": 4.98e-06, "loss": 0.0835, "step": 250 }, { "epoch": 8.387096774193548, "grad_norm": 0.08262120187282562, "learning_rate": 5.18e-06, "loss": 0.0805, "step": 260 }, { "epoch": 8.709677419354838, "grad_norm": 0.1273198425769806, "learning_rate": 5.38e-06, "loss": 0.0779, "step": 270 }, { "epoch": 9.03225806451613, "grad_norm": 0.12214396893978119, "learning_rate": 5.580000000000001e-06, "loss": 0.0757, "step": 280 }, { "epoch": 9.35483870967742, "grad_norm": 0.1394161581993103, "learning_rate": 5.78e-06, "loss": 0.0728, "step": 290 }, { "epoch": 9.67741935483871, "grad_norm": 0.12285201996564865, "learning_rate": 5.98e-06, "loss": 0.0722, "step": 300 }, { "epoch": 10.0, "grad_norm": 0.15410321950912476, "learning_rate": 6.18e-06, "loss": 0.0692, "step": 310 }, { "epoch": 10.32258064516129, "grad_norm": 0.07469774037599564, "learning_rate": 6.38e-06, "loss": 0.0676, "step": 320 }, { "epoch": 10.64516129032258, "grad_norm": 0.16188545525074005, "learning_rate": 6.58e-06, "loss": 0.0643, "step": 330 }, { "epoch": 10.967741935483872, "grad_norm": 0.07787156105041504, "learning_rate": 6.78e-06, "loss": 0.0632, "step": 340 }, { "epoch": 11.290322580645162, "grad_norm": 0.13963359594345093, "learning_rate": 6.98e-06, "loss": 0.0593, "step": 350 }, { "epoch": 11.612903225806452, "grad_norm": 0.12019138038158417, "learning_rate": 7.180000000000001e-06, "loss": 0.0594, "step": 360 }, { "epoch": 11.935483870967742, "grad_norm": 0.13999758660793304, "learning_rate": 7.3800000000000005e-06, "loss": 0.0578, "step": 370 }, { "epoch": 12.258064516129032, "grad_norm": 0.11083265393972397, "learning_rate": 7.580000000000001e-06, "loss": 0.0544, "step": 380 }, { "epoch": 12.580645161290322, "grad_norm": 0.10924753546714783, "learning_rate": 7.78e-06, "loss": 0.0529, "step": 390 }, { "epoch": 12.903225806451612, "grad_norm": 0.10007922351360321, "learning_rate": 7.98e-06, "loss": 0.0519, "step": 400 }, { "epoch": 13.225806451612904, "grad_norm": 0.09409403055906296, "learning_rate": 8.18e-06, "loss": 0.0499, "step": 410 }, { "epoch": 13.548387096774194, "grad_norm": 0.1096971333026886, "learning_rate": 8.380000000000001e-06, "loss": 0.049, "step": 420 }, { "epoch": 13.870967741935484, "grad_norm": 0.10034970194101334, "learning_rate": 8.580000000000001e-06, "loss": 0.0462, "step": 430 }, { "epoch": 14.193548387096774, "grad_norm": 0.13287365436553955, "learning_rate": 8.78e-06, "loss": 0.0454, "step": 440 }, { "epoch": 14.516129032258064, "grad_norm": 0.10294642299413681, "learning_rate": 8.98e-06, "loss": 0.0436, "step": 450 }, { "epoch": 14.838709677419354, "grad_norm": 0.11570604145526886, "learning_rate": 9.180000000000002e-06, "loss": 0.0404, "step": 460 }, { "epoch": 15.161290322580646, "grad_norm": 0.08866477012634277, "learning_rate": 9.38e-06, "loss": 0.0395, "step": 470 }, { "epoch": 15.483870967741936, "grad_norm": 0.1237388625741005, "learning_rate": 9.58e-06, "loss": 0.0377, "step": 480 }, { "epoch": 15.806451612903226, "grad_norm": 0.09724581241607666, "learning_rate": 9.78e-06, "loss": 0.0364, "step": 490 }, { "epoch": 16.129032258064516, "grad_norm": 0.1216067299246788, "learning_rate": 9.980000000000001e-06, "loss": 0.0369, "step": 500 }, { "epoch": 16.451612903225808, "grad_norm": 0.14183978736400604, "learning_rate": 1.018e-05, "loss": 0.0335, "step": 510 }, { "epoch": 16.774193548387096, "grad_norm": 0.10971418768167496, "learning_rate": 1.038e-05, "loss": 0.0337, "step": 520 }, { "epoch": 17.096774193548388, "grad_norm": 0.10205938667058945, "learning_rate": 1.058e-05, "loss": 0.0314, "step": 530 }, { "epoch": 17.419354838709676, "grad_norm": 0.08642183989286423, "learning_rate": 1.0780000000000002e-05, "loss": 0.0309, "step": 540 }, { "epoch": 17.741935483870968, "grad_norm": 0.11878576874732971, "learning_rate": 1.098e-05, "loss": 0.0313, "step": 550 }, { "epoch": 18.06451612903226, "grad_norm": 0.11662951111793518, "learning_rate": 1.118e-05, "loss": 0.03, "step": 560 }, { "epoch": 18.387096774193548, "grad_norm": 0.13768725097179413, "learning_rate": 1.1380000000000001e-05, "loss": 0.0285, "step": 570 }, { "epoch": 18.70967741935484, "grad_norm": 0.1288958489894867, "learning_rate": 1.1580000000000001e-05, "loss": 0.0288, "step": 580 }, { "epoch": 19.032258064516128, "grad_norm": 0.11418294161558151, "learning_rate": 1.178e-05, "loss": 0.0289, "step": 590 }, { "epoch": 19.35483870967742, "grad_norm": 0.12079351395368576, "learning_rate": 1.198e-05, "loss": 0.0287, "step": 600 }, { "epoch": 19.677419354838708, "grad_norm": 0.2023964524269104, "learning_rate": 1.2180000000000002e-05, "loss": 0.0279, "step": 610 }, { "epoch": 20.0, "grad_norm": 0.19349494576454163, "learning_rate": 1.238e-05, "loss": 0.0278, "step": 620 }, { "epoch": 20.322580645161292, "grad_norm": 0.17760895192623138, "learning_rate": 1.258e-05, "loss": 0.0273, "step": 630 }, { "epoch": 20.64516129032258, "grad_norm": 0.1515321135520935, "learning_rate": 1.278e-05, "loss": 0.0257, "step": 640 }, { "epoch": 20.967741935483872, "grad_norm": 0.10156595706939697, "learning_rate": 1.2980000000000001e-05, "loss": 0.0257, "step": 650 }, { "epoch": 21.29032258064516, "grad_norm": 0.12676751613616943, "learning_rate": 1.3180000000000001e-05, "loss": 0.0249, "step": 660 }, { "epoch": 21.612903225806452, "grad_norm": 0.13134433329105377, "learning_rate": 1.338e-05, "loss": 0.0239, "step": 670 }, { "epoch": 21.93548387096774, "grad_norm": 0.13195613026618958, "learning_rate": 1.358e-05, "loss": 0.0242, "step": 680 }, { "epoch": 22.258064516129032, "grad_norm": 0.12307468056678772, "learning_rate": 1.3780000000000002e-05, "loss": 0.0238, "step": 690 }, { "epoch": 22.580645161290324, "grad_norm": 0.1110624298453331, "learning_rate": 1.3980000000000002e-05, "loss": 0.0253, "step": 700 }, { "epoch": 22.903225806451612, "grad_norm": 0.13264206051826477, "learning_rate": 1.4180000000000001e-05, "loss": 0.0239, "step": 710 }, { "epoch": 23.225806451612904, "grad_norm": 0.11219354718923569, "learning_rate": 1.4380000000000001e-05, "loss": 0.0234, "step": 720 }, { "epoch": 23.548387096774192, "grad_norm": 0.14356958866119385, "learning_rate": 1.4580000000000003e-05, "loss": 0.0225, "step": 730 }, { "epoch": 23.870967741935484, "grad_norm": 0.13088181614875793, "learning_rate": 1.4779999999999999e-05, "loss": 0.0221, "step": 740 }, { "epoch": 24.193548387096776, "grad_norm": 0.15026481449604034, "learning_rate": 1.4979999999999999e-05, "loss": 0.0225, "step": 750 }, { "epoch": 24.516129032258064, "grad_norm": 0.20722658932209015, "learning_rate": 1.518e-05, "loss": 0.0225, "step": 760 }, { "epoch": 24.838709677419356, "grad_norm": 0.13866043090820312, "learning_rate": 1.538e-05, "loss": 0.0221, "step": 770 }, { "epoch": 25.161290322580644, "grad_norm": 0.11805257201194763, "learning_rate": 1.558e-05, "loss": 0.0214, "step": 780 }, { "epoch": 25.483870967741936, "grad_norm": 0.1360214352607727, "learning_rate": 1.578e-05, "loss": 0.0205, "step": 790 }, { "epoch": 25.806451612903224, "grad_norm": 0.15285663306713104, "learning_rate": 1.598e-05, "loss": 0.0215, "step": 800 }, { "epoch": 26.129032258064516, "grad_norm": 0.13831627368927002, "learning_rate": 1.618e-05, "loss": 0.021, "step": 810 }, { "epoch": 26.451612903225808, "grad_norm": 0.1219727173447609, "learning_rate": 1.6380000000000002e-05, "loss": 0.0208, "step": 820 }, { "epoch": 26.774193548387096, "grad_norm": 0.14780209958553314, "learning_rate": 1.658e-05, "loss": 0.0212, "step": 830 }, { "epoch": 27.096774193548388, "grad_norm": 0.13057461380958557, "learning_rate": 1.6780000000000002e-05, "loss": 0.0206, "step": 840 }, { "epoch": 27.419354838709676, "grad_norm": 0.13308848440647125, "learning_rate": 1.698e-05, "loss": 0.021, "step": 850 }, { "epoch": 27.741935483870968, "grad_norm": 0.1406048834323883, "learning_rate": 1.718e-05, "loss": 0.0204, "step": 860 }, { "epoch": 28.06451612903226, "grad_norm": 0.1681479662656784, "learning_rate": 1.7380000000000003e-05, "loss": 0.0203, "step": 870 }, { "epoch": 28.387096774193548, "grad_norm": 0.15317517518997192, "learning_rate": 1.758e-05, "loss": 0.0192, "step": 880 }, { "epoch": 28.70967741935484, "grad_norm": 0.17327739298343658, "learning_rate": 1.7780000000000003e-05, "loss": 0.0191, "step": 890 }, { "epoch": 29.032258064516128, "grad_norm": 0.13825540244579315, "learning_rate": 1.798e-05, "loss": 0.0194, "step": 900 }, { "epoch": 29.35483870967742, "grad_norm": 0.14294251799583435, "learning_rate": 1.818e-05, "loss": 0.0191, "step": 910 }, { "epoch": 29.677419354838708, "grad_norm": 0.1148597002029419, "learning_rate": 1.838e-05, "loss": 0.0194, "step": 920 }, { "epoch": 30.0, "grad_norm": 0.10468383878469467, "learning_rate": 1.858e-05, "loss": 0.019, "step": 930 }, { "epoch": 30.322580645161292, "grad_norm": 0.16523204743862152, "learning_rate": 1.878e-05, "loss": 0.0183, "step": 940 }, { "epoch": 30.64516129032258, "grad_norm": 0.1483301818370819, "learning_rate": 1.898e-05, "loss": 0.0183, "step": 950 }, { "epoch": 30.967741935483872, "grad_norm": 0.19092631340026855, "learning_rate": 1.918e-05, "loss": 0.0188, "step": 960 }, { "epoch": 31.29032258064516, "grad_norm": 0.19763915240764618, "learning_rate": 1.938e-05, "loss": 0.0181, "step": 970 }, { "epoch": 31.612903225806452, "grad_norm": 0.18357335031032562, "learning_rate": 1.9580000000000002e-05, "loss": 0.0179, "step": 980 }, { "epoch": 31.93548387096774, "grad_norm": 0.1674022376537323, "learning_rate": 1.978e-05, "loss": 0.0181, "step": 990 }, { "epoch": 32.25806451612903, "grad_norm": 0.1773853302001953, "learning_rate": 1.9980000000000002e-05, "loss": 0.0187, "step": 1000 }, { "epoch": 32.58064516129032, "grad_norm": 0.17372146248817444, "learning_rate": 2.0180000000000003e-05, "loss": 0.0175, "step": 1010 }, { "epoch": 32.903225806451616, "grad_norm": 0.16912385821342468, "learning_rate": 2.038e-05, "loss": 0.0177, "step": 1020 }, { "epoch": 33.225806451612904, "grad_norm": 0.1924871802330017, "learning_rate": 2.0580000000000003e-05, "loss": 0.017, "step": 1030 }, { "epoch": 33.54838709677419, "grad_norm": 0.1980140060186386, "learning_rate": 2.078e-05, "loss": 0.0174, "step": 1040 }, { "epoch": 33.87096774193548, "grad_norm": 0.15361692011356354, "learning_rate": 2.098e-05, "loss": 0.0169, "step": 1050 }, { "epoch": 34.193548387096776, "grad_norm": 0.15744690597057343, "learning_rate": 2.118e-05, "loss": 0.0174, "step": 1060 }, { "epoch": 34.516129032258064, "grad_norm": 0.18941748142242432, "learning_rate": 2.138e-05, "loss": 0.0173, "step": 1070 }, { "epoch": 34.83870967741935, "grad_norm": 0.18401823937892914, "learning_rate": 2.158e-05, "loss": 0.0176, "step": 1080 }, { "epoch": 35.16129032258065, "grad_norm": 0.1347324103116989, "learning_rate": 2.178e-05, "loss": 0.017, "step": 1090 }, { "epoch": 35.483870967741936, "grad_norm": 0.13733318448066711, "learning_rate": 2.198e-05, "loss": 0.0163, "step": 1100 }, { "epoch": 35.806451612903224, "grad_norm": 0.17490868270397186, "learning_rate": 2.218e-05, "loss": 0.0167, "step": 1110 }, { "epoch": 36.12903225806452, "grad_norm": 0.17697739601135254, "learning_rate": 2.2380000000000003e-05, "loss": 0.0173, "step": 1120 }, { "epoch": 36.45161290322581, "grad_norm": 0.19351543486118317, "learning_rate": 2.258e-05, "loss": 0.0174, "step": 1130 }, { "epoch": 36.774193548387096, "grad_norm": 0.17104071378707886, "learning_rate": 2.2780000000000002e-05, "loss": 0.0159, "step": 1140 }, { "epoch": 37.096774193548384, "grad_norm": 0.1536385416984558, "learning_rate": 2.298e-05, "loss": 0.016, "step": 1150 }, { "epoch": 37.41935483870968, "grad_norm": 0.14215905964374542, "learning_rate": 2.318e-05, "loss": 0.0167, "step": 1160 }, { "epoch": 37.74193548387097, "grad_norm": 0.17754510045051575, "learning_rate": 2.3380000000000003e-05, "loss": 0.0163, "step": 1170 }, { "epoch": 38.064516129032256, "grad_norm": 0.15013834834098816, "learning_rate": 2.358e-05, "loss": 0.0152, "step": 1180 }, { "epoch": 38.38709677419355, "grad_norm": 0.17872834205627441, "learning_rate": 2.3780000000000003e-05, "loss": 0.0152, "step": 1190 }, { "epoch": 38.70967741935484, "grad_norm": 0.16451333463191986, "learning_rate": 2.398e-05, "loss": 0.015, "step": 1200 }, { "epoch": 39.03225806451613, "grad_norm": 0.17044350504875183, "learning_rate": 2.418e-05, "loss": 0.0149, "step": 1210 }, { "epoch": 39.354838709677416, "grad_norm": 0.16590650379657745, "learning_rate": 2.438e-05, "loss": 0.0155, "step": 1220 }, { "epoch": 39.67741935483871, "grad_norm": 0.1371670961380005, "learning_rate": 2.4580000000000002e-05, "loss": 0.015, "step": 1230 }, { "epoch": 40.0, "grad_norm": 0.11834592372179031, "learning_rate": 2.478e-05, "loss": 0.0158, "step": 1240 }, { "epoch": 40.32258064516129, "grad_norm": 0.17337927222251892, "learning_rate": 2.498e-05, "loss": 0.0148, "step": 1250 }, { "epoch": 40.645161290322584, "grad_norm": 0.17919541895389557, "learning_rate": 2.5180000000000003e-05, "loss": 0.0146, "step": 1260 }, { "epoch": 40.96774193548387, "grad_norm": 0.1446114331483841, "learning_rate": 2.5380000000000004e-05, "loss": 0.0144, "step": 1270 }, { "epoch": 41.29032258064516, "grad_norm": 0.12958332896232605, "learning_rate": 2.5580000000000002e-05, "loss": 0.0143, "step": 1280 }, { "epoch": 41.61290322580645, "grad_norm": 0.208790123462677, "learning_rate": 2.5779999999999997e-05, "loss": 0.0145, "step": 1290 }, { "epoch": 41.935483870967744, "grad_norm": 0.13380534946918488, "learning_rate": 2.598e-05, "loss": 0.0144, "step": 1300 }, { "epoch": 42.25806451612903, "grad_norm": 0.202678382396698, "learning_rate": 2.618e-05, "loss": 0.0137, "step": 1310 }, { "epoch": 42.58064516129032, "grad_norm": 0.19495125114917755, "learning_rate": 2.6379999999999998e-05, "loss": 0.0147, "step": 1320 }, { "epoch": 42.903225806451616, "grad_norm": 0.17326615750789642, "learning_rate": 2.658e-05, "loss": 0.0143, "step": 1330 }, { "epoch": 43.225806451612904, "grad_norm": 0.17888951301574707, "learning_rate": 2.678e-05, "loss": 0.0143, "step": 1340 }, { "epoch": 43.54838709677419, "grad_norm": 0.1571645885705948, "learning_rate": 2.698e-05, "loss": 0.0136, "step": 1350 }, { "epoch": 43.87096774193548, "grad_norm": 0.13882234692573547, "learning_rate": 2.718e-05, "loss": 0.0136, "step": 1360 }, { "epoch": 44.193548387096776, "grad_norm": 0.12413440644741058, "learning_rate": 2.738e-05, "loss": 0.0132, "step": 1370 }, { "epoch": 44.516129032258064, "grad_norm": 0.13164107501506805, "learning_rate": 2.758e-05, "loss": 0.0138, "step": 1380 }, { "epoch": 44.83870967741935, "grad_norm": 0.14705562591552734, "learning_rate": 2.778e-05, "loss": 0.0135, "step": 1390 }, { "epoch": 45.16129032258065, "grad_norm": 0.17040109634399414, "learning_rate": 2.798e-05, "loss": 0.0136, "step": 1400 }, { "epoch": 45.483870967741936, "grad_norm": 0.17892764508724213, "learning_rate": 2.818e-05, "loss": 0.0135, "step": 1410 }, { "epoch": 45.806451612903224, "grad_norm": 0.14825959503650665, "learning_rate": 2.8380000000000003e-05, "loss": 0.013, "step": 1420 }, { "epoch": 46.12903225806452, "grad_norm": 0.11769016087055206, "learning_rate": 2.858e-05, "loss": 0.0134, "step": 1430 }, { "epoch": 46.45161290322581, "grad_norm": 0.16785204410552979, "learning_rate": 2.8780000000000002e-05, "loss": 0.0131, "step": 1440 }, { "epoch": 46.774193548387096, "grad_norm": 0.12980397045612335, "learning_rate": 2.898e-05, "loss": 0.0141, "step": 1450 }, { "epoch": 47.096774193548384, "grad_norm": 0.19282320141792297, "learning_rate": 2.9180000000000002e-05, "loss": 0.014, "step": 1460 }, { "epoch": 47.41935483870968, "grad_norm": 0.17180287837982178, "learning_rate": 2.9380000000000003e-05, "loss": 0.0133, "step": 1470 }, { "epoch": 47.74193548387097, "grad_norm": 0.1480427086353302, "learning_rate": 2.958e-05, "loss": 0.0133, "step": 1480 }, { "epoch": 48.064516129032256, "grad_norm": 0.18842412531375885, "learning_rate": 2.9780000000000003e-05, "loss": 0.0135, "step": 1490 }, { "epoch": 48.38709677419355, "grad_norm": 0.2088242620229721, "learning_rate": 2.998e-05, "loss": 0.0142, "step": 1500 }, { "epoch": 48.70967741935484, "grad_norm": 0.1986250877380371, "learning_rate": 3.0180000000000002e-05, "loss": 0.0128, "step": 1510 }, { "epoch": 49.03225806451613, "grad_norm": 0.1687200963497162, "learning_rate": 3.0380000000000004e-05, "loss": 0.0127, "step": 1520 }, { "epoch": 49.354838709677416, "grad_norm": 0.1558302789926529, "learning_rate": 3.058e-05, "loss": 0.0131, "step": 1530 }, { "epoch": 49.67741935483871, "grad_norm": 0.17784525454044342, "learning_rate": 3.078e-05, "loss": 0.0125, "step": 1540 }, { "epoch": 50.0, "grad_norm": 0.22291983664035797, "learning_rate": 3.0980000000000005e-05, "loss": 0.0121, "step": 1550 }, { "epoch": 50.32258064516129, "grad_norm": 0.21824175119400024, "learning_rate": 3.118e-05, "loss": 0.012, "step": 1560 }, { "epoch": 50.645161290322584, "grad_norm": 0.2027992457151413, "learning_rate": 3.138e-05, "loss": 0.0127, "step": 1570 }, { "epoch": 50.96774193548387, "grad_norm": 0.2533424496650696, "learning_rate": 3.1580000000000006e-05, "loss": 0.0121, "step": 1580 }, { "epoch": 51.29032258064516, "grad_norm": 0.15423865616321564, "learning_rate": 3.1780000000000004e-05, "loss": 0.0125, "step": 1590 }, { "epoch": 51.61290322580645, "grad_norm": 0.13885457813739777, "learning_rate": 3.198e-05, "loss": 0.013, "step": 1600 }, { "epoch": 51.935483870967744, "grad_norm": 0.1417354941368103, "learning_rate": 3.218e-05, "loss": 0.0119, "step": 1610 }, { "epoch": 52.25806451612903, "grad_norm": 0.17335128784179688, "learning_rate": 3.238e-05, "loss": 0.012, "step": 1620 }, { "epoch": 52.58064516129032, "grad_norm": 0.19661465287208557, "learning_rate": 3.2579999999999996e-05, "loss": 0.0119, "step": 1630 }, { "epoch": 52.903225806451616, "grad_norm": 0.2602691650390625, "learning_rate": 3.278e-05, "loss": 0.0122, "step": 1640 }, { "epoch": 53.225806451612904, "grad_norm": 0.21529579162597656, "learning_rate": 3.298e-05, "loss": 0.0117, "step": 1650 }, { "epoch": 53.54838709677419, "grad_norm": 0.1529514640569687, "learning_rate": 3.318e-05, "loss": 0.0116, "step": 1660 }, { "epoch": 53.87096774193548, "grad_norm": 0.1834433674812317, "learning_rate": 3.338e-05, "loss": 0.0125, "step": 1670 }, { "epoch": 54.193548387096776, "grad_norm": 0.24152186512947083, "learning_rate": 3.358e-05, "loss": 0.0122, "step": 1680 }, { "epoch": 54.516129032258064, "grad_norm": 0.1319815069437027, "learning_rate": 3.378e-05, "loss": 0.0116, "step": 1690 }, { "epoch": 54.83870967741935, "grad_norm": 0.16622133553028107, "learning_rate": 3.398e-05, "loss": 0.0123, "step": 1700 }, { "epoch": 55.16129032258065, "grad_norm": 0.12881875038146973, "learning_rate": 3.418e-05, "loss": 0.0115, "step": 1710 }, { "epoch": 55.483870967741936, "grad_norm": 0.16083110868930817, "learning_rate": 3.438e-05, "loss": 0.0121, "step": 1720 }, { "epoch": 55.806451612903224, "grad_norm": 0.17213866114616394, "learning_rate": 3.4580000000000004e-05, "loss": 0.0111, "step": 1730 }, { "epoch": 56.12903225806452, "grad_norm": 0.1291566640138626, "learning_rate": 3.478e-05, "loss": 0.0113, "step": 1740 }, { "epoch": 56.45161290322581, "grad_norm": 0.1637270748615265, "learning_rate": 3.498e-05, "loss": 0.011, "step": 1750 }, { "epoch": 56.774193548387096, "grad_norm": 0.1748403012752533, "learning_rate": 3.518e-05, "loss": 0.0108, "step": 1760 }, { "epoch": 57.096774193548384, "grad_norm": 0.18834678828716278, "learning_rate": 3.5380000000000003e-05, "loss": 0.0109, "step": 1770 }, { "epoch": 57.41935483870968, "grad_norm": 0.1650463044643402, "learning_rate": 3.558e-05, "loss": 0.0115, "step": 1780 }, { "epoch": 57.74193548387097, "grad_norm": 0.17039620876312256, "learning_rate": 3.578e-05, "loss": 0.0114, "step": 1790 }, { "epoch": 58.064516129032256, "grad_norm": 0.13908374309539795, "learning_rate": 3.5980000000000004e-05, "loss": 0.0107, "step": 1800 }, { "epoch": 58.38709677419355, "grad_norm": 0.1755029857158661, "learning_rate": 3.618e-05, "loss": 0.0104, "step": 1810 }, { "epoch": 58.70967741935484, "grad_norm": 0.16456303000450134, "learning_rate": 3.638e-05, "loss": 0.011, "step": 1820 }, { "epoch": 59.03225806451613, "grad_norm": 0.1215883269906044, "learning_rate": 3.6580000000000006e-05, "loss": 0.0106, "step": 1830 }, { "epoch": 59.354838709677416, "grad_norm": 0.13105925917625427, "learning_rate": 3.6780000000000004e-05, "loss": 0.0104, "step": 1840 }, { "epoch": 59.67741935483871, "grad_norm": 0.14026036858558655, "learning_rate": 3.698e-05, "loss": 0.0105, "step": 1850 }, { "epoch": 60.0, "grad_norm": 0.15687642991542816, "learning_rate": 3.7180000000000007e-05, "loss": 0.0105, "step": 1860 }, { "epoch": 60.32258064516129, "grad_norm": 0.1946769654750824, "learning_rate": 3.7380000000000005e-05, "loss": 0.011, "step": 1870 }, { "epoch": 60.645161290322584, "grad_norm": 0.2352798581123352, "learning_rate": 3.758e-05, "loss": 0.0108, "step": 1880 }, { "epoch": 60.96774193548387, "grad_norm": 0.2311507761478424, "learning_rate": 3.778000000000001e-05, "loss": 0.0119, "step": 1890 }, { "epoch": 61.29032258064516, "grad_norm": 0.24302849173545837, "learning_rate": 3.7980000000000006e-05, "loss": 0.0104, "step": 1900 }, { "epoch": 61.61290322580645, "grad_norm": 0.20787030458450317, "learning_rate": 3.818e-05, "loss": 0.011, "step": 1910 }, { "epoch": 61.935483870967744, "grad_norm": 0.21964043378829956, "learning_rate": 3.838e-05, "loss": 0.0105, "step": 1920 }, { "epoch": 62.25806451612903, "grad_norm": 0.17259149253368378, "learning_rate": 3.858e-05, "loss": 0.011, "step": 1930 }, { "epoch": 62.58064516129032, "grad_norm": 0.2735665440559387, "learning_rate": 3.878e-05, "loss": 0.0108, "step": 1940 }, { "epoch": 62.903225806451616, "grad_norm": 0.19282570481300354, "learning_rate": 3.898e-05, "loss": 0.0108, "step": 1950 }, { "epoch": 63.225806451612904, "grad_norm": 0.21735602617263794, "learning_rate": 3.918e-05, "loss": 0.0098, "step": 1960 }, { "epoch": 63.54838709677419, "grad_norm": 0.192997008562088, "learning_rate": 3.938e-05, "loss": 0.0101, "step": 1970 }, { "epoch": 63.87096774193548, "grad_norm": 0.21661442518234253, "learning_rate": 3.958e-05, "loss": 0.0107, "step": 1980 }, { "epoch": 64.19354838709677, "grad_norm": 0.23013737797737122, "learning_rate": 3.978e-05, "loss": 0.0106, "step": 1990 }, { "epoch": 64.51612903225806, "grad_norm": 0.17058345675468445, "learning_rate": 3.998e-05, "loss": 0.0101, "step": 2000 } ], "logging_steps": 10, "max_steps": 100000, "num_input_tokens_seen": 0, "num_train_epochs": 3226, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 512, "trial_name": null, "trial_params": null }