diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,26523 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 10.0, + "eval_steps": 500, + "global_step": 3310, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0030245746691871457, + "grad_norm": 2.65625, + "learning_rate": 0.0, + "loss": 1.0468, + "num_tokens": 23746.0, + "step": 1 + }, + { + "epoch": 0.0060491493383742915, + "grad_norm": 3.265625, + "learning_rate": 1.0000000000000001e-07, + "loss": 1.0817, + "num_tokens": 37443.0, + "step": 2 + }, + { + "epoch": 0.009073724007561436, + "grad_norm": 2.6875, + "learning_rate": 2.0000000000000002e-07, + "loss": 1.1353, + "num_tokens": 61002.0, + "step": 3 + }, + { + "epoch": 0.012098298676748583, + "grad_norm": 2.96875, + "learning_rate": 3.0000000000000004e-07, + "loss": 1.108, + "num_tokens": 84752.0, + "step": 4 + }, + { + "epoch": 0.015122873345935728, + "grad_norm": 2.890625, + "learning_rate": 4.0000000000000003e-07, + "loss": 1.1043, + "num_tokens": 108010.0, + "step": 5 + }, + { + "epoch": 0.018147448015122872, + "grad_norm": 2.46875, + "learning_rate": 5.000000000000001e-07, + "loss": 1.0554, + "num_tokens": 151360.0, + "step": 6 + }, + { + "epoch": 0.02117202268431002, + "grad_norm": 2.515625, + "learning_rate": 6.000000000000001e-07, + "loss": 1.0661, + "num_tokens": 189902.0, + "step": 7 + }, + { + "epoch": 0.024196597353497166, + "grad_norm": 3.921875, + "learning_rate": 7.000000000000001e-07, + "loss": 0.7904, + "num_tokens": 198464.0, + "step": 8 + }, + { + "epoch": 0.02722117202268431, + "grad_norm": 4.375, + "learning_rate": 8.000000000000001e-07, + "loss": 0.8451, + "num_tokens": 206486.0, + "step": 9 + }, + { + "epoch": 0.030245746691871456, + "grad_norm": 3.078125, + "learning_rate": 9.000000000000001e-07, + "loss": 1.1746, + "num_tokens": 229852.0, + "step": 10 + }, + { + "epoch": 0.0332703213610586, + "grad_norm": 2.828125, + "learning_rate": 1.0000000000000002e-06, + "loss": 0.9696, + "num_tokens": 245486.0, + "step": 11 + }, + { + "epoch": 0.036294896030245744, + "grad_norm": 3.125, + "learning_rate": 1.1e-06, + "loss": 1.1226, + "num_tokens": 268286.0, + "step": 12 + }, + { + "epoch": 0.03931947069943289, + "grad_norm": 4.65625, + "learning_rate": 1.2000000000000002e-06, + "loss": 0.9715, + "num_tokens": 275097.0, + "step": 13 + }, + { + "epoch": 0.04234404536862004, + "grad_norm": 4.21875, + "learning_rate": 1.3e-06, + "loss": 0.9695, + "num_tokens": 284453.0, + "step": 14 + }, + { + "epoch": 0.045368620037807186, + "grad_norm": 2.859375, + "learning_rate": 1.4000000000000001e-06, + "loss": 1.1039, + "num_tokens": 308396.0, + "step": 15 + }, + { + "epoch": 0.04839319470699433, + "grad_norm": 2.546875, + "learning_rate": 1.5e-06, + "loss": 1.0022, + "num_tokens": 331928.0, + "step": 16 + }, + { + "epoch": 0.05141776937618148, + "grad_norm": 2.640625, + "learning_rate": 1.6000000000000001e-06, + "loss": 1.0871, + "num_tokens": 361062.0, + "step": 17 + }, + { + "epoch": 0.05444234404536862, + "grad_norm": 4.25, + "learning_rate": 1.7000000000000002e-06, + "loss": 0.8392, + "num_tokens": 369445.0, + "step": 18 + }, + { + "epoch": 0.05746691871455577, + "grad_norm": 2.515625, + "learning_rate": 1.8000000000000001e-06, + "loss": 1.1283, + "num_tokens": 430994.0, + "step": 19 + }, + { + "epoch": 0.06049149338374291, + "grad_norm": 2.859375, + "learning_rate": 1.9000000000000002e-06, + "loss": 1.0474, + "num_tokens": 446540.0, + "step": 20 + }, + { + "epoch": 0.06351606805293006, + "grad_norm": 2.53125, + "learning_rate": 2.0000000000000003e-06, + "loss": 1.1249, + "num_tokens": 507041.0, + "step": 21 + }, + { + "epoch": 0.0665406427221172, + "grad_norm": 2.75, + "learning_rate": 2.1000000000000002e-06, + "loss": 1.1529, + "num_tokens": 530332.0, + "step": 22 + }, + { + "epoch": 0.06956521739130435, + "grad_norm": 2.84375, + "learning_rate": 2.2e-06, + "loss": 1.0442, + "num_tokens": 553120.0, + "step": 23 + }, + { + "epoch": 0.07258979206049149, + "grad_norm": 2.5625, + "learning_rate": 2.3000000000000004e-06, + "loss": 1.1289, + "num_tokens": 584651.0, + "step": 24 + }, + { + "epoch": 0.07561436672967864, + "grad_norm": 2.265625, + "learning_rate": 2.4000000000000003e-06, + "loss": 1.0356, + "num_tokens": 620987.0, + "step": 25 + }, + { + "epoch": 0.07863894139886578, + "grad_norm": 2.765625, + "learning_rate": 2.5e-06, + "loss": 1.1145, + "num_tokens": 646574.0, + "step": 26 + }, + { + "epoch": 0.08166351606805293, + "grad_norm": 3.0625, + "learning_rate": 2.6e-06, + "loss": 1.1062, + "num_tokens": 662134.0, + "step": 27 + }, + { + "epoch": 0.08468809073724008, + "grad_norm": 2.5, + "learning_rate": 2.7000000000000004e-06, + "loss": 1.1243, + "num_tokens": 723179.0, + "step": 28 + }, + { + "epoch": 0.08771266540642722, + "grad_norm": 2.53125, + "learning_rate": 2.8000000000000003e-06, + "loss": 1.0651, + "num_tokens": 753576.0, + "step": 29 + }, + { + "epoch": 0.09073724007561437, + "grad_norm": 2.53125, + "learning_rate": 2.9e-06, + "loss": 1.1814, + "num_tokens": 793613.0, + "step": 30 + }, + { + "epoch": 0.09376181474480151, + "grad_norm": 2.515625, + "learning_rate": 3e-06, + "loss": 1.0648, + "num_tokens": 821745.0, + "step": 31 + }, + { + "epoch": 0.09678638941398866, + "grad_norm": 3.875, + "learning_rate": 3.1000000000000004e-06, + "loss": 0.8822, + "num_tokens": 829912.0, + "step": 32 + }, + { + "epoch": 0.0998109640831758, + "grad_norm": 2.578125, + "learning_rate": 3.2000000000000003e-06, + "loss": 1.0035, + "num_tokens": 853387.0, + "step": 33 + }, + { + "epoch": 0.10283553875236295, + "grad_norm": 2.578125, + "learning_rate": 3.3000000000000006e-06, + "loss": 0.9935, + "num_tokens": 876049.0, + "step": 34 + }, + { + "epoch": 0.10586011342155009, + "grad_norm": 4.15625, + "learning_rate": 3.4000000000000005e-06, + "loss": 0.8918, + "num_tokens": 884031.0, + "step": 35 + }, + { + "epoch": 0.10888468809073724, + "grad_norm": 2.53125, + "learning_rate": 3.5e-06, + "loss": 1.0686, + "num_tokens": 915961.0, + "step": 36 + }, + { + "epoch": 0.11190926275992438, + "grad_norm": 2.328125, + "learning_rate": 3.6000000000000003e-06, + "loss": 1.1114, + "num_tokens": 967307.0, + "step": 37 + }, + { + "epoch": 0.11493383742911154, + "grad_norm": 2.375, + "learning_rate": 3.7e-06, + "loss": 1.0939, + "num_tokens": 1004272.0, + "step": 38 + }, + { + "epoch": 0.11795841209829867, + "grad_norm": 2.3125, + "learning_rate": 3.8000000000000005e-06, + "loss": 1.1418, + "num_tokens": 1038790.0, + "step": 39 + }, + { + "epoch": 0.12098298676748583, + "grad_norm": 4.375, + "learning_rate": 3.900000000000001e-06, + "loss": 0.9211, + "num_tokens": 1045663.0, + "step": 40 + }, + { + "epoch": 0.12400756143667296, + "grad_norm": 4.59375, + "learning_rate": 4.000000000000001e-06, + "loss": 0.8851, + "num_tokens": 1052878.0, + "step": 41 + }, + { + "epoch": 0.12703213610586012, + "grad_norm": 2.296875, + "learning_rate": 4.1e-06, + "loss": 1.0011, + "num_tokens": 1074839.0, + "step": 42 + }, + { + "epoch": 0.13005671077504727, + "grad_norm": 3.421875, + "learning_rate": 4.2000000000000004e-06, + "loss": 1.0592, + "num_tokens": 1085674.0, + "step": 43 + }, + { + "epoch": 0.1330812854442344, + "grad_norm": 2.296875, + "learning_rate": 4.3e-06, + "loss": 1.1155, + "num_tokens": 1135036.0, + "step": 44 + }, + { + "epoch": 0.13610586011342155, + "grad_norm": 2.640625, + "learning_rate": 4.4e-06, + "loss": 1.0414, + "num_tokens": 1155003.0, + "step": 45 + }, + { + "epoch": 0.1391304347826087, + "grad_norm": 2.234375, + "learning_rate": 4.5e-06, + "loss": 1.1091, + "num_tokens": 1193462.0, + "step": 46 + }, + { + "epoch": 0.14215500945179585, + "grad_norm": 2.515625, + "learning_rate": 4.600000000000001e-06, + "loss": 0.97, + "num_tokens": 1217342.0, + "step": 47 + }, + { + "epoch": 0.14517958412098297, + "grad_norm": 2.09375, + "learning_rate": 4.7e-06, + "loss": 1.054, + "num_tokens": 1260419.0, + "step": 48 + }, + { + "epoch": 0.14820415879017013, + "grad_norm": 5.40625, + "learning_rate": 4.800000000000001e-06, + "loss": 0.908, + "num_tokens": 1266656.0, + "step": 49 + }, + { + "epoch": 0.15122873345935728, + "grad_norm": 4.53125, + "learning_rate": 4.9000000000000005e-06, + "loss": 0.8352, + "num_tokens": 1273981.0, + "step": 50 + }, + { + "epoch": 0.15425330812854443, + "grad_norm": 2.4375, + "learning_rate": 5e-06, + "loss": 1.0796, + "num_tokens": 1297313.0, + "step": 51 + }, + { + "epoch": 0.15727788279773156, + "grad_norm": 4.40625, + "learning_rate": 5.1e-06, + "loss": 0.7765, + "num_tokens": 1305576.0, + "step": 52 + }, + { + "epoch": 0.1603024574669187, + "grad_norm": 2.328125, + "learning_rate": 5.2e-06, + "loss": 1.0755, + "num_tokens": 1334145.0, + "step": 53 + }, + { + "epoch": 0.16332703213610586, + "grad_norm": 4.125, + "learning_rate": 5.300000000000001e-06, + "loss": 0.776, + "num_tokens": 1342046.0, + "step": 54 + }, + { + "epoch": 0.166351606805293, + "grad_norm": 2.328125, + "learning_rate": 5.400000000000001e-06, + "loss": 1.0456, + "num_tokens": 1364835.0, + "step": 55 + }, + { + "epoch": 0.16937618147448016, + "grad_norm": 2.1875, + "learning_rate": 5.500000000000001e-06, + "loss": 0.9556, + "num_tokens": 1388769.0, + "step": 56 + }, + { + "epoch": 0.1724007561436673, + "grad_norm": 2.390625, + "learning_rate": 5.600000000000001e-06, + "loss": 1.0411, + "num_tokens": 1407877.0, + "step": 57 + }, + { + "epoch": 0.17542533081285444, + "grad_norm": 3.390625, + "learning_rate": 5.7e-06, + "loss": 0.749, + "num_tokens": 1416143.0, + "step": 58 + }, + { + "epoch": 0.1784499054820416, + "grad_norm": 2.15625, + "learning_rate": 5.8e-06, + "loss": 0.9762, + "num_tokens": 1439948.0, + "step": 59 + }, + { + "epoch": 0.18147448015122875, + "grad_norm": 3.15625, + "learning_rate": 5.9e-06, + "loss": 0.9186, + "num_tokens": 1449132.0, + "step": 60 + }, + { + "epoch": 0.18449905482041587, + "grad_norm": 2.421875, + "learning_rate": 6e-06, + "loss": 1.088, + "num_tokens": 1472293.0, + "step": 61 + }, + { + "epoch": 0.18752362948960302, + "grad_norm": 2.53125, + "learning_rate": 6.1e-06, + "loss": 0.9892, + "num_tokens": 1490014.0, + "step": 62 + }, + { + "epoch": 0.19054820415879017, + "grad_norm": 2.96875, + "learning_rate": 6.200000000000001e-06, + "loss": 1.1083, + "num_tokens": 1513511.0, + "step": 63 + }, + { + "epoch": 0.19357277882797733, + "grad_norm": 2.21875, + "learning_rate": 6.300000000000001e-06, + "loss": 1.0156, + "num_tokens": 1548403.0, + "step": 64 + }, + { + "epoch": 0.19659735349716445, + "grad_norm": 2.4375, + "learning_rate": 6.4000000000000006e-06, + "loss": 1.0295, + "num_tokens": 1571959.0, + "step": 65 + }, + { + "epoch": 0.1996219281663516, + "grad_norm": 3.828125, + "learning_rate": 6.5000000000000004e-06, + "loss": 0.8228, + "num_tokens": 1578179.0, + "step": 66 + }, + { + "epoch": 0.20264650283553876, + "grad_norm": 2.125, + "learning_rate": 6.600000000000001e-06, + "loss": 1.0569, + "num_tokens": 1607259.0, + "step": 67 + }, + { + "epoch": 0.2056710775047259, + "grad_norm": 3.25, + "learning_rate": 6.700000000000001e-06, + "loss": 0.7832, + "num_tokens": 1614743.0, + "step": 68 + }, + { + "epoch": 0.20869565217391303, + "grad_norm": 1.7265625, + "learning_rate": 6.800000000000001e-06, + "loss": 1.079, + "num_tokens": 1670185.0, + "step": 69 + }, + { + "epoch": 0.21172022684310018, + "grad_norm": 2.875, + "learning_rate": 6.9e-06, + "loss": 0.8326, + "num_tokens": 1677987.0, + "step": 70 + }, + { + "epoch": 0.21474480151228734, + "grad_norm": 3.5, + "learning_rate": 7e-06, + "loss": 0.7989, + "num_tokens": 1685233.0, + "step": 71 + }, + { + "epoch": 0.2177693761814745, + "grad_norm": 2.03125, + "learning_rate": 7.100000000000001e-06, + "loss": 1.0542, + "num_tokens": 1708734.0, + "step": 72 + }, + { + "epoch": 0.2207939508506616, + "grad_norm": 1.9296875, + "learning_rate": 7.2000000000000005e-06, + "loss": 1.0748, + "num_tokens": 1742967.0, + "step": 73 + }, + { + "epoch": 0.22381852551984877, + "grad_norm": 1.9765625, + "learning_rate": 7.3e-06, + "loss": 1.0216, + "num_tokens": 1765917.0, + "step": 74 + }, + { + "epoch": 0.22684310018903592, + "grad_norm": 2.71875, + "learning_rate": 7.4e-06, + "loss": 0.8749, + "num_tokens": 1776446.0, + "step": 75 + }, + { + "epoch": 0.22986767485822307, + "grad_norm": 3.265625, + "learning_rate": 7.500000000000001e-06, + "loss": 0.8232, + "num_tokens": 1782976.0, + "step": 76 + }, + { + "epoch": 0.2328922495274102, + "grad_norm": 2.71875, + "learning_rate": 7.600000000000001e-06, + "loss": 0.7087, + "num_tokens": 1790709.0, + "step": 77 + }, + { + "epoch": 0.23591682419659735, + "grad_norm": 2.109375, + "learning_rate": 7.7e-06, + "loss": 0.9672, + "num_tokens": 1803983.0, + "step": 78 + }, + { + "epoch": 0.2389413988657845, + "grad_norm": 2.5, + "learning_rate": 7.800000000000002e-06, + "loss": 0.8287, + "num_tokens": 1815981.0, + "step": 79 + }, + { + "epoch": 0.24196597353497165, + "grad_norm": 1.625, + "learning_rate": 7.9e-06, + "loss": 1.0402, + "num_tokens": 1850783.0, + "step": 80 + }, + { + "epoch": 0.2449905482041588, + "grad_norm": 1.7578125, + "learning_rate": 8.000000000000001e-06, + "loss": 0.9594, + "num_tokens": 1884143.0, + "step": 81 + }, + { + "epoch": 0.24801512287334593, + "grad_norm": 1.859375, + "learning_rate": 8.1e-06, + "loss": 0.9735, + "num_tokens": 1907950.0, + "step": 82 + }, + { + "epoch": 0.2510396975425331, + "grad_norm": 3.28125, + "learning_rate": 8.2e-06, + "loss": 0.7522, + "num_tokens": 1915348.0, + "step": 83 + }, + { + "epoch": 0.25406427221172023, + "grad_norm": 1.765625, + "learning_rate": 8.3e-06, + "loss": 1.006, + "num_tokens": 1946807.0, + "step": 84 + }, + { + "epoch": 0.2570888468809074, + "grad_norm": 1.9609375, + "learning_rate": 8.400000000000001e-06, + "loss": 1.008, + "num_tokens": 1973616.0, + "step": 85 + }, + { + "epoch": 0.26011342155009454, + "grad_norm": 1.7421875, + "learning_rate": 8.5e-06, + "loss": 1.0743, + "num_tokens": 2011629.0, + "step": 86 + }, + { + "epoch": 0.2631379962192817, + "grad_norm": 2.375, + "learning_rate": 8.6e-06, + "loss": 0.9687, + "num_tokens": 2025439.0, + "step": 87 + }, + { + "epoch": 0.2661625708884688, + "grad_norm": 1.6953125, + "learning_rate": 8.700000000000001e-06, + "loss": 1.0407, + "num_tokens": 2058035.0, + "step": 88 + }, + { + "epoch": 0.26918714555765594, + "grad_norm": 1.921875, + "learning_rate": 8.8e-06, + "loss": 1.0144, + "num_tokens": 2081303.0, + "step": 89 + }, + { + "epoch": 0.2722117202268431, + "grad_norm": 2.25, + "learning_rate": 8.900000000000001e-06, + "loss": 0.8782, + "num_tokens": 2104775.0, + "step": 90 + }, + { + "epoch": 0.27523629489603024, + "grad_norm": 2.609375, + "learning_rate": 9e-06, + "loss": 0.6739, + "num_tokens": 2113135.0, + "step": 91 + }, + { + "epoch": 0.2782608695652174, + "grad_norm": 1.71875, + "learning_rate": 9.100000000000001e-06, + "loss": 1.0268, + "num_tokens": 2146175.0, + "step": 92 + }, + { + "epoch": 0.28128544423440455, + "grad_norm": 1.8359375, + "learning_rate": 9.200000000000002e-06, + "loss": 1.0327, + "num_tokens": 2176223.0, + "step": 93 + }, + { + "epoch": 0.2843100189035917, + "grad_norm": 1.96875, + "learning_rate": 9.3e-06, + "loss": 0.9401, + "num_tokens": 2198277.0, + "step": 94 + }, + { + "epoch": 0.28733459357277885, + "grad_norm": 2.859375, + "learning_rate": 9.4e-06, + "loss": 0.7715, + "num_tokens": 2204906.0, + "step": 95 + }, + { + "epoch": 0.29035916824196595, + "grad_norm": 1.609375, + "learning_rate": 9.5e-06, + "loss": 0.8582, + "num_tokens": 2228694.0, + "step": 96 + }, + { + "epoch": 0.2933837429111531, + "grad_norm": 2.296875, + "learning_rate": 9.600000000000001e-06, + "loss": 0.8738, + "num_tokens": 2241661.0, + "step": 97 + }, + { + "epoch": 0.29640831758034025, + "grad_norm": 1.8671875, + "learning_rate": 9.7e-06, + "loss": 0.8727, + "num_tokens": 2265328.0, + "step": 98 + }, + { + "epoch": 0.2994328922495274, + "grad_norm": 2.515625, + "learning_rate": 9.800000000000001e-06, + "loss": 0.7975, + "num_tokens": 2277582.0, + "step": 99 + }, + { + "epoch": 0.30245746691871456, + "grad_norm": 1.234375, + "learning_rate": 9.9e-06, + "loss": 0.958, + "num_tokens": 2349533.0, + "step": 100 + }, + { + "epoch": 0.3054820415879017, + "grad_norm": 2.859375, + "learning_rate": 1e-05, + "loss": 0.6911, + "num_tokens": 2356643.0, + "step": 101 + }, + { + "epoch": 0.30850661625708886, + "grad_norm": 2.46875, + "learning_rate": 9.999997844876494e-06, + "loss": 0.7003, + "num_tokens": 2365209.0, + "step": 102 + }, + { + "epoch": 0.311531190926276, + "grad_norm": 1.9453125, + "learning_rate": 9.999991379508036e-06, + "loss": 1.0735, + "num_tokens": 2388576.0, + "step": 103 + }, + { + "epoch": 0.3145557655954631, + "grad_norm": 1.859375, + "learning_rate": 9.999980603900821e-06, + "loss": 0.9467, + "num_tokens": 2414595.0, + "step": 104 + }, + { + "epoch": 0.31758034026465026, + "grad_norm": 2.25, + "learning_rate": 9.999965518065171e-06, + "loss": 0.8261, + "num_tokens": 2429919.0, + "step": 105 + }, + { + "epoch": 0.3206049149338374, + "grad_norm": 2.734375, + "learning_rate": 9.999946122015535e-06, + "loss": 0.9019, + "num_tokens": 2444699.0, + "step": 106 + }, + { + "epoch": 0.32362948960302457, + "grad_norm": 2.59375, + "learning_rate": 9.99992241577049e-06, + "loss": 0.6962, + "num_tokens": 2452149.0, + "step": 107 + }, + { + "epoch": 0.3266540642722117, + "grad_norm": 2.171875, + "learning_rate": 9.999894399352742e-06, + "loss": 0.8903, + "num_tokens": 2467003.0, + "step": 108 + }, + { + "epoch": 0.32967863894139887, + "grad_norm": 2.859375, + "learning_rate": 9.999862072789131e-06, + "loss": 0.9068, + "num_tokens": 2490617.0, + "step": 109 + }, + { + "epoch": 0.332703213610586, + "grad_norm": 2.203125, + "learning_rate": 9.999825436110612e-06, + "loss": 0.6851, + "num_tokens": 2498313.0, + "step": 110 + }, + { + "epoch": 0.3357277882797732, + "grad_norm": 1.515625, + "learning_rate": 9.999784489352286e-06, + "loss": 0.9222, + "num_tokens": 2536173.0, + "step": 111 + }, + { + "epoch": 0.33875236294896033, + "grad_norm": 1.53125, + "learning_rate": 9.999739232553367e-06, + "loss": 1.0267, + "num_tokens": 2577448.0, + "step": 112 + }, + { + "epoch": 0.3417769376181474, + "grad_norm": 2.46875, + "learning_rate": 9.999689665757205e-06, + "loss": 0.8355, + "num_tokens": 2589660.0, + "step": 113 + }, + { + "epoch": 0.3448015122873346, + "grad_norm": 2.265625, + "learning_rate": 9.999635789011277e-06, + "loss": 0.859, + "num_tokens": 2612855.0, + "step": 114 + }, + { + "epoch": 0.34782608695652173, + "grad_norm": 2.203125, + "learning_rate": 9.99957760236719e-06, + "loss": 0.8669, + "num_tokens": 2626680.0, + "step": 115 + }, + { + "epoch": 0.3508506616257089, + "grad_norm": 1.7890625, + "learning_rate": 9.999515105880674e-06, + "loss": 0.9591, + "num_tokens": 2650951.0, + "step": 116 + }, + { + "epoch": 0.35387523629489603, + "grad_norm": 1.59375, + "learning_rate": 9.999448299611589e-06, + "loss": 1.0424, + "num_tokens": 2691304.0, + "step": 117 + }, + { + "epoch": 0.3568998109640832, + "grad_norm": 1.7109375, + "learning_rate": 9.999377183623929e-06, + "loss": 0.9227, + "num_tokens": 2710629.0, + "step": 118 + }, + { + "epoch": 0.35992438563327034, + "grad_norm": 1.921875, + "learning_rate": 9.999301757985807e-06, + "loss": 0.8853, + "num_tokens": 2725003.0, + "step": 119 + }, + { + "epoch": 0.3629489603024575, + "grad_norm": 2.078125, + "learning_rate": 9.99922202276947e-06, + "loss": 0.6325, + "num_tokens": 2732632.0, + "step": 120 + }, + { + "epoch": 0.3659735349716446, + "grad_norm": 1.8984375, + "learning_rate": 9.999137978051292e-06, + "loss": 0.9182, + "num_tokens": 2748957.0, + "step": 121 + }, + { + "epoch": 0.36899810964083174, + "grad_norm": 1.9453125, + "learning_rate": 9.999049623911771e-06, + "loss": 0.9456, + "num_tokens": 2772191.0, + "step": 122 + }, + { + "epoch": 0.3720226843100189, + "grad_norm": 2.046875, + "learning_rate": 9.998956960435538e-06, + "loss": 0.6343, + "num_tokens": 2779763.0, + "step": 123 + }, + { + "epoch": 0.37504725897920604, + "grad_norm": 2.328125, + "learning_rate": 9.998859987711347e-06, + "loss": 0.7665, + "num_tokens": 2786745.0, + "step": 124 + }, + { + "epoch": 0.3780718336483932, + "grad_norm": 1.5546875, + "learning_rate": 9.998758705832084e-06, + "loss": 0.9268, + "num_tokens": 2820348.0, + "step": 125 + }, + { + "epoch": 0.38109640831758035, + "grad_norm": 2.15625, + "learning_rate": 9.998653114894759e-06, + "loss": 0.7724, + "num_tokens": 2832530.0, + "step": 126 + }, + { + "epoch": 0.3841209829867675, + "grad_norm": 1.5546875, + "learning_rate": 9.998543215000508e-06, + "loss": 0.8766, + "num_tokens": 2862294.0, + "step": 127 + }, + { + "epoch": 0.38714555765595465, + "grad_norm": 1.6875, + "learning_rate": 9.998429006254601e-06, + "loss": 0.9822, + "num_tokens": 2889340.0, + "step": 128 + }, + { + "epoch": 0.39017013232514175, + "grad_norm": 2.03125, + "learning_rate": 9.998310488766428e-06, + "loss": 0.8753, + "num_tokens": 2903007.0, + "step": 129 + }, + { + "epoch": 0.3931947069943289, + "grad_norm": 1.640625, + "learning_rate": 9.99818766264951e-06, + "loss": 1.0157, + "num_tokens": 2934000.0, + "step": 130 + }, + { + "epoch": 0.39621928166351605, + "grad_norm": 1.734375, + "learning_rate": 9.998060528021493e-06, + "loss": 1.021, + "num_tokens": 2964184.0, + "step": 131 + }, + { + "epoch": 0.3992438563327032, + "grad_norm": 2.0625, + "learning_rate": 9.997929085004153e-06, + "loss": 0.7585, + "num_tokens": 2976995.0, + "step": 132 + }, + { + "epoch": 0.40226843100189036, + "grad_norm": 2.171875, + "learning_rate": 9.997793333723389e-06, + "loss": 0.6981, + "num_tokens": 2984973.0, + "step": 133 + }, + { + "epoch": 0.4052930056710775, + "grad_norm": 1.609375, + "learning_rate": 9.997653274309225e-06, + "loss": 0.9678, + "num_tokens": 3008497.0, + "step": 134 + }, + { + "epoch": 0.40831758034026466, + "grad_norm": 1.8359375, + "learning_rate": 9.997508906895819e-06, + "loss": 1.0138, + "num_tokens": 3031921.0, + "step": 135 + }, + { + "epoch": 0.4113421550094518, + "grad_norm": 1.4296875, + "learning_rate": 9.997360231621448e-06, + "loss": 0.9905, + "num_tokens": 3079067.0, + "step": 136 + }, + { + "epoch": 0.41436672967863897, + "grad_norm": 2.1875, + "learning_rate": 9.99720724862852e-06, + "loss": 0.7457, + "num_tokens": 3087376.0, + "step": 137 + }, + { + "epoch": 0.41739130434782606, + "grad_norm": 1.8125, + "learning_rate": 9.997049958063568e-06, + "loss": 0.8156, + "num_tokens": 3105117.0, + "step": 138 + }, + { + "epoch": 0.4204158790170132, + "grad_norm": 2.109375, + "learning_rate": 9.996888360077244e-06, + "loss": 0.6617, + "num_tokens": 3112310.0, + "step": 139 + }, + { + "epoch": 0.42344045368620037, + "grad_norm": 1.3984375, + "learning_rate": 9.99672245482434e-06, + "loss": 0.9457, + "num_tokens": 3153305.0, + "step": 140 + }, + { + "epoch": 0.4264650283553875, + "grad_norm": 1.453125, + "learning_rate": 9.99655224246376e-06, + "loss": 0.9472, + "num_tokens": 3188431.0, + "step": 141 + }, + { + "epoch": 0.4294896030245747, + "grad_norm": 2.125, + "learning_rate": 9.99637772315854e-06, + "loss": 0.6619, + "num_tokens": 3195427.0, + "step": 142 + }, + { + "epoch": 0.4325141776937618, + "grad_norm": 1.515625, + "learning_rate": 9.996198897075842e-06, + "loss": 0.9586, + "num_tokens": 3231587.0, + "step": 143 + }, + { + "epoch": 0.435538752362949, + "grad_norm": 1.8359375, + "learning_rate": 9.99601576438695e-06, + "loss": 0.8596, + "num_tokens": 3254138.0, + "step": 144 + }, + { + "epoch": 0.43856332703213613, + "grad_norm": 2.015625, + "learning_rate": 9.995828325267276e-06, + "loss": 0.6113, + "num_tokens": 3261146.0, + "step": 145 + }, + { + "epoch": 0.4415879017013232, + "grad_norm": 2.109375, + "learning_rate": 9.99563657989635e-06, + "loss": 0.7657, + "num_tokens": 3268251.0, + "step": 146 + }, + { + "epoch": 0.4446124763705104, + "grad_norm": 2.234375, + "learning_rate": 9.995440528457839e-06, + "loss": 0.6457, + "num_tokens": 3275330.0, + "step": 147 + }, + { + "epoch": 0.44763705103969753, + "grad_norm": 2.390625, + "learning_rate": 9.995240171139524e-06, + "loss": 0.9572, + "num_tokens": 3297793.0, + "step": 148 + }, + { + "epoch": 0.4506616257088847, + "grad_norm": 2.015625, + "learning_rate": 9.995035508133316e-06, + "loss": 0.6911, + "num_tokens": 3305895.0, + "step": 149 + }, + { + "epoch": 0.45368620037807184, + "grad_norm": 1.53125, + "learning_rate": 9.994826539635244e-06, + "loss": 0.9909, + "num_tokens": 3343754.0, + "step": 150 + }, + { + "epoch": 0.456710775047259, + "grad_norm": 1.40625, + "learning_rate": 9.994613265845468e-06, + "loss": 0.9325, + "num_tokens": 3383953.0, + "step": 151 + }, + { + "epoch": 0.45973534971644614, + "grad_norm": 1.6875, + "learning_rate": 9.994395686968267e-06, + "loss": 0.9507, + "num_tokens": 3407405.0, + "step": 152 + }, + { + "epoch": 0.4627599243856333, + "grad_norm": 1.515625, + "learning_rate": 9.994173803212046e-06, + "loss": 0.8712, + "num_tokens": 3439938.0, + "step": 153 + }, + { + "epoch": 0.4657844990548204, + "grad_norm": 2.09375, + "learning_rate": 9.993947614789332e-06, + "loss": 0.7125, + "num_tokens": 3447485.0, + "step": 154 + }, + { + "epoch": 0.46880907372400754, + "grad_norm": 1.828125, + "learning_rate": 9.993717121916778e-06, + "loss": 0.8834, + "num_tokens": 3470932.0, + "step": 155 + }, + { + "epoch": 0.4718336483931947, + "grad_norm": 2.109375, + "learning_rate": 9.993482324815152e-06, + "loss": 0.6504, + "num_tokens": 3477422.0, + "step": 156 + }, + { + "epoch": 0.47485822306238185, + "grad_norm": 2.765625, + "learning_rate": 9.993243223709357e-06, + "loss": 0.6731, + "num_tokens": 3485888.0, + "step": 157 + }, + { + "epoch": 0.477882797731569, + "grad_norm": 2.828125, + "learning_rate": 9.992999818828408e-06, + "loss": 0.7752, + "num_tokens": 3492023.0, + "step": 158 + }, + { + "epoch": 0.48090737240075615, + "grad_norm": 2.046875, + "learning_rate": 9.992752110405448e-06, + "loss": 0.6656, + "num_tokens": 3499245.0, + "step": 159 + }, + { + "epoch": 0.4839319470699433, + "grad_norm": 1.2421875, + "learning_rate": 9.992500098677737e-06, + "loss": 0.9225, + "num_tokens": 3551409.0, + "step": 160 + }, + { + "epoch": 0.48695652173913045, + "grad_norm": 1.859375, + "learning_rate": 9.992243783886663e-06, + "loss": 0.9791, + "num_tokens": 3575841.0, + "step": 161 + }, + { + "epoch": 0.4899810964083176, + "grad_norm": 2.25, + "learning_rate": 9.991983166277731e-06, + "loss": 0.7501, + "num_tokens": 3583053.0, + "step": 162 + }, + { + "epoch": 0.4930056710775047, + "grad_norm": 1.4921875, + "learning_rate": 9.991718246100573e-06, + "loss": 0.9871, + "num_tokens": 3621480.0, + "step": 163 + }, + { + "epoch": 0.49603024574669186, + "grad_norm": 1.75, + "learning_rate": 9.991449023608932e-06, + "loss": 0.799, + "num_tokens": 3634592.0, + "step": 164 + }, + { + "epoch": 0.499054820415879, + "grad_norm": 1.5078125, + "learning_rate": 9.991175499060684e-06, + "loss": 0.8656, + "num_tokens": 3656033.0, + "step": 165 + }, + { + "epoch": 0.5020793950850662, + "grad_norm": 2.0, + "learning_rate": 9.990897672717813e-06, + "loss": 0.6293, + "num_tokens": 3663109.0, + "step": 166 + }, + { + "epoch": 0.5051039697542533, + "grad_norm": 1.84375, + "learning_rate": 9.990615544846439e-06, + "loss": 0.8888, + "num_tokens": 3684178.0, + "step": 167 + }, + { + "epoch": 0.5081285444234405, + "grad_norm": 2.1875, + "learning_rate": 9.990329115716785e-06, + "loss": 0.9576, + "num_tokens": 3706442.0, + "step": 168 + }, + { + "epoch": 0.5111531190926276, + "grad_norm": 1.3125, + "learning_rate": 9.990038385603208e-06, + "loss": 0.9581, + "num_tokens": 3745652.0, + "step": 169 + }, + { + "epoch": 0.5141776937618148, + "grad_norm": 1.7265625, + "learning_rate": 9.989743354784174e-06, + "loss": 0.9439, + "num_tokens": 3768921.0, + "step": 170 + }, + { + "epoch": 0.5172022684310019, + "grad_norm": 2.0625, + "learning_rate": 9.989444023542278e-06, + "loss": 0.6765, + "num_tokens": 3777074.0, + "step": 171 + }, + { + "epoch": 0.5202268431001891, + "grad_norm": 1.7265625, + "learning_rate": 9.989140392164225e-06, + "loss": 0.9912, + "num_tokens": 3800116.0, + "step": 172 + }, + { + "epoch": 0.5232514177693762, + "grad_norm": 1.34375, + "learning_rate": 9.988832460940846e-06, + "loss": 0.9178, + "num_tokens": 3839022.0, + "step": 173 + }, + { + "epoch": 0.5262759924385634, + "grad_norm": 1.671875, + "learning_rate": 9.988520230167084e-06, + "loss": 1.0003, + "num_tokens": 3877733.0, + "step": 174 + }, + { + "epoch": 0.5293005671077504, + "grad_norm": 2.015625, + "learning_rate": 9.98820370014201e-06, + "loss": 0.7088, + "num_tokens": 3885687.0, + "step": 175 + }, + { + "epoch": 0.5323251417769376, + "grad_norm": 1.4921875, + "learning_rate": 9.987882871168801e-06, + "loss": 0.8937, + "num_tokens": 3909548.0, + "step": 176 + }, + { + "epoch": 0.5353497164461247, + "grad_norm": 2.140625, + "learning_rate": 9.98755774355476e-06, + "loss": 0.6821, + "num_tokens": 3916497.0, + "step": 177 + }, + { + "epoch": 0.5383742911153119, + "grad_norm": 1.890625, + "learning_rate": 9.987228317611305e-06, + "loss": 0.665, + "num_tokens": 3925536.0, + "step": 178 + }, + { + "epoch": 0.541398865784499, + "grad_norm": 2.109375, + "learning_rate": 9.986894593653969e-06, + "loss": 0.6184, + "num_tokens": 3932592.0, + "step": 179 + }, + { + "epoch": 0.5444234404536862, + "grad_norm": 2.0625, + "learning_rate": 9.986556572002408e-06, + "loss": 0.6561, + "num_tokens": 3939236.0, + "step": 180 + }, + { + "epoch": 0.5474480151228733, + "grad_norm": 2.1875, + "learning_rate": 9.986214252980384e-06, + "loss": 0.7677, + "num_tokens": 3946129.0, + "step": 181 + }, + { + "epoch": 0.5504725897920605, + "grad_norm": 1.5, + "learning_rate": 9.985867636915784e-06, + "loss": 0.8495, + "num_tokens": 3979042.0, + "step": 182 + }, + { + "epoch": 0.5534971644612476, + "grad_norm": 2.234375, + "learning_rate": 9.985516724140613e-06, + "loss": 0.7371, + "num_tokens": 3988506.0, + "step": 183 + }, + { + "epoch": 0.5565217391304348, + "grad_norm": 1.5703125, + "learning_rate": 9.985161514990979e-06, + "loss": 0.757, + "num_tokens": 4012192.0, + "step": 184 + }, + { + "epoch": 0.5595463137996219, + "grad_norm": 1.953125, + "learning_rate": 9.984802009807117e-06, + "loss": 0.6729, + "num_tokens": 4019617.0, + "step": 185 + }, + { + "epoch": 0.5625708884688091, + "grad_norm": 2.0625, + "learning_rate": 9.984438208933371e-06, + "loss": 0.6934, + "num_tokens": 4026500.0, + "step": 186 + }, + { + "epoch": 0.5655954631379962, + "grad_norm": 2.234375, + "learning_rate": 9.984070112718204e-06, + "loss": 0.839, + "num_tokens": 4040920.0, + "step": 187 + }, + { + "epoch": 0.5686200378071834, + "grad_norm": 1.4453125, + "learning_rate": 9.98369772151419e-06, + "loss": 0.9738, + "num_tokens": 4077400.0, + "step": 188 + }, + { + "epoch": 0.5716446124763706, + "grad_norm": 1.7265625, + "learning_rate": 9.983321035678016e-06, + "loss": 1.0429, + "num_tokens": 4100536.0, + "step": 189 + }, + { + "epoch": 0.5746691871455577, + "grad_norm": 1.46875, + "learning_rate": 9.982940055570484e-06, + "loss": 0.8931, + "num_tokens": 4124915.0, + "step": 190 + }, + { + "epoch": 0.5776937618147449, + "grad_norm": 1.6640625, + "learning_rate": 9.982554781556512e-06, + "loss": 0.8011, + "num_tokens": 4144579.0, + "step": 191 + }, + { + "epoch": 0.5807183364839319, + "grad_norm": 1.921875, + "learning_rate": 9.982165214005124e-06, + "loss": 0.8971, + "num_tokens": 4174698.0, + "step": 192 + }, + { + "epoch": 0.583742911153119, + "grad_norm": 2.015625, + "learning_rate": 9.981771353289464e-06, + "loss": 0.6302, + "num_tokens": 4182070.0, + "step": 193 + }, + { + "epoch": 0.5867674858223062, + "grad_norm": 1.7890625, + "learning_rate": 9.981373199786782e-06, + "loss": 0.9377, + "num_tokens": 4204927.0, + "step": 194 + }, + { + "epoch": 0.5897920604914934, + "grad_norm": 2.03125, + "learning_rate": 9.980970753878443e-06, + "loss": 0.6339, + "num_tokens": 4211988.0, + "step": 195 + }, + { + "epoch": 0.5928166351606805, + "grad_norm": 1.859375, + "learning_rate": 9.980564015949924e-06, + "loss": 0.804, + "num_tokens": 4226251.0, + "step": 196 + }, + { + "epoch": 0.5958412098298677, + "grad_norm": 1.3671875, + "learning_rate": 9.98015298639081e-06, + "loss": 0.9618, + "num_tokens": 4270865.0, + "step": 197 + }, + { + "epoch": 0.5988657844990548, + "grad_norm": 2.15625, + "learning_rate": 9.9797376655948e-06, + "loss": 0.6554, + "num_tokens": 4277739.0, + "step": 198 + }, + { + "epoch": 0.601890359168242, + "grad_norm": 1.375, + "learning_rate": 9.979318053959704e-06, + "loss": 0.9132, + "num_tokens": 4314114.0, + "step": 199 + }, + { + "epoch": 0.6049149338374291, + "grad_norm": 1.46875, + "learning_rate": 9.978894151887435e-06, + "loss": 0.8288, + "num_tokens": 4346849.0, + "step": 200 + }, + { + "epoch": 0.6079395085066163, + "grad_norm": 1.96875, + "learning_rate": 9.978465959784022e-06, + "loss": 0.6694, + "num_tokens": 4354671.0, + "step": 201 + }, + { + "epoch": 0.6109640831758034, + "grad_norm": 1.265625, + "learning_rate": 9.978033478059604e-06, + "loss": 0.9743, + "num_tokens": 4410722.0, + "step": 202 + }, + { + "epoch": 0.6139886578449906, + "grad_norm": 1.578125, + "learning_rate": 9.977596707128424e-06, + "loss": 0.7911, + "num_tokens": 4425383.0, + "step": 203 + }, + { + "epoch": 0.6170132325141777, + "grad_norm": 1.875, + "learning_rate": 9.977155647408834e-06, + "loss": 0.8601, + "num_tokens": 4439082.0, + "step": 204 + }, + { + "epoch": 0.6200378071833649, + "grad_norm": 1.5390625, + "learning_rate": 9.976710299323296e-06, + "loss": 0.8591, + "num_tokens": 4462254.0, + "step": 205 + }, + { + "epoch": 0.623062381852552, + "grad_norm": 1.765625, + "learning_rate": 9.976260663298384e-06, + "loss": 0.7761, + "num_tokens": 4476505.0, + "step": 206 + }, + { + "epoch": 0.6260869565217392, + "grad_norm": 1.578125, + "learning_rate": 9.975806739764766e-06, + "loss": 0.8774, + "num_tokens": 4499830.0, + "step": 207 + }, + { + "epoch": 0.6291115311909262, + "grad_norm": 1.546875, + "learning_rate": 9.97534852915723e-06, + "loss": 0.9169, + "num_tokens": 4528219.0, + "step": 208 + }, + { + "epoch": 0.6321361058601134, + "grad_norm": 1.9453125, + "learning_rate": 9.974886031914665e-06, + "loss": 0.6554, + "num_tokens": 4537279.0, + "step": 209 + }, + { + "epoch": 0.6351606805293005, + "grad_norm": 1.3828125, + "learning_rate": 9.974419248480064e-06, + "loss": 0.9438, + "num_tokens": 4576031.0, + "step": 210 + }, + { + "epoch": 0.6381852551984877, + "grad_norm": 1.3203125, + "learning_rate": 9.973948179300528e-06, + "loss": 0.8769, + "num_tokens": 4612186.0, + "step": 211 + }, + { + "epoch": 0.6412098298676748, + "grad_norm": 1.71875, + "learning_rate": 9.973472824827262e-06, + "loss": 0.9, + "num_tokens": 4636171.0, + "step": 212 + }, + { + "epoch": 0.644234404536862, + "grad_norm": 1.96875, + "learning_rate": 9.972993185515579e-06, + "loss": 0.6488, + "num_tokens": 4643725.0, + "step": 213 + }, + { + "epoch": 0.6472589792060491, + "grad_norm": 1.421875, + "learning_rate": 9.97250926182489e-06, + "loss": 0.8257, + "num_tokens": 4668185.0, + "step": 214 + }, + { + "epoch": 0.6502835538752363, + "grad_norm": 1.578125, + "learning_rate": 9.972021054218712e-06, + "loss": 0.8884, + "num_tokens": 4692899.0, + "step": 215 + }, + { + "epoch": 0.6533081285444234, + "grad_norm": 2.234375, + "learning_rate": 9.971528563164669e-06, + "loss": 0.6878, + "num_tokens": 4699533.0, + "step": 216 + }, + { + "epoch": 0.6563327032136106, + "grad_norm": 1.5703125, + "learning_rate": 9.971031789134486e-06, + "loss": 0.9377, + "num_tokens": 4739223.0, + "step": 217 + }, + { + "epoch": 0.6593572778827977, + "grad_norm": 2.3125, + "learning_rate": 9.970530732603984e-06, + "loss": 0.663, + "num_tokens": 4745988.0, + "step": 218 + }, + { + "epoch": 0.6623818525519849, + "grad_norm": 1.7265625, + "learning_rate": 9.9700253940531e-06, + "loss": 0.8811, + "num_tokens": 4769138.0, + "step": 219 + }, + { + "epoch": 0.665406427221172, + "grad_norm": 1.8671875, + "learning_rate": 9.969515773965855e-06, + "loss": 0.7817, + "num_tokens": 4786826.0, + "step": 220 + }, + { + "epoch": 0.6684310018903592, + "grad_norm": 1.9921875, + "learning_rate": 9.969001872830383e-06, + "loss": 0.6543, + "num_tokens": 4794731.0, + "step": 221 + }, + { + "epoch": 0.6714555765595464, + "grad_norm": 1.71875, + "learning_rate": 9.968483691138917e-06, + "loss": 0.9438, + "num_tokens": 4817482.0, + "step": 222 + }, + { + "epoch": 0.6744801512287335, + "grad_norm": 1.5546875, + "learning_rate": 9.967961229387784e-06, + "loss": 0.8058, + "num_tokens": 4835384.0, + "step": 223 + }, + { + "epoch": 0.6775047258979207, + "grad_norm": 1.4765625, + "learning_rate": 9.967434488077422e-06, + "loss": 1.0532, + "num_tokens": 4874389.0, + "step": 224 + }, + { + "epoch": 0.6805293005671077, + "grad_norm": 1.4921875, + "learning_rate": 9.966903467712355e-06, + "loss": 0.9105, + "num_tokens": 4904747.0, + "step": 225 + }, + { + "epoch": 0.6835538752362949, + "grad_norm": 1.3203125, + "learning_rate": 9.966368168801214e-06, + "loss": 0.9434, + "num_tokens": 4943795.0, + "step": 226 + }, + { + "epoch": 0.686578449905482, + "grad_norm": 1.2890625, + "learning_rate": 9.965828591856725e-06, + "loss": 0.9245, + "num_tokens": 4983091.0, + "step": 227 + }, + { + "epoch": 0.6896030245746692, + "grad_norm": 1.4765625, + "learning_rate": 9.965284737395714e-06, + "loss": 0.9009, + "num_tokens": 5010251.0, + "step": 228 + }, + { + "epoch": 0.6926275992438563, + "grad_norm": 1.703125, + "learning_rate": 9.964736605939098e-06, + "loss": 0.9693, + "num_tokens": 5033746.0, + "step": 229 + }, + { + "epoch": 0.6956521739130435, + "grad_norm": 1.5546875, + "learning_rate": 9.964184198011903e-06, + "loss": 0.9123, + "num_tokens": 5060627.0, + "step": 230 + }, + { + "epoch": 0.6986767485822306, + "grad_norm": 2.03125, + "learning_rate": 9.963627514143235e-06, + "loss": 0.6596, + "num_tokens": 5067837.0, + "step": 231 + }, + { + "epoch": 0.7017013232514178, + "grad_norm": 1.359375, + "learning_rate": 9.963066554866311e-06, + "loss": 0.9194, + "num_tokens": 5107700.0, + "step": 232 + }, + { + "epoch": 0.7047258979206049, + "grad_norm": 1.6953125, + "learning_rate": 9.962501320718432e-06, + "loss": 0.9108, + "num_tokens": 5131493.0, + "step": 233 + }, + { + "epoch": 0.7077504725897921, + "grad_norm": 2.453125, + "learning_rate": 9.961931812240998e-06, + "loss": 0.8554, + "num_tokens": 5145191.0, + "step": 234 + }, + { + "epoch": 0.7107750472589792, + "grad_norm": 1.640625, + "learning_rate": 9.961358029979504e-06, + "loss": 0.8794, + "num_tokens": 5182463.0, + "step": 235 + }, + { + "epoch": 0.7137996219281664, + "grad_norm": 1.8046875, + "learning_rate": 9.960779974483537e-06, + "loss": 0.8013, + "num_tokens": 5205109.0, + "step": 236 + }, + { + "epoch": 0.7168241965973535, + "grad_norm": 2.109375, + "learning_rate": 9.960197646306778e-06, + "loss": 0.7324, + "num_tokens": 5216525.0, + "step": 237 + }, + { + "epoch": 0.7198487712665407, + "grad_norm": 1.8359375, + "learning_rate": 9.959611046007e-06, + "loss": 0.752, + "num_tokens": 5232387.0, + "step": 238 + }, + { + "epoch": 0.7228733459357278, + "grad_norm": 1.4375, + "learning_rate": 9.959020174146066e-06, + "loss": 0.9267, + "num_tokens": 5265668.0, + "step": 239 + }, + { + "epoch": 0.725897920604915, + "grad_norm": 1.8828125, + "learning_rate": 9.958425031289935e-06, + "loss": 0.6535, + "num_tokens": 5274185.0, + "step": 240 + }, + { + "epoch": 0.7289224952741021, + "grad_norm": 1.8359375, + "learning_rate": 9.957825618008653e-06, + "loss": 0.8635, + "num_tokens": 5291383.0, + "step": 241 + }, + { + "epoch": 0.7319470699432892, + "grad_norm": 2.015625, + "learning_rate": 9.957221934876355e-06, + "loss": 0.6763, + "num_tokens": 5299120.0, + "step": 242 + }, + { + "epoch": 0.7349716446124763, + "grad_norm": 1.5390625, + "learning_rate": 9.956613982471274e-06, + "loss": 0.8373, + "num_tokens": 5323105.0, + "step": 243 + }, + { + "epoch": 0.7379962192816635, + "grad_norm": 2.09375, + "learning_rate": 9.956001761375723e-06, + "loss": 0.6915, + "num_tokens": 5330400.0, + "step": 244 + }, + { + "epoch": 0.7410207939508506, + "grad_norm": 1.890625, + "learning_rate": 9.955385272176108e-06, + "loss": 0.6535, + "num_tokens": 5337733.0, + "step": 245 + }, + { + "epoch": 0.7440453686200378, + "grad_norm": 1.484375, + "learning_rate": 9.954764515462921e-06, + "loss": 0.9297, + "num_tokens": 5374891.0, + "step": 246 + }, + { + "epoch": 0.7470699432892249, + "grad_norm": 1.9296875, + "learning_rate": 9.954139491830748e-06, + "loss": 0.8711, + "num_tokens": 5393999.0, + "step": 247 + }, + { + "epoch": 0.7500945179584121, + "grad_norm": 1.546875, + "learning_rate": 9.953510201878251e-06, + "loss": 0.8468, + "num_tokens": 5419669.0, + "step": 248 + }, + { + "epoch": 0.7531190926275992, + "grad_norm": 1.7421875, + "learning_rate": 9.952876646208187e-06, + "loss": 0.7886, + "num_tokens": 5432599.0, + "step": 249 + }, + { + "epoch": 0.7561436672967864, + "grad_norm": 1.703125, + "learning_rate": 9.952238825427398e-06, + "loss": 0.9167, + "num_tokens": 5464259.0, + "step": 250 + }, + { + "epoch": 0.7591682419659735, + "grad_norm": 1.4609375, + "learning_rate": 9.951596740146809e-06, + "loss": 0.9131, + "num_tokens": 5520526.0, + "step": 251 + }, + { + "epoch": 0.7621928166351607, + "grad_norm": 1.4375, + "learning_rate": 9.950950390981427e-06, + "loss": 0.8296, + "num_tokens": 5557723.0, + "step": 252 + }, + { + "epoch": 0.7652173913043478, + "grad_norm": 1.6015625, + "learning_rate": 9.95029977855035e-06, + "loss": 0.8581, + "num_tokens": 5578223.0, + "step": 253 + }, + { + "epoch": 0.768241965973535, + "grad_norm": 1.6640625, + "learning_rate": 9.949644903476752e-06, + "loss": 0.9941, + "num_tokens": 5601383.0, + "step": 254 + }, + { + "epoch": 0.7712665406427222, + "grad_norm": 2.03125, + "learning_rate": 9.948985766387899e-06, + "loss": 0.7494, + "num_tokens": 5610374.0, + "step": 255 + }, + { + "epoch": 0.7742911153119093, + "grad_norm": 1.640625, + "learning_rate": 9.948322367915131e-06, + "loss": 0.8393, + "num_tokens": 5626915.0, + "step": 256 + }, + { + "epoch": 0.7773156899810965, + "grad_norm": 1.3671875, + "learning_rate": 9.947654708693872e-06, + "loss": 0.9072, + "num_tokens": 5664745.0, + "step": 257 + }, + { + "epoch": 0.7803402646502835, + "grad_norm": 1.8203125, + "learning_rate": 9.946982789363631e-06, + "loss": 0.9734, + "num_tokens": 5688297.0, + "step": 258 + }, + { + "epoch": 0.7833648393194707, + "grad_norm": 1.5078125, + "learning_rate": 9.946306610567989e-06, + "loss": 0.9056, + "num_tokens": 5712626.0, + "step": 259 + }, + { + "epoch": 0.7863894139886578, + "grad_norm": 1.484375, + "learning_rate": 9.945626172954617e-06, + "loss": 0.8399, + "num_tokens": 5740927.0, + "step": 260 + }, + { + "epoch": 0.789413988657845, + "grad_norm": 1.3203125, + "learning_rate": 9.944941477175258e-06, + "loss": 0.8334, + "num_tokens": 5774194.0, + "step": 261 + }, + { + "epoch": 0.7924385633270321, + "grad_norm": 1.984375, + "learning_rate": 9.944252523885739e-06, + "loss": 0.9619, + "num_tokens": 5798529.0, + "step": 262 + }, + { + "epoch": 0.7954631379962193, + "grad_norm": 1.65625, + "learning_rate": 9.943559313745957e-06, + "loss": 0.8177, + "num_tokens": 5815060.0, + "step": 263 + }, + { + "epoch": 0.7984877126654064, + "grad_norm": 1.796875, + "learning_rate": 9.942861847419895e-06, + "loss": 0.6113, + "num_tokens": 5823899.0, + "step": 264 + }, + { + "epoch": 0.8015122873345936, + "grad_norm": 2.0625, + "learning_rate": 9.942160125575608e-06, + "loss": 0.699, + "num_tokens": 5831637.0, + "step": 265 + }, + { + "epoch": 0.8045368620037807, + "grad_norm": 2.359375, + "learning_rate": 9.941454148885226e-06, + "loss": 0.7698, + "num_tokens": 5844005.0, + "step": 266 + }, + { + "epoch": 0.8075614366729679, + "grad_norm": 1.7265625, + "learning_rate": 9.940743918024961e-06, + "loss": 0.8767, + "num_tokens": 5868402.0, + "step": 267 + }, + { + "epoch": 0.810586011342155, + "grad_norm": 1.3359375, + "learning_rate": 9.940029433675093e-06, + "loss": 0.9486, + "num_tokens": 5914305.0, + "step": 268 + }, + { + "epoch": 0.8136105860113422, + "grad_norm": 1.953125, + "learning_rate": 9.939310696519977e-06, + "loss": 0.7717, + "num_tokens": 5922927.0, + "step": 269 + }, + { + "epoch": 0.8166351606805293, + "grad_norm": 1.9921875, + "learning_rate": 9.938587707248045e-06, + "loss": 0.6033, + "num_tokens": 5930253.0, + "step": 270 + }, + { + "epoch": 0.8196597353497165, + "grad_norm": 2.25, + "learning_rate": 9.9378604665518e-06, + "loss": 1.0018, + "num_tokens": 5959715.0, + "step": 271 + }, + { + "epoch": 0.8226843100189036, + "grad_norm": 1.5703125, + "learning_rate": 9.937128975127814e-06, + "loss": 0.895, + "num_tokens": 5983390.0, + "step": 272 + }, + { + "epoch": 0.8257088846880908, + "grad_norm": 1.40625, + "learning_rate": 9.936393233676737e-06, + "loss": 0.8517, + "num_tokens": 6007184.0, + "step": 273 + }, + { + "epoch": 0.8287334593572779, + "grad_norm": 1.4609375, + "learning_rate": 9.935653242903283e-06, + "loss": 0.8538, + "num_tokens": 6030991.0, + "step": 274 + }, + { + "epoch": 0.831758034026465, + "grad_norm": 1.8515625, + "learning_rate": 9.93490900351624e-06, + "loss": 0.548, + "num_tokens": 6038242.0, + "step": 275 + }, + { + "epoch": 0.8347826086956521, + "grad_norm": 1.9375, + "learning_rate": 9.934160516228468e-06, + "loss": 0.662, + "num_tokens": 6045803.0, + "step": 276 + }, + { + "epoch": 0.8378071833648393, + "grad_norm": 1.9765625, + "learning_rate": 9.933407781756888e-06, + "loss": 0.9332, + "num_tokens": 6068755.0, + "step": 277 + }, + { + "epoch": 0.8408317580340264, + "grad_norm": 1.6015625, + "learning_rate": 9.9326508008225e-06, + "loss": 0.9297, + "num_tokens": 6092471.0, + "step": 278 + }, + { + "epoch": 0.8438563327032136, + "grad_norm": 1.4453125, + "learning_rate": 9.931889574150357e-06, + "loss": 0.9278, + "num_tokens": 6124407.0, + "step": 279 + }, + { + "epoch": 0.8468809073724007, + "grad_norm": 2.109375, + "learning_rate": 9.931124102469592e-06, + "loss": 0.7443, + "num_tokens": 6137972.0, + "step": 280 + }, + { + "epoch": 0.8499054820415879, + "grad_norm": 2.046875, + "learning_rate": 9.930354386513399e-06, + "loss": 0.7264, + "num_tokens": 6145700.0, + "step": 281 + }, + { + "epoch": 0.852930056710775, + "grad_norm": 1.6875, + "learning_rate": 9.929580427019034e-06, + "loss": 0.8228, + "num_tokens": 6167552.0, + "step": 282 + }, + { + "epoch": 0.8559546313799622, + "grad_norm": 1.4375, + "learning_rate": 9.928802224727822e-06, + "loss": 0.9006, + "num_tokens": 6202747.0, + "step": 283 + }, + { + "epoch": 0.8589792060491493, + "grad_norm": 1.3125, + "learning_rate": 9.928019780385152e-06, + "loss": 0.8824, + "num_tokens": 6242338.0, + "step": 284 + }, + { + "epoch": 0.8620037807183365, + "grad_norm": 1.359375, + "learning_rate": 9.927233094740473e-06, + "loss": 0.8986, + "num_tokens": 6286291.0, + "step": 285 + }, + { + "epoch": 0.8650283553875237, + "grad_norm": 1.8984375, + "learning_rate": 9.926442168547299e-06, + "loss": 0.5969, + "num_tokens": 6294315.0, + "step": 286 + }, + { + "epoch": 0.8680529300567108, + "grad_norm": 1.8046875, + "learning_rate": 9.925647002563205e-06, + "loss": 0.9017, + "num_tokens": 6312713.0, + "step": 287 + }, + { + "epoch": 0.871077504725898, + "grad_norm": 1.4921875, + "learning_rate": 9.924847597549825e-06, + "loss": 0.8578, + "num_tokens": 6335909.0, + "step": 288 + }, + { + "epoch": 0.8741020793950851, + "grad_norm": 2.109375, + "learning_rate": 9.924043954272859e-06, + "loss": 0.6787, + "num_tokens": 6343799.0, + "step": 289 + }, + { + "epoch": 0.8771266540642723, + "grad_norm": 2.125, + "learning_rate": 9.92323607350206e-06, + "loss": 0.7593, + "num_tokens": 6355386.0, + "step": 290 + }, + { + "epoch": 0.8801512287334594, + "grad_norm": 1.5546875, + "learning_rate": 9.922423956011245e-06, + "loss": 0.8506, + "num_tokens": 6383111.0, + "step": 291 + }, + { + "epoch": 0.8831758034026465, + "grad_norm": 1.7109375, + "learning_rate": 9.921607602578284e-06, + "loss": 0.8995, + "num_tokens": 6403532.0, + "step": 292 + }, + { + "epoch": 0.8862003780718336, + "grad_norm": 1.8671875, + "learning_rate": 9.920787013985106e-06, + "loss": 0.6919, + "num_tokens": 6411254.0, + "step": 293 + }, + { + "epoch": 0.8892249527410208, + "grad_norm": 1.203125, + "learning_rate": 9.919962191017703e-06, + "loss": 0.9337, + "num_tokens": 6457395.0, + "step": 294 + }, + { + "epoch": 0.8922495274102079, + "grad_norm": 2.0, + "learning_rate": 9.919133134466111e-06, + "loss": 0.6534, + "num_tokens": 6465021.0, + "step": 295 + }, + { + "epoch": 0.8952741020793951, + "grad_norm": 1.890625, + "learning_rate": 9.918299845124433e-06, + "loss": 0.99, + "num_tokens": 6487930.0, + "step": 296 + }, + { + "epoch": 0.8982986767485822, + "grad_norm": 1.9765625, + "learning_rate": 9.917462323790818e-06, + "loss": 0.659, + "num_tokens": 6494962.0, + "step": 297 + }, + { + "epoch": 0.9013232514177694, + "grad_norm": 2.015625, + "learning_rate": 9.916620571267469e-06, + "loss": 0.7453, + "num_tokens": 6508129.0, + "step": 298 + }, + { + "epoch": 0.9043478260869565, + "grad_norm": 1.8828125, + "learning_rate": 9.915774588360649e-06, + "loss": 0.9378, + "num_tokens": 6531783.0, + "step": 299 + }, + { + "epoch": 0.9073724007561437, + "grad_norm": 1.59375, + "learning_rate": 9.914924375880663e-06, + "loss": 0.8691, + "num_tokens": 6556000.0, + "step": 300 + }, + { + "epoch": 0.9103969754253308, + "grad_norm": 2.203125, + "learning_rate": 9.914069934641877e-06, + "loss": 0.6411, + "num_tokens": 6562565.0, + "step": 301 + }, + { + "epoch": 0.913421550094518, + "grad_norm": 1.4453125, + "learning_rate": 9.9132112654627e-06, + "loss": 0.7714, + "num_tokens": 6581952.0, + "step": 302 + }, + { + "epoch": 0.9164461247637051, + "grad_norm": 2.0, + "learning_rate": 9.912348369165594e-06, + "loss": 0.6831, + "num_tokens": 6589172.0, + "step": 303 + }, + { + "epoch": 0.9194706994328923, + "grad_norm": 2.140625, + "learning_rate": 9.91148124657707e-06, + "loss": 0.6686, + "num_tokens": 6596461.0, + "step": 304 + }, + { + "epoch": 0.9224952741020794, + "grad_norm": 1.7890625, + "learning_rate": 9.910609898527686e-06, + "loss": 0.9314, + "num_tokens": 6619628.0, + "step": 305 + }, + { + "epoch": 0.9255198487712666, + "grad_norm": 1.375, + "learning_rate": 9.909734325852046e-06, + "loss": 0.8958, + "num_tokens": 6658595.0, + "step": 306 + }, + { + "epoch": 0.9285444234404537, + "grad_norm": 2.03125, + "learning_rate": 9.908854529388802e-06, + "loss": 0.6646, + "num_tokens": 6665587.0, + "step": 307 + }, + { + "epoch": 0.9315689981096408, + "grad_norm": 2.3125, + "learning_rate": 9.907970509980657e-06, + "loss": 0.7265, + "num_tokens": 6676274.0, + "step": 308 + }, + { + "epoch": 0.9345935727788279, + "grad_norm": 1.5703125, + "learning_rate": 9.907082268474348e-06, + "loss": 0.8359, + "num_tokens": 6694350.0, + "step": 309 + }, + { + "epoch": 0.9376181474480151, + "grad_norm": 1.75, + "learning_rate": 9.906189805720664e-06, + "loss": 0.823, + "num_tokens": 6716968.0, + "step": 310 + }, + { + "epoch": 0.9406427221172022, + "grad_norm": 2.1875, + "learning_rate": 9.905293122574433e-06, + "loss": 0.8068, + "num_tokens": 6735785.0, + "step": 311 + }, + { + "epoch": 0.9436672967863894, + "grad_norm": 2.015625, + "learning_rate": 9.90439221989453e-06, + "loss": 0.6983, + "num_tokens": 6743447.0, + "step": 312 + }, + { + "epoch": 0.9466918714555765, + "grad_norm": 1.9921875, + "learning_rate": 9.90348709854387e-06, + "loss": 0.6764, + "num_tokens": 6749990.0, + "step": 313 + }, + { + "epoch": 0.9497164461247637, + "grad_norm": 1.296875, + "learning_rate": 9.902577759389402e-06, + "loss": 0.9368, + "num_tokens": 6797715.0, + "step": 314 + }, + { + "epoch": 0.9527410207939508, + "grad_norm": 2.015625, + "learning_rate": 9.901664203302126e-06, + "loss": 0.5811, + "num_tokens": 6805307.0, + "step": 315 + }, + { + "epoch": 0.955765595463138, + "grad_norm": 1.90625, + "learning_rate": 9.900746431157074e-06, + "loss": 0.7427, + "num_tokens": 6814138.0, + "step": 316 + }, + { + "epoch": 0.9587901701323251, + "grad_norm": 1.703125, + "learning_rate": 9.89982444383332e-06, + "loss": 0.893, + "num_tokens": 6837731.0, + "step": 317 + }, + { + "epoch": 0.9618147448015123, + "grad_norm": 1.6875, + "learning_rate": 9.898898242213971e-06, + "loss": 0.8614, + "num_tokens": 6862314.0, + "step": 318 + }, + { + "epoch": 0.9648393194706995, + "grad_norm": 1.671875, + "learning_rate": 9.897967827186175e-06, + "loss": 0.8849, + "num_tokens": 6885338.0, + "step": 319 + }, + { + "epoch": 0.9678638941398866, + "grad_norm": 1.6953125, + "learning_rate": 9.897033199641114e-06, + "loss": 0.9307, + "num_tokens": 6909227.0, + "step": 320 + }, + { + "epoch": 0.9708884688090738, + "grad_norm": 2.3125, + "learning_rate": 9.896094360474e-06, + "loss": 0.7036, + "num_tokens": 6920084.0, + "step": 321 + }, + { + "epoch": 0.9739130434782609, + "grad_norm": 1.859375, + "learning_rate": 9.895151310584092e-06, + "loss": 0.6844, + "num_tokens": 6927773.0, + "step": 322 + }, + { + "epoch": 0.9769376181474481, + "grad_norm": 1.5703125, + "learning_rate": 9.89420405087467e-06, + "loss": 0.8566, + "num_tokens": 6953439.0, + "step": 323 + }, + { + "epoch": 0.9799621928166352, + "grad_norm": 1.5546875, + "learning_rate": 9.893252582253048e-06, + "loss": 0.8922, + "num_tokens": 6984659.0, + "step": 324 + }, + { + "epoch": 0.9829867674858223, + "grad_norm": 1.375, + "learning_rate": 9.892296905630576e-06, + "loss": 0.8593, + "num_tokens": 7019144.0, + "step": 325 + }, + { + "epoch": 0.9860113421550094, + "grad_norm": 1.484375, + "learning_rate": 9.891337021922633e-06, + "loss": 0.9928, + "num_tokens": 7052547.0, + "step": 326 + }, + { + "epoch": 0.9890359168241966, + "grad_norm": 1.203125, + "learning_rate": 9.890372932048627e-06, + "loss": 0.926, + "num_tokens": 7115315.0, + "step": 327 + }, + { + "epoch": 0.9920604914933837, + "grad_norm": 2.078125, + "learning_rate": 9.889404636931991e-06, + "loss": 0.9293, + "num_tokens": 7134880.0, + "step": 328 + }, + { + "epoch": 0.9950850661625709, + "grad_norm": 1.6484375, + "learning_rate": 9.888432137500194e-06, + "loss": 0.8009, + "num_tokens": 7154697.0, + "step": 329 + }, + { + "epoch": 0.998109640831758, + "grad_norm": 1.7421875, + "learning_rate": 9.887455434684728e-06, + "loss": 0.8582, + "num_tokens": 7174499.0, + "step": 330 + }, + { + "epoch": 1.0, + "grad_norm": 1.875, + "learning_rate": 9.886474529421108e-06, + "loss": 0.9665, + "num_tokens": 7195313.0, + "step": 331 + }, + { + "epoch": 1.003024574669187, + "grad_norm": 1.421875, + "learning_rate": 9.885489422648878e-06, + "loss": 0.857, + "num_tokens": 7223685.0, + "step": 332 + }, + { + "epoch": 1.0060491493383743, + "grad_norm": 1.8125, + "learning_rate": 9.884500115311606e-06, + "loss": 0.5894, + "num_tokens": 7231269.0, + "step": 333 + }, + { + "epoch": 1.0090737240075613, + "grad_norm": 1.671875, + "learning_rate": 9.883506608356884e-06, + "loss": 0.9645, + "num_tokens": 7254908.0, + "step": 334 + }, + { + "epoch": 1.0120982986767486, + "grad_norm": 2.03125, + "learning_rate": 9.88250890273632e-06, + "loss": 0.5833, + "num_tokens": 7262995.0, + "step": 335 + }, + { + "epoch": 1.0151228733459357, + "grad_norm": 1.3125, + "learning_rate": 9.881506999405557e-06, + "loss": 0.9309, + "num_tokens": 7296788.0, + "step": 336 + }, + { + "epoch": 1.018147448015123, + "grad_norm": 1.5546875, + "learning_rate": 9.880500899324246e-06, + "loss": 0.7695, + "num_tokens": 7315317.0, + "step": 337 + }, + { + "epoch": 1.02117202268431, + "grad_norm": 1.5234375, + "learning_rate": 9.879490603456062e-06, + "loss": 0.8322, + "num_tokens": 7337254.0, + "step": 338 + }, + { + "epoch": 1.0241965973534972, + "grad_norm": 1.5703125, + "learning_rate": 9.878476112768701e-06, + "loss": 0.7743, + "num_tokens": 7360683.0, + "step": 339 + }, + { + "epoch": 1.0272211720226843, + "grad_norm": 1.578125, + "learning_rate": 9.877457428233877e-06, + "loss": 0.8903, + "num_tokens": 7383740.0, + "step": 340 + }, + { + "epoch": 1.0302457466918715, + "grad_norm": 1.234375, + "learning_rate": 9.876434550827315e-06, + "loss": 0.8976, + "num_tokens": 7424350.0, + "step": 341 + }, + { + "epoch": 1.0332703213610586, + "grad_norm": 1.546875, + "learning_rate": 9.875407481528765e-06, + "loss": 0.9265, + "num_tokens": 7448338.0, + "step": 342 + }, + { + "epoch": 1.0362948960302458, + "grad_norm": 1.421875, + "learning_rate": 9.874376221321983e-06, + "loss": 0.8085, + "num_tokens": 7476002.0, + "step": 343 + }, + { + "epoch": 1.0393194706994329, + "grad_norm": 1.6015625, + "learning_rate": 9.873340771194749e-06, + "loss": 0.6873, + "num_tokens": 7490046.0, + "step": 344 + }, + { + "epoch": 1.0423440453686201, + "grad_norm": 1.4140625, + "learning_rate": 9.872301132138845e-06, + "loss": 0.801, + "num_tokens": 7512416.0, + "step": 345 + }, + { + "epoch": 1.0453686200378072, + "grad_norm": 1.8125, + "learning_rate": 9.871257305150076e-06, + "loss": 0.703, + "num_tokens": 7523333.0, + "step": 346 + }, + { + "epoch": 1.0483931947069944, + "grad_norm": 1.875, + "learning_rate": 9.87020929122825e-06, + "loss": 0.8826, + "num_tokens": 7547202.0, + "step": 347 + }, + { + "epoch": 1.0514177693761815, + "grad_norm": 1.9609375, + "learning_rate": 9.869157091377192e-06, + "loss": 0.6259, + "num_tokens": 7554232.0, + "step": 348 + }, + { + "epoch": 1.0544423440453685, + "grad_norm": 3.34375, + "learning_rate": 9.868100706604733e-06, + "loss": 0.7082, + "num_tokens": 7564575.0, + "step": 349 + }, + { + "epoch": 1.0574669187145558, + "grad_norm": 2.1875, + "learning_rate": 9.867040137922712e-06, + "loss": 0.7078, + "num_tokens": 7571279.0, + "step": 350 + }, + { + "epoch": 1.0604914933837428, + "grad_norm": 1.8203125, + "learning_rate": 9.865975386346977e-06, + "loss": 0.6052, + "num_tokens": 7578976.0, + "step": 351 + }, + { + "epoch": 1.06351606805293, + "grad_norm": 1.21875, + "learning_rate": 9.86490645289738e-06, + "loss": 0.9606, + "num_tokens": 7634136.0, + "step": 352 + }, + { + "epoch": 1.0665406427221171, + "grad_norm": 2.21875, + "learning_rate": 9.86383333859778e-06, + "loss": 0.5874, + "num_tokens": 7640608.0, + "step": 353 + }, + { + "epoch": 1.0695652173913044, + "grad_norm": 1.21875, + "learning_rate": 9.862756044476044e-06, + "loss": 0.8338, + "num_tokens": 7676057.0, + "step": 354 + }, + { + "epoch": 1.0725897920604914, + "grad_norm": 1.3984375, + "learning_rate": 9.86167457156404e-06, + "loss": 0.892, + "num_tokens": 7706797.0, + "step": 355 + }, + { + "epoch": 1.0756143667296787, + "grad_norm": 1.6328125, + "learning_rate": 9.860588920897633e-06, + "loss": 0.9137, + "num_tokens": 7730643.0, + "step": 356 + }, + { + "epoch": 1.0786389413988657, + "grad_norm": 1.890625, + "learning_rate": 9.8594990935167e-06, + "loss": 0.6148, + "num_tokens": 7738721.0, + "step": 357 + }, + { + "epoch": 1.081663516068053, + "grad_norm": 2.0, + "learning_rate": 9.85840509046511e-06, + "loss": 0.6888, + "num_tokens": 7746001.0, + "step": 358 + }, + { + "epoch": 1.08468809073724, + "grad_norm": 1.7890625, + "learning_rate": 9.857306912790737e-06, + "loss": 0.7848, + "num_tokens": 7762007.0, + "step": 359 + }, + { + "epoch": 1.0877126654064273, + "grad_norm": 1.96875, + "learning_rate": 9.856204561545449e-06, + "loss": 0.6577, + "num_tokens": 7770016.0, + "step": 360 + }, + { + "epoch": 1.0907372400756143, + "grad_norm": 1.4375, + "learning_rate": 9.855098037785118e-06, + "loss": 0.8948, + "num_tokens": 7817083.0, + "step": 361 + }, + { + "epoch": 1.0937618147448016, + "grad_norm": 2.078125, + "learning_rate": 9.853987342569604e-06, + "loss": 0.5937, + "num_tokens": 7824321.0, + "step": 362 + }, + { + "epoch": 1.0967863894139886, + "grad_norm": 2.0, + "learning_rate": 9.852872476962769e-06, + "loss": 0.5372, + "num_tokens": 7831020.0, + "step": 363 + }, + { + "epoch": 1.0998109640831757, + "grad_norm": 2.109375, + "learning_rate": 9.851753442032467e-06, + "loss": 0.7006, + "num_tokens": 7838863.0, + "step": 364 + }, + { + "epoch": 1.102835538752363, + "grad_norm": 1.359375, + "learning_rate": 9.850630238850549e-06, + "loss": 0.9005, + "num_tokens": 7872451.0, + "step": 365 + }, + { + "epoch": 1.10586011342155, + "grad_norm": 1.5, + "learning_rate": 9.849502868492852e-06, + "loss": 0.8409, + "num_tokens": 7895057.0, + "step": 366 + }, + { + "epoch": 1.1088846880907373, + "grad_norm": 1.578125, + "learning_rate": 9.848371332039214e-06, + "loss": 0.8416, + "num_tokens": 7918665.0, + "step": 367 + }, + { + "epoch": 1.1119092627599243, + "grad_norm": 2.09375, + "learning_rate": 9.84723563057345e-06, + "loss": 0.8715, + "num_tokens": 7931536.0, + "step": 368 + }, + { + "epoch": 1.1149338374291116, + "grad_norm": 1.5703125, + "learning_rate": 9.846095765183376e-06, + "loss": 0.9599, + "num_tokens": 7954009.0, + "step": 369 + }, + { + "epoch": 1.1179584120982986, + "grad_norm": 2.078125, + "learning_rate": 9.844951736960795e-06, + "loss": 0.7092, + "num_tokens": 7961583.0, + "step": 370 + }, + { + "epoch": 1.1209829867674859, + "grad_norm": 1.5234375, + "learning_rate": 9.843803547001487e-06, + "loss": 0.9025, + "num_tokens": 7987517.0, + "step": 371 + }, + { + "epoch": 1.124007561436673, + "grad_norm": 1.6328125, + "learning_rate": 9.842651196405232e-06, + "loss": 0.7494, + "num_tokens": 8004315.0, + "step": 372 + }, + { + "epoch": 1.1270321361058602, + "grad_norm": 1.828125, + "learning_rate": 9.841494686275786e-06, + "loss": 0.8012, + "num_tokens": 8019148.0, + "step": 373 + }, + { + "epoch": 1.1300567107750472, + "grad_norm": 1.8984375, + "learning_rate": 9.840334017720896e-06, + "loss": 0.6179, + "num_tokens": 8026481.0, + "step": 374 + }, + { + "epoch": 1.1330812854442345, + "grad_norm": 1.546875, + "learning_rate": 9.839169191852286e-06, + "loss": 0.8149, + "num_tokens": 8046957.0, + "step": 375 + }, + { + "epoch": 1.1361058601134215, + "grad_norm": 1.328125, + "learning_rate": 9.838000209785663e-06, + "loss": 0.8924, + "num_tokens": 8078882.0, + "step": 376 + }, + { + "epoch": 1.1391304347826088, + "grad_norm": 2.125, + "learning_rate": 9.836827072640716e-06, + "loss": 0.6607, + "num_tokens": 8086497.0, + "step": 377 + }, + { + "epoch": 1.1421550094517958, + "grad_norm": 1.6875, + "learning_rate": 9.835649781541117e-06, + "loss": 0.9127, + "num_tokens": 8109719.0, + "step": 378 + }, + { + "epoch": 1.1451795841209829, + "grad_norm": 1.765625, + "learning_rate": 9.834468337614512e-06, + "loss": 0.6327, + "num_tokens": 8117878.0, + "step": 379 + }, + { + "epoch": 1.1482041587901701, + "grad_norm": 1.28125, + "learning_rate": 9.833282741992526e-06, + "loss": 0.9024, + "num_tokens": 8155963.0, + "step": 380 + }, + { + "epoch": 1.1512287334593574, + "grad_norm": 1.1640625, + "learning_rate": 9.832092995810763e-06, + "loss": 0.9726, + "num_tokens": 8211585.0, + "step": 381 + }, + { + "epoch": 1.1542533081285444, + "grad_norm": 2.015625, + "learning_rate": 9.830899100208799e-06, + "loss": 0.6629, + "num_tokens": 8218700.0, + "step": 382 + }, + { + "epoch": 1.1572778827977315, + "grad_norm": 1.4453125, + "learning_rate": 9.829701056330188e-06, + "loss": 0.9303, + "num_tokens": 8243203.0, + "step": 383 + }, + { + "epoch": 1.1603024574669187, + "grad_norm": 2.03125, + "learning_rate": 9.828498865322454e-06, + "loss": 0.6432, + "num_tokens": 8250565.0, + "step": 384 + }, + { + "epoch": 1.1633270321361058, + "grad_norm": 2.171875, + "learning_rate": 9.827292528337095e-06, + "loss": 0.6811, + "num_tokens": 8257180.0, + "step": 385 + }, + { + "epoch": 1.166351606805293, + "grad_norm": 1.34375, + "learning_rate": 9.826082046529581e-06, + "loss": 0.8488, + "num_tokens": 8283357.0, + "step": 386 + }, + { + "epoch": 1.16937618147448, + "grad_norm": 1.2109375, + "learning_rate": 9.82486742105935e-06, + "loss": 0.8736, + "num_tokens": 8323166.0, + "step": 387 + }, + { + "epoch": 1.1724007561436673, + "grad_norm": 1.8125, + "learning_rate": 9.82364865308981e-06, + "loss": 0.7687, + "num_tokens": 8341673.0, + "step": 388 + }, + { + "epoch": 1.1754253308128544, + "grad_norm": 2.0, + "learning_rate": 9.82242574378834e-06, + "loss": 0.7306, + "num_tokens": 8352605.0, + "step": 389 + }, + { + "epoch": 1.1784499054820416, + "grad_norm": 1.734375, + "learning_rate": 9.821198694326278e-06, + "loss": 0.8025, + "num_tokens": 8366861.0, + "step": 390 + }, + { + "epoch": 1.1814744801512287, + "grad_norm": 1.609375, + "learning_rate": 9.819967505878932e-06, + "loss": 0.8457, + "num_tokens": 8384362.0, + "step": 391 + }, + { + "epoch": 1.184499054820416, + "grad_norm": 1.4921875, + "learning_rate": 9.818732179625578e-06, + "loss": 0.9061, + "num_tokens": 8408454.0, + "step": 392 + }, + { + "epoch": 1.187523629489603, + "grad_norm": 2.203125, + "learning_rate": 9.81749271674945e-06, + "loss": 0.6724, + "num_tokens": 8416933.0, + "step": 393 + }, + { + "epoch": 1.1905482041587903, + "grad_norm": 1.6484375, + "learning_rate": 9.816249118437746e-06, + "loss": 0.7806, + "num_tokens": 8430769.0, + "step": 394 + }, + { + "epoch": 1.1935727788279773, + "grad_norm": 1.8984375, + "learning_rate": 9.815001385881624e-06, + "loss": 0.8944, + "num_tokens": 8455348.0, + "step": 395 + }, + { + "epoch": 1.1965973534971646, + "grad_norm": 1.703125, + "learning_rate": 9.813749520276205e-06, + "loss": 0.8045, + "num_tokens": 8471254.0, + "step": 396 + }, + { + "epoch": 1.1996219281663516, + "grad_norm": 1.859375, + "learning_rate": 9.812493522820567e-06, + "loss": 0.6075, + "num_tokens": 8478903.0, + "step": 397 + }, + { + "epoch": 1.2026465028355386, + "grad_norm": 1.3515625, + "learning_rate": 9.811233394717742e-06, + "loss": 0.8231, + "num_tokens": 8508866.0, + "step": 398 + }, + { + "epoch": 1.205671077504726, + "grad_norm": 1.6484375, + "learning_rate": 9.809969137174725e-06, + "loss": 0.8165, + "num_tokens": 8529086.0, + "step": 399 + }, + { + "epoch": 1.208695652173913, + "grad_norm": 1.25, + "learning_rate": 9.808700751402465e-06, + "loss": 0.9255, + "num_tokens": 8569065.0, + "step": 400 + }, + { + "epoch": 1.2117202268431002, + "grad_norm": 1.78125, + "learning_rate": 9.807428238615858e-06, + "loss": 0.5659, + "num_tokens": 8577600.0, + "step": 401 + }, + { + "epoch": 1.2147448015122873, + "grad_norm": 1.1875, + "learning_rate": 9.806151600033761e-06, + "loss": 0.8599, + "num_tokens": 8635571.0, + "step": 402 + }, + { + "epoch": 1.2177693761814745, + "grad_norm": 1.78125, + "learning_rate": 9.804870836878982e-06, + "loss": 0.8075, + "num_tokens": 8649514.0, + "step": 403 + }, + { + "epoch": 1.2207939508506616, + "grad_norm": 2.015625, + "learning_rate": 9.803585950378274e-06, + "loss": 0.6819, + "num_tokens": 8657119.0, + "step": 404 + }, + { + "epoch": 1.2238185255198488, + "grad_norm": 1.4296875, + "learning_rate": 9.802296941762347e-06, + "loss": 0.8443, + "num_tokens": 8694100.0, + "step": 405 + }, + { + "epoch": 1.2268431001890359, + "grad_norm": 1.546875, + "learning_rate": 9.801003812265853e-06, + "loss": 0.8868, + "num_tokens": 8717417.0, + "step": 406 + }, + { + "epoch": 1.2298676748582231, + "grad_norm": 1.65625, + "learning_rate": 9.799706563127395e-06, + "loss": 0.8066, + "num_tokens": 8740613.0, + "step": 407 + }, + { + "epoch": 1.2328922495274102, + "grad_norm": 1.625, + "learning_rate": 9.79840519558952e-06, + "loss": 0.8881, + "num_tokens": 8771391.0, + "step": 408 + }, + { + "epoch": 1.2359168241965974, + "grad_norm": 2.40625, + "learning_rate": 9.79709971089872e-06, + "loss": 0.8794, + "num_tokens": 8790866.0, + "step": 409 + }, + { + "epoch": 1.2389413988657845, + "grad_norm": 1.4609375, + "learning_rate": 9.795790110305431e-06, + "loss": 0.7864, + "num_tokens": 8814945.0, + "step": 410 + }, + { + "epoch": 1.2419659735349717, + "grad_norm": 2.03125, + "learning_rate": 9.794476395064032e-06, + "loss": 0.6124, + "num_tokens": 8821680.0, + "step": 411 + }, + { + "epoch": 1.2449905482041588, + "grad_norm": 1.4921875, + "learning_rate": 9.793158566432843e-06, + "loss": 0.884, + "num_tokens": 8857326.0, + "step": 412 + }, + { + "epoch": 1.2480151228733458, + "grad_norm": 1.984375, + "learning_rate": 9.79183662567412e-06, + "loss": 0.898, + "num_tokens": 8878790.0, + "step": 413 + }, + { + "epoch": 1.251039697542533, + "grad_norm": 1.8359375, + "learning_rate": 9.790510574054065e-06, + "loss": 0.5049, + "num_tokens": 8886926.0, + "step": 414 + }, + { + "epoch": 1.2540642722117203, + "grad_norm": 1.3671875, + "learning_rate": 9.789180412842813e-06, + "loss": 0.8954, + "num_tokens": 8915434.0, + "step": 415 + }, + { + "epoch": 1.2570888468809074, + "grad_norm": 1.5, + "learning_rate": 9.787846143314433e-06, + "loss": 0.8894, + "num_tokens": 8945142.0, + "step": 416 + }, + { + "epoch": 1.2601134215500944, + "grad_norm": 2.109375, + "learning_rate": 9.786507766746932e-06, + "loss": 0.6163, + "num_tokens": 8952459.0, + "step": 417 + }, + { + "epoch": 1.2631379962192817, + "grad_norm": 1.4921875, + "learning_rate": 9.785165284422256e-06, + "loss": 0.8608, + "num_tokens": 8977157.0, + "step": 418 + }, + { + "epoch": 1.2661625708884687, + "grad_norm": 1.671875, + "learning_rate": 9.783818697626273e-06, + "loss": 0.848, + "num_tokens": 8999937.0, + "step": 419 + }, + { + "epoch": 1.269187145557656, + "grad_norm": 2.15625, + "learning_rate": 9.78246800764879e-06, + "loss": 0.5848, + "num_tokens": 9007555.0, + "step": 420 + }, + { + "epoch": 1.272211720226843, + "grad_norm": 1.546875, + "learning_rate": 9.781113215783541e-06, + "loss": 0.8197, + "num_tokens": 9030121.0, + "step": 421 + }, + { + "epoch": 1.2752362948960303, + "grad_norm": 1.1484375, + "learning_rate": 9.779754323328192e-06, + "loss": 0.8156, + "num_tokens": 9069143.0, + "step": 422 + }, + { + "epoch": 1.2782608695652173, + "grad_norm": 1.71875, + "learning_rate": 9.778391331584332e-06, + "loss": 0.843, + "num_tokens": 9092538.0, + "step": 423 + }, + { + "epoch": 1.2812854442344046, + "grad_norm": 1.796875, + "learning_rate": 9.777024241857482e-06, + "loss": 0.934, + "num_tokens": 9115677.0, + "step": 424 + }, + { + "epoch": 1.2843100189035916, + "grad_norm": 1.5703125, + "learning_rate": 9.775653055457082e-06, + "loss": 0.8047, + "num_tokens": 9139706.0, + "step": 425 + }, + { + "epoch": 1.287334593572779, + "grad_norm": 1.0546875, + "learning_rate": 9.7742777736965e-06, + "loss": 0.9314, + "num_tokens": 9208422.0, + "step": 426 + }, + { + "epoch": 1.290359168241966, + "grad_norm": 2.0625, + "learning_rate": 9.772898397893028e-06, + "loss": 0.6549, + "num_tokens": 9215464.0, + "step": 427 + }, + { + "epoch": 1.293383742911153, + "grad_norm": 1.328125, + "learning_rate": 9.771514929367875e-06, + "loss": 0.9142, + "num_tokens": 9254689.0, + "step": 428 + }, + { + "epoch": 1.2964083175803403, + "grad_norm": 1.984375, + "learning_rate": 9.770127369446173e-06, + "loss": 0.5581, + "num_tokens": 9261286.0, + "step": 429 + }, + { + "epoch": 1.2994328922495275, + "grad_norm": 1.6171875, + "learning_rate": 9.76873571945697e-06, + "loss": 0.9145, + "num_tokens": 9283888.0, + "step": 430 + }, + { + "epoch": 1.3024574669187146, + "grad_norm": 2.3125, + "learning_rate": 9.76733998073324e-06, + "loss": 0.6508, + "num_tokens": 9294743.0, + "step": 431 + }, + { + "epoch": 1.3054820415879016, + "grad_norm": 1.53125, + "learning_rate": 9.76594015461186e-06, + "loss": 0.8251, + "num_tokens": 9318281.0, + "step": 432 + }, + { + "epoch": 1.3085066162570889, + "grad_norm": 1.8046875, + "learning_rate": 9.764536242433635e-06, + "loss": 0.8497, + "num_tokens": 9341548.0, + "step": 433 + }, + { + "epoch": 1.3115311909262761, + "grad_norm": 1.9921875, + "learning_rate": 9.763128245543272e-06, + "loss": 0.6394, + "num_tokens": 9349511.0, + "step": 434 + }, + { + "epoch": 1.3145557655954632, + "grad_norm": 1.515625, + "learning_rate": 9.761716165289402e-06, + "loss": 0.8561, + "num_tokens": 9372586.0, + "step": 435 + }, + { + "epoch": 1.3175803402646502, + "grad_norm": 1.8203125, + "learning_rate": 9.76030000302456e-06, + "loss": 0.5667, + "num_tokens": 9380505.0, + "step": 436 + }, + { + "epoch": 1.3206049149338375, + "grad_norm": 1.8359375, + "learning_rate": 9.75887976010519e-06, + "loss": 0.6835, + "num_tokens": 9393736.0, + "step": 437 + }, + { + "epoch": 1.3236294896030245, + "grad_norm": 1.8203125, + "learning_rate": 9.757455437891649e-06, + "loss": 0.5754, + "num_tokens": 9401267.0, + "step": 438 + }, + { + "epoch": 1.3266540642722118, + "grad_norm": 1.9765625, + "learning_rate": 9.7560270377482e-06, + "loss": 0.6337, + "num_tokens": 9408956.0, + "step": 439 + }, + { + "epoch": 1.3296786389413988, + "grad_norm": 1.5234375, + "learning_rate": 9.75459456104301e-06, + "loss": 0.7839, + "num_tokens": 9426831.0, + "step": 440 + }, + { + "epoch": 1.332703213610586, + "grad_norm": 1.484375, + "learning_rate": 9.75315800914815e-06, + "loss": 0.7879, + "num_tokens": 9445207.0, + "step": 441 + }, + { + "epoch": 1.3357277882797731, + "grad_norm": 1.328125, + "learning_rate": 9.751717383439604e-06, + "loss": 0.9078, + "num_tokens": 9485461.0, + "step": 442 + }, + { + "epoch": 1.3387523629489604, + "grad_norm": 2.03125, + "learning_rate": 9.750272685297241e-06, + "loss": 0.813, + "num_tokens": 9497847.0, + "step": 443 + }, + { + "epoch": 1.3417769376181474, + "grad_norm": 1.875, + "learning_rate": 9.748823916104847e-06, + "loss": 0.5871, + "num_tokens": 9506300.0, + "step": 444 + }, + { + "epoch": 1.3448015122873347, + "grad_norm": 1.6875, + "learning_rate": 9.747371077250096e-06, + "loss": 0.904, + "num_tokens": 9528361.0, + "step": 445 + }, + { + "epoch": 1.3478260869565217, + "grad_norm": 1.4140625, + "learning_rate": 9.745914170124568e-06, + "loss": 0.8589, + "num_tokens": 9557809.0, + "step": 446 + }, + { + "epoch": 1.3508506616257088, + "grad_norm": 1.8125, + "learning_rate": 9.744453196123733e-06, + "loss": 0.8796, + "num_tokens": 9586156.0, + "step": 447 + }, + { + "epoch": 1.353875236294896, + "grad_norm": 1.8671875, + "learning_rate": 9.742988156646962e-06, + "loss": 0.5789, + "num_tokens": 9594548.0, + "step": 448 + }, + { + "epoch": 1.3568998109640833, + "grad_norm": 1.59375, + "learning_rate": 9.741519053097516e-06, + "loss": 0.7723, + "num_tokens": 9618585.0, + "step": 449 + }, + { + "epoch": 1.3599243856332703, + "grad_norm": 1.5546875, + "learning_rate": 9.740045886882552e-06, + "loss": 0.8936, + "num_tokens": 9642522.0, + "step": 450 + }, + { + "epoch": 1.3629489603024574, + "grad_norm": 1.453125, + "learning_rate": 9.738568659413116e-06, + "loss": 0.9063, + "num_tokens": 9667168.0, + "step": 451 + }, + { + "epoch": 1.3659735349716446, + "grad_norm": 2.25, + "learning_rate": 9.737087372104143e-06, + "loss": 0.6333, + "num_tokens": 9674046.0, + "step": 452 + }, + { + "epoch": 1.3689981096408317, + "grad_norm": 1.671875, + "learning_rate": 9.73560202637446e-06, + "loss": 0.8679, + "num_tokens": 9694439.0, + "step": 453 + }, + { + "epoch": 1.372022684310019, + "grad_norm": 1.4609375, + "learning_rate": 9.734112623646783e-06, + "loss": 0.9105, + "num_tokens": 9726282.0, + "step": 454 + }, + { + "epoch": 1.375047258979206, + "grad_norm": 2.0, + "learning_rate": 9.732619165347705e-06, + "loss": 0.6189, + "num_tokens": 9733358.0, + "step": 455 + }, + { + "epoch": 1.3780718336483933, + "grad_norm": 1.3984375, + "learning_rate": 9.731121652907712e-06, + "loss": 0.8537, + "num_tokens": 9764407.0, + "step": 456 + }, + { + "epoch": 1.3810964083175803, + "grad_norm": 1.8359375, + "learning_rate": 9.729620087761167e-06, + "loss": 0.6824, + "num_tokens": 9772758.0, + "step": 457 + }, + { + "epoch": 1.3841209829867676, + "grad_norm": 1.8828125, + "learning_rate": 9.728114471346324e-06, + "loss": 0.5756, + "num_tokens": 9780402.0, + "step": 458 + }, + { + "epoch": 1.3871455576559546, + "grad_norm": 1.765625, + "learning_rate": 9.72660480510531e-06, + "loss": 0.7308, + "num_tokens": 9793329.0, + "step": 459 + }, + { + "epoch": 1.3901701323251419, + "grad_norm": 1.25, + "learning_rate": 9.725091090484128e-06, + "loss": 0.9319, + "num_tokens": 9836332.0, + "step": 460 + }, + { + "epoch": 1.393194706994329, + "grad_norm": 1.8515625, + "learning_rate": 9.723573328932669e-06, + "loss": 0.6679, + "num_tokens": 9847102.0, + "step": 461 + }, + { + "epoch": 1.396219281663516, + "grad_norm": 1.2578125, + "learning_rate": 9.722051521904689e-06, + "loss": 0.8861, + "num_tokens": 9904601.0, + "step": 462 + }, + { + "epoch": 1.3992438563327032, + "grad_norm": 1.59375, + "learning_rate": 9.72052567085783e-06, + "loss": 0.9476, + "num_tokens": 9930018.0, + "step": 463 + }, + { + "epoch": 1.4022684310018905, + "grad_norm": 1.765625, + "learning_rate": 9.718995777253598e-06, + "loss": 0.7301, + "num_tokens": 9954450.0, + "step": 464 + }, + { + "epoch": 1.4052930056710775, + "grad_norm": 1.8671875, + "learning_rate": 9.717461842557377e-06, + "loss": 0.5917, + "num_tokens": 9962159.0, + "step": 465 + }, + { + "epoch": 1.4083175803402646, + "grad_norm": 1.40625, + "learning_rate": 9.71592386823842e-06, + "loss": 0.8532, + "num_tokens": 9986183.0, + "step": 466 + }, + { + "epoch": 1.4113421550094518, + "grad_norm": 1.328125, + "learning_rate": 9.71438185576985e-06, + "loss": 0.8984, + "num_tokens": 10025195.0, + "step": 467 + }, + { + "epoch": 1.414366729678639, + "grad_norm": 1.2109375, + "learning_rate": 9.712835806628653e-06, + "loss": 0.8903, + "num_tokens": 10074128.0, + "step": 468 + }, + { + "epoch": 1.4173913043478261, + "grad_norm": 1.84375, + "learning_rate": 9.711285722295688e-06, + "loss": 0.8695, + "num_tokens": 10097434.0, + "step": 469 + }, + { + "epoch": 1.4204158790170132, + "grad_norm": 2.0, + "learning_rate": 9.709731604255675e-06, + "loss": 0.8861, + "num_tokens": 10121585.0, + "step": 470 + }, + { + "epoch": 1.4234404536862004, + "grad_norm": 1.5390625, + "learning_rate": 9.708173453997202e-06, + "loss": 0.7808, + "num_tokens": 10144043.0, + "step": 471 + }, + { + "epoch": 1.4264650283553875, + "grad_norm": 1.40625, + "learning_rate": 9.706611273012715e-06, + "loss": 0.9255, + "num_tokens": 10186476.0, + "step": 472 + }, + { + "epoch": 1.4294896030245747, + "grad_norm": 1.3046875, + "learning_rate": 9.705045062798519e-06, + "loss": 0.9554, + "num_tokens": 10222098.0, + "step": 473 + }, + { + "epoch": 1.4325141776937618, + "grad_norm": 1.96875, + "learning_rate": 9.703474824854785e-06, + "loss": 0.6763, + "num_tokens": 10229703.0, + "step": 474 + }, + { + "epoch": 1.435538752362949, + "grad_norm": 1.828125, + "learning_rate": 9.701900560685539e-06, + "loss": 0.6077, + "num_tokens": 10237448.0, + "step": 475 + }, + { + "epoch": 1.438563327032136, + "grad_norm": 1.71875, + "learning_rate": 9.700322271798657e-06, + "loss": 0.9443, + "num_tokens": 10268389.0, + "step": 476 + }, + { + "epoch": 1.4415879017013231, + "grad_norm": 1.9375, + "learning_rate": 9.698739959705882e-06, + "loss": 0.7023, + "num_tokens": 10276243.0, + "step": 477 + }, + { + "epoch": 1.4446124763705104, + "grad_norm": 1.03125, + "learning_rate": 9.6971536259228e-06, + "loss": 0.8618, + "num_tokens": 10325743.0, + "step": 478 + }, + { + "epoch": 1.4476370510396976, + "grad_norm": 1.515625, + "learning_rate": 9.695563271968853e-06, + "loss": 0.8545, + "num_tokens": 10349265.0, + "step": 479 + }, + { + "epoch": 1.4506616257088847, + "grad_norm": 2.203125, + "learning_rate": 9.69396889936734e-06, + "loss": 0.7371, + "num_tokens": 10356388.0, + "step": 480 + }, + { + "epoch": 1.4536862003780717, + "grad_norm": 1.34375, + "learning_rate": 9.692370509645396e-06, + "loss": 0.8772, + "num_tokens": 10390768.0, + "step": 481 + }, + { + "epoch": 1.456710775047259, + "grad_norm": 1.9453125, + "learning_rate": 9.690768104334015e-06, + "loss": 0.5933, + "num_tokens": 10397873.0, + "step": 482 + }, + { + "epoch": 1.4597353497164463, + "grad_norm": 2.15625, + "learning_rate": 9.68916168496803e-06, + "loss": 0.6353, + "num_tokens": 10405031.0, + "step": 483 + }, + { + "epoch": 1.4627599243856333, + "grad_norm": 1.9765625, + "learning_rate": 9.687551253086124e-06, + "loss": 0.6207, + "num_tokens": 10412132.0, + "step": 484 + }, + { + "epoch": 1.4657844990548203, + "grad_norm": 1.7421875, + "learning_rate": 9.685936810230824e-06, + "loss": 0.9409, + "num_tokens": 10435112.0, + "step": 485 + }, + { + "epoch": 1.4688090737240076, + "grad_norm": 1.8828125, + "learning_rate": 9.684318357948492e-06, + "loss": 0.5855, + "num_tokens": 10442524.0, + "step": 486 + }, + { + "epoch": 1.4718336483931946, + "grad_norm": 1.6640625, + "learning_rate": 9.682695897789334e-06, + "loss": 0.947, + "num_tokens": 10465443.0, + "step": 487 + }, + { + "epoch": 1.474858223062382, + "grad_norm": 1.6171875, + "learning_rate": 9.6810694313074e-06, + "loss": 0.8705, + "num_tokens": 10489319.0, + "step": 488 + }, + { + "epoch": 1.477882797731569, + "grad_norm": 1.2578125, + "learning_rate": 9.67943896006057e-06, + "loss": 0.8603, + "num_tokens": 10531120.0, + "step": 489 + }, + { + "epoch": 1.4809073724007562, + "grad_norm": 1.6015625, + "learning_rate": 9.677804485610558e-06, + "loss": 0.869, + "num_tokens": 10553420.0, + "step": 490 + }, + { + "epoch": 1.4839319470699432, + "grad_norm": 1.984375, + "learning_rate": 9.676166009522925e-06, + "loss": 0.837, + "num_tokens": 10569079.0, + "step": 491 + }, + { + "epoch": 1.4869565217391305, + "grad_norm": 1.2890625, + "learning_rate": 9.674523533367052e-06, + "loss": 0.8709, + "num_tokens": 10609030.0, + "step": 492 + }, + { + "epoch": 1.4899810964083176, + "grad_norm": 2.21875, + "learning_rate": 9.672877058716158e-06, + "loss": 0.7131, + "num_tokens": 10620635.0, + "step": 493 + }, + { + "epoch": 1.4930056710775048, + "grad_norm": 1.46875, + "learning_rate": 9.67122658714729e-06, + "loss": 0.845, + "num_tokens": 10643835.0, + "step": 494 + }, + { + "epoch": 1.4960302457466919, + "grad_norm": 1.78125, + "learning_rate": 9.669572120241326e-06, + "loss": 0.8442, + "num_tokens": 10659900.0, + "step": 495 + }, + { + "epoch": 1.499054820415879, + "grad_norm": 1.3125, + "learning_rate": 9.667913659582962e-06, + "loss": 0.8872, + "num_tokens": 10693187.0, + "step": 496 + }, + { + "epoch": 1.5020793950850662, + "grad_norm": 1.4609375, + "learning_rate": 9.666251206760732e-06, + "loss": 0.8352, + "num_tokens": 10720869.0, + "step": 497 + }, + { + "epoch": 1.5051039697542534, + "grad_norm": 2.140625, + "learning_rate": 9.664584763366983e-06, + "loss": 0.5827, + "num_tokens": 10727035.0, + "step": 498 + }, + { + "epoch": 1.5081285444234405, + "grad_norm": 1.90625, + "learning_rate": 9.662914330997894e-06, + "loss": 0.6202, + "num_tokens": 10734832.0, + "step": 499 + }, + { + "epoch": 1.5111531190926275, + "grad_norm": 1.7265625, + "learning_rate": 9.661239911253457e-06, + "loss": 0.6368, + "num_tokens": 10745779.0, + "step": 500 + }, + { + "epoch": 1.5141776937618148, + "grad_norm": 1.234375, + "learning_rate": 9.659561505737486e-06, + "loss": 0.8189, + "num_tokens": 10784760.0, + "step": 501 + }, + { + "epoch": 1.517202268431002, + "grad_norm": 1.7265625, + "learning_rate": 9.657879116057612e-06, + "loss": 0.8604, + "num_tokens": 10807768.0, + "step": 502 + }, + { + "epoch": 1.520226843100189, + "grad_norm": 1.2578125, + "learning_rate": 9.656192743825283e-06, + "loss": 0.9376, + "num_tokens": 10847237.0, + "step": 503 + }, + { + "epoch": 1.5232514177693761, + "grad_norm": 1.4765625, + "learning_rate": 9.654502390655762e-06, + "loss": 0.8189, + "num_tokens": 10870666.0, + "step": 504 + }, + { + "epoch": 1.5262759924385634, + "grad_norm": 2.046875, + "learning_rate": 9.652808058168125e-06, + "loss": 0.6395, + "num_tokens": 10878007.0, + "step": 505 + }, + { + "epoch": 1.5293005671077504, + "grad_norm": 1.640625, + "learning_rate": 9.651109747985257e-06, + "loss": 0.8526, + "num_tokens": 10902996.0, + "step": 506 + }, + { + "epoch": 1.5323251417769375, + "grad_norm": 2.15625, + "learning_rate": 9.649407461733857e-06, + "loss": 0.6729, + "num_tokens": 10909889.0, + "step": 507 + }, + { + "epoch": 1.5353497164461247, + "grad_norm": 1.4375, + "learning_rate": 9.647701201044429e-06, + "loss": 0.8199, + "num_tokens": 10934700.0, + "step": 508 + }, + { + "epoch": 1.538374291115312, + "grad_norm": 1.84375, + "learning_rate": 9.645990967551287e-06, + "loss": 0.6187, + "num_tokens": 10943015.0, + "step": 509 + }, + { + "epoch": 1.541398865784499, + "grad_norm": 2.125, + "learning_rate": 9.644276762892548e-06, + "loss": 0.6536, + "num_tokens": 10950477.0, + "step": 510 + }, + { + "epoch": 1.544423440453686, + "grad_norm": 1.3203125, + "learning_rate": 9.64255858871013e-06, + "loss": 0.9708, + "num_tokens": 10989834.0, + "step": 511 + }, + { + "epoch": 1.5474480151228733, + "grad_norm": 1.2578125, + "learning_rate": 9.640836446649761e-06, + "loss": 0.8405, + "num_tokens": 11040992.0, + "step": 512 + }, + { + "epoch": 1.5504725897920606, + "grad_norm": 1.4140625, + "learning_rate": 9.639110338360963e-06, + "loss": 0.782, + "num_tokens": 11063234.0, + "step": 513 + }, + { + "epoch": 1.5534971644612476, + "grad_norm": 2.09375, + "learning_rate": 9.637380265497057e-06, + "loss": 0.9152, + "num_tokens": 11082452.0, + "step": 514 + }, + { + "epoch": 1.5565217391304347, + "grad_norm": 1.5703125, + "learning_rate": 9.635646229715168e-06, + "loss": 0.9293, + "num_tokens": 11105265.0, + "step": 515 + }, + { + "epoch": 1.559546313799622, + "grad_norm": 1.65625, + "learning_rate": 9.633908232676206e-06, + "loss": 0.6698, + "num_tokens": 11114532.0, + "step": 516 + }, + { + "epoch": 1.5625708884688092, + "grad_norm": 1.3984375, + "learning_rate": 9.632166276044888e-06, + "loss": 0.8454, + "num_tokens": 11147447.0, + "step": 517 + }, + { + "epoch": 1.5655954631379962, + "grad_norm": 1.390625, + "learning_rate": 9.630420361489711e-06, + "loss": 0.7368, + "num_tokens": 11168631.0, + "step": 518 + }, + { + "epoch": 1.5686200378071833, + "grad_norm": 2.125, + "learning_rate": 9.628670490682974e-06, + "loss": 0.7578, + "num_tokens": 11183747.0, + "step": 519 + }, + { + "epoch": 1.5716446124763706, + "grad_norm": 1.8515625, + "learning_rate": 9.626916665300756e-06, + "loss": 0.6072, + "num_tokens": 11191384.0, + "step": 520 + }, + { + "epoch": 1.5746691871455578, + "grad_norm": 1.8203125, + "learning_rate": 9.62515888702293e-06, + "loss": 0.744, + "num_tokens": 11203922.0, + "step": 521 + }, + { + "epoch": 1.5776937618147449, + "grad_norm": 1.40625, + "learning_rate": 9.623397157533155e-06, + "loss": 0.8266, + "num_tokens": 11227742.0, + "step": 522 + }, + { + "epoch": 1.580718336483932, + "grad_norm": 1.5859375, + "learning_rate": 9.62163147851887e-06, + "loss": 0.8348, + "num_tokens": 11255602.0, + "step": 523 + }, + { + "epoch": 1.5837429111531192, + "grad_norm": 1.90625, + "learning_rate": 9.619861851671306e-06, + "loss": 0.6302, + "num_tokens": 11263179.0, + "step": 524 + }, + { + "epoch": 1.5867674858223062, + "grad_norm": 1.375, + "learning_rate": 9.618088278685462e-06, + "loss": 0.8591, + "num_tokens": 11301147.0, + "step": 525 + }, + { + "epoch": 1.5897920604914932, + "grad_norm": 1.9296875, + "learning_rate": 9.616310761260128e-06, + "loss": 0.7745, + "num_tokens": 11314215.0, + "step": 526 + }, + { + "epoch": 1.5928166351606805, + "grad_norm": 1.890625, + "learning_rate": 9.614529301097867e-06, + "loss": 0.5925, + "num_tokens": 11321527.0, + "step": 527 + }, + { + "epoch": 1.5958412098298678, + "grad_norm": 2.140625, + "learning_rate": 9.612743899905021e-06, + "loss": 0.9237, + "num_tokens": 11351376.0, + "step": 528 + }, + { + "epoch": 1.5988657844990548, + "grad_norm": 1.515625, + "learning_rate": 9.610954559391704e-06, + "loss": 0.8973, + "num_tokens": 11376876.0, + "step": 529 + }, + { + "epoch": 1.6018903591682419, + "grad_norm": 1.7265625, + "learning_rate": 9.609161281271808e-06, + "loss": 0.8345, + "num_tokens": 11395499.0, + "step": 530 + }, + { + "epoch": 1.6049149338374291, + "grad_norm": 1.78125, + "learning_rate": 9.607364067262988e-06, + "loss": 0.6124, + "num_tokens": 11403634.0, + "step": 531 + }, + { + "epoch": 1.6079395085066164, + "grad_norm": 1.421875, + "learning_rate": 9.60556291908668e-06, + "loss": 0.8729, + "num_tokens": 11433323.0, + "step": 532 + }, + { + "epoch": 1.6109640831758034, + "grad_norm": 1.2421875, + "learning_rate": 9.603757838468079e-06, + "loss": 0.8594, + "num_tokens": 11472561.0, + "step": 533 + }, + { + "epoch": 1.6139886578449905, + "grad_norm": 1.390625, + "learning_rate": 9.601948827136153e-06, + "loss": 0.9113, + "num_tokens": 11505457.0, + "step": 534 + }, + { + "epoch": 1.6170132325141777, + "grad_norm": 1.0390625, + "learning_rate": 9.600135886823627e-06, + "loss": 0.8749, + "num_tokens": 11560406.0, + "step": 535 + }, + { + "epoch": 1.620037807183365, + "grad_norm": 1.734375, + "learning_rate": 9.598319019267e-06, + "loss": 0.9418, + "num_tokens": 11587046.0, + "step": 536 + }, + { + "epoch": 1.623062381852552, + "grad_norm": 1.8671875, + "learning_rate": 9.596498226206525e-06, + "loss": 0.5581, + "num_tokens": 11595024.0, + "step": 537 + }, + { + "epoch": 1.626086956521739, + "grad_norm": 1.4765625, + "learning_rate": 9.594673509386217e-06, + "loss": 0.8676, + "num_tokens": 11632665.0, + "step": 538 + }, + { + "epoch": 1.6291115311909263, + "grad_norm": 1.8515625, + "learning_rate": 9.592844870553849e-06, + "loss": 0.7354, + "num_tokens": 11644416.0, + "step": 539 + }, + { + "epoch": 1.6321361058601134, + "grad_norm": 1.6328125, + "learning_rate": 9.591012311460952e-06, + "loss": 0.8259, + "num_tokens": 11665088.0, + "step": 540 + }, + { + "epoch": 1.6351606805293004, + "grad_norm": 1.390625, + "learning_rate": 9.58917583386281e-06, + "loss": 0.9554, + "num_tokens": 11703915.0, + "step": 541 + }, + { + "epoch": 1.6381852551984877, + "grad_norm": 1.671875, + "learning_rate": 9.58733543951846e-06, + "loss": 0.9412, + "num_tokens": 11727726.0, + "step": 542 + }, + { + "epoch": 1.641209829867675, + "grad_norm": 1.296875, + "learning_rate": 9.585491130190696e-06, + "loss": 0.9481, + "num_tokens": 11780626.0, + "step": 543 + }, + { + "epoch": 1.644234404536862, + "grad_norm": 1.796875, + "learning_rate": 9.583642907646053e-06, + "loss": 0.9121, + "num_tokens": 11812903.0, + "step": 544 + }, + { + "epoch": 1.647258979206049, + "grad_norm": 2.265625, + "learning_rate": 9.581790773654821e-06, + "loss": 0.8127, + "num_tokens": 11824945.0, + "step": 545 + }, + { + "epoch": 1.6502835538752363, + "grad_norm": 1.421875, + "learning_rate": 9.579934729991037e-06, + "loss": 0.839, + "num_tokens": 11849652.0, + "step": 546 + }, + { + "epoch": 1.6533081285444235, + "grad_norm": 1.234375, + "learning_rate": 9.578074778432475e-06, + "loss": 0.95, + "num_tokens": 11909649.0, + "step": 547 + }, + { + "epoch": 1.6563327032136106, + "grad_norm": 1.671875, + "learning_rate": 9.576210920760662e-06, + "loss": 0.8287, + "num_tokens": 11932235.0, + "step": 548 + }, + { + "epoch": 1.6593572778827976, + "grad_norm": 1.859375, + "learning_rate": 9.57434315876086e-06, + "loss": 0.5939, + "num_tokens": 11939686.0, + "step": 549 + }, + { + "epoch": 1.662381852551985, + "grad_norm": 1.3125, + "learning_rate": 9.572471494222072e-06, + "loss": 0.9001, + "num_tokens": 11978387.0, + "step": 550 + }, + { + "epoch": 1.6654064272211722, + "grad_norm": 1.2421875, + "learning_rate": 9.57059592893704e-06, + "loss": 0.8463, + "num_tokens": 12008728.0, + "step": 551 + }, + { + "epoch": 1.6684310018903592, + "grad_norm": 1.8125, + "learning_rate": 9.56871646470224e-06, + "loss": 0.7588, + "num_tokens": 12019514.0, + "step": 552 + }, + { + "epoch": 1.6714555765595462, + "grad_norm": 1.53125, + "learning_rate": 9.566833103317887e-06, + "loss": 0.878, + "num_tokens": 12050801.0, + "step": 553 + }, + { + "epoch": 1.6744801512287335, + "grad_norm": 1.4375, + "learning_rate": 9.564945846587925e-06, + "loss": 0.8096, + "num_tokens": 12072925.0, + "step": 554 + }, + { + "epoch": 1.6775047258979208, + "grad_norm": 1.2578125, + "learning_rate": 9.56305469632003e-06, + "loss": 0.8754, + "num_tokens": 12121822.0, + "step": 555 + }, + { + "epoch": 1.6805293005671076, + "grad_norm": 2.015625, + "learning_rate": 9.561159654325607e-06, + "loss": 0.6118, + "num_tokens": 12129020.0, + "step": 556 + }, + { + "epoch": 1.6835538752362949, + "grad_norm": 1.5546875, + "learning_rate": 9.55926072241979e-06, + "loss": 0.8727, + "num_tokens": 12152265.0, + "step": 557 + }, + { + "epoch": 1.6865784499054821, + "grad_norm": 1.953125, + "learning_rate": 9.557357902421438e-06, + "loss": 0.7693, + "num_tokens": 12167048.0, + "step": 558 + }, + { + "epoch": 1.6896030245746692, + "grad_norm": 1.984375, + "learning_rate": 9.555451196153133e-06, + "loss": 0.6624, + "num_tokens": 12174962.0, + "step": 559 + }, + { + "epoch": 1.6926275992438562, + "grad_norm": 1.890625, + "learning_rate": 9.553540605441182e-06, + "loss": 0.6233, + "num_tokens": 12183059.0, + "step": 560 + }, + { + "epoch": 1.6956521739130435, + "grad_norm": 1.375, + "learning_rate": 9.55162613211561e-06, + "loss": 0.8391, + "num_tokens": 12212536.0, + "step": 561 + }, + { + "epoch": 1.6986767485822307, + "grad_norm": 1.5078125, + "learning_rate": 9.549707778010164e-06, + "loss": 0.7498, + "num_tokens": 12233453.0, + "step": 562 + }, + { + "epoch": 1.7017013232514178, + "grad_norm": 1.5703125, + "learning_rate": 9.547785544962303e-06, + "loss": 0.7604, + "num_tokens": 12256486.0, + "step": 563 + }, + { + "epoch": 1.7047258979206048, + "grad_norm": 2.125, + "learning_rate": 9.545859434813209e-06, + "loss": 0.7831, + "num_tokens": 12269798.0, + "step": 564 + }, + { + "epoch": 1.707750472589792, + "grad_norm": 1.6484375, + "learning_rate": 9.543929449407768e-06, + "loss": 0.9422, + "num_tokens": 12293906.0, + "step": 565 + }, + { + "epoch": 1.7107750472589793, + "grad_norm": 1.5703125, + "learning_rate": 9.541995590594589e-06, + "loss": 0.935, + "num_tokens": 12317484.0, + "step": 566 + }, + { + "epoch": 1.7137996219281664, + "grad_norm": 1.9296875, + "learning_rate": 9.54005786022598e-06, + "loss": 0.6388, + "num_tokens": 12324700.0, + "step": 567 + }, + { + "epoch": 1.7168241965973534, + "grad_norm": 1.3046875, + "learning_rate": 9.538116260157965e-06, + "loss": 0.8854, + "num_tokens": 12369368.0, + "step": 568 + }, + { + "epoch": 1.7198487712665407, + "grad_norm": 1.2890625, + "learning_rate": 9.53617079225027e-06, + "loss": 0.8744, + "num_tokens": 12407148.0, + "step": 569 + }, + { + "epoch": 1.722873345935728, + "grad_norm": 1.5625, + "learning_rate": 9.534221458366331e-06, + "loss": 0.8081, + "num_tokens": 12427469.0, + "step": 570 + }, + { + "epoch": 1.725897920604915, + "grad_norm": 1.6640625, + "learning_rate": 9.53226826037328e-06, + "loss": 0.8113, + "num_tokens": 12449664.0, + "step": 571 + }, + { + "epoch": 1.728922495274102, + "grad_norm": 1.515625, + "learning_rate": 9.530311200141957e-06, + "loss": 0.8573, + "num_tokens": 12473499.0, + "step": 572 + }, + { + "epoch": 1.7319470699432893, + "grad_norm": 1.875, + "learning_rate": 9.528350279546897e-06, + "loss": 0.6223, + "num_tokens": 12481631.0, + "step": 573 + }, + { + "epoch": 1.7349716446124763, + "grad_norm": 1.5078125, + "learning_rate": 9.526385500466332e-06, + "loss": 0.7358, + "num_tokens": 12511380.0, + "step": 574 + }, + { + "epoch": 1.7379962192816634, + "grad_norm": 1.3671875, + "learning_rate": 9.524416864782196e-06, + "loss": 0.846, + "num_tokens": 12547767.0, + "step": 575 + }, + { + "epoch": 1.7410207939508506, + "grad_norm": 1.96875, + "learning_rate": 9.522444374380107e-06, + "loss": 0.5425, + "num_tokens": 12554843.0, + "step": 576 + }, + { + "epoch": 1.744045368620038, + "grad_norm": 1.640625, + "learning_rate": 9.520468031149385e-06, + "loss": 0.815, + "num_tokens": 12577150.0, + "step": 577 + }, + { + "epoch": 1.747069943289225, + "grad_norm": 1.34375, + "learning_rate": 9.518487836983035e-06, + "loss": 0.9105, + "num_tokens": 12611376.0, + "step": 578 + }, + { + "epoch": 1.750094517958412, + "grad_norm": 2.15625, + "learning_rate": 9.51650379377775e-06, + "loss": 0.6649, + "num_tokens": 12617880.0, + "step": 579 + }, + { + "epoch": 1.7531190926275992, + "grad_norm": 2.140625, + "learning_rate": 9.514515903433915e-06, + "loss": 0.9624, + "num_tokens": 12641035.0, + "step": 580 + }, + { + "epoch": 1.7561436672967865, + "grad_norm": 1.234375, + "learning_rate": 9.51252416785559e-06, + "loss": 0.8544, + "num_tokens": 12696214.0, + "step": 581 + }, + { + "epoch": 1.7591682419659735, + "grad_norm": 2.421875, + "learning_rate": 9.510528588950533e-06, + "loss": 0.8727, + "num_tokens": 12707607.0, + "step": 582 + }, + { + "epoch": 1.7621928166351606, + "grad_norm": 2.03125, + "learning_rate": 9.50852916863017e-06, + "loss": 0.6043, + "num_tokens": 12714302.0, + "step": 583 + }, + { + "epoch": 1.7652173913043478, + "grad_norm": 1.25, + "learning_rate": 9.50652590880961e-06, + "loss": 0.8499, + "num_tokens": 12760445.0, + "step": 584 + }, + { + "epoch": 1.7682419659735351, + "grad_norm": 2.109375, + "learning_rate": 9.50451881140764e-06, + "loss": 0.6398, + "num_tokens": 12767992.0, + "step": 585 + }, + { + "epoch": 1.7712665406427222, + "grad_norm": 1.546875, + "learning_rate": 9.502507878346729e-06, + "loss": 0.7958, + "num_tokens": 12789366.0, + "step": 586 + }, + { + "epoch": 1.7742911153119092, + "grad_norm": 2.03125, + "learning_rate": 9.500493111553007e-06, + "loss": 0.6877, + "num_tokens": 12799764.0, + "step": 587 + }, + { + "epoch": 1.7773156899810965, + "grad_norm": 1.8359375, + "learning_rate": 9.498474512956291e-06, + "loss": 0.7888, + "num_tokens": 12819930.0, + "step": 588 + }, + { + "epoch": 1.7803402646502835, + "grad_norm": 1.4140625, + "learning_rate": 9.496452084490054e-06, + "loss": 0.7977, + "num_tokens": 12843729.0, + "step": 589 + }, + { + "epoch": 1.7833648393194705, + "grad_norm": 1.984375, + "learning_rate": 9.49442582809145e-06, + "loss": 0.6019, + "num_tokens": 12851165.0, + "step": 590 + }, + { + "epoch": 1.7863894139886578, + "grad_norm": 1.890625, + "learning_rate": 9.492395745701287e-06, + "loss": 0.6249, + "num_tokens": 12858874.0, + "step": 591 + }, + { + "epoch": 1.789413988657845, + "grad_norm": 1.9921875, + "learning_rate": 9.49036183926405e-06, + "loss": 0.6618, + "num_tokens": 12866076.0, + "step": 592 + }, + { + "epoch": 1.792438563327032, + "grad_norm": 2.25, + "learning_rate": 9.488324110727878e-06, + "loss": 0.6398, + "num_tokens": 12872538.0, + "step": 593 + }, + { + "epoch": 1.7954631379962191, + "grad_norm": 1.96875, + "learning_rate": 9.486282562044577e-06, + "loss": 0.6357, + "num_tokens": 12879739.0, + "step": 594 + }, + { + "epoch": 1.7984877126654064, + "grad_norm": 1.4921875, + "learning_rate": 9.484237195169606e-06, + "loss": 0.7764, + "num_tokens": 12896230.0, + "step": 595 + }, + { + "epoch": 1.8015122873345937, + "grad_norm": 1.4296875, + "learning_rate": 9.482188012062084e-06, + "loss": 0.8252, + "num_tokens": 12926319.0, + "step": 596 + }, + { + "epoch": 1.8045368620037807, + "grad_norm": 1.5390625, + "learning_rate": 9.480135014684787e-06, + "loss": 0.8621, + "num_tokens": 12953099.0, + "step": 597 + }, + { + "epoch": 1.8075614366729678, + "grad_norm": 1.953125, + "learning_rate": 9.478078205004142e-06, + "loss": 0.6083, + "num_tokens": 12960885.0, + "step": 598 + }, + { + "epoch": 1.810586011342155, + "grad_norm": 1.3671875, + "learning_rate": 9.476017584990229e-06, + "loss": 0.876, + "num_tokens": 12988983.0, + "step": 599 + }, + { + "epoch": 1.8136105860113423, + "grad_norm": 1.7890625, + "learning_rate": 9.473953156616777e-06, + "loss": 0.8051, + "num_tokens": 13011090.0, + "step": 600 + }, + { + "epoch": 1.8166351606805293, + "grad_norm": 2.015625, + "learning_rate": 9.471884921861165e-06, + "loss": 0.6572, + "num_tokens": 13018247.0, + "step": 601 + }, + { + "epoch": 1.8196597353497164, + "grad_norm": 1.53125, + "learning_rate": 9.469812882704413e-06, + "loss": 0.8909, + "num_tokens": 13041644.0, + "step": 602 + }, + { + "epoch": 1.8226843100189036, + "grad_norm": 1.703125, + "learning_rate": 9.467737041131188e-06, + "loss": 0.9712, + "num_tokens": 13065027.0, + "step": 603 + }, + { + "epoch": 1.825708884688091, + "grad_norm": 1.546875, + "learning_rate": 9.465657399129801e-06, + "loss": 0.8185, + "num_tokens": 13088580.0, + "step": 604 + }, + { + "epoch": 1.828733459357278, + "grad_norm": 1.625, + "learning_rate": 9.4635739586922e-06, + "loss": 0.886, + "num_tokens": 13115574.0, + "step": 605 + }, + { + "epoch": 1.831758034026465, + "grad_norm": 1.359375, + "learning_rate": 9.461486721813972e-06, + "loss": 0.9191, + "num_tokens": 13153198.0, + "step": 606 + }, + { + "epoch": 1.8347826086956522, + "grad_norm": 1.25, + "learning_rate": 9.45939569049434e-06, + "loss": 0.9226, + "num_tokens": 13198093.0, + "step": 607 + }, + { + "epoch": 1.8378071833648393, + "grad_norm": 1.34375, + "learning_rate": 9.45730086673616e-06, + "loss": 0.8273, + "num_tokens": 13227337.0, + "step": 608 + }, + { + "epoch": 1.8408317580340263, + "grad_norm": 1.828125, + "learning_rate": 9.45520225254593e-06, + "loss": 0.8565, + "num_tokens": 13251103.0, + "step": 609 + }, + { + "epoch": 1.8438563327032136, + "grad_norm": 1.2421875, + "learning_rate": 9.453099849933766e-06, + "loss": 0.811, + "num_tokens": 13294594.0, + "step": 610 + }, + { + "epoch": 1.8468809073724008, + "grad_norm": 1.640625, + "learning_rate": 9.450993660913418e-06, + "loss": 0.8664, + "num_tokens": 13318421.0, + "step": 611 + }, + { + "epoch": 1.849905482041588, + "grad_norm": 1.1953125, + "learning_rate": 9.448883687502261e-06, + "loss": 0.8987, + "num_tokens": 13363368.0, + "step": 612 + }, + { + "epoch": 1.852930056710775, + "grad_norm": 1.90625, + "learning_rate": 9.446769931721301e-06, + "loss": 0.6336, + "num_tokens": 13371394.0, + "step": 613 + }, + { + "epoch": 1.8559546313799622, + "grad_norm": 1.3515625, + "learning_rate": 9.444652395595159e-06, + "loss": 0.8493, + "num_tokens": 13404943.0, + "step": 614 + }, + { + "epoch": 1.8589792060491495, + "grad_norm": 1.6171875, + "learning_rate": 9.442531081152082e-06, + "loss": 0.7491, + "num_tokens": 13428307.0, + "step": 615 + }, + { + "epoch": 1.8620037807183365, + "grad_norm": 1.7578125, + "learning_rate": 9.440405990423933e-06, + "loss": 0.7385, + "num_tokens": 13438894.0, + "step": 616 + }, + { + "epoch": 1.8650283553875235, + "grad_norm": 1.9296875, + "learning_rate": 9.438277125446194e-06, + "loss": 0.7044, + "num_tokens": 13446897.0, + "step": 617 + }, + { + "epoch": 1.8680529300567108, + "grad_norm": 2.09375, + "learning_rate": 9.436144488257963e-06, + "loss": 0.7403, + "num_tokens": 13459492.0, + "step": 618 + }, + { + "epoch": 1.871077504725898, + "grad_norm": 1.6171875, + "learning_rate": 9.434008080901946e-06, + "loss": 0.8486, + "num_tokens": 13482777.0, + "step": 619 + }, + { + "epoch": 1.874102079395085, + "grad_norm": 2.125, + "learning_rate": 9.431867905424466e-06, + "loss": 0.6361, + "num_tokens": 13489311.0, + "step": 620 + }, + { + "epoch": 1.8771266540642721, + "grad_norm": 1.4140625, + "learning_rate": 9.429723963875454e-06, + "loss": 0.9128, + "num_tokens": 13520345.0, + "step": 621 + }, + { + "epoch": 1.8801512287334594, + "grad_norm": 1.921875, + "learning_rate": 9.427576258308446e-06, + "loss": 0.6995, + "num_tokens": 13530296.0, + "step": 622 + }, + { + "epoch": 1.8831758034026465, + "grad_norm": 1.34375, + "learning_rate": 9.425424790780581e-06, + "loss": 0.862, + "num_tokens": 13562179.0, + "step": 623 + }, + { + "epoch": 1.8862003780718335, + "grad_norm": 1.7578125, + "learning_rate": 9.423269563352611e-06, + "loss": 0.8138, + "num_tokens": 13576564.0, + "step": 624 + }, + { + "epoch": 1.8892249527410208, + "grad_norm": 1.3203125, + "learning_rate": 9.421110578088883e-06, + "loss": 0.8215, + "num_tokens": 13605160.0, + "step": 625 + }, + { + "epoch": 1.892249527410208, + "grad_norm": 2.140625, + "learning_rate": 9.418947837057338e-06, + "loss": 0.5705, + "num_tokens": 13611531.0, + "step": 626 + }, + { + "epoch": 1.895274102079395, + "grad_norm": 1.8125, + "learning_rate": 9.416781342329523e-06, + "loss": 0.6939, + "num_tokens": 13624048.0, + "step": 627 + }, + { + "epoch": 1.898298676748582, + "grad_norm": 2.078125, + "learning_rate": 9.41461109598058e-06, + "loss": 0.6731, + "num_tokens": 13631252.0, + "step": 628 + }, + { + "epoch": 1.9013232514177694, + "grad_norm": 3.265625, + "learning_rate": 9.412437100089236e-06, + "loss": 0.9868, + "num_tokens": 13655379.0, + "step": 629 + }, + { + "epoch": 1.9043478260869566, + "grad_norm": 1.1640625, + "learning_rate": 9.410259356737822e-06, + "loss": 0.8839, + "num_tokens": 13705232.0, + "step": 630 + }, + { + "epoch": 1.9073724007561437, + "grad_norm": 1.828125, + "learning_rate": 9.408077868012248e-06, + "loss": 0.7711, + "num_tokens": 13722104.0, + "step": 631 + }, + { + "epoch": 1.9103969754253307, + "grad_norm": 1.4453125, + "learning_rate": 9.405892636002016e-06, + "loss": 0.8027, + "num_tokens": 13746039.0, + "step": 632 + }, + { + "epoch": 1.913421550094518, + "grad_norm": 1.9921875, + "learning_rate": 9.403703662800213e-06, + "loss": 0.7428, + "num_tokens": 13755910.0, + "step": 633 + }, + { + "epoch": 1.9164461247637052, + "grad_norm": 1.515625, + "learning_rate": 9.401510950503506e-06, + "loss": 0.8145, + "num_tokens": 13782533.0, + "step": 634 + }, + { + "epoch": 1.9194706994328923, + "grad_norm": 1.3359375, + "learning_rate": 9.39931450121215e-06, + "loss": 0.8151, + "num_tokens": 13811848.0, + "step": 635 + }, + { + "epoch": 1.9224952741020793, + "grad_norm": 1.6015625, + "learning_rate": 9.397114317029975e-06, + "loss": 0.8439, + "num_tokens": 13835344.0, + "step": 636 + }, + { + "epoch": 1.9255198487712666, + "grad_norm": 2.03125, + "learning_rate": 9.394910400064387e-06, + "loss": 0.5652, + "num_tokens": 13842151.0, + "step": 637 + }, + { + "epoch": 1.9285444234404538, + "grad_norm": 1.8203125, + "learning_rate": 9.392702752426377e-06, + "loss": 0.627, + "num_tokens": 13850055.0, + "step": 638 + }, + { + "epoch": 1.9315689981096407, + "grad_norm": 1.53125, + "learning_rate": 9.390491376230491e-06, + "loss": 0.8769, + "num_tokens": 13873448.0, + "step": 639 + }, + { + "epoch": 1.934593572778828, + "grad_norm": 2.0, + "learning_rate": 9.388276273594865e-06, + "loss": 0.6612, + "num_tokens": 13881063.0, + "step": 640 + }, + { + "epoch": 1.9376181474480152, + "grad_norm": 1.1640625, + "learning_rate": 9.386057446641195e-06, + "loss": 0.9563, + "num_tokens": 13945097.0, + "step": 641 + }, + { + "epoch": 1.9406427221172022, + "grad_norm": 1.6640625, + "learning_rate": 9.383834897494747e-06, + "loss": 0.8555, + "num_tokens": 13968911.0, + "step": 642 + }, + { + "epoch": 1.9436672967863893, + "grad_norm": 1.546875, + "learning_rate": 9.38160862828435e-06, + "loss": 0.7759, + "num_tokens": 13995346.0, + "step": 643 + }, + { + "epoch": 1.9466918714555765, + "grad_norm": 2.046875, + "learning_rate": 9.379378641142394e-06, + "loss": 0.6935, + "num_tokens": 14005661.0, + "step": 644 + }, + { + "epoch": 1.9497164461247638, + "grad_norm": 2.046875, + "learning_rate": 9.37714493820484e-06, + "loss": 0.6241, + "num_tokens": 14015108.0, + "step": 645 + }, + { + "epoch": 1.9527410207939508, + "grad_norm": 2.046875, + "learning_rate": 9.374907521611198e-06, + "loss": 0.7575, + "num_tokens": 14027292.0, + "step": 646 + }, + { + "epoch": 1.9557655954631379, + "grad_norm": 2.0, + "learning_rate": 9.372666393504537e-06, + "loss": 0.6833, + "num_tokens": 14035410.0, + "step": 647 + }, + { + "epoch": 1.9587901701323251, + "grad_norm": 1.609375, + "learning_rate": 9.370421556031487e-06, + "loss": 0.7732, + "num_tokens": 14052868.0, + "step": 648 + }, + { + "epoch": 1.9618147448015124, + "grad_norm": 1.1328125, + "learning_rate": 9.368173011342224e-06, + "loss": 0.9583, + "num_tokens": 14115475.0, + "step": 649 + }, + { + "epoch": 1.9648393194706995, + "grad_norm": 1.4375, + "learning_rate": 9.365920761590478e-06, + "loss": 0.8014, + "num_tokens": 14138266.0, + "step": 650 + }, + { + "epoch": 1.9678638941398865, + "grad_norm": 1.734375, + "learning_rate": 9.363664808933529e-06, + "loss": 0.9068, + "num_tokens": 14161685.0, + "step": 651 + }, + { + "epoch": 1.9708884688090738, + "grad_norm": 1.1015625, + "learning_rate": 9.3614051555322e-06, + "loss": 0.8471, + "num_tokens": 14212612.0, + "step": 652 + }, + { + "epoch": 1.973913043478261, + "grad_norm": 1.6796875, + "learning_rate": 9.35914180355086e-06, + "loss": 0.8478, + "num_tokens": 14232989.0, + "step": 653 + }, + { + "epoch": 1.976937618147448, + "grad_norm": 1.5234375, + "learning_rate": 9.356874755157424e-06, + "loss": 0.8438, + "num_tokens": 14257681.0, + "step": 654 + }, + { + "epoch": 1.979962192816635, + "grad_norm": 2.078125, + "learning_rate": 9.354604012523344e-06, + "loss": 0.657, + "num_tokens": 14264759.0, + "step": 655 + }, + { + "epoch": 1.9829867674858224, + "grad_norm": 1.578125, + "learning_rate": 9.352329577823613e-06, + "loss": 0.8488, + "num_tokens": 14289266.0, + "step": 656 + }, + { + "epoch": 1.9860113421550094, + "grad_norm": 1.28125, + "learning_rate": 9.350051453236755e-06, + "loss": 1.0032, + "num_tokens": 14344238.0, + "step": 657 + }, + { + "epoch": 1.9890359168241964, + "grad_norm": 1.8671875, + "learning_rate": 9.347769640944834e-06, + "loss": 0.5617, + "num_tokens": 14352942.0, + "step": 658 + }, + { + "epoch": 1.9920604914933837, + "grad_norm": 1.8359375, + "learning_rate": 9.345484143133447e-06, + "loss": 0.5818, + "num_tokens": 14360150.0, + "step": 659 + }, + { + "epoch": 1.995085066162571, + "grad_norm": 1.6796875, + "learning_rate": 9.343194961991716e-06, + "loss": 0.7118, + "num_tokens": 14378108.0, + "step": 660 + }, + { + "epoch": 1.998109640831758, + "grad_norm": 1.90625, + "learning_rate": 9.340902099712294e-06, + "loss": 0.6401, + "num_tokens": 14385617.0, + "step": 661 + }, + { + "epoch": 2.0, + "grad_norm": 2.265625, + "learning_rate": 9.33860555849136e-06, + "loss": 0.6405, + "num_tokens": 14390626.0, + "step": 662 + }, + { + "epoch": 2.0030245746691873, + "grad_norm": 2.0625, + "learning_rate": 9.336305340528613e-06, + "loss": 0.5857, + "num_tokens": 14398288.0, + "step": 663 + }, + { + "epoch": 2.006049149338374, + "grad_norm": 1.8984375, + "learning_rate": 9.33400144802728e-06, + "loss": 0.6372, + "num_tokens": 14405593.0, + "step": 664 + }, + { + "epoch": 2.0090737240075613, + "grad_norm": 1.375, + "learning_rate": 9.331693883194105e-06, + "loss": 0.7523, + "num_tokens": 14427279.0, + "step": 665 + }, + { + "epoch": 2.0120982986767486, + "grad_norm": 1.5546875, + "learning_rate": 9.329382648239348e-06, + "loss": 0.8092, + "num_tokens": 14448500.0, + "step": 666 + }, + { + "epoch": 2.015122873345936, + "grad_norm": 1.2421875, + "learning_rate": 9.327067745376785e-06, + "loss": 0.858, + "num_tokens": 14484671.0, + "step": 667 + }, + { + "epoch": 2.0181474480151227, + "grad_norm": 1.6484375, + "learning_rate": 9.324749176823704e-06, + "loss": 0.6795, + "num_tokens": 14497582.0, + "step": 668 + }, + { + "epoch": 2.02117202268431, + "grad_norm": 1.4921875, + "learning_rate": 9.322426944800909e-06, + "loss": 0.8187, + "num_tokens": 14521209.0, + "step": 669 + }, + { + "epoch": 2.024196597353497, + "grad_norm": 1.7734375, + "learning_rate": 9.320101051532709e-06, + "loss": 0.5385, + "num_tokens": 14529012.0, + "step": 670 + }, + { + "epoch": 2.0272211720226845, + "grad_norm": 2.125, + "learning_rate": 9.317771499246918e-06, + "loss": 0.6371, + "num_tokens": 14537356.0, + "step": 671 + }, + { + "epoch": 2.0302457466918713, + "grad_norm": 1.6171875, + "learning_rate": 9.31543829017486e-06, + "loss": 1.0007, + "num_tokens": 14576992.0, + "step": 672 + }, + { + "epoch": 2.0332703213610586, + "grad_norm": 2.0, + "learning_rate": 9.313101426551357e-06, + "loss": 0.6694, + "num_tokens": 14584486.0, + "step": 673 + }, + { + "epoch": 2.036294896030246, + "grad_norm": 1.6015625, + "learning_rate": 9.310760910614736e-06, + "loss": 0.8072, + "num_tokens": 14604858.0, + "step": 674 + }, + { + "epoch": 2.039319470699433, + "grad_norm": 1.7421875, + "learning_rate": 9.308416744606816e-06, + "loss": 0.6181, + "num_tokens": 14612838.0, + "step": 675 + }, + { + "epoch": 2.04234404536862, + "grad_norm": 2.0625, + "learning_rate": 9.306068930772918e-06, + "loss": 0.6372, + "num_tokens": 14620160.0, + "step": 676 + }, + { + "epoch": 2.045368620037807, + "grad_norm": 1.953125, + "learning_rate": 9.303717471361855e-06, + "loss": 0.6643, + "num_tokens": 14628083.0, + "step": 677 + }, + { + "epoch": 2.0483931947069944, + "grad_norm": 1.4765625, + "learning_rate": 9.301362368625931e-06, + "loss": 0.9108, + "num_tokens": 14652645.0, + "step": 678 + }, + { + "epoch": 2.0514177693761813, + "grad_norm": 2.03125, + "learning_rate": 9.29900362482094e-06, + "loss": 0.6652, + "num_tokens": 14660270.0, + "step": 679 + }, + { + "epoch": 2.0544423440453685, + "grad_norm": 1.390625, + "learning_rate": 9.296641242206165e-06, + "loss": 0.7433, + "num_tokens": 14682586.0, + "step": 680 + }, + { + "epoch": 2.057466918714556, + "grad_norm": 1.859375, + "learning_rate": 9.294275223044372e-06, + "loss": 0.5065, + "num_tokens": 14689710.0, + "step": 681 + }, + { + "epoch": 2.060491493383743, + "grad_norm": 1.8671875, + "learning_rate": 9.291905569601812e-06, + "loss": 0.5662, + "num_tokens": 14697253.0, + "step": 682 + }, + { + "epoch": 2.06351606805293, + "grad_norm": 1.1484375, + "learning_rate": 9.289532284148218e-06, + "loss": 0.8541, + "num_tokens": 14739585.0, + "step": 683 + }, + { + "epoch": 2.066540642722117, + "grad_norm": 2.046875, + "learning_rate": 9.287155368956798e-06, + "loss": 0.6007, + "num_tokens": 14745744.0, + "step": 684 + }, + { + "epoch": 2.0695652173913044, + "grad_norm": 1.671875, + "learning_rate": 9.284774826304239e-06, + "loss": 0.9339, + "num_tokens": 14768931.0, + "step": 685 + }, + { + "epoch": 2.0725897920604917, + "grad_norm": 1.5234375, + "learning_rate": 9.282390658470703e-06, + "loss": 0.72, + "num_tokens": 14785933.0, + "step": 686 + }, + { + "epoch": 2.0756143667296785, + "grad_norm": 1.734375, + "learning_rate": 9.280002867739826e-06, + "loss": 0.7041, + "num_tokens": 14799996.0, + "step": 687 + }, + { + "epoch": 2.0786389413988657, + "grad_norm": 1.625, + "learning_rate": 9.277611456398709e-06, + "loss": 0.8718, + "num_tokens": 14819832.0, + "step": 688 + }, + { + "epoch": 2.081663516068053, + "grad_norm": 1.515625, + "learning_rate": 9.275216426737924e-06, + "loss": 0.7887, + "num_tokens": 14840823.0, + "step": 689 + }, + { + "epoch": 2.0846880907372403, + "grad_norm": 1.3984375, + "learning_rate": 9.272817781051512e-06, + "loss": 0.7479, + "num_tokens": 14864201.0, + "step": 690 + }, + { + "epoch": 2.087712665406427, + "grad_norm": 1.25, + "learning_rate": 9.270415521636971e-06, + "loss": 0.849, + "num_tokens": 14896543.0, + "step": 691 + }, + { + "epoch": 2.0907372400756143, + "grad_norm": 1.46875, + "learning_rate": 9.268009650795264e-06, + "loss": 0.8651, + "num_tokens": 14921123.0, + "step": 692 + }, + { + "epoch": 2.0937618147448016, + "grad_norm": 1.75, + "learning_rate": 9.265600170830813e-06, + "loss": 0.5965, + "num_tokens": 14929233.0, + "step": 693 + }, + { + "epoch": 2.096786389413989, + "grad_norm": 1.6015625, + "learning_rate": 9.263187084051499e-06, + "loss": 0.8831, + "num_tokens": 14951751.0, + "step": 694 + }, + { + "epoch": 2.0998109640831757, + "grad_norm": 1.953125, + "learning_rate": 9.260770392768652e-06, + "loss": 0.6623, + "num_tokens": 14959868.0, + "step": 695 + }, + { + "epoch": 2.102835538752363, + "grad_norm": 1.9921875, + "learning_rate": 9.258350099297057e-06, + "loss": 1.0105, + "num_tokens": 14982825.0, + "step": 696 + }, + { + "epoch": 2.10586011342155, + "grad_norm": 1.2578125, + "learning_rate": 9.255926205954953e-06, + "loss": 0.8969, + "num_tokens": 15019553.0, + "step": 697 + }, + { + "epoch": 2.108884688090737, + "grad_norm": 1.4921875, + "learning_rate": 9.253498715064025e-06, + "loss": 0.8048, + "num_tokens": 15042063.0, + "step": 698 + }, + { + "epoch": 2.1119092627599243, + "grad_norm": 1.375, + "learning_rate": 9.251067628949402e-06, + "loss": 0.7426, + "num_tokens": 15067881.0, + "step": 699 + }, + { + "epoch": 2.1149338374291116, + "grad_norm": 1.6015625, + "learning_rate": 9.248632949939655e-06, + "loss": 0.6659, + "num_tokens": 15080673.0, + "step": 700 + }, + { + "epoch": 2.117958412098299, + "grad_norm": 1.125, + "learning_rate": 9.246194680366802e-06, + "loss": 0.8402, + "num_tokens": 15139245.0, + "step": 701 + }, + { + "epoch": 2.1209829867674856, + "grad_norm": 1.90625, + "learning_rate": 9.243752822566297e-06, + "loss": 0.6257, + "num_tokens": 15145964.0, + "step": 702 + }, + { + "epoch": 2.124007561436673, + "grad_norm": 1.3203125, + "learning_rate": 9.24130737887703e-06, + "loss": 0.8193, + "num_tokens": 15179165.0, + "step": 703 + }, + { + "epoch": 2.12703213610586, + "grad_norm": 1.703125, + "learning_rate": 9.23885835164133e-06, + "loss": 0.7552, + "num_tokens": 15198559.0, + "step": 704 + }, + { + "epoch": 2.1300567107750474, + "grad_norm": 1.96875, + "learning_rate": 9.23640574320495e-06, + "loss": 0.7539, + "num_tokens": 15214211.0, + "step": 705 + }, + { + "epoch": 2.1330812854442343, + "grad_norm": 1.828125, + "learning_rate": 9.233949555917084e-06, + "loss": 0.5905, + "num_tokens": 15222292.0, + "step": 706 + }, + { + "epoch": 2.1361058601134215, + "grad_norm": 1.4375, + "learning_rate": 9.231489792130343e-06, + "loss": 0.8755, + "num_tokens": 15254760.0, + "step": 707 + }, + { + "epoch": 2.139130434782609, + "grad_norm": 1.9609375, + "learning_rate": 9.229026454200773e-06, + "loss": 0.6595, + "num_tokens": 15262577.0, + "step": 708 + }, + { + "epoch": 2.142155009451796, + "grad_norm": 1.40625, + "learning_rate": 9.226559544487839e-06, + "loss": 0.9554, + "num_tokens": 15294831.0, + "step": 709 + }, + { + "epoch": 2.145179584120983, + "grad_norm": 1.46875, + "learning_rate": 9.224089065354428e-06, + "loss": 0.8515, + "num_tokens": 15318319.0, + "step": 710 + }, + { + "epoch": 2.14820415879017, + "grad_norm": 1.4375, + "learning_rate": 9.221615019166844e-06, + "loss": 0.8626, + "num_tokens": 15350489.0, + "step": 711 + }, + { + "epoch": 2.1512287334593574, + "grad_norm": 1.8515625, + "learning_rate": 9.219137408294808e-06, + "loss": 0.6397, + "num_tokens": 15358659.0, + "step": 712 + }, + { + "epoch": 2.1542533081285447, + "grad_norm": 1.8359375, + "learning_rate": 9.216656235111463e-06, + "loss": 0.5818, + "num_tokens": 15366744.0, + "step": 713 + }, + { + "epoch": 2.1572778827977315, + "grad_norm": 1.703125, + "learning_rate": 9.214171501993349e-06, + "loss": 0.8959, + "num_tokens": 15389000.0, + "step": 714 + }, + { + "epoch": 2.1603024574669187, + "grad_norm": 1.296875, + "learning_rate": 9.211683211320431e-06, + "loss": 0.9422, + "num_tokens": 15422866.0, + "step": 715 + }, + { + "epoch": 2.163327032136106, + "grad_norm": 1.4375, + "learning_rate": 9.209191365476074e-06, + "loss": 0.8465, + "num_tokens": 15446270.0, + "step": 716 + }, + { + "epoch": 2.166351606805293, + "grad_norm": 2.234375, + "learning_rate": 9.206695966847048e-06, + "loss": 0.6541, + "num_tokens": 15453802.0, + "step": 717 + }, + { + "epoch": 2.16937618147448, + "grad_norm": 1.6953125, + "learning_rate": 9.204197017823526e-06, + "loss": 0.8361, + "num_tokens": 15476022.0, + "step": 718 + }, + { + "epoch": 2.1724007561436673, + "grad_norm": 2.046875, + "learning_rate": 9.201694520799086e-06, + "loss": 0.6193, + "num_tokens": 15483485.0, + "step": 719 + }, + { + "epoch": 2.1754253308128546, + "grad_norm": 1.953125, + "learning_rate": 9.199188478170699e-06, + "loss": 0.6235, + "num_tokens": 15491947.0, + "step": 720 + }, + { + "epoch": 2.1784499054820414, + "grad_norm": 1.75, + "learning_rate": 9.19667889233874e-06, + "loss": 0.7877, + "num_tokens": 15514910.0, + "step": 721 + }, + { + "epoch": 2.1814744801512287, + "grad_norm": 1.53125, + "learning_rate": 9.194165765706963e-06, + "loss": 0.8697, + "num_tokens": 15539251.0, + "step": 722 + }, + { + "epoch": 2.184499054820416, + "grad_norm": 1.7578125, + "learning_rate": 9.19164910068253e-06, + "loss": 0.6743, + "num_tokens": 15548992.0, + "step": 723 + }, + { + "epoch": 2.187523629489603, + "grad_norm": 1.9453125, + "learning_rate": 9.189128899675982e-06, + "loss": 0.5935, + "num_tokens": 15556228.0, + "step": 724 + }, + { + "epoch": 2.19054820415879, + "grad_norm": 1.578125, + "learning_rate": 9.186605165101253e-06, + "loss": 0.7712, + "num_tokens": 15573542.0, + "step": 725 + }, + { + "epoch": 2.1935727788279773, + "grad_norm": 1.7109375, + "learning_rate": 9.184077899375653e-06, + "loss": 0.8621, + "num_tokens": 15595306.0, + "step": 726 + }, + { + "epoch": 2.1965973534971646, + "grad_norm": 1.515625, + "learning_rate": 9.181547104919884e-06, + "loss": 0.8812, + "num_tokens": 15619050.0, + "step": 727 + }, + { + "epoch": 2.1996219281663514, + "grad_norm": 1.484375, + "learning_rate": 9.179012784158023e-06, + "loss": 0.8366, + "num_tokens": 15648121.0, + "step": 728 + }, + { + "epoch": 2.2026465028355386, + "grad_norm": 1.5859375, + "learning_rate": 9.176474939517524e-06, + "loss": 0.7923, + "num_tokens": 15668639.0, + "step": 729 + }, + { + "epoch": 2.205671077504726, + "grad_norm": 1.5390625, + "learning_rate": 9.173933573429217e-06, + "loss": 0.8377, + "num_tokens": 15693007.0, + "step": 730 + }, + { + "epoch": 2.208695652173913, + "grad_norm": 1.4296875, + "learning_rate": 9.171388688327307e-06, + "loss": 0.8182, + "num_tokens": 15721011.0, + "step": 731 + }, + { + "epoch": 2.2117202268431, + "grad_norm": 1.9140625, + "learning_rate": 9.16884028664937e-06, + "loss": 0.5473, + "num_tokens": 15727902.0, + "step": 732 + }, + { + "epoch": 2.2147448015122873, + "grad_norm": 1.3515625, + "learning_rate": 9.166288370836343e-06, + "loss": 0.7499, + "num_tokens": 15755835.0, + "step": 733 + }, + { + "epoch": 2.2177693761814745, + "grad_norm": 1.625, + "learning_rate": 9.163732943332539e-06, + "loss": 0.5619, + "num_tokens": 15764692.0, + "step": 734 + }, + { + "epoch": 2.220793950850662, + "grad_norm": 1.3671875, + "learning_rate": 9.161174006585629e-06, + "loss": 0.8145, + "num_tokens": 15797297.0, + "step": 735 + }, + { + "epoch": 2.2238185255198486, + "grad_norm": 2.25, + "learning_rate": 9.158611563046645e-06, + "loss": 0.6409, + "num_tokens": 15803285.0, + "step": 736 + }, + { + "epoch": 2.226843100189036, + "grad_norm": 1.3828125, + "learning_rate": 9.156045615169978e-06, + "loss": 0.8399, + "num_tokens": 15831766.0, + "step": 737 + }, + { + "epoch": 2.229867674858223, + "grad_norm": 1.140625, + "learning_rate": 9.153476165413379e-06, + "loss": 0.8773, + "num_tokens": 15878362.0, + "step": 738 + }, + { + "epoch": 2.2328922495274104, + "grad_norm": 1.578125, + "learning_rate": 9.15090321623795e-06, + "loss": 0.8096, + "num_tokens": 15900371.0, + "step": 739 + }, + { + "epoch": 2.235916824196597, + "grad_norm": 1.7734375, + "learning_rate": 9.148326770108147e-06, + "loss": 0.7331, + "num_tokens": 15917119.0, + "step": 740 + }, + { + "epoch": 2.2389413988657845, + "grad_norm": 1.0234375, + "learning_rate": 9.145746829491775e-06, + "loss": 0.9085, + "num_tokens": 15995555.0, + "step": 741 + }, + { + "epoch": 2.2419659735349717, + "grad_norm": 1.7890625, + "learning_rate": 9.14316339685998e-06, + "loss": 0.5865, + "num_tokens": 16004842.0, + "step": 742 + }, + { + "epoch": 2.244990548204159, + "grad_norm": 1.2890625, + "learning_rate": 9.140576474687263e-06, + "loss": 0.8386, + "num_tokens": 16043376.0, + "step": 743 + }, + { + "epoch": 2.248015122873346, + "grad_norm": 1.9140625, + "learning_rate": 9.137986065451464e-06, + "loss": 0.5767, + "num_tokens": 16050324.0, + "step": 744 + }, + { + "epoch": 2.251039697542533, + "grad_norm": 1.3046875, + "learning_rate": 9.135392171633756e-06, + "loss": 0.8326, + "num_tokens": 16089338.0, + "step": 745 + }, + { + "epoch": 2.2540642722117203, + "grad_norm": 1.15625, + "learning_rate": 9.132794795718662e-06, + "loss": 0.785, + "num_tokens": 16126064.0, + "step": 746 + }, + { + "epoch": 2.257088846880907, + "grad_norm": 1.71875, + "learning_rate": 9.130193940194027e-06, + "loss": 0.8736, + "num_tokens": 16150569.0, + "step": 747 + }, + { + "epoch": 2.2601134215500944, + "grad_norm": 1.6875, + "learning_rate": 9.127589607551038e-06, + "loss": 0.8734, + "num_tokens": 16173044.0, + "step": 748 + }, + { + "epoch": 2.2631379962192817, + "grad_norm": 1.7109375, + "learning_rate": 9.12498180028421e-06, + "loss": 0.7498, + "num_tokens": 16190480.0, + "step": 749 + }, + { + "epoch": 2.266162570888469, + "grad_norm": 2.015625, + "learning_rate": 9.122370520891387e-06, + "loss": 0.6074, + "num_tokens": 16198063.0, + "step": 750 + }, + { + "epoch": 2.2691871455576558, + "grad_norm": 1.28125, + "learning_rate": 9.119755771873737e-06, + "loss": 0.846, + "num_tokens": 16229012.0, + "step": 751 + }, + { + "epoch": 2.272211720226843, + "grad_norm": 1.328125, + "learning_rate": 9.117137555735753e-06, + "loss": 0.9073, + "num_tokens": 16268549.0, + "step": 752 + }, + { + "epoch": 2.2752362948960303, + "grad_norm": 1.359375, + "learning_rate": 9.114515874985245e-06, + "loss": 0.828, + "num_tokens": 16299229.0, + "step": 753 + }, + { + "epoch": 2.2782608695652176, + "grad_norm": 1.2734375, + "learning_rate": 9.111890732133346e-06, + "loss": 0.8783, + "num_tokens": 16335499.0, + "step": 754 + }, + { + "epoch": 2.2812854442344044, + "grad_norm": 1.328125, + "learning_rate": 9.109262129694506e-06, + "loss": 0.8516, + "num_tokens": 16363368.0, + "step": 755 + }, + { + "epoch": 2.2843100189035916, + "grad_norm": 1.2421875, + "learning_rate": 9.106630070186482e-06, + "loss": 0.7893, + "num_tokens": 16400874.0, + "step": 756 + }, + { + "epoch": 2.287334593572779, + "grad_norm": 1.5, + "learning_rate": 9.103994556130349e-06, + "loss": 0.7896, + "num_tokens": 16424055.0, + "step": 757 + }, + { + "epoch": 2.2903591682419657, + "grad_norm": 1.6796875, + "learning_rate": 9.101355590050489e-06, + "loss": 0.734, + "num_tokens": 16440796.0, + "step": 758 + }, + { + "epoch": 2.293383742911153, + "grad_norm": 2.09375, + "learning_rate": 9.098713174474588e-06, + "loss": 0.6839, + "num_tokens": 16448503.0, + "step": 759 + }, + { + "epoch": 2.2964083175803403, + "grad_norm": 1.6640625, + "learning_rate": 9.096067311933641e-06, + "loss": 0.7677, + "num_tokens": 16464677.0, + "step": 760 + }, + { + "epoch": 2.2994328922495275, + "grad_norm": 2.3125, + "learning_rate": 9.093418004961939e-06, + "loss": 0.7, + "num_tokens": 16471272.0, + "step": 761 + }, + { + "epoch": 2.302457466918715, + "grad_norm": 2.109375, + "learning_rate": 9.090765256097075e-06, + "loss": 0.6345, + "num_tokens": 16477940.0, + "step": 762 + }, + { + "epoch": 2.3054820415879016, + "grad_norm": 1.625, + "learning_rate": 9.08810906787994e-06, + "loss": 0.7066, + "num_tokens": 16492783.0, + "step": 763 + }, + { + "epoch": 2.308506616257089, + "grad_norm": 1.703125, + "learning_rate": 9.085449442854715e-06, + "loss": 0.7518, + "num_tokens": 16506733.0, + "step": 764 + }, + { + "epoch": 2.311531190926276, + "grad_norm": 1.734375, + "learning_rate": 9.082786383568877e-06, + "loss": 0.7467, + "num_tokens": 16521439.0, + "step": 765 + }, + { + "epoch": 2.314555765595463, + "grad_norm": 1.34375, + "learning_rate": 9.080119892573194e-06, + "loss": 0.8679, + "num_tokens": 16560324.0, + "step": 766 + }, + { + "epoch": 2.31758034026465, + "grad_norm": 2.1875, + "learning_rate": 9.077449972421716e-06, + "loss": 0.5733, + "num_tokens": 16566691.0, + "step": 767 + }, + { + "epoch": 2.3206049149338375, + "grad_norm": 1.8671875, + "learning_rate": 9.074776625671777e-06, + "loss": 0.6297, + "num_tokens": 16574794.0, + "step": 768 + }, + { + "epoch": 2.3236294896030247, + "grad_norm": 1.765625, + "learning_rate": 9.072099854884002e-06, + "loss": 0.9068, + "num_tokens": 16597984.0, + "step": 769 + }, + { + "epoch": 2.3266540642722116, + "grad_norm": 2.09375, + "learning_rate": 9.069419662622284e-06, + "loss": 0.5642, + "num_tokens": 16605048.0, + "step": 770 + }, + { + "epoch": 2.329678638941399, + "grad_norm": 1.9296875, + "learning_rate": 9.066736051453802e-06, + "loss": 0.638, + "num_tokens": 16612503.0, + "step": 771 + }, + { + "epoch": 2.332703213610586, + "grad_norm": 1.8984375, + "learning_rate": 9.064049023949003e-06, + "loss": 0.6997, + "num_tokens": 16622499.0, + "step": 772 + }, + { + "epoch": 2.3357277882797733, + "grad_norm": 1.5234375, + "learning_rate": 9.061358582681614e-06, + "loss": 0.8366, + "num_tokens": 16647143.0, + "step": 773 + }, + { + "epoch": 2.33875236294896, + "grad_norm": 1.7265625, + "learning_rate": 9.058664730228625e-06, + "loss": 0.8424, + "num_tokens": 16670768.0, + "step": 774 + }, + { + "epoch": 2.3417769376181474, + "grad_norm": 1.8828125, + "learning_rate": 9.055967469170297e-06, + "loss": 0.6468, + "num_tokens": 16679393.0, + "step": 775 + }, + { + "epoch": 2.3448015122873347, + "grad_norm": 1.328125, + "learning_rate": 9.053266802090152e-06, + "loss": 0.8827, + "num_tokens": 16711692.0, + "step": 776 + }, + { + "epoch": 2.3478260869565215, + "grad_norm": 1.515625, + "learning_rate": 9.050562731574983e-06, + "loss": 0.936, + "num_tokens": 16739059.0, + "step": 777 + }, + { + "epoch": 2.3508506616257088, + "grad_norm": 1.5625, + "learning_rate": 9.047855260214829e-06, + "loss": 0.7708, + "num_tokens": 16758924.0, + "step": 778 + }, + { + "epoch": 2.353875236294896, + "grad_norm": 1.65625, + "learning_rate": 9.045144390603e-06, + "loss": 0.8988, + "num_tokens": 16782314.0, + "step": 779 + }, + { + "epoch": 2.3568998109640833, + "grad_norm": 1.8125, + "learning_rate": 9.042430125336057e-06, + "loss": 0.6254, + "num_tokens": 16790290.0, + "step": 780 + }, + { + "epoch": 2.3599243856332706, + "grad_norm": 1.875, + "learning_rate": 9.039712467013808e-06, + "loss": 0.8239, + "num_tokens": 16812996.0, + "step": 781 + }, + { + "epoch": 2.3629489603024574, + "grad_norm": 1.6015625, + "learning_rate": 9.036991418239316e-06, + "loss": 0.771, + "num_tokens": 16834616.0, + "step": 782 + }, + { + "epoch": 2.3659735349716446, + "grad_norm": 1.9140625, + "learning_rate": 9.03426698161889e-06, + "loss": 0.643, + "num_tokens": 16842810.0, + "step": 783 + }, + { + "epoch": 2.368998109640832, + "grad_norm": 1.125, + "learning_rate": 9.031539159762086e-06, + "loss": 0.9352, + "num_tokens": 16896764.0, + "step": 784 + }, + { + "epoch": 2.3720226843100187, + "grad_norm": 1.953125, + "learning_rate": 9.028807955281701e-06, + "loss": 0.6122, + "num_tokens": 16904646.0, + "step": 785 + }, + { + "epoch": 2.375047258979206, + "grad_norm": 2.046875, + "learning_rate": 9.02607337079377e-06, + "loss": 0.6263, + "num_tokens": 16911496.0, + "step": 786 + }, + { + "epoch": 2.3780718336483933, + "grad_norm": 1.9140625, + "learning_rate": 9.023335408917566e-06, + "loss": 0.6735, + "num_tokens": 16919775.0, + "step": 787 + }, + { + "epoch": 2.3810964083175805, + "grad_norm": 1.1484375, + "learning_rate": 9.020594072275602e-06, + "loss": 0.9113, + "num_tokens": 16972806.0, + "step": 788 + }, + { + "epoch": 2.3841209829867673, + "grad_norm": 1.375, + "learning_rate": 9.01784936349362e-06, + "loss": 0.9241, + "num_tokens": 17012949.0, + "step": 789 + }, + { + "epoch": 2.3871455576559546, + "grad_norm": 1.7890625, + "learning_rate": 9.015101285200586e-06, + "loss": 0.7532, + "num_tokens": 17029354.0, + "step": 790 + }, + { + "epoch": 2.390170132325142, + "grad_norm": 2.078125, + "learning_rate": 9.012349840028707e-06, + "loss": 0.5943, + "num_tokens": 17036458.0, + "step": 791 + }, + { + "epoch": 2.393194706994329, + "grad_norm": 1.859375, + "learning_rate": 9.009595030613403e-06, + "loss": 0.6081, + "num_tokens": 17044068.0, + "step": 792 + }, + { + "epoch": 2.396219281663516, + "grad_norm": 1.9140625, + "learning_rate": 9.00683685959332e-06, + "loss": 0.7287, + "num_tokens": 17057432.0, + "step": 793 + }, + { + "epoch": 2.399243856332703, + "grad_norm": 1.4140625, + "learning_rate": 9.004075329610324e-06, + "loss": 0.758, + "num_tokens": 17080225.0, + "step": 794 + }, + { + "epoch": 2.4022684310018905, + "grad_norm": 1.9140625, + "learning_rate": 9.001310443309503e-06, + "loss": 0.6007, + "num_tokens": 17087440.0, + "step": 795 + }, + { + "epoch": 2.4052930056710773, + "grad_norm": 1.9609375, + "learning_rate": 8.998542203339148e-06, + "loss": 0.6001, + "num_tokens": 17095639.0, + "step": 796 + }, + { + "epoch": 2.4083175803402646, + "grad_norm": 1.6328125, + "learning_rate": 8.995770612350778e-06, + "loss": 0.8106, + "num_tokens": 17117690.0, + "step": 797 + }, + { + "epoch": 2.411342155009452, + "grad_norm": 1.640625, + "learning_rate": 8.992995672999105e-06, + "loss": 0.8172, + "num_tokens": 17136472.0, + "step": 798 + }, + { + "epoch": 2.414366729678639, + "grad_norm": 1.2578125, + "learning_rate": 8.990217387942061e-06, + "loss": 0.8795, + "num_tokens": 17176514.0, + "step": 799 + }, + { + "epoch": 2.417391304347826, + "grad_norm": 1.5234375, + "learning_rate": 8.987435759840782e-06, + "loss": 0.6903, + "num_tokens": 17195404.0, + "step": 800 + }, + { + "epoch": 2.420415879017013, + "grad_norm": 1.5859375, + "learning_rate": 8.984650791359596e-06, + "loss": 0.8308, + "num_tokens": 17225740.0, + "step": 801 + }, + { + "epoch": 2.4234404536862004, + "grad_norm": 1.8203125, + "learning_rate": 8.98186248516604e-06, + "loss": 0.5717, + "num_tokens": 17233159.0, + "step": 802 + }, + { + "epoch": 2.4264650283553877, + "grad_norm": 1.3046875, + "learning_rate": 8.979070843930841e-06, + "loss": 0.8977, + "num_tokens": 17271204.0, + "step": 803 + }, + { + "epoch": 2.4294896030245745, + "grad_norm": 1.46875, + "learning_rate": 8.976275870327928e-06, + "loss": 0.8226, + "num_tokens": 17300085.0, + "step": 804 + }, + { + "epoch": 2.4325141776937618, + "grad_norm": 2.1875, + "learning_rate": 8.97347756703442e-06, + "loss": 0.6224, + "num_tokens": 17307928.0, + "step": 805 + }, + { + "epoch": 2.435538752362949, + "grad_norm": 1.3515625, + "learning_rate": 8.970675936730618e-06, + "loss": 0.8602, + "num_tokens": 17343876.0, + "step": 806 + }, + { + "epoch": 2.4385633270321363, + "grad_norm": 1.625, + "learning_rate": 8.967870982100017e-06, + "loss": 0.85, + "num_tokens": 17381189.0, + "step": 807 + }, + { + "epoch": 2.441587901701323, + "grad_norm": 1.75, + "learning_rate": 8.965062705829295e-06, + "loss": 0.7603, + "num_tokens": 17396401.0, + "step": 808 + }, + { + "epoch": 2.4446124763705104, + "grad_norm": 1.296875, + "learning_rate": 8.96225111060831e-06, + "loss": 0.8506, + "num_tokens": 17429578.0, + "step": 809 + }, + { + "epoch": 2.4476370510396976, + "grad_norm": 1.453125, + "learning_rate": 8.959436199130102e-06, + "loss": 0.7296, + "num_tokens": 17449484.0, + "step": 810 + }, + { + "epoch": 2.450661625708885, + "grad_norm": 1.75, + "learning_rate": 8.956617974090879e-06, + "loss": 0.7459, + "num_tokens": 17461209.0, + "step": 811 + }, + { + "epoch": 2.4536862003780717, + "grad_norm": 2.234375, + "learning_rate": 8.953796438190038e-06, + "loss": 0.7686, + "num_tokens": 17468051.0, + "step": 812 + }, + { + "epoch": 2.456710775047259, + "grad_norm": 1.6328125, + "learning_rate": 8.950971594130131e-06, + "loss": 0.7106, + "num_tokens": 17481294.0, + "step": 813 + }, + { + "epoch": 2.4597353497164463, + "grad_norm": 1.4609375, + "learning_rate": 8.948143444616889e-06, + "loss": 0.8622, + "num_tokens": 17520917.0, + "step": 814 + }, + { + "epoch": 2.462759924385633, + "grad_norm": 1.6875, + "learning_rate": 8.945311992359206e-06, + "loss": 0.7075, + "num_tokens": 17540647.0, + "step": 815 + }, + { + "epoch": 2.4657844990548203, + "grad_norm": 1.2109375, + "learning_rate": 8.942477240069137e-06, + "loss": 0.8601, + "num_tokens": 17578186.0, + "step": 816 + }, + { + "epoch": 2.4688090737240076, + "grad_norm": 1.9921875, + "learning_rate": 8.939639190461904e-06, + "loss": 0.5788, + "num_tokens": 17586453.0, + "step": 817 + }, + { + "epoch": 2.471833648393195, + "grad_norm": 1.703125, + "learning_rate": 8.936797846255882e-06, + "loss": 0.6872, + "num_tokens": 17598347.0, + "step": 818 + }, + { + "epoch": 2.4748582230623817, + "grad_norm": 1.7734375, + "learning_rate": 8.933953210172604e-06, + "loss": 0.6241, + "num_tokens": 17606722.0, + "step": 819 + }, + { + "epoch": 2.477882797731569, + "grad_norm": 1.984375, + "learning_rate": 8.931105284936752e-06, + "loss": 0.5709, + "num_tokens": 17614213.0, + "step": 820 + }, + { + "epoch": 2.480907372400756, + "grad_norm": 2.46875, + "learning_rate": 8.928254073276166e-06, + "loss": 0.6838, + "num_tokens": 17623056.0, + "step": 821 + }, + { + "epoch": 2.4839319470699435, + "grad_norm": 1.1328125, + "learning_rate": 8.925399577921827e-06, + "loss": 0.8416, + "num_tokens": 17670528.0, + "step": 822 + }, + { + "epoch": 2.4869565217391303, + "grad_norm": 1.3984375, + "learning_rate": 8.922541801607866e-06, + "loss": 0.7973, + "num_tokens": 17696898.0, + "step": 823 + }, + { + "epoch": 2.4899810964083176, + "grad_norm": 1.7578125, + "learning_rate": 8.919680747071554e-06, + "loss": 0.8836, + "num_tokens": 17720108.0, + "step": 824 + }, + { + "epoch": 2.493005671077505, + "grad_norm": 1.796875, + "learning_rate": 8.916816417053302e-06, + "loss": 0.5703, + "num_tokens": 17727662.0, + "step": 825 + }, + { + "epoch": 2.4960302457466916, + "grad_norm": 1.8125, + "learning_rate": 8.913948814296658e-06, + "loss": 0.6288, + "num_tokens": 17736130.0, + "step": 826 + }, + { + "epoch": 2.499054820415879, + "grad_norm": 1.5859375, + "learning_rate": 8.911077941548306e-06, + "loss": 0.8714, + "num_tokens": 17756943.0, + "step": 827 + }, + { + "epoch": 2.502079395085066, + "grad_norm": 1.6640625, + "learning_rate": 8.908203801558065e-06, + "loss": 0.9315, + "num_tokens": 17779900.0, + "step": 828 + }, + { + "epoch": 2.5051039697542534, + "grad_norm": 1.921875, + "learning_rate": 8.905326397078873e-06, + "loss": 0.6679, + "num_tokens": 17790468.0, + "step": 829 + }, + { + "epoch": 2.5081285444234407, + "grad_norm": 1.515625, + "learning_rate": 8.90244573086681e-06, + "loss": 0.8741, + "num_tokens": 17813433.0, + "step": 830 + }, + { + "epoch": 2.5111531190926275, + "grad_norm": 1.171875, + "learning_rate": 8.899561805681067e-06, + "loss": 0.8645, + "num_tokens": 17860899.0, + "step": 831 + }, + { + "epoch": 2.5141776937618148, + "grad_norm": 1.9609375, + "learning_rate": 8.896674624283965e-06, + "loss": 0.6793, + "num_tokens": 17873252.0, + "step": 832 + }, + { + "epoch": 2.517202268431002, + "grad_norm": 2.0625, + "learning_rate": 8.893784189440937e-06, + "loss": 0.6255, + "num_tokens": 17880494.0, + "step": 833 + }, + { + "epoch": 2.520226843100189, + "grad_norm": 1.9921875, + "learning_rate": 8.890890503920538e-06, + "loss": 0.6319, + "num_tokens": 17890417.0, + "step": 834 + }, + { + "epoch": 2.523251417769376, + "grad_norm": 1.8203125, + "learning_rate": 8.887993570494436e-06, + "loss": 0.6292, + "num_tokens": 17898557.0, + "step": 835 + }, + { + "epoch": 2.5262759924385634, + "grad_norm": 1.2578125, + "learning_rate": 8.885093391937406e-06, + "loss": 0.8854, + "num_tokens": 17932460.0, + "step": 836 + }, + { + "epoch": 2.52930056710775, + "grad_norm": 1.0234375, + "learning_rate": 8.882189971027335e-06, + "loss": 0.887, + "num_tokens": 17989325.0, + "step": 837 + }, + { + "epoch": 2.5323251417769375, + "grad_norm": 1.2265625, + "learning_rate": 8.879283310545214e-06, + "loss": 0.8395, + "num_tokens": 18025112.0, + "step": 838 + }, + { + "epoch": 2.5353497164461247, + "grad_norm": 1.96875, + "learning_rate": 8.876373413275139e-06, + "loss": 0.5875, + "num_tokens": 18032483.0, + "step": 839 + }, + { + "epoch": 2.538374291115312, + "grad_norm": 1.796875, + "learning_rate": 8.873460282004302e-06, + "loss": 0.782, + "num_tokens": 18046987.0, + "step": 840 + }, + { + "epoch": 2.5413988657844993, + "grad_norm": 1.6171875, + "learning_rate": 8.870543919522996e-06, + "loss": 0.9036, + "num_tokens": 18076468.0, + "step": 841 + }, + { + "epoch": 2.544423440453686, + "grad_norm": 1.9609375, + "learning_rate": 8.86762432862461e-06, + "loss": 0.5755, + "num_tokens": 18084100.0, + "step": 842 + }, + { + "epoch": 2.5474480151228733, + "grad_norm": 1.765625, + "learning_rate": 8.86470151210562e-06, + "loss": 0.7118, + "num_tokens": 18098291.0, + "step": 843 + }, + { + "epoch": 2.5504725897920606, + "grad_norm": 1.109375, + "learning_rate": 8.861775472765599e-06, + "loss": 0.9195, + "num_tokens": 18154750.0, + "step": 844 + }, + { + "epoch": 2.5534971644612474, + "grad_norm": 2.21875, + "learning_rate": 8.858846213407201e-06, + "loss": 0.7337, + "num_tokens": 18165394.0, + "step": 845 + }, + { + "epoch": 2.5565217391304347, + "grad_norm": 1.9296875, + "learning_rate": 8.855913736836168e-06, + "loss": 0.5722, + "num_tokens": 18172282.0, + "step": 846 + }, + { + "epoch": 2.559546313799622, + "grad_norm": 1.4921875, + "learning_rate": 8.85297804586132e-06, + "loss": 0.8082, + "num_tokens": 18205176.0, + "step": 847 + }, + { + "epoch": 2.562570888468809, + "grad_norm": 1.78125, + "learning_rate": 8.85003914329456e-06, + "loss": 0.6086, + "num_tokens": 18213867.0, + "step": 848 + }, + { + "epoch": 2.5655954631379965, + "grad_norm": 1.9140625, + "learning_rate": 8.84709703195086e-06, + "loss": 0.6079, + "num_tokens": 18221822.0, + "step": 849 + }, + { + "epoch": 2.5686200378071833, + "grad_norm": 1.015625, + "learning_rate": 8.844151714648274e-06, + "loss": 0.8661, + "num_tokens": 18289615.0, + "step": 850 + }, + { + "epoch": 2.5716446124763706, + "grad_norm": 1.4375, + "learning_rate": 8.841203194207925e-06, + "loss": 0.7843, + "num_tokens": 18314269.0, + "step": 851 + }, + { + "epoch": 2.574669187145558, + "grad_norm": 1.8984375, + "learning_rate": 8.838251473453998e-06, + "loss": 0.6222, + "num_tokens": 18322565.0, + "step": 852 + }, + { + "epoch": 2.5776937618147446, + "grad_norm": 1.984375, + "learning_rate": 8.835296555213753e-06, + "loss": 0.7054, + "num_tokens": 18330632.0, + "step": 853 + }, + { + "epoch": 2.580718336483932, + "grad_norm": 1.4140625, + "learning_rate": 8.832338442317499e-06, + "loss": 0.8176, + "num_tokens": 18356947.0, + "step": 854 + }, + { + "epoch": 2.583742911153119, + "grad_norm": 2.078125, + "learning_rate": 8.829377137598621e-06, + "loss": 0.6131, + "num_tokens": 18364073.0, + "step": 855 + }, + { + "epoch": 2.586767485822306, + "grad_norm": 1.7578125, + "learning_rate": 8.82641264389355e-06, + "loss": 0.942, + "num_tokens": 18386924.0, + "step": 856 + }, + { + "epoch": 2.5897920604914932, + "grad_norm": 1.9453125, + "learning_rate": 8.823444964041777e-06, + "loss": 0.697, + "num_tokens": 18394583.0, + "step": 857 + }, + { + "epoch": 2.5928166351606805, + "grad_norm": 1.171875, + "learning_rate": 8.820474100885839e-06, + "loss": 0.8945, + "num_tokens": 18447045.0, + "step": 858 + }, + { + "epoch": 2.5958412098298678, + "grad_norm": 2.03125, + "learning_rate": 8.817500057271328e-06, + "loss": 0.6542, + "num_tokens": 18454404.0, + "step": 859 + }, + { + "epoch": 2.598865784499055, + "grad_norm": 1.875, + "learning_rate": 8.81452283604688e-06, + "loss": 0.6266, + "num_tokens": 18462503.0, + "step": 860 + }, + { + "epoch": 2.601890359168242, + "grad_norm": 1.265625, + "learning_rate": 8.811542440064174e-06, + "loss": 0.9295, + "num_tokens": 18504359.0, + "step": 861 + }, + { + "epoch": 2.604914933837429, + "grad_norm": 1.1875, + "learning_rate": 8.80855887217793e-06, + "loss": 0.8943, + "num_tokens": 18553563.0, + "step": 862 + }, + { + "epoch": 2.6079395085066164, + "grad_norm": 1.6171875, + "learning_rate": 8.805572135245911e-06, + "loss": 0.8894, + "num_tokens": 18576585.0, + "step": 863 + }, + { + "epoch": 2.610964083175803, + "grad_norm": 1.6328125, + "learning_rate": 8.802582232128908e-06, + "loss": 0.7094, + "num_tokens": 18592317.0, + "step": 864 + }, + { + "epoch": 2.6139886578449905, + "grad_norm": 1.890625, + "learning_rate": 8.799589165690749e-06, + "loss": 0.6036, + "num_tokens": 18600298.0, + "step": 865 + }, + { + "epoch": 2.6170132325141777, + "grad_norm": 2.09375, + "learning_rate": 8.796592938798287e-06, + "loss": 0.6303, + "num_tokens": 18607678.0, + "step": 866 + }, + { + "epoch": 2.620037807183365, + "grad_norm": 1.765625, + "learning_rate": 8.793593554321413e-06, + "loss": 0.7312, + "num_tokens": 18622374.0, + "step": 867 + }, + { + "epoch": 2.6230623818525522, + "grad_norm": 1.515625, + "learning_rate": 8.790591015133029e-06, + "loss": 0.8304, + "num_tokens": 18645433.0, + "step": 868 + }, + { + "epoch": 2.626086956521739, + "grad_norm": 1.5859375, + "learning_rate": 8.787585324109067e-06, + "loss": 0.7725, + "num_tokens": 18661745.0, + "step": 869 + }, + { + "epoch": 2.6291115311909263, + "grad_norm": 1.8515625, + "learning_rate": 8.784576484128477e-06, + "loss": 0.607, + "num_tokens": 18669439.0, + "step": 870 + }, + { + "epoch": 2.6321361058601136, + "grad_norm": 1.578125, + "learning_rate": 8.781564498073226e-06, + "loss": 0.8146, + "num_tokens": 18692554.0, + "step": 871 + }, + { + "epoch": 2.6351606805293004, + "grad_norm": 2.078125, + "learning_rate": 8.77854936882829e-06, + "loss": 0.618, + "num_tokens": 18699334.0, + "step": 872 + }, + { + "epoch": 2.6381852551984877, + "grad_norm": 1.984375, + "learning_rate": 8.775531099281657e-06, + "loss": 0.6613, + "num_tokens": 18706608.0, + "step": 873 + }, + { + "epoch": 2.641209829867675, + "grad_norm": 1.703125, + "learning_rate": 8.772509692324325e-06, + "loss": 0.7398, + "num_tokens": 18720515.0, + "step": 874 + }, + { + "epoch": 2.6442344045368618, + "grad_norm": 1.78125, + "learning_rate": 8.7694851508503e-06, + "loss": 0.6747, + "num_tokens": 18729298.0, + "step": 875 + }, + { + "epoch": 2.647258979206049, + "grad_norm": 1.8203125, + "learning_rate": 8.766457477756583e-06, + "loss": 0.5998, + "num_tokens": 18736652.0, + "step": 876 + }, + { + "epoch": 2.6502835538752363, + "grad_norm": 1.90625, + "learning_rate": 8.76342667594318e-06, + "loss": 0.6172, + "num_tokens": 18744613.0, + "step": 877 + }, + { + "epoch": 2.6533081285444235, + "grad_norm": 1.3984375, + "learning_rate": 8.760392748313092e-06, + "loss": 0.7778, + "num_tokens": 18767603.0, + "step": 878 + }, + { + "epoch": 2.656332703213611, + "grad_norm": 1.5859375, + "learning_rate": 8.75735569777231e-06, + "loss": 0.8384, + "num_tokens": 18793281.0, + "step": 879 + }, + { + "epoch": 2.6593572778827976, + "grad_norm": 1.4765625, + "learning_rate": 8.754315527229825e-06, + "loss": 0.8452, + "num_tokens": 18817806.0, + "step": 880 + }, + { + "epoch": 2.662381852551985, + "grad_norm": 1.2265625, + "learning_rate": 8.751272239597612e-06, + "loss": 0.82, + "num_tokens": 18853945.0, + "step": 881 + }, + { + "epoch": 2.665406427221172, + "grad_norm": 1.125, + "learning_rate": 8.748225837790626e-06, + "loss": 0.8467, + "num_tokens": 18900282.0, + "step": 882 + }, + { + "epoch": 2.668431001890359, + "grad_norm": 0.984375, + "learning_rate": 8.745176324726816e-06, + "loss": 0.8729, + "num_tokens": 18962836.0, + "step": 883 + }, + { + "epoch": 2.6714555765595462, + "grad_norm": 1.765625, + "learning_rate": 8.7421237033271e-06, + "loss": 0.7722, + "num_tokens": 18978959.0, + "step": 884 + }, + { + "epoch": 2.6744801512287335, + "grad_norm": 1.484375, + "learning_rate": 8.739067976515385e-06, + "loss": 0.8165, + "num_tokens": 19002388.0, + "step": 885 + }, + { + "epoch": 2.6775047258979208, + "grad_norm": 2.03125, + "learning_rate": 8.73600914721854e-06, + "loss": 0.6615, + "num_tokens": 19010260.0, + "step": 886 + }, + { + "epoch": 2.6805293005671076, + "grad_norm": 1.5859375, + "learning_rate": 8.732947218366414e-06, + "loss": 0.875, + "num_tokens": 19033780.0, + "step": 887 + }, + { + "epoch": 2.683553875236295, + "grad_norm": 1.2109375, + "learning_rate": 8.729882192891823e-06, + "loss": 0.8508, + "num_tokens": 19073153.0, + "step": 888 + }, + { + "epoch": 2.686578449905482, + "grad_norm": 1.8046875, + "learning_rate": 8.726814073730548e-06, + "loss": 0.6463, + "num_tokens": 19080909.0, + "step": 889 + }, + { + "epoch": 2.6896030245746694, + "grad_norm": 1.109375, + "learning_rate": 8.723742863821332e-06, + "loss": 0.8852, + "num_tokens": 19123685.0, + "step": 890 + }, + { + "epoch": 2.692627599243856, + "grad_norm": 1.1015625, + "learning_rate": 8.720668566105883e-06, + "loss": 0.8624, + "num_tokens": 19174155.0, + "step": 891 + }, + { + "epoch": 2.6956521739130435, + "grad_norm": 1.5234375, + "learning_rate": 8.717591183528865e-06, + "loss": 0.8824, + "num_tokens": 19197805.0, + "step": 892 + }, + { + "epoch": 2.6986767485822307, + "grad_norm": 1.140625, + "learning_rate": 8.71451071903789e-06, + "loss": 0.8442, + "num_tokens": 19241596.0, + "step": 893 + }, + { + "epoch": 2.7017013232514175, + "grad_norm": 1.9921875, + "learning_rate": 8.711427175583531e-06, + "loss": 0.6795, + "num_tokens": 19253945.0, + "step": 894 + }, + { + "epoch": 2.704725897920605, + "grad_norm": 1.703125, + "learning_rate": 8.708340556119307e-06, + "loss": 0.9465, + "num_tokens": 19276847.0, + "step": 895 + }, + { + "epoch": 2.707750472589792, + "grad_norm": 1.53125, + "learning_rate": 8.705250863601684e-06, + "loss": 0.818, + "num_tokens": 19298213.0, + "step": 896 + }, + { + "epoch": 2.7107750472589793, + "grad_norm": 1.421875, + "learning_rate": 8.702158100990066e-06, + "loss": 0.7499, + "num_tokens": 19318729.0, + "step": 897 + }, + { + "epoch": 2.7137996219281666, + "grad_norm": 1.71875, + "learning_rate": 8.699062271246806e-06, + "loss": 0.83, + "num_tokens": 19342218.0, + "step": 898 + }, + { + "epoch": 2.7168241965973534, + "grad_norm": 1.1796875, + "learning_rate": 8.695963377337191e-06, + "loss": 0.8496, + "num_tokens": 19381359.0, + "step": 899 + }, + { + "epoch": 2.7198487712665407, + "grad_norm": 1.4375, + "learning_rate": 8.69286142222944e-06, + "loss": 0.7658, + "num_tokens": 19404261.0, + "step": 900 + }, + { + "epoch": 2.722873345935728, + "grad_norm": 1.5078125, + "learning_rate": 8.689756408894709e-06, + "loss": 0.8525, + "num_tokens": 19427386.0, + "step": 901 + }, + { + "epoch": 2.7258979206049148, + "grad_norm": 2.21875, + "learning_rate": 8.686648340307081e-06, + "loss": 0.8624, + "num_tokens": 19451218.0, + "step": 902 + }, + { + "epoch": 2.728922495274102, + "grad_norm": 2.109375, + "learning_rate": 8.683537219443565e-06, + "loss": 0.6239, + "num_tokens": 19457974.0, + "step": 903 + }, + { + "epoch": 2.7319470699432893, + "grad_norm": 1.4375, + "learning_rate": 8.680423049284096e-06, + "loss": 0.7467, + "num_tokens": 19481187.0, + "step": 904 + }, + { + "epoch": 2.734971644612476, + "grad_norm": 1.4140625, + "learning_rate": 8.677305832811524e-06, + "loss": 0.8669, + "num_tokens": 19509835.0, + "step": 905 + }, + { + "epoch": 2.7379962192816634, + "grad_norm": 1.4765625, + "learning_rate": 8.674185573011625e-06, + "loss": 0.7749, + "num_tokens": 19534317.0, + "step": 906 + }, + { + "epoch": 2.7410207939508506, + "grad_norm": 1.4375, + "learning_rate": 8.671062272873084e-06, + "loss": 0.7467, + "num_tokens": 19553434.0, + "step": 907 + }, + { + "epoch": 2.744045368620038, + "grad_norm": 1.234375, + "learning_rate": 8.667935935387497e-06, + "loss": 0.9858, + "num_tokens": 19607751.0, + "step": 908 + }, + { + "epoch": 2.747069943289225, + "grad_norm": 1.125, + "learning_rate": 8.664806563549379e-06, + "loss": 0.9213, + "num_tokens": 19657734.0, + "step": 909 + }, + { + "epoch": 2.750094517958412, + "grad_norm": 1.2265625, + "learning_rate": 8.66167416035614e-06, + "loss": 0.8379, + "num_tokens": 19699514.0, + "step": 910 + }, + { + "epoch": 2.7531190926275992, + "grad_norm": 1.8984375, + "learning_rate": 8.658538728808097e-06, + "loss": 0.626, + "num_tokens": 19707300.0, + "step": 911 + }, + { + "epoch": 2.7561436672967865, + "grad_norm": 1.6484375, + "learning_rate": 8.655400271908472e-06, + "loss": 0.8955, + "num_tokens": 19729651.0, + "step": 912 + }, + { + "epoch": 2.7591682419659733, + "grad_norm": 1.953125, + "learning_rate": 8.65225879266338e-06, + "loss": 0.7085, + "num_tokens": 19741403.0, + "step": 913 + }, + { + "epoch": 2.7621928166351606, + "grad_norm": 1.46875, + "learning_rate": 8.649114294081833e-06, + "loss": 0.81, + "num_tokens": 19762339.0, + "step": 914 + }, + { + "epoch": 2.765217391304348, + "grad_norm": 1.09375, + "learning_rate": 8.645966779175734e-06, + "loss": 0.9062, + "num_tokens": 19820316.0, + "step": 915 + }, + { + "epoch": 2.768241965973535, + "grad_norm": 1.25, + "learning_rate": 8.642816250959874e-06, + "loss": 0.8476, + "num_tokens": 19851731.0, + "step": 916 + }, + { + "epoch": 2.7712665406427224, + "grad_norm": 1.9140625, + "learning_rate": 8.639662712451935e-06, + "loss": 0.6836, + "num_tokens": 19866116.0, + "step": 917 + }, + { + "epoch": 2.774291115311909, + "grad_norm": 1.9375, + "learning_rate": 8.636506166672478e-06, + "loss": 0.6042, + "num_tokens": 19873786.0, + "step": 918 + }, + { + "epoch": 2.7773156899810965, + "grad_norm": 1.28125, + "learning_rate": 8.633346616644944e-06, + "loss": 0.8475, + "num_tokens": 19906619.0, + "step": 919 + }, + { + "epoch": 2.7803402646502837, + "grad_norm": 1.2890625, + "learning_rate": 8.630184065395655e-06, + "loss": 0.8938, + "num_tokens": 19946795.0, + "step": 920 + }, + { + "epoch": 2.7833648393194705, + "grad_norm": 2.09375, + "learning_rate": 8.627018515953807e-06, + "loss": 0.6462, + "num_tokens": 19953749.0, + "step": 921 + }, + { + "epoch": 2.786389413988658, + "grad_norm": 1.5859375, + "learning_rate": 8.623849971351464e-06, + "loss": 0.8361, + "num_tokens": 19978367.0, + "step": 922 + }, + { + "epoch": 2.789413988657845, + "grad_norm": 1.9375, + "learning_rate": 8.620678434623563e-06, + "loss": 0.5886, + "num_tokens": 19985957.0, + "step": 923 + }, + { + "epoch": 2.792438563327032, + "grad_norm": 2.0, + "learning_rate": 8.617503908807905e-06, + "loss": 0.6739, + "num_tokens": 19993847.0, + "step": 924 + }, + { + "epoch": 2.795463137996219, + "grad_norm": 1.4296875, + "learning_rate": 8.614326396945154e-06, + "loss": 0.8264, + "num_tokens": 20023993.0, + "step": 925 + }, + { + "epoch": 2.7984877126654064, + "grad_norm": 1.4375, + "learning_rate": 8.611145902078838e-06, + "loss": 0.8949, + "num_tokens": 20050438.0, + "step": 926 + }, + { + "epoch": 2.8015122873345937, + "grad_norm": 1.5859375, + "learning_rate": 8.607962427255334e-06, + "loss": 0.8123, + "num_tokens": 20073525.0, + "step": 927 + }, + { + "epoch": 2.804536862003781, + "grad_norm": 1.1953125, + "learning_rate": 8.60477597552388e-06, + "loss": 0.9069, + "num_tokens": 20121335.0, + "step": 928 + }, + { + "epoch": 2.8075614366729678, + "grad_norm": 1.46875, + "learning_rate": 8.601586549936567e-06, + "loss": 0.815, + "num_tokens": 20142456.0, + "step": 929 + }, + { + "epoch": 2.810586011342155, + "grad_norm": 1.265625, + "learning_rate": 8.598394153548328e-06, + "loss": 0.8769, + "num_tokens": 20181987.0, + "step": 930 + }, + { + "epoch": 2.8136105860113423, + "grad_norm": 1.6484375, + "learning_rate": 8.595198789416943e-06, + "loss": 0.6946, + "num_tokens": 20199701.0, + "step": 931 + }, + { + "epoch": 2.816635160680529, + "grad_norm": 1.1171875, + "learning_rate": 8.592000460603039e-06, + "loss": 0.852, + "num_tokens": 20254561.0, + "step": 932 + }, + { + "epoch": 2.8196597353497164, + "grad_norm": 1.5078125, + "learning_rate": 8.588799170170077e-06, + "loss": 0.8157, + "num_tokens": 20278796.0, + "step": 933 + }, + { + "epoch": 2.8226843100189036, + "grad_norm": 1.7109375, + "learning_rate": 8.58559492118436e-06, + "loss": 0.7763, + "num_tokens": 20294707.0, + "step": 934 + }, + { + "epoch": 2.825708884688091, + "grad_norm": 1.5859375, + "learning_rate": 8.582387716715021e-06, + "loss": 0.6387, + "num_tokens": 20308652.0, + "step": 935 + }, + { + "epoch": 2.828733459357278, + "grad_norm": 1.609375, + "learning_rate": 8.579177559834027e-06, + "loss": 0.778, + "num_tokens": 20332673.0, + "step": 936 + }, + { + "epoch": 2.831758034026465, + "grad_norm": 1.6796875, + "learning_rate": 8.57596445361617e-06, + "loss": 0.7394, + "num_tokens": 20345851.0, + "step": 937 + }, + { + "epoch": 2.8347826086956522, + "grad_norm": 1.6328125, + "learning_rate": 8.572748401139066e-06, + "loss": 0.7595, + "num_tokens": 20368398.0, + "step": 938 + }, + { + "epoch": 2.8378071833648395, + "grad_norm": 1.7109375, + "learning_rate": 8.56952940548316e-06, + "loss": 0.8938, + "num_tokens": 20391023.0, + "step": 939 + }, + { + "epoch": 2.8408317580340263, + "grad_norm": 1.5625, + "learning_rate": 8.566307469731706e-06, + "loss": 0.8255, + "num_tokens": 20415290.0, + "step": 940 + }, + { + "epoch": 2.8438563327032136, + "grad_norm": 1.9921875, + "learning_rate": 8.563082596970785e-06, + "loss": 0.5908, + "num_tokens": 20422787.0, + "step": 941 + }, + { + "epoch": 2.846880907372401, + "grad_norm": 2.046875, + "learning_rate": 8.559854790289278e-06, + "loss": 0.8192, + "num_tokens": 20435713.0, + "step": 942 + }, + { + "epoch": 2.8499054820415877, + "grad_norm": 2.09375, + "learning_rate": 8.55662405277889e-06, + "loss": 0.637, + "num_tokens": 20442845.0, + "step": 943 + }, + { + "epoch": 2.852930056710775, + "grad_norm": 1.53125, + "learning_rate": 8.553390387534126e-06, + "loss": 0.7975, + "num_tokens": 20467633.0, + "step": 944 + }, + { + "epoch": 2.855954631379962, + "grad_norm": 1.171875, + "learning_rate": 8.550153797652292e-06, + "loss": 0.8423, + "num_tokens": 20507501.0, + "step": 945 + }, + { + "epoch": 2.8589792060491495, + "grad_norm": 1.4921875, + "learning_rate": 8.546914286233505e-06, + "loss": 0.7567, + "num_tokens": 20530471.0, + "step": 946 + }, + { + "epoch": 2.8620037807183367, + "grad_norm": 1.5546875, + "learning_rate": 8.543671856380672e-06, + "loss": 0.8628, + "num_tokens": 20554311.0, + "step": 947 + }, + { + "epoch": 2.8650283553875235, + "grad_norm": 1.2109375, + "learning_rate": 8.540426511199498e-06, + "loss": 0.9008, + "num_tokens": 20605810.0, + "step": 948 + }, + { + "epoch": 2.868052930056711, + "grad_norm": 2.609375, + "learning_rate": 8.537178253798483e-06, + "loss": 0.6708, + "num_tokens": 20611481.0, + "step": 949 + }, + { + "epoch": 2.871077504725898, + "grad_norm": 1.0078125, + "learning_rate": 8.53392708728891e-06, + "loss": 0.8133, + "num_tokens": 20663060.0, + "step": 950 + }, + { + "epoch": 2.874102079395085, + "grad_norm": 1.7734375, + "learning_rate": 8.530673014784857e-06, + "loss": 0.6787, + "num_tokens": 20671982.0, + "step": 951 + }, + { + "epoch": 2.877126654064272, + "grad_norm": 1.703125, + "learning_rate": 8.527416039403178e-06, + "loss": 0.741, + "num_tokens": 20686117.0, + "step": 952 + }, + { + "epoch": 2.8801512287334594, + "grad_norm": 1.2890625, + "learning_rate": 8.524156164263509e-06, + "loss": 0.8334, + "num_tokens": 20722668.0, + "step": 953 + }, + { + "epoch": 2.8831758034026462, + "grad_norm": 1.171875, + "learning_rate": 8.52089339248827e-06, + "loss": 0.8377, + "num_tokens": 20774132.0, + "step": 954 + }, + { + "epoch": 2.8862003780718335, + "grad_norm": 1.296875, + "learning_rate": 8.517627727202647e-06, + "loss": 0.9398, + "num_tokens": 20815432.0, + "step": 955 + }, + { + "epoch": 2.8892249527410208, + "grad_norm": 1.2421875, + "learning_rate": 8.514359171534603e-06, + "loss": 0.9225, + "num_tokens": 20862599.0, + "step": 956 + }, + { + "epoch": 2.892249527410208, + "grad_norm": 2.03125, + "learning_rate": 8.511087728614863e-06, + "loss": 0.6942, + "num_tokens": 20870312.0, + "step": 957 + }, + { + "epoch": 2.8952741020793953, + "grad_norm": 1.28125, + "learning_rate": 8.507813401576925e-06, + "loss": 0.8676, + "num_tokens": 20904893.0, + "step": 958 + }, + { + "epoch": 2.898298676748582, + "grad_norm": 1.703125, + "learning_rate": 8.504536193557049e-06, + "loss": 0.7192, + "num_tokens": 20918946.0, + "step": 959 + }, + { + "epoch": 2.9013232514177694, + "grad_norm": 2.03125, + "learning_rate": 8.501256107694246e-06, + "loss": 0.5749, + "num_tokens": 20925668.0, + "step": 960 + }, + { + "epoch": 2.9043478260869566, + "grad_norm": 1.9921875, + "learning_rate": 8.497973147130296e-06, + "loss": 0.6326, + "num_tokens": 20933191.0, + "step": 961 + }, + { + "epoch": 2.9073724007561434, + "grad_norm": 2.015625, + "learning_rate": 8.494687315009718e-06, + "loss": 0.7004, + "num_tokens": 20940503.0, + "step": 962 + }, + { + "epoch": 2.9103969754253307, + "grad_norm": 1.5234375, + "learning_rate": 8.491398614479795e-06, + "loss": 0.8186, + "num_tokens": 20963535.0, + "step": 963 + }, + { + "epoch": 2.913421550094518, + "grad_norm": 1.4140625, + "learning_rate": 8.488107048690554e-06, + "loss": 0.8802, + "num_tokens": 20994600.0, + "step": 964 + }, + { + "epoch": 2.9164461247637052, + "grad_norm": 1.90625, + "learning_rate": 8.484812620794757e-06, + "loss": 0.6752, + "num_tokens": 21003254.0, + "step": 965 + }, + { + "epoch": 2.9194706994328925, + "grad_norm": 1.5078125, + "learning_rate": 8.481515333947921e-06, + "loss": 0.8777, + "num_tokens": 21027011.0, + "step": 966 + }, + { + "epoch": 2.9224952741020793, + "grad_norm": 1.3984375, + "learning_rate": 8.478215191308289e-06, + "loss": 0.8479, + "num_tokens": 21057840.0, + "step": 967 + }, + { + "epoch": 2.9255198487712666, + "grad_norm": 1.96875, + "learning_rate": 8.474912196036852e-06, + "loss": 0.6616, + "num_tokens": 21065606.0, + "step": 968 + }, + { + "epoch": 2.928544423440454, + "grad_norm": 2.09375, + "learning_rate": 8.471606351297323e-06, + "loss": 0.6193, + "num_tokens": 21072439.0, + "step": 969 + }, + { + "epoch": 2.9315689981096407, + "grad_norm": 1.6796875, + "learning_rate": 8.468297660256148e-06, + "loss": 0.6083, + "num_tokens": 21083802.0, + "step": 970 + }, + { + "epoch": 2.934593572778828, + "grad_norm": 1.640625, + "learning_rate": 8.4649861260825e-06, + "loss": 0.8795, + "num_tokens": 21107780.0, + "step": 971 + }, + { + "epoch": 2.937618147448015, + "grad_norm": 1.5234375, + "learning_rate": 8.461671751948277e-06, + "loss": 0.8797, + "num_tokens": 21141253.0, + "step": 972 + }, + { + "epoch": 2.940642722117202, + "grad_norm": 1.515625, + "learning_rate": 8.458354541028091e-06, + "loss": 0.8017, + "num_tokens": 21165438.0, + "step": 973 + }, + { + "epoch": 2.9436672967863893, + "grad_norm": 1.1484375, + "learning_rate": 8.455034496499276e-06, + "loss": 0.8821, + "num_tokens": 21219262.0, + "step": 974 + }, + { + "epoch": 2.9466918714555765, + "grad_norm": 1.8125, + "learning_rate": 8.451711621541881e-06, + "loss": 0.5991, + "num_tokens": 21226860.0, + "step": 975 + }, + { + "epoch": 2.949716446124764, + "grad_norm": 1.890625, + "learning_rate": 8.448385919338664e-06, + "loss": 0.7797, + "num_tokens": 21239867.0, + "step": 976 + }, + { + "epoch": 2.952741020793951, + "grad_norm": 1.5703125, + "learning_rate": 8.445057393075088e-06, + "loss": 0.8903, + "num_tokens": 21263274.0, + "step": 977 + }, + { + "epoch": 2.955765595463138, + "grad_norm": 1.890625, + "learning_rate": 8.441726045939327e-06, + "loss": 0.6035, + "num_tokens": 21271205.0, + "step": 978 + }, + { + "epoch": 2.958790170132325, + "grad_norm": 1.7890625, + "learning_rate": 8.438391881122255e-06, + "loss": 0.635, + "num_tokens": 21279376.0, + "step": 979 + }, + { + "epoch": 2.9618147448015124, + "grad_norm": 1.9296875, + "learning_rate": 8.435054901817438e-06, + "loss": 0.5676, + "num_tokens": 21286669.0, + "step": 980 + }, + { + "epoch": 2.9648393194706992, + "grad_norm": 1.8984375, + "learning_rate": 8.431715111221152e-06, + "loss": 0.5781, + "num_tokens": 21294080.0, + "step": 981 + }, + { + "epoch": 2.9678638941398865, + "grad_norm": 2.140625, + "learning_rate": 8.428372512532354e-06, + "loss": 0.5619, + "num_tokens": 21301797.0, + "step": 982 + }, + { + "epoch": 2.9708884688090738, + "grad_norm": 1.9921875, + "learning_rate": 8.425027108952693e-06, + "loss": 0.6354, + "num_tokens": 21308827.0, + "step": 983 + }, + { + "epoch": 2.973913043478261, + "grad_norm": 1.6875, + "learning_rate": 8.421678903686508e-06, + "loss": 0.865, + "num_tokens": 21332626.0, + "step": 984 + }, + { + "epoch": 2.9769376181474483, + "grad_norm": 1.2578125, + "learning_rate": 8.418327899940816e-06, + "loss": 0.7954, + "num_tokens": 21368850.0, + "step": 985 + }, + { + "epoch": 2.979962192816635, + "grad_norm": 1.6484375, + "learning_rate": 8.414974100925321e-06, + "loss": 0.7877, + "num_tokens": 21389237.0, + "step": 986 + }, + { + "epoch": 2.9829867674858224, + "grad_norm": 1.21875, + "learning_rate": 8.411617509852401e-06, + "loss": 0.858, + "num_tokens": 21428447.0, + "step": 987 + }, + { + "epoch": 2.9860113421550096, + "grad_norm": 1.171875, + "learning_rate": 8.408258129937108e-06, + "loss": 0.8524, + "num_tokens": 21481148.0, + "step": 988 + }, + { + "epoch": 2.9890359168241964, + "grad_norm": 1.7890625, + "learning_rate": 8.404895964397166e-06, + "loss": 0.8918, + "num_tokens": 21504264.0, + "step": 989 + }, + { + "epoch": 2.9920604914933837, + "grad_norm": 1.765625, + "learning_rate": 8.401531016452967e-06, + "loss": 0.6336, + "num_tokens": 21512638.0, + "step": 990 + }, + { + "epoch": 2.995085066162571, + "grad_norm": 1.8359375, + "learning_rate": 8.398163289327568e-06, + "loss": 0.8587, + "num_tokens": 21535693.0, + "step": 991 + }, + { + "epoch": 2.998109640831758, + "grad_norm": 1.2109375, + "learning_rate": 8.394792786246686e-06, + "loss": 0.8569, + "num_tokens": 21574768.0, + "step": 992 + }, + { + "epoch": 3.0, + "grad_norm": 1.9765625, + "learning_rate": 8.391419510438703e-06, + "loss": 0.7115, + "num_tokens": 21585939.0, + "step": 993 + }, + { + "epoch": 3.0030245746691873, + "grad_norm": 1.296875, + "learning_rate": 8.388043465134649e-06, + "loss": 0.814, + "num_tokens": 21619763.0, + "step": 994 + }, + { + "epoch": 3.006049149338374, + "grad_norm": 1.7578125, + "learning_rate": 8.384664653568213e-06, + "loss": 0.7716, + "num_tokens": 21635612.0, + "step": 995 + }, + { + "epoch": 3.0090737240075613, + "grad_norm": 0.98828125, + "learning_rate": 8.38128307897573e-06, + "loss": 0.8898, + "num_tokens": 21692955.0, + "step": 996 + }, + { + "epoch": 3.0120982986767486, + "grad_norm": 1.859375, + "learning_rate": 8.377898744596185e-06, + "loss": 0.6568, + "num_tokens": 21700946.0, + "step": 997 + }, + { + "epoch": 3.015122873345936, + "grad_norm": 1.1953125, + "learning_rate": 8.3745116536712e-06, + "loss": 0.8047, + "num_tokens": 21739792.0, + "step": 998 + }, + { + "epoch": 3.0181474480151227, + "grad_norm": 1.9375, + "learning_rate": 8.371121809445043e-06, + "loss": 0.6095, + "num_tokens": 21747333.0, + "step": 999 + }, + { + "epoch": 3.02117202268431, + "grad_norm": 2.046875, + "learning_rate": 8.367729215164618e-06, + "loss": 0.6778, + "num_tokens": 21759022.0, + "step": 1000 + }, + { + "epoch": 3.024196597353497, + "grad_norm": 1.8046875, + "learning_rate": 8.364333874079462e-06, + "loss": 0.701, + "num_tokens": 21772453.0, + "step": 1001 + }, + { + "epoch": 3.0272211720226845, + "grad_norm": 1.328125, + "learning_rate": 8.360935789441746e-06, + "loss": 0.7577, + "num_tokens": 21803250.0, + "step": 1002 + }, + { + "epoch": 3.0302457466918713, + "grad_norm": 1.609375, + "learning_rate": 8.357534964506262e-06, + "loss": 0.834, + "num_tokens": 21827088.0, + "step": 1003 + }, + { + "epoch": 3.0332703213610586, + "grad_norm": 1.9375, + "learning_rate": 8.354131402530436e-06, + "loss": 0.6907, + "num_tokens": 21836409.0, + "step": 1004 + }, + { + "epoch": 3.036294896030246, + "grad_norm": 1.0546875, + "learning_rate": 8.350725106774308e-06, + "loss": 0.7807, + "num_tokens": 21886503.0, + "step": 1005 + }, + { + "epoch": 3.039319470699433, + "grad_norm": 1.75, + "learning_rate": 8.34731608050054e-06, + "loss": 0.871, + "num_tokens": 21909384.0, + "step": 1006 + }, + { + "epoch": 3.04234404536862, + "grad_norm": 2.046875, + "learning_rate": 8.343904326974409e-06, + "loss": 0.5204, + "num_tokens": 21915769.0, + "step": 1007 + }, + { + "epoch": 3.045368620037807, + "grad_norm": 1.578125, + "learning_rate": 8.340489849463804e-06, + "loss": 0.8012, + "num_tokens": 21937876.0, + "step": 1008 + }, + { + "epoch": 3.0483931947069944, + "grad_norm": 1.4921875, + "learning_rate": 8.33707265123922e-06, + "loss": 0.8004, + "num_tokens": 21958986.0, + "step": 1009 + }, + { + "epoch": 3.0514177693761813, + "grad_norm": 1.3828125, + "learning_rate": 8.333652735573767e-06, + "loss": 0.9114, + "num_tokens": 21989248.0, + "step": 1010 + }, + { + "epoch": 3.0544423440453685, + "grad_norm": 1.4296875, + "learning_rate": 8.330230105743147e-06, + "loss": 0.7355, + "num_tokens": 22011895.0, + "step": 1011 + }, + { + "epoch": 3.057466918714556, + "grad_norm": 1.7421875, + "learning_rate": 8.32680476502567e-06, + "loss": 0.749, + "num_tokens": 22024238.0, + "step": 1012 + }, + { + "epoch": 3.060491493383743, + "grad_norm": 1.28125, + "learning_rate": 8.323376716702236e-06, + "loss": 0.8301, + "num_tokens": 22066756.0, + "step": 1013 + }, + { + "epoch": 3.06351606805293, + "grad_norm": 2.078125, + "learning_rate": 8.319945964056346e-06, + "loss": 0.7064, + "num_tokens": 22074759.0, + "step": 1014 + }, + { + "epoch": 3.066540642722117, + "grad_norm": 1.625, + "learning_rate": 8.316512510374083e-06, + "loss": 0.8789, + "num_tokens": 22098563.0, + "step": 1015 + }, + { + "epoch": 3.0695652173913044, + "grad_norm": 1.484375, + "learning_rate": 8.313076358944122e-06, + "loss": 0.7521, + "num_tokens": 22120252.0, + "step": 1016 + }, + { + "epoch": 3.0725897920604917, + "grad_norm": 1.671875, + "learning_rate": 8.309637513057723e-06, + "loss": 0.7726, + "num_tokens": 22138301.0, + "step": 1017 + }, + { + "epoch": 3.0756143667296785, + "grad_norm": 1.3125, + "learning_rate": 8.306195976008723e-06, + "loss": 0.9114, + "num_tokens": 22179277.0, + "step": 1018 + }, + { + "epoch": 3.0786389413988657, + "grad_norm": 1.8828125, + "learning_rate": 8.302751751093539e-06, + "loss": 0.6234, + "num_tokens": 22189300.0, + "step": 1019 + }, + { + "epoch": 3.081663516068053, + "grad_norm": 1.234375, + "learning_rate": 8.299304841611162e-06, + "loss": 0.8349, + "num_tokens": 22225578.0, + "step": 1020 + }, + { + "epoch": 3.0846880907372403, + "grad_norm": 1.5546875, + "learning_rate": 8.295855250863156e-06, + "loss": 0.7223, + "num_tokens": 22243033.0, + "step": 1021 + }, + { + "epoch": 3.087712665406427, + "grad_norm": 1.953125, + "learning_rate": 8.29240298215365e-06, + "loss": 0.551, + "num_tokens": 22250457.0, + "step": 1022 + }, + { + "epoch": 3.0907372400756143, + "grad_norm": 1.921875, + "learning_rate": 8.288948038789341e-06, + "loss": 0.6245, + "num_tokens": 22258284.0, + "step": 1023 + }, + { + "epoch": 3.0937618147448016, + "grad_norm": 1.390625, + "learning_rate": 8.285490424079486e-06, + "loss": 0.9023, + "num_tokens": 22297355.0, + "step": 1024 + }, + { + "epoch": 3.096786389413989, + "grad_norm": 1.2578125, + "learning_rate": 8.282030141335899e-06, + "loss": 0.8662, + "num_tokens": 22342487.0, + "step": 1025 + }, + { + "epoch": 3.0998109640831757, + "grad_norm": 1.15625, + "learning_rate": 8.278567193872955e-06, + "loss": 0.8287, + "num_tokens": 22377443.0, + "step": 1026 + }, + { + "epoch": 3.102835538752363, + "grad_norm": 1.578125, + "learning_rate": 8.275101585007578e-06, + "loss": 0.8369, + "num_tokens": 22401080.0, + "step": 1027 + }, + { + "epoch": 3.10586011342155, + "grad_norm": 1.7734375, + "learning_rate": 8.271633318059241e-06, + "loss": 0.595, + "num_tokens": 22409111.0, + "step": 1028 + }, + { + "epoch": 3.108884688090737, + "grad_norm": 1.4609375, + "learning_rate": 8.268162396349963e-06, + "loss": 0.8882, + "num_tokens": 22433664.0, + "step": 1029 + }, + { + "epoch": 3.1119092627599243, + "grad_norm": 2.1875, + "learning_rate": 8.264688823204307e-06, + "loss": 0.6208, + "num_tokens": 22440627.0, + "step": 1030 + }, + { + "epoch": 3.1149338374291116, + "grad_norm": 1.203125, + "learning_rate": 8.261212601949374e-06, + "loss": 0.8799, + "num_tokens": 22479801.0, + "step": 1031 + }, + { + "epoch": 3.117958412098299, + "grad_norm": 1.75, + "learning_rate": 8.2577337359148e-06, + "loss": 0.8083, + "num_tokens": 22502946.0, + "step": 1032 + }, + { + "epoch": 3.1209829867674856, + "grad_norm": 1.515625, + "learning_rate": 8.254252228432762e-06, + "loss": 0.6805, + "num_tokens": 22522049.0, + "step": 1033 + }, + { + "epoch": 3.124007561436673, + "grad_norm": 1.90625, + "learning_rate": 8.250768082837957e-06, + "loss": 0.5852, + "num_tokens": 22529228.0, + "step": 1034 + }, + { + "epoch": 3.12703213610586, + "grad_norm": 1.9609375, + "learning_rate": 8.247281302467616e-06, + "loss": 0.6133, + "num_tokens": 22536244.0, + "step": 1035 + }, + { + "epoch": 3.1300567107750474, + "grad_norm": 2.046875, + "learning_rate": 8.24379189066149e-06, + "loss": 0.6613, + "num_tokens": 22543701.0, + "step": 1036 + }, + { + "epoch": 3.1330812854442343, + "grad_norm": 1.4296875, + "learning_rate": 8.240299850761851e-06, + "loss": 0.8474, + "num_tokens": 22570284.0, + "step": 1037 + }, + { + "epoch": 3.1361058601134215, + "grad_norm": 1.484375, + "learning_rate": 8.236805186113492e-06, + "loss": 0.8006, + "num_tokens": 22593078.0, + "step": 1038 + }, + { + "epoch": 3.139130434782609, + "grad_norm": 1.5703125, + "learning_rate": 8.233307900063714e-06, + "loss": 0.7791, + "num_tokens": 22615612.0, + "step": 1039 + }, + { + "epoch": 3.142155009451796, + "grad_norm": 1.5, + "learning_rate": 8.229807995962332e-06, + "loss": 0.7893, + "num_tokens": 22638666.0, + "step": 1040 + }, + { + "epoch": 3.145179584120983, + "grad_norm": 1.71875, + "learning_rate": 8.226305477161669e-06, + "loss": 0.7411, + "num_tokens": 22657159.0, + "step": 1041 + }, + { + "epoch": 3.14820415879017, + "grad_norm": 1.2734375, + "learning_rate": 8.222800347016553e-06, + "loss": 0.883, + "num_tokens": 22691572.0, + "step": 1042 + }, + { + "epoch": 3.1512287334593574, + "grad_norm": 1.4921875, + "learning_rate": 8.219292608884309e-06, + "loss": 0.8291, + "num_tokens": 22713368.0, + "step": 1043 + }, + { + "epoch": 3.1542533081285447, + "grad_norm": 1.5859375, + "learning_rate": 8.215782266124766e-06, + "loss": 0.6966, + "num_tokens": 22736685.0, + "step": 1044 + }, + { + "epoch": 3.1572778827977315, + "grad_norm": 1.4453125, + "learning_rate": 8.212269322100244e-06, + "loss": 0.8869, + "num_tokens": 22770044.0, + "step": 1045 + }, + { + "epoch": 3.1603024574669187, + "grad_norm": 2.09375, + "learning_rate": 8.208753780175557e-06, + "loss": 0.6648, + "num_tokens": 22777493.0, + "step": 1046 + }, + { + "epoch": 3.163327032136106, + "grad_norm": 1.171875, + "learning_rate": 8.205235643718004e-06, + "loss": 0.8793, + "num_tokens": 22820722.0, + "step": 1047 + }, + { + "epoch": 3.166351606805293, + "grad_norm": 1.2109375, + "learning_rate": 8.201714916097372e-06, + "loss": 0.7419, + "num_tokens": 22852179.0, + "step": 1048 + }, + { + "epoch": 3.16937618147448, + "grad_norm": 1.3046875, + "learning_rate": 8.198191600685931e-06, + "loss": 0.8352, + "num_tokens": 22889174.0, + "step": 1049 + }, + { + "epoch": 3.1724007561436673, + "grad_norm": 1.2734375, + "learning_rate": 8.194665700858422e-06, + "loss": 0.8318, + "num_tokens": 22920413.0, + "step": 1050 + }, + { + "epoch": 3.1754253308128546, + "grad_norm": 2.125, + "learning_rate": 8.191137219992073e-06, + "loss": 0.5999, + "num_tokens": 22928113.0, + "step": 1051 + }, + { + "epoch": 3.1784499054820414, + "grad_norm": 1.96875, + "learning_rate": 8.187606161466575e-06, + "loss": 0.5718, + "num_tokens": 22935404.0, + "step": 1052 + }, + { + "epoch": 3.1814744801512287, + "grad_norm": 1.828125, + "learning_rate": 8.184072528664091e-06, + "loss": 0.6311, + "num_tokens": 22944552.0, + "step": 1053 + }, + { + "epoch": 3.184499054820416, + "grad_norm": 1.53125, + "learning_rate": 8.180536324969251e-06, + "loss": 0.9149, + "num_tokens": 22967978.0, + "step": 1054 + }, + { + "epoch": 3.187523629489603, + "grad_norm": 2.359375, + "learning_rate": 8.176997553769146e-06, + "loss": 0.9797, + "num_tokens": 22990847.0, + "step": 1055 + }, + { + "epoch": 3.19054820415879, + "grad_norm": 1.9921875, + "learning_rate": 8.17345621845333e-06, + "loss": 0.6058, + "num_tokens": 22998061.0, + "step": 1056 + }, + { + "epoch": 3.1935727788279773, + "grad_norm": 1.3046875, + "learning_rate": 8.169912322413803e-06, + "loss": 0.7437, + "num_tokens": 23028355.0, + "step": 1057 + }, + { + "epoch": 3.1965973534971646, + "grad_norm": 1.8203125, + "learning_rate": 8.166365869045029e-06, + "loss": 0.5505, + "num_tokens": 23037908.0, + "step": 1058 + }, + { + "epoch": 3.1996219281663514, + "grad_norm": 1.9609375, + "learning_rate": 8.162816861743917e-06, + "loss": 0.6, + "num_tokens": 23045270.0, + "step": 1059 + }, + { + "epoch": 3.2026465028355386, + "grad_norm": 1.9453125, + "learning_rate": 8.159265303909818e-06, + "loss": 0.597, + "num_tokens": 23052419.0, + "step": 1060 + }, + { + "epoch": 3.205671077504726, + "grad_norm": 1.265625, + "learning_rate": 8.155711198944536e-06, + "loss": 0.912, + "num_tokens": 23089514.0, + "step": 1061 + }, + { + "epoch": 3.208695652173913, + "grad_norm": 1.8046875, + "learning_rate": 8.152154550252305e-06, + "loss": 0.5763, + "num_tokens": 23097360.0, + "step": 1062 + }, + { + "epoch": 3.2117202268431, + "grad_norm": 1.546875, + "learning_rate": 8.1485953612398e-06, + "loss": 0.7743, + "num_tokens": 23114730.0, + "step": 1063 + }, + { + "epoch": 3.2147448015122873, + "grad_norm": 1.59375, + "learning_rate": 8.14503363531613e-06, + "loss": 0.8313, + "num_tokens": 23137036.0, + "step": 1064 + }, + { + "epoch": 3.2177693761814745, + "grad_norm": 1.5390625, + "learning_rate": 8.14146937589283e-06, + "loss": 0.7861, + "num_tokens": 23161413.0, + "step": 1065 + }, + { + "epoch": 3.220793950850662, + "grad_norm": 1.125, + "learning_rate": 8.137902586383866e-06, + "loss": 0.8154, + "num_tokens": 23204513.0, + "step": 1066 + }, + { + "epoch": 3.2238185255198486, + "grad_norm": 1.8359375, + "learning_rate": 8.134333270205624e-06, + "loss": 0.695, + "num_tokens": 23216758.0, + "step": 1067 + }, + { + "epoch": 3.226843100189036, + "grad_norm": 1.578125, + "learning_rate": 8.130761430776915e-06, + "loss": 0.9625, + "num_tokens": 23239596.0, + "step": 1068 + }, + { + "epoch": 3.229867674858223, + "grad_norm": 2.046875, + "learning_rate": 8.12718707151896e-06, + "loss": 0.6027, + "num_tokens": 23246156.0, + "step": 1069 + }, + { + "epoch": 3.2328922495274104, + "grad_norm": 1.859375, + "learning_rate": 8.123610195855395e-06, + "loss": 0.8401, + "num_tokens": 23263930.0, + "step": 1070 + }, + { + "epoch": 3.235916824196597, + "grad_norm": 0.98828125, + "learning_rate": 8.120030807212276e-06, + "loss": 0.8825, + "num_tokens": 23328489.0, + "step": 1071 + }, + { + "epoch": 3.2389413988657845, + "grad_norm": 1.6796875, + "learning_rate": 8.116448909018051e-06, + "loss": 0.7173, + "num_tokens": 23341902.0, + "step": 1072 + }, + { + "epoch": 3.2419659735349717, + "grad_norm": 1.6015625, + "learning_rate": 8.112864504703582e-06, + "loss": 0.7741, + "num_tokens": 23358728.0, + "step": 1073 + }, + { + "epoch": 3.244990548204159, + "grad_norm": 1.8359375, + "learning_rate": 8.10927759770213e-06, + "loss": 0.601, + "num_tokens": 23366960.0, + "step": 1074 + }, + { + "epoch": 3.248015122873346, + "grad_norm": 1.2265625, + "learning_rate": 8.10568819144935e-06, + "loss": 0.825, + "num_tokens": 23407308.0, + "step": 1075 + }, + { + "epoch": 3.251039697542533, + "grad_norm": 1.828125, + "learning_rate": 8.102096289383293e-06, + "loss": 0.655, + "num_tokens": 23417993.0, + "step": 1076 + }, + { + "epoch": 3.2540642722117203, + "grad_norm": 1.3984375, + "learning_rate": 8.0985018949444e-06, + "loss": 0.7812, + "num_tokens": 23443237.0, + "step": 1077 + }, + { + "epoch": 3.257088846880907, + "grad_norm": 1.4765625, + "learning_rate": 8.094905011575496e-06, + "loss": 0.7638, + "num_tokens": 23468012.0, + "step": 1078 + }, + { + "epoch": 3.2601134215500944, + "grad_norm": 1.8125, + "learning_rate": 8.0913056427218e-06, + "loss": 0.5275, + "num_tokens": 23475406.0, + "step": 1079 + }, + { + "epoch": 3.2631379962192817, + "grad_norm": 1.8359375, + "learning_rate": 8.087703791830902e-06, + "loss": 0.5864, + "num_tokens": 23482863.0, + "step": 1080 + }, + { + "epoch": 3.266162570888469, + "grad_norm": 1.5625, + "learning_rate": 8.084099462352772e-06, + "loss": 0.8628, + "num_tokens": 23505059.0, + "step": 1081 + }, + { + "epoch": 3.2691871455576558, + "grad_norm": 1.734375, + "learning_rate": 8.080492657739754e-06, + "loss": 0.7549, + "num_tokens": 23519928.0, + "step": 1082 + }, + { + "epoch": 3.272211720226843, + "grad_norm": 2.0625, + "learning_rate": 8.076883381446564e-06, + "loss": 0.6021, + "num_tokens": 23526694.0, + "step": 1083 + }, + { + "epoch": 3.2752362948960303, + "grad_norm": 1.65625, + "learning_rate": 8.073271636930287e-06, + "loss": 0.8317, + "num_tokens": 23545442.0, + "step": 1084 + }, + { + "epoch": 3.2782608695652176, + "grad_norm": 1.671875, + "learning_rate": 8.069657427650364e-06, + "loss": 0.7595, + "num_tokens": 23563177.0, + "step": 1085 + }, + { + "epoch": 3.2812854442344044, + "grad_norm": 1.671875, + "learning_rate": 8.06604075706861e-06, + "loss": 0.8831, + "num_tokens": 23586412.0, + "step": 1086 + }, + { + "epoch": 3.2843100189035916, + "grad_norm": 2.0, + "learning_rate": 8.062421628649187e-06, + "loss": 0.6098, + "num_tokens": 23593342.0, + "step": 1087 + }, + { + "epoch": 3.287334593572779, + "grad_norm": 1.3828125, + "learning_rate": 8.058800045858613e-06, + "loss": 0.8559, + "num_tokens": 23622406.0, + "step": 1088 + }, + { + "epoch": 3.2903591682419657, + "grad_norm": 1.875, + "learning_rate": 8.055176012165758e-06, + "loss": 0.6131, + "num_tokens": 23630388.0, + "step": 1089 + }, + { + "epoch": 3.293383742911153, + "grad_norm": 1.3046875, + "learning_rate": 8.051549531041843e-06, + "loss": 0.8986, + "num_tokens": 23671878.0, + "step": 1090 + }, + { + "epoch": 3.2964083175803403, + "grad_norm": 1.9453125, + "learning_rate": 8.047920605960428e-06, + "loss": 0.5341, + "num_tokens": 23678333.0, + "step": 1091 + }, + { + "epoch": 3.2994328922495275, + "grad_norm": 1.109375, + "learning_rate": 8.044289240397415e-06, + "loss": 0.8641, + "num_tokens": 23723856.0, + "step": 1092 + }, + { + "epoch": 3.302457466918715, + "grad_norm": 1.5546875, + "learning_rate": 8.040655437831047e-06, + "loss": 0.8639, + "num_tokens": 23747387.0, + "step": 1093 + }, + { + "epoch": 3.3054820415879016, + "grad_norm": 1.328125, + "learning_rate": 8.037019201741901e-06, + "loss": 0.8903, + "num_tokens": 23791806.0, + "step": 1094 + }, + { + "epoch": 3.308506616257089, + "grad_norm": 1.3828125, + "learning_rate": 8.033380535612875e-06, + "loss": 0.8519, + "num_tokens": 23823097.0, + "step": 1095 + }, + { + "epoch": 3.311531190926276, + "grad_norm": 1.53125, + "learning_rate": 8.02973944292921e-06, + "loss": 0.6401, + "num_tokens": 23836252.0, + "step": 1096 + }, + { + "epoch": 3.314555765595463, + "grad_norm": 1.828125, + "learning_rate": 8.026095927178458e-06, + "loss": 0.5972, + "num_tokens": 23843529.0, + "step": 1097 + }, + { + "epoch": 3.31758034026465, + "grad_norm": 1.7578125, + "learning_rate": 8.022449991850501e-06, + "loss": 0.7433, + "num_tokens": 23857262.0, + "step": 1098 + }, + { + "epoch": 3.3206049149338375, + "grad_norm": 1.421875, + "learning_rate": 8.018801640437535e-06, + "loss": 0.8108, + "num_tokens": 23880329.0, + "step": 1099 + }, + { + "epoch": 3.3236294896030247, + "grad_norm": 1.3515625, + "learning_rate": 8.015150876434066e-06, + "loss": 0.8907, + "num_tokens": 23918721.0, + "step": 1100 + }, + { + "epoch": 3.3266540642722116, + "grad_norm": 1.7578125, + "learning_rate": 8.011497703336921e-06, + "loss": 0.6789, + "num_tokens": 23930206.0, + "step": 1101 + }, + { + "epoch": 3.329678638941399, + "grad_norm": 1.8671875, + "learning_rate": 8.007842124645225e-06, + "loss": 0.618, + "num_tokens": 23939068.0, + "step": 1102 + }, + { + "epoch": 3.332703213610586, + "grad_norm": 1.46875, + "learning_rate": 8.004184143860408e-06, + "loss": 0.8127, + "num_tokens": 23959040.0, + "step": 1103 + }, + { + "epoch": 3.3357277882797733, + "grad_norm": 1.1171875, + "learning_rate": 8.00052376448621e-06, + "loss": 0.8127, + "num_tokens": 23999436.0, + "step": 1104 + }, + { + "epoch": 3.33875236294896, + "grad_norm": 1.71875, + "learning_rate": 7.996860990028658e-06, + "loss": 0.8037, + "num_tokens": 24022637.0, + "step": 1105 + }, + { + "epoch": 3.3417769376181474, + "grad_norm": 1.765625, + "learning_rate": 7.993195823996077e-06, + "loss": 0.8107, + "num_tokens": 24041468.0, + "step": 1106 + }, + { + "epoch": 3.3448015122873347, + "grad_norm": 1.453125, + "learning_rate": 7.989528269899085e-06, + "loss": 0.7298, + "num_tokens": 24061639.0, + "step": 1107 + }, + { + "epoch": 3.3478260869565215, + "grad_norm": 1.96875, + "learning_rate": 7.98585833125058e-06, + "loss": 0.5972, + "num_tokens": 24069076.0, + "step": 1108 + }, + { + "epoch": 3.3508506616257088, + "grad_norm": 1.4140625, + "learning_rate": 7.982186011565755e-06, + "loss": 0.8806, + "num_tokens": 24093459.0, + "step": 1109 + }, + { + "epoch": 3.353875236294896, + "grad_norm": 1.609375, + "learning_rate": 7.978511314362077e-06, + "loss": 0.8864, + "num_tokens": 24117450.0, + "step": 1110 + }, + { + "epoch": 3.3568998109640833, + "grad_norm": 1.421875, + "learning_rate": 7.974834243159288e-06, + "loss": 0.7627, + "num_tokens": 24142559.0, + "step": 1111 + }, + { + "epoch": 3.3599243856332706, + "grad_norm": 1.34375, + "learning_rate": 7.971154801479409e-06, + "loss": 0.836, + "num_tokens": 24175121.0, + "step": 1112 + }, + { + "epoch": 3.3629489603024574, + "grad_norm": 1.546875, + "learning_rate": 7.96747299284673e-06, + "loss": 0.8202, + "num_tokens": 24198712.0, + "step": 1113 + }, + { + "epoch": 3.3659735349716446, + "grad_norm": 1.4765625, + "learning_rate": 7.963788820787805e-06, + "loss": 0.8036, + "num_tokens": 24222741.0, + "step": 1114 + }, + { + "epoch": 3.368998109640832, + "grad_norm": 1.6953125, + "learning_rate": 7.960102288831454e-06, + "loss": 0.615, + "num_tokens": 24233406.0, + "step": 1115 + }, + { + "epoch": 3.3720226843100187, + "grad_norm": 1.265625, + "learning_rate": 7.956413400508763e-06, + "loss": 0.8271, + "num_tokens": 24266756.0, + "step": 1116 + }, + { + "epoch": 3.375047258979206, + "grad_norm": 1.75, + "learning_rate": 7.952722159353065e-06, + "loss": 0.7714, + "num_tokens": 24283246.0, + "step": 1117 + }, + { + "epoch": 3.3780718336483933, + "grad_norm": 1.9921875, + "learning_rate": 7.949028568899951e-06, + "loss": 0.635, + "num_tokens": 24290621.0, + "step": 1118 + }, + { + "epoch": 3.3810964083175805, + "grad_norm": 1.796875, + "learning_rate": 7.945332632687268e-06, + "loss": 0.607, + "num_tokens": 24298955.0, + "step": 1119 + }, + { + "epoch": 3.3841209829867673, + "grad_norm": 1.9921875, + "learning_rate": 7.941634354255098e-06, + "loss": 0.6304, + "num_tokens": 24306317.0, + "step": 1120 + }, + { + "epoch": 3.3871455576559546, + "grad_norm": 2.03125, + "learning_rate": 7.937933737145777e-06, + "loss": 0.6373, + "num_tokens": 24314107.0, + "step": 1121 + }, + { + "epoch": 3.390170132325142, + "grad_norm": 1.6015625, + "learning_rate": 7.934230784903875e-06, + "loss": 0.7044, + "num_tokens": 24329078.0, + "step": 1122 + }, + { + "epoch": 3.393194706994329, + "grad_norm": 1.125, + "learning_rate": 7.9305255010762e-06, + "loss": 0.8215, + "num_tokens": 24372739.0, + "step": 1123 + }, + { + "epoch": 3.396219281663516, + "grad_norm": 1.8515625, + "learning_rate": 7.926817889211798e-06, + "loss": 0.5573, + "num_tokens": 24379977.0, + "step": 1124 + }, + { + "epoch": 3.399243856332703, + "grad_norm": 2.0625, + "learning_rate": 7.923107952861938e-06, + "loss": 0.5946, + "num_tokens": 24386500.0, + "step": 1125 + }, + { + "epoch": 3.4022684310018905, + "grad_norm": 1.1328125, + "learning_rate": 7.919395695580114e-06, + "loss": 0.8417, + "num_tokens": 24433603.0, + "step": 1126 + }, + { + "epoch": 3.4052930056710773, + "grad_norm": 1.6796875, + "learning_rate": 7.915681120922055e-06, + "loss": 0.8496, + "num_tokens": 24454174.0, + "step": 1127 + }, + { + "epoch": 3.4083175803402646, + "grad_norm": 1.9609375, + "learning_rate": 7.911964232445698e-06, + "loss": 0.5546, + "num_tokens": 24461311.0, + "step": 1128 + }, + { + "epoch": 3.411342155009452, + "grad_norm": 1.984375, + "learning_rate": 7.908245033711201e-06, + "loss": 0.6468, + "num_tokens": 24468935.0, + "step": 1129 + }, + { + "epoch": 3.414366729678639, + "grad_norm": 2.234375, + "learning_rate": 7.904523528280932e-06, + "loss": 0.6463, + "num_tokens": 24476046.0, + "step": 1130 + }, + { + "epoch": 3.417391304347826, + "grad_norm": 1.203125, + "learning_rate": 7.900799719719473e-06, + "loss": 0.8307, + "num_tokens": 24515209.0, + "step": 1131 + }, + { + "epoch": 3.420415879017013, + "grad_norm": 1.6796875, + "learning_rate": 7.89707361159361e-06, + "loss": 0.7641, + "num_tokens": 24532561.0, + "step": 1132 + }, + { + "epoch": 3.4234404536862004, + "grad_norm": 1.2421875, + "learning_rate": 7.893345207472329e-06, + "loss": 0.8257, + "num_tokens": 24570184.0, + "step": 1133 + }, + { + "epoch": 3.4264650283553877, + "grad_norm": 1.2578125, + "learning_rate": 7.88961451092682e-06, + "loss": 0.8736, + "num_tokens": 24609637.0, + "step": 1134 + }, + { + "epoch": 3.4294896030245745, + "grad_norm": 1.921875, + "learning_rate": 7.88588152553046e-06, + "loss": 0.6278, + "num_tokens": 24617368.0, + "step": 1135 + }, + { + "epoch": 3.4325141776937618, + "grad_norm": 1.484375, + "learning_rate": 7.882146254858832e-06, + "loss": 0.8413, + "num_tokens": 24639814.0, + "step": 1136 + }, + { + "epoch": 3.435538752362949, + "grad_norm": 1.5546875, + "learning_rate": 7.878408702489695e-06, + "loss": 0.8721, + "num_tokens": 24663449.0, + "step": 1137 + }, + { + "epoch": 3.4385633270321363, + "grad_norm": 1.375, + "learning_rate": 7.874668872003e-06, + "loss": 0.8309, + "num_tokens": 24690652.0, + "step": 1138 + }, + { + "epoch": 3.441587901701323, + "grad_norm": 2.0, + "learning_rate": 7.870926766980879e-06, + "loss": 0.6454, + "num_tokens": 24697799.0, + "step": 1139 + }, + { + "epoch": 3.4446124763705104, + "grad_norm": 1.8984375, + "learning_rate": 7.867182391007644e-06, + "loss": 0.6142, + "num_tokens": 24705205.0, + "step": 1140 + }, + { + "epoch": 3.4476370510396976, + "grad_norm": 1.3203125, + "learning_rate": 7.863435747669781e-06, + "loss": 0.8389, + "num_tokens": 24735542.0, + "step": 1141 + }, + { + "epoch": 3.450661625708885, + "grad_norm": 1.90625, + "learning_rate": 7.859686840555944e-06, + "loss": 0.5596, + "num_tokens": 24743615.0, + "step": 1142 + }, + { + "epoch": 3.4536862003780717, + "grad_norm": 1.9765625, + "learning_rate": 7.85593567325696e-06, + "loss": 0.6713, + "num_tokens": 24751430.0, + "step": 1143 + }, + { + "epoch": 3.456710775047259, + "grad_norm": 2.0, + "learning_rate": 7.85218224936582e-06, + "loss": 0.6523, + "num_tokens": 24759539.0, + "step": 1144 + }, + { + "epoch": 3.4597353497164463, + "grad_norm": 2.0625, + "learning_rate": 7.848426572477677e-06, + "loss": 0.8091, + "num_tokens": 24770573.0, + "step": 1145 + }, + { + "epoch": 3.462759924385633, + "grad_norm": 1.828125, + "learning_rate": 7.844668646189842e-06, + "loss": 0.6257, + "num_tokens": 24778721.0, + "step": 1146 + }, + { + "epoch": 3.4657844990548203, + "grad_norm": 1.828125, + "learning_rate": 7.840908474101774e-06, + "loss": 0.5983, + "num_tokens": 24786506.0, + "step": 1147 + }, + { + "epoch": 3.4688090737240076, + "grad_norm": 1.3671875, + "learning_rate": 7.837146059815096e-06, + "loss": 0.8854, + "num_tokens": 24818355.0, + "step": 1148 + }, + { + "epoch": 3.471833648393195, + "grad_norm": 1.2578125, + "learning_rate": 7.833381406933564e-06, + "loss": 0.8606, + "num_tokens": 24859215.0, + "step": 1149 + }, + { + "epoch": 3.4748582230623817, + "grad_norm": 1.40625, + "learning_rate": 7.829614519063093e-06, + "loss": 0.8235, + "num_tokens": 24883090.0, + "step": 1150 + }, + { + "epoch": 3.477882797731569, + "grad_norm": 1.2421875, + "learning_rate": 7.825845399811723e-06, + "loss": 0.8555, + "num_tokens": 24922945.0, + "step": 1151 + }, + { + "epoch": 3.480907372400756, + "grad_norm": 1.96875, + "learning_rate": 7.822074052789648e-06, + "loss": 0.5723, + "num_tokens": 24930693.0, + "step": 1152 + }, + { + "epoch": 3.4839319470699435, + "grad_norm": 2.078125, + "learning_rate": 7.818300481609184e-06, + "loss": 0.6104, + "num_tokens": 24937038.0, + "step": 1153 + }, + { + "epoch": 3.4869565217391303, + "grad_norm": 1.546875, + "learning_rate": 7.814524689884776e-06, + "loss": 0.8943, + "num_tokens": 24960296.0, + "step": 1154 + }, + { + "epoch": 3.4899810964083176, + "grad_norm": 1.1640625, + "learning_rate": 7.810746681233008e-06, + "loss": 0.7859, + "num_tokens": 24998175.0, + "step": 1155 + }, + { + "epoch": 3.493005671077505, + "grad_norm": 1.25, + "learning_rate": 7.806966459272577e-06, + "loss": 0.8781, + "num_tokens": 25039857.0, + "step": 1156 + }, + { + "epoch": 3.4960302457466916, + "grad_norm": 1.3203125, + "learning_rate": 7.8031840276243e-06, + "loss": 0.8377, + "num_tokens": 25069975.0, + "step": 1157 + }, + { + "epoch": 3.499054820415879, + "grad_norm": 2.03125, + "learning_rate": 7.799399389911122e-06, + "loss": 0.563, + "num_tokens": 25076909.0, + "step": 1158 + }, + { + "epoch": 3.502079395085066, + "grad_norm": 1.7734375, + "learning_rate": 7.795612549758083e-06, + "loss": 0.5304, + "num_tokens": 25084323.0, + "step": 1159 + }, + { + "epoch": 3.5051039697542534, + "grad_norm": 1.5859375, + "learning_rate": 7.791823510792348e-06, + "loss": 0.8594, + "num_tokens": 25107395.0, + "step": 1160 + }, + { + "epoch": 3.5081285444234407, + "grad_norm": 1.9765625, + "learning_rate": 7.788032276643181e-06, + "loss": 0.8501, + "num_tokens": 25120122.0, + "step": 1161 + }, + { + "epoch": 3.5111531190926275, + "grad_norm": 1.453125, + "learning_rate": 7.784238850941948e-06, + "loss": 0.797, + "num_tokens": 25144809.0, + "step": 1162 + }, + { + "epoch": 3.5141776937618148, + "grad_norm": 1.359375, + "learning_rate": 7.78044323732212e-06, + "loss": 0.7702, + "num_tokens": 25170554.0, + "step": 1163 + }, + { + "epoch": 3.517202268431002, + "grad_norm": 1.5625, + "learning_rate": 7.776645439419259e-06, + "loss": 0.7303, + "num_tokens": 25191939.0, + "step": 1164 + }, + { + "epoch": 3.520226843100189, + "grad_norm": 0.96875, + "learning_rate": 7.772845460871016e-06, + "loss": 0.8626, + "num_tokens": 25266048.0, + "step": 1165 + }, + { + "epoch": 3.523251417769376, + "grad_norm": 1.8359375, + "learning_rate": 7.769043305317138e-06, + "loss": 0.7305, + "num_tokens": 25278638.0, + "step": 1166 + }, + { + "epoch": 3.5262759924385634, + "grad_norm": 1.4375, + "learning_rate": 7.765238976399451e-06, + "loss": 0.836, + "num_tokens": 25303354.0, + "step": 1167 + }, + { + "epoch": 3.52930056710775, + "grad_norm": 1.7890625, + "learning_rate": 7.761432477761869e-06, + "loss": 0.645, + "num_tokens": 25314184.0, + "step": 1168 + }, + { + "epoch": 3.5323251417769375, + "grad_norm": 1.8203125, + "learning_rate": 7.75762381305038e-06, + "loss": 0.5799, + "num_tokens": 25322332.0, + "step": 1169 + }, + { + "epoch": 3.5353497164461247, + "grad_norm": 1.6640625, + "learning_rate": 7.753812985913043e-06, + "loss": 0.6954, + "num_tokens": 25333904.0, + "step": 1170 + }, + { + "epoch": 3.538374291115312, + "grad_norm": 1.4375, + "learning_rate": 7.75e-06, + "loss": 0.8582, + "num_tokens": 25364224.0, + "step": 1171 + }, + { + "epoch": 3.5413988657844993, + "grad_norm": 1.7421875, + "learning_rate": 7.74618485896345e-06, + "loss": 0.7431, + "num_tokens": 25378132.0, + "step": 1172 + }, + { + "epoch": 3.544423440453686, + "grad_norm": 1.234375, + "learning_rate": 7.742367566457656e-06, + "loss": 0.8881, + "num_tokens": 25423288.0, + "step": 1173 + }, + { + "epoch": 3.5474480151228733, + "grad_norm": 1.8828125, + "learning_rate": 7.738548126138954e-06, + "loss": 0.6124, + "num_tokens": 25430855.0, + "step": 1174 + }, + { + "epoch": 3.5504725897920606, + "grad_norm": 2.03125, + "learning_rate": 7.734726541665722e-06, + "loss": 0.5142, + "num_tokens": 25437420.0, + "step": 1175 + }, + { + "epoch": 3.5534971644612474, + "grad_norm": 1.4296875, + "learning_rate": 7.7309028166984e-06, + "loss": 0.8578, + "num_tokens": 25467089.0, + "step": 1176 + }, + { + "epoch": 3.5565217391304347, + "grad_norm": 1.9453125, + "learning_rate": 7.727076954899475e-06, + "loss": 0.5649, + "num_tokens": 25474018.0, + "step": 1177 + }, + { + "epoch": 3.559546313799622, + "grad_norm": 1.671875, + "learning_rate": 7.723248959933486e-06, + "loss": 0.7991, + "num_tokens": 25494907.0, + "step": 1178 + }, + { + "epoch": 3.562570888468809, + "grad_norm": 1.359375, + "learning_rate": 7.719418835467009e-06, + "loss": 0.882, + "num_tokens": 25535278.0, + "step": 1179 + }, + { + "epoch": 3.5655954631379965, + "grad_norm": 1.2109375, + "learning_rate": 7.715586585168666e-06, + "loss": 0.9151, + "num_tokens": 25576406.0, + "step": 1180 + }, + { + "epoch": 3.5686200378071833, + "grad_norm": 2.171875, + "learning_rate": 7.711752212709106e-06, + "loss": 0.5828, + "num_tokens": 25583495.0, + "step": 1181 + }, + { + "epoch": 3.5716446124763706, + "grad_norm": 1.78125, + "learning_rate": 7.70791572176102e-06, + "loss": 0.7003, + "num_tokens": 25594304.0, + "step": 1182 + }, + { + "epoch": 3.574669187145558, + "grad_norm": 1.90625, + "learning_rate": 7.704077115999124e-06, + "loss": 0.5444, + "num_tokens": 25601623.0, + "step": 1183 + }, + { + "epoch": 3.5776937618147446, + "grad_norm": 1.3046875, + "learning_rate": 7.700236399100159e-06, + "loss": 0.7954, + "num_tokens": 25631804.0, + "step": 1184 + }, + { + "epoch": 3.580718336483932, + "grad_norm": 1.1953125, + "learning_rate": 7.696393574742889e-06, + "loss": 0.9049, + "num_tokens": 25671616.0, + "step": 1185 + }, + { + "epoch": 3.583742911153119, + "grad_norm": 1.7578125, + "learning_rate": 7.692548646608099e-06, + "loss": 0.6248, + "num_tokens": 25680004.0, + "step": 1186 + }, + { + "epoch": 3.586767485822306, + "grad_norm": 1.265625, + "learning_rate": 7.688701618378583e-06, + "loss": 0.8469, + "num_tokens": 25718409.0, + "step": 1187 + }, + { + "epoch": 3.5897920604914932, + "grad_norm": 1.453125, + "learning_rate": 7.684852493739153e-06, + "loss": 0.8531, + "num_tokens": 25741917.0, + "step": 1188 + }, + { + "epoch": 3.5928166351606805, + "grad_norm": 1.75, + "learning_rate": 7.681001276376625e-06, + "loss": 0.7676, + "num_tokens": 25764984.0, + "step": 1189 + }, + { + "epoch": 3.5958412098298678, + "grad_norm": 1.8359375, + "learning_rate": 7.677147969979824e-06, + "loss": 0.5997, + "num_tokens": 25774251.0, + "step": 1190 + }, + { + "epoch": 3.598865784499055, + "grad_norm": 2.09375, + "learning_rate": 7.673292578239569e-06, + "loss": 0.6529, + "num_tokens": 25782276.0, + "step": 1191 + }, + { + "epoch": 3.601890359168242, + "grad_norm": 2.125, + "learning_rate": 7.66943510484868e-06, + "loss": 0.647, + "num_tokens": 25788817.0, + "step": 1192 + }, + { + "epoch": 3.604914933837429, + "grad_norm": 1.5390625, + "learning_rate": 7.665575553501973e-06, + "loss": 0.7828, + "num_tokens": 25809814.0, + "step": 1193 + }, + { + "epoch": 3.6079395085066164, + "grad_norm": 2.03125, + "learning_rate": 7.661713927896254e-06, + "loss": 0.6725, + "num_tokens": 25818158.0, + "step": 1194 + }, + { + "epoch": 3.610964083175803, + "grad_norm": 1.5078125, + "learning_rate": 7.65785023173031e-06, + "loss": 0.8691, + "num_tokens": 25842375.0, + "step": 1195 + }, + { + "epoch": 3.6139886578449905, + "grad_norm": 1.2109375, + "learning_rate": 7.653984468704918e-06, + "loss": 0.8951, + "num_tokens": 25889684.0, + "step": 1196 + }, + { + "epoch": 3.6170132325141777, + "grad_norm": 1.4453125, + "learning_rate": 7.650116642522831e-06, + "loss": 0.8376, + "num_tokens": 25914377.0, + "step": 1197 + }, + { + "epoch": 3.620037807183365, + "grad_norm": 1.765625, + "learning_rate": 7.646246756888781e-06, + "loss": 0.6281, + "num_tokens": 25922999.0, + "step": 1198 + }, + { + "epoch": 3.6230623818525522, + "grad_norm": 1.3984375, + "learning_rate": 7.64237481550947e-06, + "loss": 0.8277, + "num_tokens": 25947989.0, + "step": 1199 + }, + { + "epoch": 3.626086956521739, + "grad_norm": 1.46875, + "learning_rate": 7.638500822093566e-06, + "loss": 0.7573, + "num_tokens": 25969996.0, + "step": 1200 + }, + { + "epoch": 3.6291115311909263, + "grad_norm": 1.671875, + "learning_rate": 7.634624780351715e-06, + "loss": 0.5417, + "num_tokens": 25978622.0, + "step": 1201 + }, + { + "epoch": 3.6321361058601136, + "grad_norm": 1.1484375, + "learning_rate": 7.63074669399651e-06, + "loss": 0.8362, + "num_tokens": 26018012.0, + "step": 1202 + }, + { + "epoch": 3.6351606805293004, + "grad_norm": 1.5625, + "learning_rate": 7.626866566742508e-06, + "loss": 0.8551, + "num_tokens": 26040455.0, + "step": 1203 + }, + { + "epoch": 3.6381852551984877, + "grad_norm": 1.6484375, + "learning_rate": 7.622984402306226e-06, + "loss": 0.7315, + "num_tokens": 26062560.0, + "step": 1204 + }, + { + "epoch": 3.641209829867675, + "grad_norm": 1.515625, + "learning_rate": 7.619100204406127e-06, + "loss": 0.8505, + "num_tokens": 26085469.0, + "step": 1205 + }, + { + "epoch": 3.6442344045368618, + "grad_norm": 1.984375, + "learning_rate": 7.615213976762619e-06, + "loss": 0.6818, + "num_tokens": 26092572.0, + "step": 1206 + }, + { + "epoch": 3.647258979206049, + "grad_norm": 1.7890625, + "learning_rate": 7.6113257230980615e-06, + "loss": 0.5791, + "num_tokens": 26100673.0, + "step": 1207 + }, + { + "epoch": 3.6502835538752363, + "grad_norm": 1.8671875, + "learning_rate": 7.60743544713675e-06, + "loss": 0.6104, + "num_tokens": 26108276.0, + "step": 1208 + }, + { + "epoch": 3.6533081285444235, + "grad_norm": 1.8046875, + "learning_rate": 7.603543152604918e-06, + "loss": 0.5859, + "num_tokens": 26116016.0, + "step": 1209 + }, + { + "epoch": 3.656332703213611, + "grad_norm": 1.3125, + "learning_rate": 7.59964884323073e-06, + "loss": 0.807, + "num_tokens": 26150388.0, + "step": 1210 + }, + { + "epoch": 3.6593572778827976, + "grad_norm": 1.515625, + "learning_rate": 7.595752522744287e-06, + "loss": 0.8252, + "num_tokens": 26173960.0, + "step": 1211 + }, + { + "epoch": 3.662381852551985, + "grad_norm": 1.4453125, + "learning_rate": 7.591854194877609e-06, + "loss": 0.8289, + "num_tokens": 26199950.0, + "step": 1212 + }, + { + "epoch": 3.665406427221172, + "grad_norm": 1.7421875, + "learning_rate": 7.587953863364642e-06, + "loss": 0.8731, + "num_tokens": 26222894.0, + "step": 1213 + }, + { + "epoch": 3.668431001890359, + "grad_norm": 1.3671875, + "learning_rate": 7.584051531941252e-06, + "loss": 0.7974, + "num_tokens": 26252736.0, + "step": 1214 + }, + { + "epoch": 3.6714555765595462, + "grad_norm": 1.5546875, + "learning_rate": 7.5801472043452195e-06, + "loss": 0.7954, + "num_tokens": 26275143.0, + "step": 1215 + }, + { + "epoch": 3.6744801512287335, + "grad_norm": 1.265625, + "learning_rate": 7.5762408843162364e-06, + "loss": 0.8237, + "num_tokens": 26314770.0, + "step": 1216 + }, + { + "epoch": 3.6775047258979208, + "grad_norm": 2.15625, + "learning_rate": 7.572332575595904e-06, + "loss": 0.5858, + "num_tokens": 26322154.0, + "step": 1217 + }, + { + "epoch": 3.6805293005671076, + "grad_norm": 1.4921875, + "learning_rate": 7.5684222819277265e-06, + "loss": 0.8656, + "num_tokens": 26344857.0, + "step": 1218 + }, + { + "epoch": 3.683553875236295, + "grad_norm": 1.34375, + "learning_rate": 7.564510007057114e-06, + "loss": 0.8701, + "num_tokens": 26374171.0, + "step": 1219 + }, + { + "epoch": 3.686578449905482, + "grad_norm": 1.9921875, + "learning_rate": 7.5605957547313695e-06, + "loss": 0.7676, + "num_tokens": 26382298.0, + "step": 1220 + }, + { + "epoch": 3.6896030245746694, + "grad_norm": 1.953125, + "learning_rate": 7.556679528699692e-06, + "loss": 0.6162, + "num_tokens": 26389596.0, + "step": 1221 + }, + { + "epoch": 3.692627599243856, + "grad_norm": 1.2421875, + "learning_rate": 7.5527613327131685e-06, + "loss": 0.8183, + "num_tokens": 26424828.0, + "step": 1222 + }, + { + "epoch": 3.6956521739130435, + "grad_norm": 1.5703125, + "learning_rate": 7.548841170524779e-06, + "loss": 0.7174, + "num_tokens": 26440234.0, + "step": 1223 + }, + { + "epoch": 3.6986767485822307, + "grad_norm": 1.7734375, + "learning_rate": 7.544919045889382e-06, + "loss": 0.5905, + "num_tokens": 26448229.0, + "step": 1224 + }, + { + "epoch": 3.7017013232514175, + "grad_norm": 1.796875, + "learning_rate": 7.540994962563714e-06, + "loss": 0.6975, + "num_tokens": 26458881.0, + "step": 1225 + }, + { + "epoch": 3.704725897920605, + "grad_norm": 1.875, + "learning_rate": 7.537068924306392e-06, + "loss": 0.5906, + "num_tokens": 26466682.0, + "step": 1226 + }, + { + "epoch": 3.707750472589792, + "grad_norm": 1.4140625, + "learning_rate": 7.5331409348779035e-06, + "loss": 0.8332, + "num_tokens": 26496167.0, + "step": 1227 + }, + { + "epoch": 3.7107750472589793, + "grad_norm": 2.1875, + "learning_rate": 7.529210998040604e-06, + "loss": 0.6203, + "num_tokens": 26502062.0, + "step": 1228 + }, + { + "epoch": 3.7137996219281666, + "grad_norm": 1.3046875, + "learning_rate": 7.525279117558719e-06, + "loss": 0.893, + "num_tokens": 26544652.0, + "step": 1229 + }, + { + "epoch": 3.7168241965973534, + "grad_norm": 2.0, + "learning_rate": 7.521345297198327e-06, + "loss": 0.6169, + "num_tokens": 26552391.0, + "step": 1230 + }, + { + "epoch": 3.7198487712665407, + "grad_norm": 2.109375, + "learning_rate": 7.5174095407273765e-06, + "loss": 0.542, + "num_tokens": 26558757.0, + "step": 1231 + }, + { + "epoch": 3.722873345935728, + "grad_norm": 1.21875, + "learning_rate": 7.513471851915657e-06, + "loss": 0.8977, + "num_tokens": 26597511.0, + "step": 1232 + }, + { + "epoch": 3.7258979206049148, + "grad_norm": 1.6328125, + "learning_rate": 7.509532234534818e-06, + "loss": 0.855, + "num_tokens": 26617109.0, + "step": 1233 + }, + { + "epoch": 3.728922495274102, + "grad_norm": 1.6015625, + "learning_rate": 7.505590692358353e-06, + "loss": 0.8836, + "num_tokens": 26639835.0, + "step": 1234 + }, + { + "epoch": 3.7319470699432893, + "grad_norm": 1.59375, + "learning_rate": 7.501647229161599e-06, + "loss": 0.8001, + "num_tokens": 26658531.0, + "step": 1235 + }, + { + "epoch": 3.734971644612476, + "grad_norm": 1.21875, + "learning_rate": 7.497701848721738e-06, + "loss": 0.7716, + "num_tokens": 26695700.0, + "step": 1236 + }, + { + "epoch": 3.7379962192816634, + "grad_norm": 1.796875, + "learning_rate": 7.493754554817782e-06, + "loss": 0.6702, + "num_tokens": 26707340.0, + "step": 1237 + }, + { + "epoch": 3.7410207939508506, + "grad_norm": 1.59375, + "learning_rate": 7.489805351230575e-06, + "loss": 0.7139, + "num_tokens": 26727537.0, + "step": 1238 + }, + { + "epoch": 3.744045368620038, + "grad_norm": 1.4765625, + "learning_rate": 7.485854241742799e-06, + "loss": 0.7937, + "num_tokens": 26753727.0, + "step": 1239 + }, + { + "epoch": 3.747069943289225, + "grad_norm": 1.71875, + "learning_rate": 7.4819012301389514e-06, + "loss": 0.8487, + "num_tokens": 26777560.0, + "step": 1240 + }, + { + "epoch": 3.750094517958412, + "grad_norm": 1.7578125, + "learning_rate": 7.477946320205358e-06, + "loss": 0.6975, + "num_tokens": 26793409.0, + "step": 1241 + }, + { + "epoch": 3.7531190926275992, + "grad_norm": 1.125, + "learning_rate": 7.473989515730159e-06, + "loss": 0.9109, + "num_tokens": 26847727.0, + "step": 1242 + }, + { + "epoch": 3.7561436672967865, + "grad_norm": 1.2109375, + "learning_rate": 7.470030820503312e-06, + "loss": 0.8456, + "num_tokens": 26884714.0, + "step": 1243 + }, + { + "epoch": 3.7591682419659733, + "grad_norm": 1.5234375, + "learning_rate": 7.466070238316584e-06, + "loss": 0.7834, + "num_tokens": 26908524.0, + "step": 1244 + }, + { + "epoch": 3.7621928166351606, + "grad_norm": 1.578125, + "learning_rate": 7.4621077729635516e-06, + "loss": 0.7912, + "num_tokens": 26931299.0, + "step": 1245 + }, + { + "epoch": 3.765217391304348, + "grad_norm": 1.65625, + "learning_rate": 7.458143428239593e-06, + "loss": 0.8805, + "num_tokens": 26954654.0, + "step": 1246 + }, + { + "epoch": 3.768241965973535, + "grad_norm": 1.78125, + "learning_rate": 7.454177207941884e-06, + "loss": 0.6834, + "num_tokens": 26968476.0, + "step": 1247 + }, + { + "epoch": 3.7712665406427224, + "grad_norm": 1.28125, + "learning_rate": 7.450209115869407e-06, + "loss": 0.8012, + "num_tokens": 27002787.0, + "step": 1248 + }, + { + "epoch": 3.774291115311909, + "grad_norm": 1.453125, + "learning_rate": 7.446239155822921e-06, + "loss": 0.8306, + "num_tokens": 27029039.0, + "step": 1249 + }, + { + "epoch": 3.7773156899810965, + "grad_norm": 1.8515625, + "learning_rate": 7.44226733160499e-06, + "loss": 0.8507, + "num_tokens": 27052303.0, + "step": 1250 + }, + { + "epoch": 3.7803402646502837, + "grad_norm": 1.6484375, + "learning_rate": 7.438293647019958e-06, + "loss": 0.8816, + "num_tokens": 27076033.0, + "step": 1251 + }, + { + "epoch": 3.7833648393194705, + "grad_norm": 1.421875, + "learning_rate": 7.434318105873943e-06, + "loss": 0.6734, + "num_tokens": 27096505.0, + "step": 1252 + }, + { + "epoch": 3.786389413988658, + "grad_norm": 1.984375, + "learning_rate": 7.430340711974855e-06, + "loss": 0.5527, + "num_tokens": 27103779.0, + "step": 1253 + }, + { + "epoch": 3.789413988657845, + "grad_norm": 1.7109375, + "learning_rate": 7.426361469132367e-06, + "loss": 0.7663, + "num_tokens": 27118683.0, + "step": 1254 + }, + { + "epoch": 3.792438563327032, + "grad_norm": 1.84375, + "learning_rate": 7.422380381157931e-06, + "loss": 0.6562, + "num_tokens": 27127216.0, + "step": 1255 + }, + { + "epoch": 3.795463137996219, + "grad_norm": 1.9453125, + "learning_rate": 7.418397451864763e-06, + "loss": 0.6263, + "num_tokens": 27134483.0, + "step": 1256 + }, + { + "epoch": 3.7984877126654064, + "grad_norm": 1.1953125, + "learning_rate": 7.4144126850678396e-06, + "loss": 0.7895, + "num_tokens": 27170015.0, + "step": 1257 + }, + { + "epoch": 3.8015122873345937, + "grad_norm": 1.296875, + "learning_rate": 7.410426084583906e-06, + "loss": 0.8609, + "num_tokens": 27209607.0, + "step": 1258 + }, + { + "epoch": 3.804536862003781, + "grad_norm": 2.015625, + "learning_rate": 7.406437654231453e-06, + "loss": 0.6118, + "num_tokens": 27216087.0, + "step": 1259 + }, + { + "epoch": 3.8075614366729678, + "grad_norm": 1.625, + "learning_rate": 7.402447397830732e-06, + "loss": 0.8242, + "num_tokens": 27239694.0, + "step": 1260 + }, + { + "epoch": 3.810586011342155, + "grad_norm": 1.46875, + "learning_rate": 7.398455319203741e-06, + "loss": 0.7637, + "num_tokens": 27261791.0, + "step": 1261 + }, + { + "epoch": 3.8136105860113423, + "grad_norm": 1.984375, + "learning_rate": 7.394461422174224e-06, + "loss": 0.6063, + "num_tokens": 27268684.0, + "step": 1262 + }, + { + "epoch": 3.816635160680529, + "grad_norm": 1.53125, + "learning_rate": 7.390465710567664e-06, + "loss": 0.8382, + "num_tokens": 27297460.0, + "step": 1263 + }, + { + "epoch": 3.8196597353497164, + "grad_norm": 1.4140625, + "learning_rate": 7.386468188211285e-06, + "loss": 0.8188, + "num_tokens": 27325762.0, + "step": 1264 + }, + { + "epoch": 3.8226843100189036, + "grad_norm": 1.4609375, + "learning_rate": 7.382468858934046e-06, + "loss": 0.7702, + "num_tokens": 27350658.0, + "step": 1265 + }, + { + "epoch": 3.825708884688091, + "grad_norm": 1.5234375, + "learning_rate": 7.378467726566632e-06, + "loss": 0.8587, + "num_tokens": 27374248.0, + "step": 1266 + }, + { + "epoch": 3.828733459357278, + "grad_norm": 1.65625, + "learning_rate": 7.374464794941461e-06, + "loss": 0.8573, + "num_tokens": 27397914.0, + "step": 1267 + }, + { + "epoch": 3.831758034026465, + "grad_norm": 1.6484375, + "learning_rate": 7.370460067892672e-06, + "loss": 0.6041, + "num_tokens": 27409632.0, + "step": 1268 + }, + { + "epoch": 3.8347826086956522, + "grad_norm": 1.9453125, + "learning_rate": 7.36645354925612e-06, + "loss": 0.5551, + "num_tokens": 27418233.0, + "step": 1269 + }, + { + "epoch": 3.8378071833648395, + "grad_norm": 1.6171875, + "learning_rate": 7.362445242869384e-06, + "loss": 0.7749, + "num_tokens": 27438391.0, + "step": 1270 + }, + { + "epoch": 3.8408317580340263, + "grad_norm": 1.28125, + "learning_rate": 7.358435152571749e-06, + "loss": 0.9156, + "num_tokens": 27478090.0, + "step": 1271 + }, + { + "epoch": 3.8438563327032136, + "grad_norm": 1.5, + "learning_rate": 7.3544232822042084e-06, + "loss": 0.8443, + "num_tokens": 27501710.0, + "step": 1272 + }, + { + "epoch": 3.846880907372401, + "grad_norm": 1.640625, + "learning_rate": 7.350409635609466e-06, + "loss": 0.6994, + "num_tokens": 27515626.0, + "step": 1273 + }, + { + "epoch": 3.8499054820415877, + "grad_norm": 1.5703125, + "learning_rate": 7.34639421663192e-06, + "loss": 0.8384, + "num_tokens": 27543011.0, + "step": 1274 + }, + { + "epoch": 3.852930056710775, + "grad_norm": 2.046875, + "learning_rate": 7.342377029117673e-06, + "loss": 0.6145, + "num_tokens": 27550160.0, + "step": 1275 + }, + { + "epoch": 3.855954631379962, + "grad_norm": 1.9921875, + "learning_rate": 7.338358076914517e-06, + "loss": 0.5527, + "num_tokens": 27556665.0, + "step": 1276 + }, + { + "epoch": 3.8589792060491495, + "grad_norm": 1.453125, + "learning_rate": 7.334337363871936e-06, + "loss": 0.8201, + "num_tokens": 27580355.0, + "step": 1277 + }, + { + "epoch": 3.8620037807183367, + "grad_norm": 1.0546875, + "learning_rate": 7.330314893841102e-06, + "loss": 0.8073, + "num_tokens": 27632821.0, + "step": 1278 + }, + { + "epoch": 3.8650283553875235, + "grad_norm": 1.359375, + "learning_rate": 7.326290670674864e-06, + "loss": 0.8894, + "num_tokens": 27665057.0, + "step": 1279 + }, + { + "epoch": 3.868052930056711, + "grad_norm": 1.8046875, + "learning_rate": 7.322264698227759e-06, + "loss": 0.616, + "num_tokens": 27672912.0, + "step": 1280 + }, + { + "epoch": 3.871077504725898, + "grad_norm": 1.171875, + "learning_rate": 7.318236980355993e-06, + "loss": 0.8628, + "num_tokens": 27717214.0, + "step": 1281 + }, + { + "epoch": 3.874102079395085, + "grad_norm": 1.4765625, + "learning_rate": 7.314207520917445e-06, + "loss": 0.8361, + "num_tokens": 27740578.0, + "step": 1282 + }, + { + "epoch": 3.877126654064272, + "grad_norm": 1.8984375, + "learning_rate": 7.310176323771663e-06, + "loss": 0.6327, + "num_tokens": 27748895.0, + "step": 1283 + }, + { + "epoch": 3.8801512287334594, + "grad_norm": 1.2265625, + "learning_rate": 7.306143392779862e-06, + "loss": 0.7828, + "num_tokens": 27788443.0, + "step": 1284 + }, + { + "epoch": 3.8831758034026462, + "grad_norm": 1.40625, + "learning_rate": 7.30210873180491e-06, + "loss": 0.742, + "num_tokens": 27811053.0, + "step": 1285 + }, + { + "epoch": 3.8862003780718335, + "grad_norm": 1.0, + "learning_rate": 7.298072344711342e-06, + "loss": 0.8687, + "num_tokens": 27876730.0, + "step": 1286 + }, + { + "epoch": 3.8892249527410208, + "grad_norm": 1.7890625, + "learning_rate": 7.2940342353653394e-06, + "loss": 0.6011, + "num_tokens": 27884638.0, + "step": 1287 + }, + { + "epoch": 3.892249527410208, + "grad_norm": 1.4375, + "learning_rate": 7.289994407634735e-06, + "loss": 0.7465, + "num_tokens": 27907612.0, + "step": 1288 + }, + { + "epoch": 3.8952741020793953, + "grad_norm": 1.3671875, + "learning_rate": 7.285952865389007e-06, + "loss": 0.863, + "num_tokens": 27934960.0, + "step": 1289 + }, + { + "epoch": 3.898298676748582, + "grad_norm": 1.5, + "learning_rate": 7.28190961249928e-06, + "loss": 0.8508, + "num_tokens": 27966498.0, + "step": 1290 + }, + { + "epoch": 3.9013232514177694, + "grad_norm": 1.375, + "learning_rate": 7.2778646528383105e-06, + "loss": 0.8032, + "num_tokens": 27990841.0, + "step": 1291 + }, + { + "epoch": 3.9043478260869566, + "grad_norm": 1.109375, + "learning_rate": 7.273817990280494e-06, + "loss": 0.8904, + "num_tokens": 28045383.0, + "step": 1292 + }, + { + "epoch": 3.9073724007561434, + "grad_norm": 1.5078125, + "learning_rate": 7.269769628701855e-06, + "loss": 0.8562, + "num_tokens": 28069317.0, + "step": 1293 + }, + { + "epoch": 3.9103969754253307, + "grad_norm": 1.453125, + "learning_rate": 7.265719571980048e-06, + "loss": 0.7036, + "num_tokens": 28087689.0, + "step": 1294 + }, + { + "epoch": 3.913421550094518, + "grad_norm": 2.015625, + "learning_rate": 7.261667823994351e-06, + "loss": 0.5634, + "num_tokens": 28094489.0, + "step": 1295 + }, + { + "epoch": 3.9164461247637052, + "grad_norm": 1.46875, + "learning_rate": 7.257614388625656e-06, + "loss": 0.822, + "num_tokens": 28117779.0, + "step": 1296 + }, + { + "epoch": 3.9194706994328925, + "grad_norm": 1.84375, + "learning_rate": 7.253559269756481e-06, + "loss": 0.6822, + "num_tokens": 28127322.0, + "step": 1297 + }, + { + "epoch": 3.9224952741020793, + "grad_norm": 1.6171875, + "learning_rate": 7.249502471270947e-06, + "loss": 0.8311, + "num_tokens": 28150536.0, + "step": 1298 + }, + { + "epoch": 3.9255198487712666, + "grad_norm": 1.375, + "learning_rate": 7.245443997054791e-06, + "loss": 0.759, + "num_tokens": 28174389.0, + "step": 1299 + }, + { + "epoch": 3.928544423440454, + "grad_norm": 1.1171875, + "learning_rate": 7.2413838509953504e-06, + "loss": 0.7728, + "num_tokens": 28213954.0, + "step": 1300 + }, + { + "epoch": 3.9315689981096407, + "grad_norm": 1.3046875, + "learning_rate": 7.237322036981568e-06, + "loss": 0.9305, + "num_tokens": 28254167.0, + "step": 1301 + }, + { + "epoch": 3.934593572778828, + "grad_norm": 1.265625, + "learning_rate": 7.23325855890398e-06, + "loss": 0.8311, + "num_tokens": 28287086.0, + "step": 1302 + }, + { + "epoch": 3.937618147448015, + "grad_norm": 1.078125, + "learning_rate": 7.22919342065472e-06, + "loss": 0.9392, + "num_tokens": 28341928.0, + "step": 1303 + }, + { + "epoch": 3.940642722117202, + "grad_norm": 1.90625, + "learning_rate": 7.225126626127507e-06, + "loss": 0.6917, + "num_tokens": 28351460.0, + "step": 1304 + }, + { + "epoch": 3.9436672967863893, + "grad_norm": 1.4609375, + "learning_rate": 7.221058179217653e-06, + "loss": 0.8242, + "num_tokens": 28375068.0, + "step": 1305 + }, + { + "epoch": 3.9466918714555765, + "grad_norm": 1.4296875, + "learning_rate": 7.216988083822048e-06, + "loss": 0.739, + "num_tokens": 28398084.0, + "step": 1306 + }, + { + "epoch": 3.949716446124764, + "grad_norm": 1.7421875, + "learning_rate": 7.212916343839163e-06, + "loss": 0.7461, + "num_tokens": 28418288.0, + "step": 1307 + }, + { + "epoch": 3.952741020793951, + "grad_norm": 1.5859375, + "learning_rate": 7.208842963169044e-06, + "loss": 0.856, + "num_tokens": 28441814.0, + "step": 1308 + }, + { + "epoch": 3.955765595463138, + "grad_norm": 1.921875, + "learning_rate": 7.204767945713304e-06, + "loss": 0.5911, + "num_tokens": 28449312.0, + "step": 1309 + }, + { + "epoch": 3.958790170132325, + "grad_norm": 1.296875, + "learning_rate": 7.200691295375134e-06, + "loss": 0.7969, + "num_tokens": 28478933.0, + "step": 1310 + }, + { + "epoch": 3.9618147448015124, + "grad_norm": 1.8046875, + "learning_rate": 7.196613016059281e-06, + "loss": 0.9228, + "num_tokens": 28502165.0, + "step": 1311 + }, + { + "epoch": 3.9648393194706992, + "grad_norm": 1.640625, + "learning_rate": 7.19253311167205e-06, + "loss": 0.8414, + "num_tokens": 28525403.0, + "step": 1312 + }, + { + "epoch": 3.9678638941398865, + "grad_norm": 1.4140625, + "learning_rate": 7.188451586121312e-06, + "loss": 0.7931, + "num_tokens": 28550866.0, + "step": 1313 + }, + { + "epoch": 3.9708884688090738, + "grad_norm": 1.6953125, + "learning_rate": 7.184368443316481e-06, + "loss": 0.7741, + "num_tokens": 28566955.0, + "step": 1314 + }, + { + "epoch": 3.973913043478261, + "grad_norm": 2.0, + "learning_rate": 7.18028368716853e-06, + "loss": 0.5834, + "num_tokens": 28574707.0, + "step": 1315 + }, + { + "epoch": 3.9769376181474483, + "grad_norm": 1.6484375, + "learning_rate": 7.176197321589967e-06, + "loss": 0.7953, + "num_tokens": 28590081.0, + "step": 1316 + }, + { + "epoch": 3.979962192816635, + "grad_norm": 1.5234375, + "learning_rate": 7.172109350494849e-06, + "loss": 0.8436, + "num_tokens": 28618278.0, + "step": 1317 + }, + { + "epoch": 3.9829867674858224, + "grad_norm": 1.9453125, + "learning_rate": 7.168019777798767e-06, + "loss": 0.6058, + "num_tokens": 28625008.0, + "step": 1318 + }, + { + "epoch": 3.9860113421550096, + "grad_norm": 1.4765625, + "learning_rate": 7.163928607418849e-06, + "loss": 0.7958, + "num_tokens": 28648506.0, + "step": 1319 + }, + { + "epoch": 3.9890359168241964, + "grad_norm": 1.6640625, + "learning_rate": 7.159835843273748e-06, + "loss": 0.904, + "num_tokens": 28671640.0, + "step": 1320 + }, + { + "epoch": 3.9920604914933837, + "grad_norm": 1.1875, + "learning_rate": 7.155741489283653e-06, + "loss": 0.9319, + "num_tokens": 28716857.0, + "step": 1321 + }, + { + "epoch": 3.995085066162571, + "grad_norm": 1.546875, + "learning_rate": 7.1516455493702654e-06, + "loss": 0.8767, + "num_tokens": 28740053.0, + "step": 1322 + }, + { + "epoch": 3.998109640831758, + "grad_norm": 1.1953125, + "learning_rate": 7.147548027456811e-06, + "loss": 0.7994, + "num_tokens": 28776183.0, + "step": 1323 + }, + { + "epoch": 4.0, + "grad_norm": 2.234375, + "learning_rate": 7.143448927468031e-06, + "loss": 0.5814, + "num_tokens": 28781252.0, + "step": 1324 + }, + { + "epoch": 4.003024574669187, + "grad_norm": 1.546875, + "learning_rate": 7.139348253330177e-06, + "loss": 0.8584, + "num_tokens": 28804440.0, + "step": 1325 + }, + { + "epoch": 4.0060491493383745, + "grad_norm": 2.0625, + "learning_rate": 7.135246008971009e-06, + "loss": 0.6473, + "num_tokens": 28812686.0, + "step": 1326 + }, + { + "epoch": 4.009073724007561, + "grad_norm": 1.6484375, + "learning_rate": 7.131142198319791e-06, + "loss": 0.8637, + "num_tokens": 28835266.0, + "step": 1327 + }, + { + "epoch": 4.012098298676748, + "grad_norm": 1.0078125, + "learning_rate": 7.1270368253072855e-06, + "loss": 0.85, + "num_tokens": 28896265.0, + "step": 1328 + }, + { + "epoch": 4.015122873345936, + "grad_norm": 0.98828125, + "learning_rate": 7.122929893865754e-06, + "loss": 0.8931, + "num_tokens": 28951475.0, + "step": 1329 + }, + { + "epoch": 4.018147448015123, + "grad_norm": 1.875, + "learning_rate": 7.1188214079289476e-06, + "loss": 0.6092, + "num_tokens": 28959145.0, + "step": 1330 + }, + { + "epoch": 4.02117202268431, + "grad_norm": 1.2109375, + "learning_rate": 7.114711371432113e-06, + "loss": 0.815, + "num_tokens": 28998815.0, + "step": 1331 + }, + { + "epoch": 4.024196597353497, + "grad_norm": 1.2421875, + "learning_rate": 7.110599788311972e-06, + "loss": 0.842, + "num_tokens": 29035978.0, + "step": 1332 + }, + { + "epoch": 4.027221172022684, + "grad_norm": 1.921875, + "learning_rate": 7.106486662506737e-06, + "loss": 0.5375, + "num_tokens": 29042782.0, + "step": 1333 + }, + { + "epoch": 4.030245746691872, + "grad_norm": 1.9375, + "learning_rate": 7.1023719979560934e-06, + "loss": 0.5847, + "num_tokens": 29050271.0, + "step": 1334 + }, + { + "epoch": 4.033270321361059, + "grad_norm": 1.3984375, + "learning_rate": 7.098255798601199e-06, + "loss": 0.7817, + "num_tokens": 29074102.0, + "step": 1335 + }, + { + "epoch": 4.036294896030245, + "grad_norm": 1.78125, + "learning_rate": 7.0941380683846885e-06, + "loss": 0.6347, + "num_tokens": 29082505.0, + "step": 1336 + }, + { + "epoch": 4.039319470699433, + "grad_norm": 1.9609375, + "learning_rate": 7.090018811250653e-06, + "loss": 0.789, + "num_tokens": 29095594.0, + "step": 1337 + }, + { + "epoch": 4.04234404536862, + "grad_norm": 1.5703125, + "learning_rate": 7.0858980311446575e-06, + "loss": 0.8989, + "num_tokens": 29118649.0, + "step": 1338 + }, + { + "epoch": 4.045368620037808, + "grad_norm": 1.03125, + "learning_rate": 7.081775732013714e-06, + "loss": 0.8736, + "num_tokens": 29171612.0, + "step": 1339 + }, + { + "epoch": 4.048393194706994, + "grad_norm": 1.0546875, + "learning_rate": 7.0776519178063e-06, + "loss": 0.8013, + "num_tokens": 29217911.0, + "step": 1340 + }, + { + "epoch": 4.051417769376181, + "grad_norm": 1.4453125, + "learning_rate": 7.073526592472337e-06, + "loss": 0.8597, + "num_tokens": 29245674.0, + "step": 1341 + }, + { + "epoch": 4.054442344045369, + "grad_norm": 1.515625, + "learning_rate": 7.069399759963196e-06, + "loss": 0.7059, + "num_tokens": 29261547.0, + "step": 1342 + }, + { + "epoch": 4.057466918714556, + "grad_norm": 1.96875, + "learning_rate": 7.065271424231694e-06, + "loss": 0.6469, + "num_tokens": 29269528.0, + "step": 1343 + }, + { + "epoch": 4.060491493383743, + "grad_norm": 1.390625, + "learning_rate": 7.0611415892320845e-06, + "loss": 0.7618, + "num_tokens": 29291941.0, + "step": 1344 + }, + { + "epoch": 4.06351606805293, + "grad_norm": 1.1015625, + "learning_rate": 7.057010258920057e-06, + "loss": 0.8304, + "num_tokens": 29331044.0, + "step": 1345 + }, + { + "epoch": 4.066540642722117, + "grad_norm": 1.40625, + "learning_rate": 7.052877437252739e-06, + "loss": 0.814, + "num_tokens": 29355504.0, + "step": 1346 + }, + { + "epoch": 4.069565217391304, + "grad_norm": 1.28125, + "learning_rate": 7.048743128188676e-06, + "loss": 0.8829, + "num_tokens": 29389473.0, + "step": 1347 + }, + { + "epoch": 4.072589792060492, + "grad_norm": 1.8515625, + "learning_rate": 7.044607335687851e-06, + "loss": 0.58, + "num_tokens": 29396838.0, + "step": 1348 + }, + { + "epoch": 4.0756143667296785, + "grad_norm": 1.25, + "learning_rate": 7.040470063711655e-06, + "loss": 0.8737, + "num_tokens": 29435007.0, + "step": 1349 + }, + { + "epoch": 4.078638941398866, + "grad_norm": 1.5078125, + "learning_rate": 7.036331316222907e-06, + "loss": 0.8035, + "num_tokens": 29457639.0, + "step": 1350 + }, + { + "epoch": 4.081663516068053, + "grad_norm": 1.0625, + "learning_rate": 7.032191097185832e-06, + "loss": 0.8915, + "num_tokens": 29507581.0, + "step": 1351 + }, + { + "epoch": 4.08468809073724, + "grad_norm": 1.4609375, + "learning_rate": 7.028049410566068e-06, + "loss": 0.7869, + "num_tokens": 29530268.0, + "step": 1352 + }, + { + "epoch": 4.0877126654064275, + "grad_norm": 1.390625, + "learning_rate": 7.023906260330656e-06, + "loss": 0.8883, + "num_tokens": 29570369.0, + "step": 1353 + }, + { + "epoch": 4.090737240075614, + "grad_norm": 1.703125, + "learning_rate": 7.019761650448044e-06, + "loss": 0.7408, + "num_tokens": 29585603.0, + "step": 1354 + }, + { + "epoch": 4.093761814744801, + "grad_norm": 1.9609375, + "learning_rate": 7.015615584888072e-06, + "loss": 0.5987, + "num_tokens": 29595764.0, + "step": 1355 + }, + { + "epoch": 4.096786389413989, + "grad_norm": 2.0, + "learning_rate": 7.011468067621977e-06, + "loss": 0.6706, + "num_tokens": 29603160.0, + "step": 1356 + }, + { + "epoch": 4.099810964083176, + "grad_norm": 2.09375, + "learning_rate": 7.007319102622389e-06, + "loss": 0.5545, + "num_tokens": 29610052.0, + "step": 1357 + }, + { + "epoch": 4.1028355387523625, + "grad_norm": 1.421875, + "learning_rate": 7.003168693863319e-06, + "loss": 0.7807, + "num_tokens": 29634040.0, + "step": 1358 + }, + { + "epoch": 4.10586011342155, + "grad_norm": 1.8671875, + "learning_rate": 6.999016845320168e-06, + "loss": 0.5799, + "num_tokens": 29641504.0, + "step": 1359 + }, + { + "epoch": 4.108884688090737, + "grad_norm": 1.7109375, + "learning_rate": 6.9948635609697076e-06, + "loss": 0.8562, + "num_tokens": 29664701.0, + "step": 1360 + }, + { + "epoch": 4.111909262759925, + "grad_norm": 1.3984375, + "learning_rate": 6.990708844790093e-06, + "loss": 0.8819, + "num_tokens": 29692912.0, + "step": 1361 + }, + { + "epoch": 4.114933837429112, + "grad_norm": 1.3203125, + "learning_rate": 6.986552700760843e-06, + "loss": 0.8203, + "num_tokens": 29731221.0, + "step": 1362 + }, + { + "epoch": 4.117958412098298, + "grad_norm": 1.9765625, + "learning_rate": 6.9823951328628555e-06, + "loss": 0.5683, + "num_tokens": 29737842.0, + "step": 1363 + }, + { + "epoch": 4.120982986767486, + "grad_norm": 1.78125, + "learning_rate": 6.9782361450783765e-06, + "loss": 0.5155, + "num_tokens": 29746256.0, + "step": 1364 + }, + { + "epoch": 4.124007561436673, + "grad_norm": 1.5625, + "learning_rate": 6.974075741391024e-06, + "loss": 0.7742, + "num_tokens": 29764254.0, + "step": 1365 + }, + { + "epoch": 4.12703213610586, + "grad_norm": 1.921875, + "learning_rate": 6.969913925785772e-06, + "loss": 0.6498, + "num_tokens": 29774229.0, + "step": 1366 + }, + { + "epoch": 4.130056710775047, + "grad_norm": 1.921875, + "learning_rate": 6.965750702248936e-06, + "loss": 0.5688, + "num_tokens": 29781696.0, + "step": 1367 + }, + { + "epoch": 4.133081285444234, + "grad_norm": 1.125, + "learning_rate": 6.961586074768194e-06, + "loss": 0.8312, + "num_tokens": 29821208.0, + "step": 1368 + }, + { + "epoch": 4.136105860113422, + "grad_norm": 2.0625, + "learning_rate": 6.957420047332558e-06, + "loss": 0.7071, + "num_tokens": 29833826.0, + "step": 1369 + }, + { + "epoch": 4.139130434782609, + "grad_norm": 2.0625, + "learning_rate": 6.953252623932388e-06, + "loss": 0.6033, + "num_tokens": 29841370.0, + "step": 1370 + }, + { + "epoch": 4.142155009451796, + "grad_norm": 1.8203125, + "learning_rate": 6.949083808559378e-06, + "loss": 0.7084, + "num_tokens": 29853456.0, + "step": 1371 + }, + { + "epoch": 4.145179584120983, + "grad_norm": 1.578125, + "learning_rate": 6.9449136052065535e-06, + "loss": 0.7496, + "num_tokens": 29872465.0, + "step": 1372 + }, + { + "epoch": 4.14820415879017, + "grad_norm": 1.7734375, + "learning_rate": 6.940742017868274e-06, + "loss": 0.5878, + "num_tokens": 29879968.0, + "step": 1373 + }, + { + "epoch": 4.151228733459357, + "grad_norm": 1.7109375, + "learning_rate": 6.936569050540221e-06, + "loss": 0.6653, + "num_tokens": 29893480.0, + "step": 1374 + }, + { + "epoch": 4.154253308128545, + "grad_norm": 1.3671875, + "learning_rate": 6.932394707219396e-06, + "loss": 0.7968, + "num_tokens": 29925968.0, + "step": 1375 + }, + { + "epoch": 4.1572778827977315, + "grad_norm": 1.4375, + "learning_rate": 6.928218991904128e-06, + "loss": 0.8402, + "num_tokens": 29949196.0, + "step": 1376 + }, + { + "epoch": 4.160302457466918, + "grad_norm": 1.296875, + "learning_rate": 6.924041908594048e-06, + "loss": 0.8938, + "num_tokens": 29989486.0, + "step": 1377 + }, + { + "epoch": 4.163327032136106, + "grad_norm": 1.8984375, + "learning_rate": 6.919863461290105e-06, + "loss": 0.5807, + "num_tokens": 29997216.0, + "step": 1378 + }, + { + "epoch": 4.166351606805293, + "grad_norm": 1.8046875, + "learning_rate": 6.9156836539945535e-06, + "loss": 0.5759, + "num_tokens": 30005239.0, + "step": 1379 + }, + { + "epoch": 4.1693761814744805, + "grad_norm": 1.25, + "learning_rate": 6.911502490710946e-06, + "loss": 0.8729, + "num_tokens": 30044616.0, + "step": 1380 + }, + { + "epoch": 4.172400756143667, + "grad_norm": 1.3984375, + "learning_rate": 6.90731997544414e-06, + "loss": 0.7746, + "num_tokens": 30065923.0, + "step": 1381 + }, + { + "epoch": 4.175425330812854, + "grad_norm": 2.109375, + "learning_rate": 6.903136112200285e-06, + "loss": 0.6144, + "num_tokens": 30072859.0, + "step": 1382 + }, + { + "epoch": 4.178449905482042, + "grad_norm": 1.2890625, + "learning_rate": 6.898950904986821e-06, + "loss": 0.771, + "num_tokens": 30102499.0, + "step": 1383 + }, + { + "epoch": 4.181474480151229, + "grad_norm": 1.3671875, + "learning_rate": 6.894764357812476e-06, + "loss": 0.7587, + "num_tokens": 30128761.0, + "step": 1384 + }, + { + "epoch": 4.1844990548204155, + "grad_norm": 1.171875, + "learning_rate": 6.890576474687264e-06, + "loss": 0.8702, + "num_tokens": 30175160.0, + "step": 1385 + }, + { + "epoch": 4.187523629489603, + "grad_norm": 1.9375, + "learning_rate": 6.886387259622474e-06, + "loss": 0.5403, + "num_tokens": 30182367.0, + "step": 1386 + }, + { + "epoch": 4.19054820415879, + "grad_norm": 2.015625, + "learning_rate": 6.882196716630674e-06, + "loss": 0.5781, + "num_tokens": 30189005.0, + "step": 1387 + }, + { + "epoch": 4.193572778827978, + "grad_norm": 2.0625, + "learning_rate": 6.878004849725702e-06, + "loss": 0.6249, + "num_tokens": 30195622.0, + "step": 1388 + }, + { + "epoch": 4.196597353497165, + "grad_norm": 1.5859375, + "learning_rate": 6.873811662922666e-06, + "loss": 0.8212, + "num_tokens": 30215705.0, + "step": 1389 + }, + { + "epoch": 4.199621928166351, + "grad_norm": 1.984375, + "learning_rate": 6.869617160237937e-06, + "loss": 0.627, + "num_tokens": 30222916.0, + "step": 1390 + }, + { + "epoch": 4.202646502835539, + "grad_norm": 1.3046875, + "learning_rate": 6.865421345689147e-06, + "loss": 0.7823, + "num_tokens": 30253309.0, + "step": 1391 + }, + { + "epoch": 4.205671077504726, + "grad_norm": 1.9375, + "learning_rate": 6.861224223295183e-06, + "loss": 0.5748, + "num_tokens": 30260769.0, + "step": 1392 + }, + { + "epoch": 4.208695652173913, + "grad_norm": 1.2890625, + "learning_rate": 6.857025797076189e-06, + "loss": 0.7227, + "num_tokens": 30287538.0, + "step": 1393 + }, + { + "epoch": 4.2117202268431, + "grad_norm": 1.078125, + "learning_rate": 6.8528260710535495e-06, + "loss": 0.8711, + "num_tokens": 30335144.0, + "step": 1394 + }, + { + "epoch": 4.214744801512287, + "grad_norm": 1.984375, + "learning_rate": 6.8486250492499066e-06, + "loss": 0.59, + "num_tokens": 30342327.0, + "step": 1395 + }, + { + "epoch": 4.217769376181474, + "grad_norm": 1.78125, + "learning_rate": 6.84442273568913e-06, + "loss": 0.7711, + "num_tokens": 30357349.0, + "step": 1396 + }, + { + "epoch": 4.220793950850662, + "grad_norm": 1.375, + "learning_rate": 6.840219134396334e-06, + "loss": 0.827, + "num_tokens": 30391941.0, + "step": 1397 + }, + { + "epoch": 4.223818525519849, + "grad_norm": 2.078125, + "learning_rate": 6.836014249397869e-06, + "loss": 0.6723, + "num_tokens": 30399425.0, + "step": 1398 + }, + { + "epoch": 4.226843100189036, + "grad_norm": 1.4453125, + "learning_rate": 6.831808084721305e-06, + "loss": 0.8219, + "num_tokens": 30422879.0, + "step": 1399 + }, + { + "epoch": 4.229867674858223, + "grad_norm": 1.296875, + "learning_rate": 6.82760064439545e-06, + "loss": 0.9012, + "num_tokens": 30462205.0, + "step": 1400 + }, + { + "epoch": 4.23289224952741, + "grad_norm": 1.6484375, + "learning_rate": 6.8233919324503265e-06, + "loss": 0.8614, + "num_tokens": 30485406.0, + "step": 1401 + }, + { + "epoch": 4.235916824196598, + "grad_norm": 1.65625, + "learning_rate": 6.819181952917172e-06, + "loss": 0.8827, + "num_tokens": 30508397.0, + "step": 1402 + }, + { + "epoch": 4.2389413988657845, + "grad_norm": 1.8984375, + "learning_rate": 6.814970709828448e-06, + "loss": 0.5837, + "num_tokens": 30516082.0, + "step": 1403 + }, + { + "epoch": 4.241965973534971, + "grad_norm": 1.9296875, + "learning_rate": 6.810758207217817e-06, + "loss": 0.6041, + "num_tokens": 30523261.0, + "step": 1404 + }, + { + "epoch": 4.244990548204159, + "grad_norm": 1.40625, + "learning_rate": 6.806544449120151e-06, + "loss": 0.8189, + "num_tokens": 30547028.0, + "step": 1405 + }, + { + "epoch": 4.248015122873346, + "grad_norm": 1.3203125, + "learning_rate": 6.802329439571531e-06, + "loss": 0.89, + "num_tokens": 30584229.0, + "step": 1406 + }, + { + "epoch": 4.251039697542533, + "grad_norm": 1.6015625, + "learning_rate": 6.798113182609225e-06, + "loss": 0.8535, + "num_tokens": 30607567.0, + "step": 1407 + }, + { + "epoch": 4.25406427221172, + "grad_norm": 1.609375, + "learning_rate": 6.793895682271702e-06, + "loss": 0.7122, + "num_tokens": 30623957.0, + "step": 1408 + }, + { + "epoch": 4.257088846880907, + "grad_norm": 1.421875, + "learning_rate": 6.789676942598626e-06, + "loss": 0.8014, + "num_tokens": 30644951.0, + "step": 1409 + }, + { + "epoch": 4.260113421550095, + "grad_norm": 1.4140625, + "learning_rate": 6.7854569676308404e-06, + "loss": 0.717, + "num_tokens": 30666789.0, + "step": 1410 + }, + { + "epoch": 4.263137996219282, + "grad_norm": 1.8359375, + "learning_rate": 6.781235761410374e-06, + "loss": 0.6598, + "num_tokens": 30679111.0, + "step": 1411 + }, + { + "epoch": 4.2661625708884685, + "grad_norm": 1.375, + "learning_rate": 6.777013327980438e-06, + "loss": 0.97, + "num_tokens": 30724062.0, + "step": 1412 + }, + { + "epoch": 4.269187145557656, + "grad_norm": 1.6015625, + "learning_rate": 6.772789671385417e-06, + "loss": 0.7398, + "num_tokens": 30739071.0, + "step": 1413 + }, + { + "epoch": 4.272211720226843, + "grad_norm": 1.046875, + "learning_rate": 6.768564795670866e-06, + "loss": 0.8209, + "num_tokens": 30789163.0, + "step": 1414 + }, + { + "epoch": 4.27523629489603, + "grad_norm": 1.21875, + "learning_rate": 6.764338704883511e-06, + "loss": 0.8156, + "num_tokens": 30828474.0, + "step": 1415 + }, + { + "epoch": 4.278260869565218, + "grad_norm": 1.515625, + "learning_rate": 6.760111403071239e-06, + "loss": 0.8683, + "num_tokens": 30851851.0, + "step": 1416 + }, + { + "epoch": 4.281285444234404, + "grad_norm": 1.7578125, + "learning_rate": 6.755882894283097e-06, + "loss": 0.6335, + "num_tokens": 30860779.0, + "step": 1417 + }, + { + "epoch": 4.284310018903592, + "grad_norm": 1.859375, + "learning_rate": 6.751653182569289e-06, + "loss": 0.6183, + "num_tokens": 30868080.0, + "step": 1418 + }, + { + "epoch": 4.287334593572779, + "grad_norm": 1.2109375, + "learning_rate": 6.747422271981171e-06, + "loss": 0.8042, + "num_tokens": 30906478.0, + "step": 1419 + }, + { + "epoch": 4.290359168241966, + "grad_norm": 1.203125, + "learning_rate": 6.7431901665712485e-06, + "loss": 0.8129, + "num_tokens": 30945265.0, + "step": 1420 + }, + { + "epoch": 4.293383742911153, + "grad_norm": 1.234375, + "learning_rate": 6.73895687039317e-06, + "loss": 0.741, + "num_tokens": 30974091.0, + "step": 1421 + }, + { + "epoch": 4.29640831758034, + "grad_norm": 2.0, + "learning_rate": 6.734722387501724e-06, + "loss": 0.5692, + "num_tokens": 30981641.0, + "step": 1422 + }, + { + "epoch": 4.299432892249527, + "grad_norm": 1.4375, + "learning_rate": 6.730486721952838e-06, + "loss": 0.8583, + "num_tokens": 31014098.0, + "step": 1423 + }, + { + "epoch": 4.302457466918715, + "grad_norm": 1.7265625, + "learning_rate": 6.726249877803568e-06, + "loss": 0.7337, + "num_tokens": 31027245.0, + "step": 1424 + }, + { + "epoch": 4.305482041587902, + "grad_norm": 1.96875, + "learning_rate": 6.722011859112105e-06, + "loss": 0.626, + "num_tokens": 31034802.0, + "step": 1425 + }, + { + "epoch": 4.308506616257089, + "grad_norm": 1.78125, + "learning_rate": 6.717772669937761e-06, + "loss": 0.5716, + "num_tokens": 31043244.0, + "step": 1426 + }, + { + "epoch": 4.311531190926276, + "grad_norm": 1.6796875, + "learning_rate": 6.713532314340968e-06, + "loss": 0.6333, + "num_tokens": 31056966.0, + "step": 1427 + }, + { + "epoch": 4.314555765595463, + "grad_norm": 1.125, + "learning_rate": 6.70929079638328e-06, + "loss": 0.8689, + "num_tokens": 31109355.0, + "step": 1428 + }, + { + "epoch": 4.317580340264651, + "grad_norm": 1.1640625, + "learning_rate": 6.705048120127357e-06, + "loss": 0.8265, + "num_tokens": 31154797.0, + "step": 1429 + }, + { + "epoch": 4.3206049149338375, + "grad_norm": 1.265625, + "learning_rate": 6.700804289636977e-06, + "loss": 0.8827, + "num_tokens": 31193719.0, + "step": 1430 + }, + { + "epoch": 4.323629489603024, + "grad_norm": 2.015625, + "learning_rate": 6.696559308977021e-06, + "loss": 0.6392, + "num_tokens": 31200480.0, + "step": 1431 + }, + { + "epoch": 4.326654064272212, + "grad_norm": 2.09375, + "learning_rate": 6.692313182213463e-06, + "loss": 0.5879, + "num_tokens": 31207736.0, + "step": 1432 + }, + { + "epoch": 4.329678638941399, + "grad_norm": 1.546875, + "learning_rate": 6.688065913413391e-06, + "loss": 0.7776, + "num_tokens": 31230680.0, + "step": 1433 + }, + { + "epoch": 4.332703213610586, + "grad_norm": 1.984375, + "learning_rate": 6.68381750664497e-06, + "loss": 0.6208, + "num_tokens": 31238155.0, + "step": 1434 + }, + { + "epoch": 4.335727788279773, + "grad_norm": 1.875, + "learning_rate": 6.679567965977466e-06, + "loss": 0.6207, + "num_tokens": 31245723.0, + "step": 1435 + }, + { + "epoch": 4.33875236294896, + "grad_norm": 2.1875, + "learning_rate": 6.67531729548123e-06, + "loss": 0.5785, + "num_tokens": 31252855.0, + "step": 1436 + }, + { + "epoch": 4.341776937618148, + "grad_norm": 1.3828125, + "learning_rate": 6.67106549922769e-06, + "loss": 0.7782, + "num_tokens": 31277204.0, + "step": 1437 + }, + { + "epoch": 4.344801512287335, + "grad_norm": 1.9609375, + "learning_rate": 6.666812581289356e-06, + "loss": 0.5675, + "num_tokens": 31284333.0, + "step": 1438 + }, + { + "epoch": 4.3478260869565215, + "grad_norm": 1.484375, + "learning_rate": 6.662558545739812e-06, + "loss": 0.8247, + "num_tokens": 31320830.0, + "step": 1439 + }, + { + "epoch": 4.350850661625709, + "grad_norm": 1.8046875, + "learning_rate": 6.658303396653712e-06, + "loss": 0.6037, + "num_tokens": 31330631.0, + "step": 1440 + }, + { + "epoch": 4.353875236294896, + "grad_norm": 1.8984375, + "learning_rate": 6.6540471381067785e-06, + "loss": 0.6179, + "num_tokens": 31338442.0, + "step": 1441 + }, + { + "epoch": 4.356899810964083, + "grad_norm": 1.3828125, + "learning_rate": 6.649789774175792e-06, + "loss": 0.9163, + "num_tokens": 31374482.0, + "step": 1442 + }, + { + "epoch": 4.359924385633271, + "grad_norm": 1.40625, + "learning_rate": 6.645531308938596e-06, + "loss": 0.8293, + "num_tokens": 31404560.0, + "step": 1443 + }, + { + "epoch": 4.362948960302457, + "grad_norm": 1.7578125, + "learning_rate": 6.64127174647409e-06, + "loss": 0.5646, + "num_tokens": 31412792.0, + "step": 1444 + }, + { + "epoch": 4.365973534971644, + "grad_norm": 1.71875, + "learning_rate": 6.637011090862219e-06, + "loss": 0.8024, + "num_tokens": 31431163.0, + "step": 1445 + }, + { + "epoch": 4.368998109640832, + "grad_norm": 1.5234375, + "learning_rate": 6.632749346183979e-06, + "loss": 0.8514, + "num_tokens": 31458532.0, + "step": 1446 + }, + { + "epoch": 4.372022684310019, + "grad_norm": 1.8515625, + "learning_rate": 6.628486516521409e-06, + "loss": 0.5621, + "num_tokens": 31466230.0, + "step": 1447 + }, + { + "epoch": 4.375047258979206, + "grad_norm": 1.2109375, + "learning_rate": 6.624222605957586e-06, + "loss": 0.8238, + "num_tokens": 31505347.0, + "step": 1448 + }, + { + "epoch": 4.378071833648393, + "grad_norm": 1.8671875, + "learning_rate": 6.619957618576622e-06, + "loss": 0.6355, + "num_tokens": 31513305.0, + "step": 1449 + }, + { + "epoch": 4.38109640831758, + "grad_norm": 1.9921875, + "learning_rate": 6.6156915584636626e-06, + "loss": 0.6345, + "num_tokens": 31520628.0, + "step": 1450 + }, + { + "epoch": 4.384120982986768, + "grad_norm": 1.484375, + "learning_rate": 6.611424429704879e-06, + "loss": 0.8348, + "num_tokens": 31543699.0, + "step": 1451 + }, + { + "epoch": 4.387145557655955, + "grad_norm": 1.6796875, + "learning_rate": 6.607156236387466e-06, + "loss": 0.648, + "num_tokens": 31560694.0, + "step": 1452 + }, + { + "epoch": 4.390170132325141, + "grad_norm": 1.5390625, + "learning_rate": 6.602886982599642e-06, + "loss": 0.9383, + "num_tokens": 31596245.0, + "step": 1453 + }, + { + "epoch": 4.393194706994329, + "grad_norm": 1.328125, + "learning_rate": 6.598616672430631e-06, + "loss": 0.8257, + "num_tokens": 31626900.0, + "step": 1454 + }, + { + "epoch": 4.396219281663516, + "grad_norm": 1.2890625, + "learning_rate": 6.594345309970681e-06, + "loss": 0.8311, + "num_tokens": 31661187.0, + "step": 1455 + }, + { + "epoch": 4.399243856332703, + "grad_norm": 1.84375, + "learning_rate": 6.590072899311043e-06, + "loss": 0.7152, + "num_tokens": 31674161.0, + "step": 1456 + }, + { + "epoch": 4.4022684310018905, + "grad_norm": 1.3125, + "learning_rate": 6.585799444543967e-06, + "loss": 0.8021, + "num_tokens": 31701123.0, + "step": 1457 + }, + { + "epoch": 4.405293005671077, + "grad_norm": 1.65625, + "learning_rate": 6.581524949762711e-06, + "loss": 0.8881, + "num_tokens": 31724550.0, + "step": 1458 + }, + { + "epoch": 4.408317580340265, + "grad_norm": 1.609375, + "learning_rate": 6.577249419061524e-06, + "loss": 0.8331, + "num_tokens": 31746908.0, + "step": 1459 + }, + { + "epoch": 4.411342155009452, + "grad_norm": 1.8125, + "learning_rate": 6.5729728565356495e-06, + "loss": 0.5555, + "num_tokens": 31754837.0, + "step": 1460 + }, + { + "epoch": 4.414366729678639, + "grad_norm": 1.515625, + "learning_rate": 6.5686952662813215e-06, + "loss": 0.8614, + "num_tokens": 31778906.0, + "step": 1461 + }, + { + "epoch": 4.417391304347826, + "grad_norm": 1.5390625, + "learning_rate": 6.564416652395752e-06, + "loss": 0.8515, + "num_tokens": 31802926.0, + "step": 1462 + }, + { + "epoch": 4.420415879017013, + "grad_norm": 1.28125, + "learning_rate": 6.560137018977139e-06, + "loss": 0.8881, + "num_tokens": 31842092.0, + "step": 1463 + }, + { + "epoch": 4.4234404536862, + "grad_norm": 1.6640625, + "learning_rate": 6.5558563701246546e-06, + "loss": 0.7736, + "num_tokens": 31859609.0, + "step": 1464 + }, + { + "epoch": 4.426465028355388, + "grad_norm": 1.4765625, + "learning_rate": 6.551574709938443e-06, + "loss": 0.8157, + "num_tokens": 31880642.0, + "step": 1465 + }, + { + "epoch": 4.4294896030245745, + "grad_norm": 1.3984375, + "learning_rate": 6.547292042519623e-06, + "loss": 0.7063, + "num_tokens": 31901087.0, + "step": 1466 + }, + { + "epoch": 4.432514177693762, + "grad_norm": 1.59375, + "learning_rate": 6.543008371970267e-06, + "loss": 0.8046, + "num_tokens": 31922496.0, + "step": 1467 + }, + { + "epoch": 4.435538752362949, + "grad_norm": 1.3515625, + "learning_rate": 6.538723702393417e-06, + "loss": 0.7668, + "num_tokens": 31947051.0, + "step": 1468 + }, + { + "epoch": 4.438563327032136, + "grad_norm": 1.5234375, + "learning_rate": 6.53443803789307e-06, + "loss": 0.7947, + "num_tokens": 31967737.0, + "step": 1469 + }, + { + "epoch": 4.441587901701324, + "grad_norm": 1.0625, + "learning_rate": 6.530151382574177e-06, + "loss": 0.9615, + "num_tokens": 32029253.0, + "step": 1470 + }, + { + "epoch": 4.44461247637051, + "grad_norm": 1.6328125, + "learning_rate": 6.525863740542633e-06, + "loss": 0.8931, + "num_tokens": 32052769.0, + "step": 1471 + }, + { + "epoch": 4.447637051039697, + "grad_norm": 1.25, + "learning_rate": 6.521575115905284e-06, + "loss": 0.7618, + "num_tokens": 32082755.0, + "step": 1472 + }, + { + "epoch": 4.450661625708885, + "grad_norm": 1.1953125, + "learning_rate": 6.517285512769915e-06, + "loss": 0.8502, + "num_tokens": 32129946.0, + "step": 1473 + }, + { + "epoch": 4.453686200378072, + "grad_norm": 1.828125, + "learning_rate": 6.5129949352452475e-06, + "loss": 0.564, + "num_tokens": 32137877.0, + "step": 1474 + }, + { + "epoch": 4.456710775047259, + "grad_norm": 2.078125, + "learning_rate": 6.5087033874409354e-06, + "loss": 0.5891, + "num_tokens": 32145390.0, + "step": 1475 + }, + { + "epoch": 4.459735349716446, + "grad_norm": 1.515625, + "learning_rate": 6.504410873467566e-06, + "loss": 0.7634, + "num_tokens": 32168563.0, + "step": 1476 + }, + { + "epoch": 4.462759924385633, + "grad_norm": 1.8359375, + "learning_rate": 6.500117397436647e-06, + "loss": 0.6044, + "num_tokens": 32176895.0, + "step": 1477 + }, + { + "epoch": 4.465784499054821, + "grad_norm": 1.9921875, + "learning_rate": 6.4958229634606115e-06, + "loss": 0.6796, + "num_tokens": 32184701.0, + "step": 1478 + }, + { + "epoch": 4.468809073724008, + "grad_norm": 1.6484375, + "learning_rate": 6.4915275756528085e-06, + "loss": 0.8259, + "num_tokens": 32208227.0, + "step": 1479 + }, + { + "epoch": 4.471833648393194, + "grad_norm": 1.171875, + "learning_rate": 6.487231238127499e-06, + "loss": 0.7854, + "num_tokens": 32244660.0, + "step": 1480 + }, + { + "epoch": 4.474858223062382, + "grad_norm": 1.1796875, + "learning_rate": 6.482933954999858e-06, + "loss": 0.8384, + "num_tokens": 32282507.0, + "step": 1481 + }, + { + "epoch": 4.477882797731569, + "grad_norm": 2.03125, + "learning_rate": 6.478635730385964e-06, + "loss": 0.6443, + "num_tokens": 32290327.0, + "step": 1482 + }, + { + "epoch": 4.480907372400756, + "grad_norm": 1.625, + "learning_rate": 6.474336568402795e-06, + "loss": 0.71, + "num_tokens": 32305916.0, + "step": 1483 + }, + { + "epoch": 4.4839319470699435, + "grad_norm": 1.765625, + "learning_rate": 6.470036473168228e-06, + "loss": 0.6665, + "num_tokens": 32317038.0, + "step": 1484 + }, + { + "epoch": 4.48695652173913, + "grad_norm": 1.328125, + "learning_rate": 6.465735448801037e-06, + "loss": 0.844, + "num_tokens": 32348767.0, + "step": 1485 + }, + { + "epoch": 4.489981096408318, + "grad_norm": 2.15625, + "learning_rate": 6.461433499420882e-06, + "loss": 0.6613, + "num_tokens": 32356026.0, + "step": 1486 + }, + { + "epoch": 4.493005671077505, + "grad_norm": 2.03125, + "learning_rate": 6.457130629148312e-06, + "loss": 0.5326, + "num_tokens": 32362113.0, + "step": 1487 + }, + { + "epoch": 4.496030245746692, + "grad_norm": 1.9375, + "learning_rate": 6.452826842104757e-06, + "loss": 0.6657, + "num_tokens": 32369679.0, + "step": 1488 + }, + { + "epoch": 4.499054820415879, + "grad_norm": 1.5859375, + "learning_rate": 6.448522142412522e-06, + "loss": 0.7697, + "num_tokens": 32386959.0, + "step": 1489 + }, + { + "epoch": 4.502079395085066, + "grad_norm": 2.078125, + "learning_rate": 6.444216534194791e-06, + "loss": 0.5918, + "num_tokens": 32393463.0, + "step": 1490 + }, + { + "epoch": 4.505103969754253, + "grad_norm": 1.421875, + "learning_rate": 6.439910021575617e-06, + "loss": 0.7738, + "num_tokens": 32416496.0, + "step": 1491 + }, + { + "epoch": 4.508128544423441, + "grad_norm": 1.6015625, + "learning_rate": 6.4356026086799176e-06, + "loss": 0.9132, + "num_tokens": 32440392.0, + "step": 1492 + }, + { + "epoch": 4.5111531190926275, + "grad_norm": 1.9140625, + "learning_rate": 6.431294299633473e-06, + "loss": 0.5377, + "num_tokens": 32447172.0, + "step": 1493 + }, + { + "epoch": 4.514177693761814, + "grad_norm": 1.9375, + "learning_rate": 6.426985098562924e-06, + "loss": 0.6011, + "num_tokens": 32454281.0, + "step": 1494 + }, + { + "epoch": 4.517202268431002, + "grad_norm": 1.203125, + "learning_rate": 6.422675009595762e-06, + "loss": 0.8536, + "num_tokens": 32491110.0, + "step": 1495 + }, + { + "epoch": 4.520226843100189, + "grad_norm": 1.9140625, + "learning_rate": 6.418364036860334e-06, + "loss": 0.6313, + "num_tokens": 32502426.0, + "step": 1496 + }, + { + "epoch": 4.523251417769377, + "grad_norm": 1.46875, + "learning_rate": 6.414052184485827e-06, + "loss": 0.8026, + "num_tokens": 32526737.0, + "step": 1497 + }, + { + "epoch": 4.526275992438563, + "grad_norm": 1.7109375, + "learning_rate": 6.409739456602277e-06, + "loss": 0.7479, + "num_tokens": 32540601.0, + "step": 1498 + }, + { + "epoch": 4.52930056710775, + "grad_norm": 1.765625, + "learning_rate": 6.405425857340554e-06, + "loss": 0.5877, + "num_tokens": 32549067.0, + "step": 1499 + }, + { + "epoch": 4.532325141776938, + "grad_norm": 1.4296875, + "learning_rate": 6.4011113908323645e-06, + "loss": 0.7248, + "num_tokens": 32573015.0, + "step": 1500 + }, + { + "epoch": 4.535349716446125, + "grad_norm": 1.984375, + "learning_rate": 6.396796061210245e-06, + "loss": 0.6255, + "num_tokens": 32580221.0, + "step": 1501 + }, + { + "epoch": 4.5383742911153115, + "grad_norm": 1.8828125, + "learning_rate": 6.39247987260756e-06, + "loss": 0.5646, + "num_tokens": 32587725.0, + "step": 1502 + }, + { + "epoch": 4.541398865784499, + "grad_norm": 1.1484375, + "learning_rate": 6.388162829158497e-06, + "loss": 0.8084, + "num_tokens": 32625609.0, + "step": 1503 + }, + { + "epoch": 4.544423440453686, + "grad_norm": 1.40625, + "learning_rate": 6.383844934998058e-06, + "loss": 0.745, + "num_tokens": 32650519.0, + "step": 1504 + }, + { + "epoch": 4.547448015122873, + "grad_norm": 1.46875, + "learning_rate": 6.3795261942620665e-06, + "loss": 0.699, + "num_tokens": 32674569.0, + "step": 1505 + }, + { + "epoch": 4.550472589792061, + "grad_norm": 1.1796875, + "learning_rate": 6.375206611087154e-06, + "loss": 0.7608, + "num_tokens": 32713278.0, + "step": 1506 + }, + { + "epoch": 4.553497164461247, + "grad_norm": 1.625, + "learning_rate": 6.370886189610754e-06, + "loss": 0.7214, + "num_tokens": 32729357.0, + "step": 1507 + }, + { + "epoch": 4.556521739130435, + "grad_norm": 1.40625, + "learning_rate": 6.366564933971111e-06, + "loss": 0.7984, + "num_tokens": 32754736.0, + "step": 1508 + }, + { + "epoch": 4.559546313799622, + "grad_norm": 1.8203125, + "learning_rate": 6.362242848307263e-06, + "loss": 0.6063, + "num_tokens": 32762605.0, + "step": 1509 + }, + { + "epoch": 4.562570888468809, + "grad_norm": 1.984375, + "learning_rate": 6.357919936759046e-06, + "loss": 0.6137, + "num_tokens": 32769331.0, + "step": 1510 + }, + { + "epoch": 4.5655954631379965, + "grad_norm": 1.2109375, + "learning_rate": 6.353596203467085e-06, + "loss": 0.8412, + "num_tokens": 32811489.0, + "step": 1511 + }, + { + "epoch": 4.568620037807183, + "grad_norm": 1.5, + "learning_rate": 6.349271652572794e-06, + "loss": 0.93, + "num_tokens": 32835577.0, + "step": 1512 + }, + { + "epoch": 4.571644612476371, + "grad_norm": 1.2265625, + "learning_rate": 6.344946288218369e-06, + "loss": 0.8662, + "num_tokens": 32874661.0, + "step": 1513 + }, + { + "epoch": 4.574669187145558, + "grad_norm": 1.96875, + "learning_rate": 6.340620114546784e-06, + "loss": 0.6003, + "num_tokens": 32882018.0, + "step": 1514 + }, + { + "epoch": 4.577693761814745, + "grad_norm": 1.53125, + "learning_rate": 6.336293135701791e-06, + "loss": 0.8251, + "num_tokens": 32904675.0, + "step": 1515 + }, + { + "epoch": 4.5807183364839315, + "grad_norm": 1.296875, + "learning_rate": 6.3319653558279115e-06, + "loss": 0.875, + "num_tokens": 32946606.0, + "step": 1516 + }, + { + "epoch": 4.583742911153119, + "grad_norm": 1.2421875, + "learning_rate": 6.3276367790704315e-06, + "loss": 0.8871, + "num_tokens": 32985132.0, + "step": 1517 + }, + { + "epoch": 4.586767485822306, + "grad_norm": 1.234375, + "learning_rate": 6.32330740957541e-06, + "loss": 0.8192, + "num_tokens": 33017621.0, + "step": 1518 + }, + { + "epoch": 4.589792060491494, + "grad_norm": 1.234375, + "learning_rate": 6.318977251489651e-06, + "loss": 0.8661, + "num_tokens": 33056357.0, + "step": 1519 + }, + { + "epoch": 4.5928166351606805, + "grad_norm": 1.4765625, + "learning_rate": 6.314646308960725e-06, + "loss": 0.825, + "num_tokens": 33080360.0, + "step": 1520 + }, + { + "epoch": 4.595841209829867, + "grad_norm": 0.96484375, + "learning_rate": 6.3103145861369505e-06, + "loss": 0.8936, + "num_tokens": 33151017.0, + "step": 1521 + }, + { + "epoch": 4.598865784499055, + "grad_norm": 1.5078125, + "learning_rate": 6.305982087167392e-06, + "loss": 0.8335, + "num_tokens": 33174548.0, + "step": 1522 + }, + { + "epoch": 4.601890359168242, + "grad_norm": 1.15625, + "learning_rate": 6.30164881620186e-06, + "loss": 0.7979, + "num_tokens": 33213997.0, + "step": 1523 + }, + { + "epoch": 4.60491493383743, + "grad_norm": 1.6484375, + "learning_rate": 6.297314777390902e-06, + "loss": 0.747, + "num_tokens": 33230797.0, + "step": 1524 + }, + { + "epoch": 4.607939508506616, + "grad_norm": 1.921875, + "learning_rate": 6.292979974885801e-06, + "loss": 0.6393, + "num_tokens": 33239895.0, + "step": 1525 + }, + { + "epoch": 4.610964083175803, + "grad_norm": 1.5546875, + "learning_rate": 6.2886444128385745e-06, + "loss": 0.7794, + "num_tokens": 33262521.0, + "step": 1526 + }, + { + "epoch": 4.613988657844991, + "grad_norm": 1.5859375, + "learning_rate": 6.284308095401964e-06, + "loss": 0.8792, + "num_tokens": 33286138.0, + "step": 1527 + }, + { + "epoch": 4.617013232514178, + "grad_norm": 2.0, + "learning_rate": 6.279971026729437e-06, + "loss": 0.5573, + "num_tokens": 33293498.0, + "step": 1528 + }, + { + "epoch": 4.6200378071833645, + "grad_norm": 1.84375, + "learning_rate": 6.275633210975179e-06, + "loss": 0.5844, + "num_tokens": 33301182.0, + "step": 1529 + }, + { + "epoch": 4.623062381852552, + "grad_norm": 1.8203125, + "learning_rate": 6.271294652294091e-06, + "loss": 0.7919, + "num_tokens": 33317068.0, + "step": 1530 + }, + { + "epoch": 4.626086956521739, + "grad_norm": 1.203125, + "learning_rate": 6.266955354841788e-06, + "loss": 0.8983, + "num_tokens": 33355604.0, + "step": 1531 + }, + { + "epoch": 4.629111531190926, + "grad_norm": 1.390625, + "learning_rate": 6.26261532277459e-06, + "loss": 0.816, + "num_tokens": 33383877.0, + "step": 1532 + }, + { + "epoch": 4.632136105860114, + "grad_norm": 1.4921875, + "learning_rate": 6.25827456024952e-06, + "loss": 0.7561, + "num_tokens": 33405411.0, + "step": 1533 + }, + { + "epoch": 4.6351606805293, + "grad_norm": 1.953125, + "learning_rate": 6.253933071424304e-06, + "loss": 0.5286, + "num_tokens": 33412653.0, + "step": 1534 + }, + { + "epoch": 4.638185255198488, + "grad_norm": 1.03125, + "learning_rate": 6.249590860457362e-06, + "loss": 0.8788, + "num_tokens": 33474136.0, + "step": 1535 + }, + { + "epoch": 4.641209829867675, + "grad_norm": 1.765625, + "learning_rate": 6.245247931507804e-06, + "loss": 0.6518, + "num_tokens": 33485269.0, + "step": 1536 + }, + { + "epoch": 4.644234404536862, + "grad_norm": 1.2109375, + "learning_rate": 6.24090428873543e-06, + "loss": 0.7871, + "num_tokens": 33517545.0, + "step": 1537 + }, + { + "epoch": 4.6472589792060495, + "grad_norm": 1.3203125, + "learning_rate": 6.236559936300723e-06, + "loss": 0.8973, + "num_tokens": 33556437.0, + "step": 1538 + }, + { + "epoch": 4.650283553875236, + "grad_norm": 1.15625, + "learning_rate": 6.232214878364846e-06, + "loss": 0.8626, + "num_tokens": 33598135.0, + "step": 1539 + }, + { + "epoch": 4.653308128544423, + "grad_norm": 2.15625, + "learning_rate": 6.227869119089638e-06, + "loss": 0.5871, + "num_tokens": 33605952.0, + "step": 1540 + }, + { + "epoch": 4.656332703213611, + "grad_norm": 1.4375, + "learning_rate": 6.2235226626376075e-06, + "loss": 0.8084, + "num_tokens": 33630042.0, + "step": 1541 + }, + { + "epoch": 4.659357277882798, + "grad_norm": 1.53125, + "learning_rate": 6.219175513171933e-06, + "loss": 0.8902, + "num_tokens": 33653355.0, + "step": 1542 + }, + { + "epoch": 4.6623818525519845, + "grad_norm": 1.3125, + "learning_rate": 6.214827674856458e-06, + "loss": 0.8104, + "num_tokens": 33687887.0, + "step": 1543 + }, + { + "epoch": 4.665406427221172, + "grad_norm": 1.1640625, + "learning_rate": 6.210479151855681e-06, + "loss": 0.7738, + "num_tokens": 33726132.0, + "step": 1544 + }, + { + "epoch": 4.668431001890359, + "grad_norm": 1.4140625, + "learning_rate": 6.206129948334762e-06, + "loss": 0.8299, + "num_tokens": 33751046.0, + "step": 1545 + }, + { + "epoch": 4.671455576559547, + "grad_norm": 1.5234375, + "learning_rate": 6.201780068459511e-06, + "loss": 0.7362, + "num_tokens": 33772538.0, + "step": 1546 + }, + { + "epoch": 4.6744801512287335, + "grad_norm": 1.5625, + "learning_rate": 6.19742951639638e-06, + "loss": 0.6858, + "num_tokens": 33786554.0, + "step": 1547 + }, + { + "epoch": 4.67750472589792, + "grad_norm": 1.203125, + "learning_rate": 6.193078296312476e-06, + "loss": 0.8876, + "num_tokens": 33825937.0, + "step": 1548 + }, + { + "epoch": 4.680529300567108, + "grad_norm": 1.8203125, + "learning_rate": 6.188726412375536e-06, + "loss": 0.7448, + "num_tokens": 33839108.0, + "step": 1549 + }, + { + "epoch": 4.683553875236295, + "grad_norm": 1.640625, + "learning_rate": 6.184373868753938e-06, + "loss": 0.6681, + "num_tokens": 33852889.0, + "step": 1550 + }, + { + "epoch": 4.686578449905482, + "grad_norm": 1.75, + "learning_rate": 6.180020669616689e-06, + "loss": 0.6067, + "num_tokens": 33863744.0, + "step": 1551 + }, + { + "epoch": 4.689603024574669, + "grad_norm": 1.3671875, + "learning_rate": 6.175666819133427e-06, + "loss": 0.8075, + "num_tokens": 33895652.0, + "step": 1552 + }, + { + "epoch": 4.692627599243856, + "grad_norm": 1.6015625, + "learning_rate": 6.171312321474413e-06, + "loss": 0.8619, + "num_tokens": 33915780.0, + "step": 1553 + }, + { + "epoch": 4.695652173913043, + "grad_norm": 1.515625, + "learning_rate": 6.166957180810525e-06, + "loss": 0.8443, + "num_tokens": 33938878.0, + "step": 1554 + }, + { + "epoch": 4.698676748582231, + "grad_norm": 1.8671875, + "learning_rate": 6.162601401313259e-06, + "loss": 0.5716, + "num_tokens": 33946404.0, + "step": 1555 + }, + { + "epoch": 4.7017013232514175, + "grad_norm": 1.421875, + "learning_rate": 6.1582449871547245e-06, + "loss": 0.8733, + "num_tokens": 33974651.0, + "step": 1556 + }, + { + "epoch": 4.704725897920605, + "grad_norm": 1.671875, + "learning_rate": 6.1538879425076355e-06, + "loss": 0.7504, + "num_tokens": 33992056.0, + "step": 1557 + }, + { + "epoch": 4.707750472589792, + "grad_norm": 1.3984375, + "learning_rate": 6.149530271545314e-06, + "loss": 0.847, + "num_tokens": 34019437.0, + "step": 1558 + }, + { + "epoch": 4.710775047258979, + "grad_norm": 1.296875, + "learning_rate": 6.1451719784416775e-06, + "loss": 0.7526, + "num_tokens": 34053503.0, + "step": 1559 + }, + { + "epoch": 4.713799621928167, + "grad_norm": 1.828125, + "learning_rate": 6.140813067371243e-06, + "loss": 0.7106, + "num_tokens": 34064803.0, + "step": 1560 + }, + { + "epoch": 4.716824196597353, + "grad_norm": 1.8671875, + "learning_rate": 6.1364535425091175e-06, + "loss": 0.5652, + "num_tokens": 34072609.0, + "step": 1561 + }, + { + "epoch": 4.719848771266541, + "grad_norm": 1.4140625, + "learning_rate": 6.132093408030995e-06, + "loss": 0.7746, + "num_tokens": 34096349.0, + "step": 1562 + }, + { + "epoch": 4.722873345935728, + "grad_norm": 1.640625, + "learning_rate": 6.127732668113158e-06, + "loss": 0.8541, + "num_tokens": 34116431.0, + "step": 1563 + }, + { + "epoch": 4.725897920604915, + "grad_norm": 1.53125, + "learning_rate": 6.123371326932461e-06, + "loss": 0.8514, + "num_tokens": 34140262.0, + "step": 1564 + }, + { + "epoch": 4.7289224952741025, + "grad_norm": 2.03125, + "learning_rate": 6.119009388666344e-06, + "loss": 0.6429, + "num_tokens": 34148722.0, + "step": 1565 + }, + { + "epoch": 4.731947069943289, + "grad_norm": 1.65625, + "learning_rate": 6.11464685749281e-06, + "loss": 0.7482, + "num_tokens": 34170852.0, + "step": 1566 + }, + { + "epoch": 4.734971644612476, + "grad_norm": 1.765625, + "learning_rate": 6.110283737590436e-06, + "loss": 0.7791, + "num_tokens": 34186330.0, + "step": 1567 + }, + { + "epoch": 4.737996219281664, + "grad_norm": 1.5625, + "learning_rate": 6.105920033138361e-06, + "loss": 0.8598, + "num_tokens": 34209254.0, + "step": 1568 + }, + { + "epoch": 4.741020793950851, + "grad_norm": 1.8828125, + "learning_rate": 6.101555748316281e-06, + "loss": 0.6021, + "num_tokens": 34217063.0, + "step": 1569 + }, + { + "epoch": 4.7440453686200375, + "grad_norm": 1.5078125, + "learning_rate": 6.097190887304455e-06, + "loss": 0.7827, + "num_tokens": 34237838.0, + "step": 1570 + }, + { + "epoch": 4.747069943289225, + "grad_norm": 1.5078125, + "learning_rate": 6.0928254542836855e-06, + "loss": 0.7906, + "num_tokens": 34259608.0, + "step": 1571 + }, + { + "epoch": 4.750094517958412, + "grad_norm": 1.3203125, + "learning_rate": 6.088459453435328e-06, + "loss": 0.8412, + "num_tokens": 34292566.0, + "step": 1572 + }, + { + "epoch": 4.7531190926276, + "grad_norm": 1.2265625, + "learning_rate": 6.084092888941282e-06, + "loss": 0.887, + "num_tokens": 34329741.0, + "step": 1573 + }, + { + "epoch": 4.7561436672967865, + "grad_norm": 1.890625, + "learning_rate": 6.079725764983985e-06, + "loss": 0.5817, + "num_tokens": 34337781.0, + "step": 1574 + }, + { + "epoch": 4.759168241965973, + "grad_norm": 2.015625, + "learning_rate": 6.07535808574641e-06, + "loss": 0.6566, + "num_tokens": 34344812.0, + "step": 1575 + }, + { + "epoch": 4.762192816635161, + "grad_norm": 1.484375, + "learning_rate": 6.070989855412065e-06, + "loss": 0.8547, + "num_tokens": 34374632.0, + "step": 1576 + }, + { + "epoch": 4.765217391304348, + "grad_norm": 1.875, + "learning_rate": 6.066621078164979e-06, + "loss": 0.6513, + "num_tokens": 34382252.0, + "step": 1577 + }, + { + "epoch": 4.768241965973535, + "grad_norm": 1.4453125, + "learning_rate": 6.062251758189717e-06, + "loss": 0.7699, + "num_tokens": 34404708.0, + "step": 1578 + }, + { + "epoch": 4.771266540642722, + "grad_norm": 1.6640625, + "learning_rate": 6.0578818996713496e-06, + "loss": 0.7725, + "num_tokens": 34427782.0, + "step": 1579 + }, + { + "epoch": 4.774291115311909, + "grad_norm": 1.8984375, + "learning_rate": 6.053511506795474e-06, + "loss": 0.6211, + "num_tokens": 34436343.0, + "step": 1580 + }, + { + "epoch": 4.777315689981096, + "grad_norm": 1.875, + "learning_rate": 6.049140583748193e-06, + "loss": 0.6562, + "num_tokens": 34446685.0, + "step": 1581 + }, + { + "epoch": 4.780340264650284, + "grad_norm": 1.5859375, + "learning_rate": 6.044769134716121e-06, + "loss": 0.7838, + "num_tokens": 34467222.0, + "step": 1582 + }, + { + "epoch": 4.7833648393194705, + "grad_norm": 1.375, + "learning_rate": 6.040397163886376e-06, + "loss": 0.8201, + "num_tokens": 34492431.0, + "step": 1583 + }, + { + "epoch": 4.786389413988658, + "grad_norm": 1.453125, + "learning_rate": 6.036024675446571e-06, + "loss": 0.7079, + "num_tokens": 34514209.0, + "step": 1584 + }, + { + "epoch": 4.789413988657845, + "grad_norm": 2.109375, + "learning_rate": 6.031651673584819e-06, + "loss": 0.6422, + "num_tokens": 34521458.0, + "step": 1585 + }, + { + "epoch": 4.792438563327032, + "grad_norm": 1.8671875, + "learning_rate": 6.0272781624897266e-06, + "loss": 0.6115, + "num_tokens": 34529450.0, + "step": 1586 + }, + { + "epoch": 4.79546313799622, + "grad_norm": 1.421875, + "learning_rate": 6.022904146350384e-06, + "loss": 0.8797, + "num_tokens": 34569095.0, + "step": 1587 + }, + { + "epoch": 4.798487712665406, + "grad_norm": 1.34375, + "learning_rate": 6.018529629356367e-06, + "loss": 0.9002, + "num_tokens": 34608476.0, + "step": 1588 + }, + { + "epoch": 4.801512287334593, + "grad_norm": 1.9375, + "learning_rate": 6.014154615697729e-06, + "loss": 0.5842, + "num_tokens": 34616438.0, + "step": 1589 + }, + { + "epoch": 4.804536862003781, + "grad_norm": 1.515625, + "learning_rate": 6.009779109565005e-06, + "loss": 0.7629, + "num_tokens": 34640189.0, + "step": 1590 + }, + { + "epoch": 4.807561436672968, + "grad_norm": 1.578125, + "learning_rate": 6.005403115149194e-06, + "loss": 0.6899, + "num_tokens": 34663896.0, + "step": 1591 + }, + { + "epoch": 4.810586011342155, + "grad_norm": 1.7421875, + "learning_rate": 6.001026636641768e-06, + "loss": 0.6888, + "num_tokens": 34680036.0, + "step": 1592 + }, + { + "epoch": 4.813610586011342, + "grad_norm": 1.453125, + "learning_rate": 5.996649678234663e-06, + "loss": 0.7307, + "num_tokens": 34699025.0, + "step": 1593 + }, + { + "epoch": 4.816635160680529, + "grad_norm": 1.375, + "learning_rate": 5.9922722441202676e-06, + "loss": 0.8334, + "num_tokens": 34737419.0, + "step": 1594 + }, + { + "epoch": 4.819659735349717, + "grad_norm": 1.234375, + "learning_rate": 5.987894338491438e-06, + "loss": 0.8579, + "num_tokens": 34776490.0, + "step": 1595 + }, + { + "epoch": 4.822684310018904, + "grad_norm": 1.328125, + "learning_rate": 5.983515965541467e-06, + "loss": 0.841, + "num_tokens": 34805156.0, + "step": 1596 + }, + { + "epoch": 4.8257088846880904, + "grad_norm": 1.96875, + "learning_rate": 5.979137129464111e-06, + "loss": 0.5871, + "num_tokens": 34812624.0, + "step": 1597 + }, + { + "epoch": 4.828733459357278, + "grad_norm": 2.0, + "learning_rate": 5.974757834453559e-06, + "loss": 0.6368, + "num_tokens": 34820039.0, + "step": 1598 + }, + { + "epoch": 4.831758034026465, + "grad_norm": 1.4453125, + "learning_rate": 5.970378084704441e-06, + "loss": 0.7311, + "num_tokens": 34839161.0, + "step": 1599 + }, + { + "epoch": 4.834782608695652, + "grad_norm": 1.796875, + "learning_rate": 5.965997884411828e-06, + "loss": 0.5355, + "num_tokens": 34847988.0, + "step": 1600 + }, + { + "epoch": 4.8378071833648395, + "grad_norm": 1.734375, + "learning_rate": 5.961617237771217e-06, + "loss": 0.8624, + "num_tokens": 34869582.0, + "step": 1601 + }, + { + "epoch": 4.840831758034026, + "grad_norm": 1.5390625, + "learning_rate": 5.957236148978534e-06, + "loss": 0.8668, + "num_tokens": 34892597.0, + "step": 1602 + }, + { + "epoch": 4.843856332703213, + "grad_norm": 2.15625, + "learning_rate": 5.9528546222301305e-06, + "loss": 0.6, + "num_tokens": 34899436.0, + "step": 1603 + }, + { + "epoch": 4.846880907372401, + "grad_norm": 1.3828125, + "learning_rate": 5.948472661722774e-06, + "loss": 0.7878, + "num_tokens": 34925092.0, + "step": 1604 + }, + { + "epoch": 4.849905482041588, + "grad_norm": 1.90625, + "learning_rate": 5.944090271653652e-06, + "loss": 0.6402, + "num_tokens": 34935590.0, + "step": 1605 + }, + { + "epoch": 4.852930056710775, + "grad_norm": 1.9609375, + "learning_rate": 5.939707456220358e-06, + "loss": 0.6096, + "num_tokens": 34945832.0, + "step": 1606 + }, + { + "epoch": 4.855954631379962, + "grad_norm": 1.3671875, + "learning_rate": 5.935324219620897e-06, + "loss": 0.8369, + "num_tokens": 34978551.0, + "step": 1607 + }, + { + "epoch": 4.858979206049149, + "grad_norm": 1.90625, + "learning_rate": 5.930940566053679e-06, + "loss": 0.6797, + "num_tokens": 34987035.0, + "step": 1608 + }, + { + "epoch": 4.862003780718337, + "grad_norm": 1.671875, + "learning_rate": 5.926556499717506e-06, + "loss": 0.7679, + "num_tokens": 35001502.0, + "step": 1609 + }, + { + "epoch": 4.8650283553875235, + "grad_norm": 1.8359375, + "learning_rate": 5.922172024811583e-06, + "loss": 0.8647, + "num_tokens": 35018506.0, + "step": 1610 + }, + { + "epoch": 4.868052930056711, + "grad_norm": 1.7265625, + "learning_rate": 5.917787145535502e-06, + "loss": 0.766, + "num_tokens": 35034685.0, + "step": 1611 + }, + { + "epoch": 4.871077504725898, + "grad_norm": 2.078125, + "learning_rate": 5.913401866089245e-06, + "loss": 0.5622, + "num_tokens": 35040921.0, + "step": 1612 + }, + { + "epoch": 4.874102079395085, + "grad_norm": 1.5859375, + "learning_rate": 5.909016190673173e-06, + "loss": 0.7448, + "num_tokens": 35062815.0, + "step": 1613 + }, + { + "epoch": 4.877126654064273, + "grad_norm": 2.1875, + "learning_rate": 5.904630123488031e-06, + "loss": 0.5753, + "num_tokens": 35069353.0, + "step": 1614 + }, + { + "epoch": 4.880151228733459, + "grad_norm": 1.09375, + "learning_rate": 5.900243668734939e-06, + "loss": 0.8904, + "num_tokens": 35117311.0, + "step": 1615 + }, + { + "epoch": 4.883175803402646, + "grad_norm": 1.9296875, + "learning_rate": 5.895856830615384e-06, + "loss": 0.6412, + "num_tokens": 35128694.0, + "step": 1616 + }, + { + "epoch": 4.886200378071834, + "grad_norm": 1.3203125, + "learning_rate": 5.891469613331225e-06, + "loss": 0.8676, + "num_tokens": 35160758.0, + "step": 1617 + }, + { + "epoch": 4.889224952741021, + "grad_norm": 1.65625, + "learning_rate": 5.88708202108468e-06, + "loss": 0.6946, + "num_tokens": 35174841.0, + "step": 1618 + }, + { + "epoch": 4.892249527410208, + "grad_norm": 1.4453125, + "learning_rate": 5.88269405807833e-06, + "loss": 0.8494, + "num_tokens": 35200724.0, + "step": 1619 + }, + { + "epoch": 4.895274102079395, + "grad_norm": 1.5078125, + "learning_rate": 5.878305728515109e-06, + "loss": 0.7946, + "num_tokens": 35222796.0, + "step": 1620 + }, + { + "epoch": 4.898298676748582, + "grad_norm": 1.4140625, + "learning_rate": 5.873917036598301e-06, + "loss": 0.6796, + "num_tokens": 35248392.0, + "step": 1621 + }, + { + "epoch": 4.90132325141777, + "grad_norm": 1.3125, + "learning_rate": 5.86952798653154e-06, + "loss": 0.8469, + "num_tokens": 35285469.0, + "step": 1622 + }, + { + "epoch": 4.904347826086957, + "grad_norm": 1.5703125, + "learning_rate": 5.865138582518802e-06, + "loss": 0.8808, + "num_tokens": 35309038.0, + "step": 1623 + }, + { + "epoch": 4.9073724007561434, + "grad_norm": 1.4453125, + "learning_rate": 5.8607488287644e-06, + "loss": 0.7053, + "num_tokens": 35331750.0, + "step": 1624 + }, + { + "epoch": 4.910396975425331, + "grad_norm": 2.109375, + "learning_rate": 5.856358729472984e-06, + "loss": 0.5528, + "num_tokens": 35338456.0, + "step": 1625 + }, + { + "epoch": 4.913421550094518, + "grad_norm": 1.78125, + "learning_rate": 5.851968288849535e-06, + "loss": 0.7162, + "num_tokens": 35350707.0, + "step": 1626 + }, + { + "epoch": 4.916446124763705, + "grad_norm": 2.140625, + "learning_rate": 5.84757751109936e-06, + "loss": 0.5759, + "num_tokens": 35357943.0, + "step": 1627 + }, + { + "epoch": 4.9194706994328925, + "grad_norm": 1.3671875, + "learning_rate": 5.8431864004280905e-06, + "loss": 0.9034, + "num_tokens": 35397318.0, + "step": 1628 + }, + { + "epoch": 4.922495274102079, + "grad_norm": 1.4609375, + "learning_rate": 5.838794961041673e-06, + "loss": 0.7743, + "num_tokens": 35422401.0, + "step": 1629 + }, + { + "epoch": 4.925519848771266, + "grad_norm": 1.671875, + "learning_rate": 5.834403197146374e-06, + "loss": 0.7711, + "num_tokens": 35439577.0, + "step": 1630 + }, + { + "epoch": 4.928544423440454, + "grad_norm": 1.375, + "learning_rate": 5.830011112948768e-06, + "loss": 0.888, + "num_tokens": 35470715.0, + "step": 1631 + }, + { + "epoch": 4.931568998109641, + "grad_norm": 1.2421875, + "learning_rate": 5.825618712655737e-06, + "loss": 0.8766, + "num_tokens": 35510807.0, + "step": 1632 + }, + { + "epoch": 4.934593572778828, + "grad_norm": 1.2734375, + "learning_rate": 5.821226000474466e-06, + "loss": 0.82, + "num_tokens": 35549643.0, + "step": 1633 + }, + { + "epoch": 4.937618147448015, + "grad_norm": 1.453125, + "learning_rate": 5.816832980612435e-06, + "loss": 0.8282, + "num_tokens": 35573005.0, + "step": 1634 + }, + { + "epoch": 4.940642722117202, + "grad_norm": 1.2578125, + "learning_rate": 5.812439657277427e-06, + "loss": 0.8246, + "num_tokens": 35611843.0, + "step": 1635 + }, + { + "epoch": 4.94366729678639, + "grad_norm": 1.9375, + "learning_rate": 5.808046034677506e-06, + "loss": 0.6141, + "num_tokens": 35619266.0, + "step": 1636 + }, + { + "epoch": 4.9466918714555765, + "grad_norm": 1.8046875, + "learning_rate": 5.803652117021029e-06, + "loss": 0.6789, + "num_tokens": 35630359.0, + "step": 1637 + }, + { + "epoch": 4.949716446124763, + "grad_norm": 1.6171875, + "learning_rate": 5.799257908516636e-06, + "loss": 0.7788, + "num_tokens": 35646114.0, + "step": 1638 + }, + { + "epoch": 4.952741020793951, + "grad_norm": 1.3203125, + "learning_rate": 5.7948634133732395e-06, + "loss": 0.8746, + "num_tokens": 35688303.0, + "step": 1639 + }, + { + "epoch": 4.955765595463138, + "grad_norm": 1.1015625, + "learning_rate": 5.7904686358000325e-06, + "loss": 0.9178, + "num_tokens": 35744204.0, + "step": 1640 + }, + { + "epoch": 4.958790170132325, + "grad_norm": 2.109375, + "learning_rate": 5.786073580006478e-06, + "loss": 0.6657, + "num_tokens": 35750975.0, + "step": 1641 + }, + { + "epoch": 4.961814744801512, + "grad_norm": 1.4921875, + "learning_rate": 5.7816782502023025e-06, + "loss": 0.8176, + "num_tokens": 35773426.0, + "step": 1642 + }, + { + "epoch": 4.964839319470699, + "grad_norm": 1.8828125, + "learning_rate": 5.777282650597496e-06, + "loss": 0.7589, + "num_tokens": 35786089.0, + "step": 1643 + }, + { + "epoch": 4.967863894139887, + "grad_norm": 1.5703125, + "learning_rate": 5.772886785402308e-06, + "loss": 0.8267, + "num_tokens": 35806447.0, + "step": 1644 + }, + { + "epoch": 4.970888468809074, + "grad_norm": 1.265625, + "learning_rate": 5.768490658827242e-06, + "loss": 0.7782, + "num_tokens": 35839054.0, + "step": 1645 + }, + { + "epoch": 4.973913043478261, + "grad_norm": 1.9296875, + "learning_rate": 5.764094275083051e-06, + "loss": 0.5606, + "num_tokens": 35846601.0, + "step": 1646 + }, + { + "epoch": 4.976937618147448, + "grad_norm": 1.4921875, + "learning_rate": 5.759697638380736e-06, + "loss": 0.6928, + "num_tokens": 35864752.0, + "step": 1647 + }, + { + "epoch": 4.979962192816635, + "grad_norm": 1.609375, + "learning_rate": 5.755300752931537e-06, + "loss": 0.6941, + "num_tokens": 35882279.0, + "step": 1648 + }, + { + "epoch": 4.982986767485822, + "grad_norm": 2.0625, + "learning_rate": 5.750903622946938e-06, + "loss": 0.651, + "num_tokens": 35888879.0, + "step": 1649 + }, + { + "epoch": 4.98601134215501, + "grad_norm": 1.296875, + "learning_rate": 5.74650625263865e-06, + "loss": 0.7636, + "num_tokens": 35917273.0, + "step": 1650 + }, + { + "epoch": 4.9890359168241964, + "grad_norm": 1.96875, + "learning_rate": 5.742108646218622e-06, + "loss": 0.5739, + "num_tokens": 35925077.0, + "step": 1651 + }, + { + "epoch": 4.992060491493383, + "grad_norm": 1.5078125, + "learning_rate": 5.737710807899023e-06, + "loss": 0.7309, + "num_tokens": 35944112.0, + "step": 1652 + }, + { + "epoch": 4.995085066162571, + "grad_norm": 1.7109375, + "learning_rate": 5.733312741892245e-06, + "loss": 0.6903, + "num_tokens": 35956041.0, + "step": 1653 + }, + { + "epoch": 4.998109640831758, + "grad_norm": 1.8359375, + "learning_rate": 5.728914452410902e-06, + "loss": 0.6096, + "num_tokens": 35963741.0, + "step": 1654 + }, + { + "epoch": 5.0, + "grad_norm": 1.8359375, + "learning_rate": 5.724515943667818e-06, + "loss": 0.7217, + "num_tokens": 35976565.0, + "step": 1655 + }, + { + "epoch": 5.003024574669187, + "grad_norm": 1.890625, + "learning_rate": 5.720117219876025e-06, + "loss": 0.6646, + "num_tokens": 35985013.0, + "step": 1656 + }, + { + "epoch": 5.0060491493383745, + "grad_norm": 1.109375, + "learning_rate": 5.7157182852487705e-06, + "loss": 0.9183, + "num_tokens": 36033401.0, + "step": 1657 + }, + { + "epoch": 5.009073724007561, + "grad_norm": 1.828125, + "learning_rate": 5.711319143999494e-06, + "loss": 0.6978, + "num_tokens": 36044110.0, + "step": 1658 + }, + { + "epoch": 5.012098298676748, + "grad_norm": 1.09375, + "learning_rate": 5.706919800341837e-06, + "loss": 0.7915, + "num_tokens": 36088936.0, + "step": 1659 + }, + { + "epoch": 5.015122873345936, + "grad_norm": 1.59375, + "learning_rate": 5.702520258489636e-06, + "loss": 0.7337, + "num_tokens": 36110427.0, + "step": 1660 + }, + { + "epoch": 5.018147448015123, + "grad_norm": 1.59375, + "learning_rate": 5.698120522656916e-06, + "loss": 0.7341, + "num_tokens": 36129114.0, + "step": 1661 + }, + { + "epoch": 5.02117202268431, + "grad_norm": 1.453125, + "learning_rate": 5.693720597057886e-06, + "loss": 0.8482, + "num_tokens": 36152536.0, + "step": 1662 + }, + { + "epoch": 5.024196597353497, + "grad_norm": 1.8359375, + "learning_rate": 5.68932048590694e-06, + "loss": 0.61, + "num_tokens": 36160278.0, + "step": 1663 + }, + { + "epoch": 5.027221172022684, + "grad_norm": 1.390625, + "learning_rate": 5.684920193418646e-06, + "loss": 0.7152, + "num_tokens": 36183709.0, + "step": 1664 + }, + { + "epoch": 5.030245746691872, + "grad_norm": 1.203125, + "learning_rate": 5.6805197238077515e-06, + "loss": 0.8173, + "num_tokens": 36220457.0, + "step": 1665 + }, + { + "epoch": 5.033270321361059, + "grad_norm": 1.4140625, + "learning_rate": 5.676119081289167e-06, + "loss": 0.83, + "num_tokens": 36246658.0, + "step": 1666 + }, + { + "epoch": 5.036294896030245, + "grad_norm": 1.5625, + "learning_rate": 5.671718270077971e-06, + "loss": 0.8532, + "num_tokens": 36269774.0, + "step": 1667 + }, + { + "epoch": 5.039319470699433, + "grad_norm": 1.359375, + "learning_rate": 5.667317294389408e-06, + "loss": 0.732, + "num_tokens": 36293398.0, + "step": 1668 + }, + { + "epoch": 5.04234404536862, + "grad_norm": 1.5859375, + "learning_rate": 5.662916158438874e-06, + "loss": 0.8987, + "num_tokens": 36315890.0, + "step": 1669 + }, + { + "epoch": 5.045368620037808, + "grad_norm": 1.921875, + "learning_rate": 5.65851486644192e-06, + "loss": 0.5845, + "num_tokens": 36323217.0, + "step": 1670 + }, + { + "epoch": 5.048393194706994, + "grad_norm": 1.625, + "learning_rate": 5.654113422614249e-06, + "loss": 0.7332, + "num_tokens": 36338155.0, + "step": 1671 + }, + { + "epoch": 5.051417769376181, + "grad_norm": 1.578125, + "learning_rate": 5.649711831171707e-06, + "loss": 0.7788, + "num_tokens": 36359057.0, + "step": 1672 + }, + { + "epoch": 5.054442344045369, + "grad_norm": 1.4453125, + "learning_rate": 5.645310096330281e-06, + "loss": 0.7806, + "num_tokens": 36383659.0, + "step": 1673 + }, + { + "epoch": 5.057466918714556, + "grad_norm": 1.171875, + "learning_rate": 5.640908222306098e-06, + "loss": 0.8403, + "num_tokens": 36426175.0, + "step": 1674 + }, + { + "epoch": 5.060491493383743, + "grad_norm": 1.2890625, + "learning_rate": 5.636506213315417e-06, + "loss": 0.8614, + "num_tokens": 36458610.0, + "step": 1675 + }, + { + "epoch": 5.06351606805293, + "grad_norm": 1.328125, + "learning_rate": 5.632104073574624e-06, + "loss": 0.7533, + "num_tokens": 36483871.0, + "step": 1676 + }, + { + "epoch": 5.066540642722117, + "grad_norm": 1.5703125, + "learning_rate": 5.627701807300234e-06, + "loss": 0.7965, + "num_tokens": 36507010.0, + "step": 1677 + }, + { + "epoch": 5.069565217391304, + "grad_norm": 1.9296875, + "learning_rate": 5.623299418708882e-06, + "loss": 0.622, + "num_tokens": 36514619.0, + "step": 1678 + }, + { + "epoch": 5.072589792060492, + "grad_norm": 1.3828125, + "learning_rate": 5.618896912017318e-06, + "loss": 0.7333, + "num_tokens": 36537168.0, + "step": 1679 + }, + { + "epoch": 5.0756143667296785, + "grad_norm": 1.453125, + "learning_rate": 5.614494291442407e-06, + "loss": 0.8219, + "num_tokens": 36561496.0, + "step": 1680 + }, + { + "epoch": 5.078638941398866, + "grad_norm": 1.4296875, + "learning_rate": 5.610091561201123e-06, + "loss": 0.78, + "num_tokens": 36587852.0, + "step": 1681 + }, + { + "epoch": 5.081663516068053, + "grad_norm": 1.390625, + "learning_rate": 5.605688725510545e-06, + "loss": 0.8367, + "num_tokens": 36617267.0, + "step": 1682 + }, + { + "epoch": 5.08468809073724, + "grad_norm": 1.7265625, + "learning_rate": 5.601285788587853e-06, + "loss": 0.5944, + "num_tokens": 36628968.0, + "step": 1683 + }, + { + "epoch": 5.0877126654064275, + "grad_norm": 1.40625, + "learning_rate": 5.5968827546503244e-06, + "loss": 0.8366, + "num_tokens": 36659675.0, + "step": 1684 + }, + { + "epoch": 5.090737240075614, + "grad_norm": 1.578125, + "learning_rate": 5.592479627915329e-06, + "loss": 0.6992, + "num_tokens": 36677810.0, + "step": 1685 + }, + { + "epoch": 5.093761814744801, + "grad_norm": 1.9375, + "learning_rate": 5.588076412600321e-06, + "loss": 0.5965, + "num_tokens": 36685232.0, + "step": 1686 + }, + { + "epoch": 5.096786389413989, + "grad_norm": 1.546875, + "learning_rate": 5.58367311292285e-06, + "loss": 0.7429, + "num_tokens": 36703989.0, + "step": 1687 + }, + { + "epoch": 5.099810964083176, + "grad_norm": 1.9921875, + "learning_rate": 5.57926973310054e-06, + "loss": 0.5743, + "num_tokens": 36711534.0, + "step": 1688 + }, + { + "epoch": 5.1028355387523625, + "grad_norm": 1.9296875, + "learning_rate": 5.574866277351087e-06, + "loss": 0.5318, + "num_tokens": 36718993.0, + "step": 1689 + }, + { + "epoch": 5.10586011342155, + "grad_norm": 1.5390625, + "learning_rate": 5.5704627498922685e-06, + "loss": 0.8583, + "num_tokens": 36741864.0, + "step": 1690 + }, + { + "epoch": 5.108884688090737, + "grad_norm": 1.6796875, + "learning_rate": 5.566059154941925e-06, + "loss": 0.694, + "num_tokens": 36754657.0, + "step": 1691 + }, + { + "epoch": 5.111909262759925, + "grad_norm": 2.109375, + "learning_rate": 5.561655496717966e-06, + "loss": 0.5803, + "num_tokens": 36761761.0, + "step": 1692 + }, + { + "epoch": 5.114933837429112, + "grad_norm": 1.5546875, + "learning_rate": 5.557251779438359e-06, + "loss": 0.738, + "num_tokens": 36778131.0, + "step": 1693 + }, + { + "epoch": 5.117958412098298, + "grad_norm": 1.4140625, + "learning_rate": 5.552848007321125e-06, + "loss": 0.7793, + "num_tokens": 36806407.0, + "step": 1694 + }, + { + "epoch": 5.120982986767486, + "grad_norm": 1.78125, + "learning_rate": 5.5484441845843455e-06, + "loss": 0.617, + "num_tokens": 36814730.0, + "step": 1695 + }, + { + "epoch": 5.124007561436673, + "grad_norm": 1.75, + "learning_rate": 5.544040315446142e-06, + "loss": 0.5309, + "num_tokens": 36822477.0, + "step": 1696 + }, + { + "epoch": 5.12703213610586, + "grad_norm": 1.4609375, + "learning_rate": 5.539636404124684e-06, + "loss": 0.849, + "num_tokens": 36847779.0, + "step": 1697 + }, + { + "epoch": 5.130056710775047, + "grad_norm": 1.9765625, + "learning_rate": 5.535232454838185e-06, + "loss": 0.6357, + "num_tokens": 36855385.0, + "step": 1698 + }, + { + "epoch": 5.133081285444234, + "grad_norm": 1.828125, + "learning_rate": 5.53082847180489e-06, + "loss": 0.564, + "num_tokens": 36863145.0, + "step": 1699 + }, + { + "epoch": 5.136105860113422, + "grad_norm": 1.6640625, + "learning_rate": 5.526424459243077e-06, + "loss": 0.7304, + "num_tokens": 36879012.0, + "step": 1700 + }, + { + "epoch": 5.139130434782609, + "grad_norm": 2.109375, + "learning_rate": 5.522020421371053e-06, + "loss": 0.6462, + "num_tokens": 36886004.0, + "step": 1701 + }, + { + "epoch": 5.142155009451796, + "grad_norm": 1.59375, + "learning_rate": 5.517616362407151e-06, + "loss": 0.7612, + "num_tokens": 36903132.0, + "step": 1702 + }, + { + "epoch": 5.145179584120983, + "grad_norm": 1.8046875, + "learning_rate": 5.513212286569721e-06, + "loss": 0.6813, + "num_tokens": 36917018.0, + "step": 1703 + }, + { + "epoch": 5.14820415879017, + "grad_norm": 1.53125, + "learning_rate": 5.508808198077132e-06, + "loss": 0.8363, + "num_tokens": 36940516.0, + "step": 1704 + }, + { + "epoch": 5.151228733459357, + "grad_norm": 1.375, + "learning_rate": 5.504404101147762e-06, + "loss": 0.6917, + "num_tokens": 36961421.0, + "step": 1705 + }, + { + "epoch": 5.154253308128545, + "grad_norm": 1.171875, + "learning_rate": 5.500000000000001e-06, + "loss": 0.8151, + "num_tokens": 37000907.0, + "step": 1706 + }, + { + "epoch": 5.1572778827977315, + "grad_norm": 1.046875, + "learning_rate": 5.495595898852241e-06, + "loss": 0.8667, + "num_tokens": 37055684.0, + "step": 1707 + }, + { + "epoch": 5.160302457466918, + "grad_norm": 1.6484375, + "learning_rate": 5.49119180192287e-06, + "loss": 0.8875, + "num_tokens": 37079463.0, + "step": 1708 + }, + { + "epoch": 5.163327032136106, + "grad_norm": 1.75, + "learning_rate": 5.48678771343028e-06, + "loss": 0.5747, + "num_tokens": 37087985.0, + "step": 1709 + }, + { + "epoch": 5.166351606805293, + "grad_norm": 1.5546875, + "learning_rate": 5.482383637592851e-06, + "loss": 0.696, + "num_tokens": 37105107.0, + "step": 1710 + }, + { + "epoch": 5.1693761814744805, + "grad_norm": 1.234375, + "learning_rate": 5.477979578628949e-06, + "loss": 0.9022, + "num_tokens": 37145427.0, + "step": 1711 + }, + { + "epoch": 5.172400756143667, + "grad_norm": 1.59375, + "learning_rate": 5.473575540756925e-06, + "loss": 0.8506, + "num_tokens": 37168847.0, + "step": 1712 + }, + { + "epoch": 5.175425330812854, + "grad_norm": 1.7265625, + "learning_rate": 5.469171528195112e-06, + "loss": 0.5227, + "num_tokens": 37176888.0, + "step": 1713 + }, + { + "epoch": 5.178449905482042, + "grad_norm": 1.8125, + "learning_rate": 5.464767545161817e-06, + "loss": 0.5183, + "num_tokens": 37184517.0, + "step": 1714 + }, + { + "epoch": 5.181474480151229, + "grad_norm": 1.4453125, + "learning_rate": 5.4603635958753175e-06, + "loss": 0.8856, + "num_tokens": 37218416.0, + "step": 1715 + }, + { + "epoch": 5.1844990548204155, + "grad_norm": 1.7890625, + "learning_rate": 5.455959684553861e-06, + "loss": 0.563, + "num_tokens": 37226034.0, + "step": 1716 + }, + { + "epoch": 5.187523629489603, + "grad_norm": 1.171875, + "learning_rate": 5.451555815415658e-06, + "loss": 0.862, + "num_tokens": 37269578.0, + "step": 1717 + }, + { + "epoch": 5.19054820415879, + "grad_norm": 1.359375, + "learning_rate": 5.447151992678876e-06, + "loss": 0.8459, + "num_tokens": 37309090.0, + "step": 1718 + }, + { + "epoch": 5.193572778827978, + "grad_norm": 1.6796875, + "learning_rate": 5.442748220561643e-06, + "loss": 0.8016, + "num_tokens": 37328920.0, + "step": 1719 + }, + { + "epoch": 5.196597353497165, + "grad_norm": 1.46875, + "learning_rate": 5.438344503282036e-06, + "loss": 0.7929, + "num_tokens": 37355676.0, + "step": 1720 + }, + { + "epoch": 5.199621928166351, + "grad_norm": 1.4296875, + "learning_rate": 5.433940845058076e-06, + "loss": 0.828, + "num_tokens": 37380138.0, + "step": 1721 + }, + { + "epoch": 5.202646502835539, + "grad_norm": 2.125, + "learning_rate": 5.429537250107733e-06, + "loss": 0.6597, + "num_tokens": 37386865.0, + "step": 1722 + }, + { + "epoch": 5.205671077504726, + "grad_norm": 1.9609375, + "learning_rate": 5.425133722648915e-06, + "loss": 0.6764, + "num_tokens": 37395165.0, + "step": 1723 + }, + { + "epoch": 5.208695652173913, + "grad_norm": 1.796875, + "learning_rate": 5.420730266899463e-06, + "loss": 0.8559, + "num_tokens": 37418130.0, + "step": 1724 + }, + { + "epoch": 5.2117202268431, + "grad_norm": 1.3203125, + "learning_rate": 5.41632688707715e-06, + "loss": 0.8479, + "num_tokens": 37449740.0, + "step": 1725 + }, + { + "epoch": 5.214744801512287, + "grad_norm": 1.453125, + "learning_rate": 5.41192358739968e-06, + "loss": 0.7788, + "num_tokens": 37473752.0, + "step": 1726 + }, + { + "epoch": 5.217769376181474, + "grad_norm": 1.15625, + "learning_rate": 5.407520372084675e-06, + "loss": 0.805, + "num_tokens": 37510932.0, + "step": 1727 + }, + { + "epoch": 5.220793950850662, + "grad_norm": 1.578125, + "learning_rate": 5.403117245349676e-06, + "loss": 0.8714, + "num_tokens": 37534480.0, + "step": 1728 + }, + { + "epoch": 5.223818525519849, + "grad_norm": 1.328125, + "learning_rate": 5.398714211412148e-06, + "loss": 0.8555, + "num_tokens": 37568306.0, + "step": 1729 + }, + { + "epoch": 5.226843100189036, + "grad_norm": 2.15625, + "learning_rate": 5.3943112744894565e-06, + "loss": 0.6273, + "num_tokens": 37574759.0, + "step": 1730 + }, + { + "epoch": 5.229867674858223, + "grad_norm": 1.4765625, + "learning_rate": 5.38990843879888e-06, + "loss": 0.8665, + "num_tokens": 37597671.0, + "step": 1731 + }, + { + "epoch": 5.23289224952741, + "grad_norm": 1.21875, + "learning_rate": 5.3855057085575945e-06, + "loss": 0.9099, + "num_tokens": 37645522.0, + "step": 1732 + }, + { + "epoch": 5.235916824196598, + "grad_norm": 1.65625, + "learning_rate": 5.381103087982684e-06, + "loss": 0.7202, + "num_tokens": 37659118.0, + "step": 1733 + }, + { + "epoch": 5.2389413988657845, + "grad_norm": 1.90625, + "learning_rate": 5.376700581291121e-06, + "loss": 0.5801, + "num_tokens": 37666404.0, + "step": 1734 + }, + { + "epoch": 5.241965973534971, + "grad_norm": 1.0625, + "learning_rate": 5.3722981926997665e-06, + "loss": 0.788, + "num_tokens": 37708006.0, + "step": 1735 + }, + { + "epoch": 5.244990548204159, + "grad_norm": 1.8515625, + "learning_rate": 5.367895926425377e-06, + "loss": 0.5865, + "num_tokens": 37715763.0, + "step": 1736 + }, + { + "epoch": 5.248015122873346, + "grad_norm": 1.4765625, + "learning_rate": 5.363493786684586e-06, + "loss": 0.7754, + "num_tokens": 37739198.0, + "step": 1737 + }, + { + "epoch": 5.251039697542533, + "grad_norm": 1.140625, + "learning_rate": 5.359091777693902e-06, + "loss": 0.8761, + "num_tokens": 37785348.0, + "step": 1738 + }, + { + "epoch": 5.25406427221172, + "grad_norm": 1.8515625, + "learning_rate": 5.354689903669721e-06, + "loss": 0.6158, + "num_tokens": 37793161.0, + "step": 1739 + }, + { + "epoch": 5.257088846880907, + "grad_norm": 1.3203125, + "learning_rate": 5.3502881688282946e-06, + "loss": 0.938, + "num_tokens": 37832404.0, + "step": 1740 + }, + { + "epoch": 5.260113421550095, + "grad_norm": 1.4140625, + "learning_rate": 5.345886577385754e-06, + "loss": 0.7912, + "num_tokens": 37855448.0, + "step": 1741 + }, + { + "epoch": 5.263137996219282, + "grad_norm": 1.609375, + "learning_rate": 5.341485133558082e-06, + "loss": 0.8915, + "num_tokens": 37878476.0, + "step": 1742 + }, + { + "epoch": 5.2661625708884685, + "grad_norm": 1.890625, + "learning_rate": 5.337083841561128e-06, + "loss": 0.783, + "num_tokens": 37891463.0, + "step": 1743 + }, + { + "epoch": 5.269187145557656, + "grad_norm": 1.7109375, + "learning_rate": 5.332682705610596e-06, + "loss": 0.8052, + "num_tokens": 37915738.0, + "step": 1744 + }, + { + "epoch": 5.272211720226843, + "grad_norm": 1.8359375, + "learning_rate": 5.3282817299220305e-06, + "loss": 0.5405, + "num_tokens": 37922869.0, + "step": 1745 + }, + { + "epoch": 5.27523629489603, + "grad_norm": 1.390625, + "learning_rate": 5.3238809187108355e-06, + "loss": 0.8166, + "num_tokens": 37953268.0, + "step": 1746 + }, + { + "epoch": 5.278260869565218, + "grad_norm": 1.25, + "learning_rate": 5.319480276192252e-06, + "loss": 0.7707, + "num_tokens": 37984577.0, + "step": 1747 + }, + { + "epoch": 5.281285444234404, + "grad_norm": 1.15625, + "learning_rate": 5.315079806581355e-06, + "loss": 0.7779, + "num_tokens": 38022581.0, + "step": 1748 + }, + { + "epoch": 5.284310018903592, + "grad_norm": 2.046875, + "learning_rate": 5.310679514093061e-06, + "loss": 0.621, + "num_tokens": 38030068.0, + "step": 1749 + }, + { + "epoch": 5.287334593572779, + "grad_norm": 1.4375, + "learning_rate": 5.306279402942116e-06, + "loss": 0.7764, + "num_tokens": 38053836.0, + "step": 1750 + }, + { + "epoch": 5.290359168241966, + "grad_norm": 2.078125, + "learning_rate": 5.301879477343086e-06, + "loss": 0.5138, + "num_tokens": 38060021.0, + "step": 1751 + }, + { + "epoch": 5.293383742911153, + "grad_norm": 1.390625, + "learning_rate": 5.297479741510364e-06, + "loss": 0.8465, + "num_tokens": 38087101.0, + "step": 1752 + }, + { + "epoch": 5.29640831758034, + "grad_norm": 1.8828125, + "learning_rate": 5.293080199658164e-06, + "loss": 0.8361, + "num_tokens": 38105014.0, + "step": 1753 + }, + { + "epoch": 5.299432892249527, + "grad_norm": 1.9453125, + "learning_rate": 5.288680856000507e-06, + "loss": 0.6083, + "num_tokens": 38112356.0, + "step": 1754 + }, + { + "epoch": 5.302457466918715, + "grad_norm": 1.1484375, + "learning_rate": 5.28428171475123e-06, + "loss": 0.8921, + "num_tokens": 38158625.0, + "step": 1755 + }, + { + "epoch": 5.305482041587902, + "grad_norm": 1.609375, + "learning_rate": 5.279882780123976e-06, + "loss": 0.746, + "num_tokens": 38182561.0, + "step": 1756 + }, + { + "epoch": 5.308506616257089, + "grad_norm": 2.09375, + "learning_rate": 5.2754840563321855e-06, + "loss": 0.5846, + "num_tokens": 38189302.0, + "step": 1757 + }, + { + "epoch": 5.311531190926276, + "grad_norm": 1.984375, + "learning_rate": 5.2710855475891e-06, + "loss": 0.5732, + "num_tokens": 38196270.0, + "step": 1758 + }, + { + "epoch": 5.314555765595463, + "grad_norm": 2.140625, + "learning_rate": 5.266687258107756e-06, + "loss": 0.6581, + "num_tokens": 38204058.0, + "step": 1759 + }, + { + "epoch": 5.317580340264651, + "grad_norm": 1.9921875, + "learning_rate": 5.262289192100979e-06, + "loss": 0.5981, + "num_tokens": 38211287.0, + "step": 1760 + }, + { + "epoch": 5.3206049149338375, + "grad_norm": 1.4921875, + "learning_rate": 5.257891353781381e-06, + "loss": 0.8406, + "num_tokens": 38238876.0, + "step": 1761 + }, + { + "epoch": 5.323629489603024, + "grad_norm": 2.1875, + "learning_rate": 5.25349374736135e-06, + "loss": 0.6359, + "num_tokens": 38245667.0, + "step": 1762 + }, + { + "epoch": 5.326654064272212, + "grad_norm": 1.8828125, + "learning_rate": 5.249096377053064e-06, + "loss": 0.5936, + "num_tokens": 38253770.0, + "step": 1763 + }, + { + "epoch": 5.329678638941399, + "grad_norm": 1.390625, + "learning_rate": 5.244699247068465e-06, + "loss": 0.8276, + "num_tokens": 38281680.0, + "step": 1764 + }, + { + "epoch": 5.332703213610586, + "grad_norm": 1.9453125, + "learning_rate": 5.2403023616192675e-06, + "loss": 0.5929, + "num_tokens": 38289289.0, + "step": 1765 + }, + { + "epoch": 5.335727788279773, + "grad_norm": 1.3828125, + "learning_rate": 5.235905724916952e-06, + "loss": 0.7895, + "num_tokens": 38311164.0, + "step": 1766 + }, + { + "epoch": 5.33875236294896, + "grad_norm": 1.125, + "learning_rate": 5.231509341172761e-06, + "loss": 0.9025, + "num_tokens": 38361493.0, + "step": 1767 + }, + { + "epoch": 5.341776937618148, + "grad_norm": 1.2421875, + "learning_rate": 5.227113214597693e-06, + "loss": 0.882, + "num_tokens": 38403395.0, + "step": 1768 + }, + { + "epoch": 5.344801512287335, + "grad_norm": 1.8515625, + "learning_rate": 5.222717349402506e-06, + "loss": 0.5834, + "num_tokens": 38410868.0, + "step": 1769 + }, + { + "epoch": 5.3478260869565215, + "grad_norm": 1.796875, + "learning_rate": 5.218321749797699e-06, + "loss": 0.5313, + "num_tokens": 38418701.0, + "step": 1770 + }, + { + "epoch": 5.350850661625709, + "grad_norm": 1.7734375, + "learning_rate": 5.213926419993524e-06, + "loss": 0.527, + "num_tokens": 38426443.0, + "step": 1771 + }, + { + "epoch": 5.353875236294896, + "grad_norm": 1.265625, + "learning_rate": 5.209531364199968e-06, + "loss": 0.8347, + "num_tokens": 38463968.0, + "step": 1772 + }, + { + "epoch": 5.356899810964083, + "grad_norm": 1.59375, + "learning_rate": 5.205136586626762e-06, + "loss": 0.7929, + "num_tokens": 38480697.0, + "step": 1773 + }, + { + "epoch": 5.359924385633271, + "grad_norm": 1.1484375, + "learning_rate": 5.200742091483367e-06, + "loss": 0.7094, + "num_tokens": 38520401.0, + "step": 1774 + }, + { + "epoch": 5.362948960302457, + "grad_norm": 1.90625, + "learning_rate": 5.196347882978971e-06, + "loss": 0.5942, + "num_tokens": 38527701.0, + "step": 1775 + }, + { + "epoch": 5.365973534971644, + "grad_norm": 1.5625, + "learning_rate": 5.1919539653224944e-06, + "loss": 0.7747, + "num_tokens": 38551851.0, + "step": 1776 + }, + { + "epoch": 5.368998109640832, + "grad_norm": 1.34375, + "learning_rate": 5.187560342722576e-06, + "loss": 0.7018, + "num_tokens": 38574480.0, + "step": 1777 + }, + { + "epoch": 5.372022684310019, + "grad_norm": 1.96875, + "learning_rate": 5.183167019387565e-06, + "loss": 0.6282, + "num_tokens": 38581206.0, + "step": 1778 + }, + { + "epoch": 5.375047258979206, + "grad_norm": 1.9140625, + "learning_rate": 5.178773999525536e-06, + "loss": 0.621, + "num_tokens": 38588865.0, + "step": 1779 + }, + { + "epoch": 5.378071833648393, + "grad_norm": 1.6953125, + "learning_rate": 5.174381287344264e-06, + "loss": 0.7393, + "num_tokens": 38603615.0, + "step": 1780 + }, + { + "epoch": 5.38109640831758, + "grad_norm": 1.609375, + "learning_rate": 5.169988887051234e-06, + "loss": 0.8481, + "num_tokens": 38626427.0, + "step": 1781 + }, + { + "epoch": 5.384120982986768, + "grad_norm": 1.3828125, + "learning_rate": 5.165596802853627e-06, + "loss": 0.7535, + "num_tokens": 38649094.0, + "step": 1782 + }, + { + "epoch": 5.387145557655955, + "grad_norm": 1.2109375, + "learning_rate": 5.161205038958329e-06, + "loss": 0.786, + "num_tokens": 38689033.0, + "step": 1783 + }, + { + "epoch": 5.390170132325141, + "grad_norm": 1.9453125, + "learning_rate": 5.156813599571913e-06, + "loss": 0.5719, + "num_tokens": 38696024.0, + "step": 1784 + }, + { + "epoch": 5.393194706994329, + "grad_norm": 1.296875, + "learning_rate": 5.152422488900641e-06, + "loss": 0.8606, + "num_tokens": 38727827.0, + "step": 1785 + }, + { + "epoch": 5.396219281663516, + "grad_norm": 1.390625, + "learning_rate": 5.148031711150468e-06, + "loss": 0.8636, + "num_tokens": 38759870.0, + "step": 1786 + }, + { + "epoch": 5.399243856332703, + "grad_norm": 1.9140625, + "learning_rate": 5.143641270527018e-06, + "loss": 0.5716, + "num_tokens": 38767367.0, + "step": 1787 + }, + { + "epoch": 5.4022684310018905, + "grad_norm": 1.375, + "learning_rate": 5.139251171235601e-06, + "loss": 0.8403, + "num_tokens": 38795262.0, + "step": 1788 + }, + { + "epoch": 5.405293005671077, + "grad_norm": 1.96875, + "learning_rate": 5.134861417481199e-06, + "loss": 0.7069, + "num_tokens": 38805657.0, + "step": 1789 + }, + { + "epoch": 5.408317580340265, + "grad_norm": 1.4375, + "learning_rate": 5.130472013468461e-06, + "loss": 0.8155, + "num_tokens": 38829599.0, + "step": 1790 + }, + { + "epoch": 5.411342155009452, + "grad_norm": 1.390625, + "learning_rate": 5.1260829634017e-06, + "loss": 0.8517, + "num_tokens": 38862001.0, + "step": 1791 + }, + { + "epoch": 5.414366729678639, + "grad_norm": 2.25, + "learning_rate": 5.121694271484893e-06, + "loss": 0.5431, + "num_tokens": 38868911.0, + "step": 1792 + }, + { + "epoch": 5.417391304347826, + "grad_norm": 1.4453125, + "learning_rate": 5.117305941921672e-06, + "loss": 0.8078, + "num_tokens": 38892677.0, + "step": 1793 + }, + { + "epoch": 5.420415879017013, + "grad_norm": 1.9609375, + "learning_rate": 5.112917978915323e-06, + "loss": 0.6367, + "num_tokens": 38900249.0, + "step": 1794 + }, + { + "epoch": 5.4234404536862, + "grad_norm": 1.3984375, + "learning_rate": 5.108530386668778e-06, + "loss": 0.7521, + "num_tokens": 38924485.0, + "step": 1795 + }, + { + "epoch": 5.426465028355388, + "grad_norm": 1.109375, + "learning_rate": 5.1041431693846175e-06, + "loss": 0.8751, + "num_tokens": 38972855.0, + "step": 1796 + }, + { + "epoch": 5.4294896030245745, + "grad_norm": 1.8671875, + "learning_rate": 5.0997563312650645e-06, + "loss": 0.623, + "num_tokens": 38980484.0, + "step": 1797 + }, + { + "epoch": 5.432514177693762, + "grad_norm": 1.1484375, + "learning_rate": 5.0953698765119695e-06, + "loss": 0.8438, + "num_tokens": 39019624.0, + "step": 1798 + }, + { + "epoch": 5.435538752362949, + "grad_norm": 2.125, + "learning_rate": 5.0909838093268294e-06, + "loss": 0.6029, + "num_tokens": 39026346.0, + "step": 1799 + }, + { + "epoch": 5.438563327032136, + "grad_norm": 1.5859375, + "learning_rate": 5.086598133910758e-06, + "loss": 0.7975, + "num_tokens": 39048766.0, + "step": 1800 + }, + { + "epoch": 5.441587901701324, + "grad_norm": 1.46875, + "learning_rate": 5.0822128544645015e-06, + "loss": 0.8546, + "num_tokens": 39072395.0, + "step": 1801 + }, + { + "epoch": 5.44461247637051, + "grad_norm": 1.046875, + "learning_rate": 5.0778279751884185e-06, + "loss": 0.8405, + "num_tokens": 39130613.0, + "step": 1802 + }, + { + "epoch": 5.447637051039697, + "grad_norm": 1.4765625, + "learning_rate": 5.073443500282496e-06, + "loss": 0.8472, + "num_tokens": 39161142.0, + "step": 1803 + }, + { + "epoch": 5.450661625708885, + "grad_norm": 1.8046875, + "learning_rate": 5.069059433946324e-06, + "loss": 0.6113, + "num_tokens": 39171476.0, + "step": 1804 + }, + { + "epoch": 5.453686200378072, + "grad_norm": 1.4609375, + "learning_rate": 5.064675780379104e-06, + "loss": 0.7816, + "num_tokens": 39194408.0, + "step": 1805 + }, + { + "epoch": 5.456710775047259, + "grad_norm": 0.90234375, + "learning_rate": 5.060292543779643e-06, + "loss": 0.8442, + "num_tokens": 39257961.0, + "step": 1806 + }, + { + "epoch": 5.459735349716446, + "grad_norm": 1.1875, + "learning_rate": 5.055909728346351e-06, + "loss": 0.9171, + "num_tokens": 39304909.0, + "step": 1807 + }, + { + "epoch": 5.462759924385633, + "grad_norm": 1.8125, + "learning_rate": 5.051527338277227e-06, + "loss": 0.5799, + "num_tokens": 39312884.0, + "step": 1808 + }, + { + "epoch": 5.465784499054821, + "grad_norm": 1.7578125, + "learning_rate": 5.047145377769871e-06, + "loss": 0.8783, + "num_tokens": 39335925.0, + "step": 1809 + }, + { + "epoch": 5.468809073724008, + "grad_norm": 1.9765625, + "learning_rate": 5.0427638510214686e-06, + "loss": 0.6638, + "num_tokens": 39345557.0, + "step": 1810 + }, + { + "epoch": 5.471833648393194, + "grad_norm": 1.953125, + "learning_rate": 5.038382762228786e-06, + "loss": 0.5823, + "num_tokens": 39352745.0, + "step": 1811 + }, + { + "epoch": 5.474858223062382, + "grad_norm": 1.3984375, + "learning_rate": 5.034002115588173e-06, + "loss": 0.7109, + "num_tokens": 39376595.0, + "step": 1812 + }, + { + "epoch": 5.477882797731569, + "grad_norm": 1.53125, + "learning_rate": 5.02962191529556e-06, + "loss": 0.666, + "num_tokens": 39394795.0, + "step": 1813 + }, + { + "epoch": 5.480907372400756, + "grad_norm": 1.421875, + "learning_rate": 5.025242165546443e-06, + "loss": 0.7588, + "num_tokens": 39416907.0, + "step": 1814 + }, + { + "epoch": 5.4839319470699435, + "grad_norm": 1.5546875, + "learning_rate": 5.02086287053589e-06, + "loss": 0.7552, + "num_tokens": 39435892.0, + "step": 1815 + }, + { + "epoch": 5.48695652173913, + "grad_norm": 1.765625, + "learning_rate": 5.016484034458534e-06, + "loss": 0.7005, + "num_tokens": 39447581.0, + "step": 1816 + }, + { + "epoch": 5.489981096408318, + "grad_norm": 1.5390625, + "learning_rate": 5.012105661508566e-06, + "loss": 0.8608, + "num_tokens": 39471582.0, + "step": 1817 + }, + { + "epoch": 5.493005671077505, + "grad_norm": 1.4140625, + "learning_rate": 5.007727755879732e-06, + "loss": 0.7411, + "num_tokens": 39494804.0, + "step": 1818 + }, + { + "epoch": 5.496030245746692, + "grad_norm": 2.078125, + "learning_rate": 5.00335032176534e-06, + "loss": 0.6158, + "num_tokens": 39501674.0, + "step": 1819 + }, + { + "epoch": 5.499054820415879, + "grad_norm": 1.3359375, + "learning_rate": 4.998973363358233e-06, + "loss": 0.7957, + "num_tokens": 39530879.0, + "step": 1820 + }, + { + "epoch": 5.502079395085066, + "grad_norm": 1.9140625, + "learning_rate": 4.994596884850809e-06, + "loss": 0.5997, + "num_tokens": 39539492.0, + "step": 1821 + }, + { + "epoch": 5.505103969754253, + "grad_norm": 1.5546875, + "learning_rate": 4.990220890434997e-06, + "loss": 0.9501, + "num_tokens": 39574975.0, + "step": 1822 + }, + { + "epoch": 5.508128544423441, + "grad_norm": 2.046875, + "learning_rate": 4.985845384302271e-06, + "loss": 0.5578, + "num_tokens": 39582530.0, + "step": 1823 + }, + { + "epoch": 5.5111531190926275, + "grad_norm": 1.109375, + "learning_rate": 4.981470370643636e-06, + "loss": 0.881, + "num_tokens": 39632437.0, + "step": 1824 + }, + { + "epoch": 5.514177693761814, + "grad_norm": 1.21875, + "learning_rate": 4.977095853649617e-06, + "loss": 0.8728, + "num_tokens": 39671976.0, + "step": 1825 + }, + { + "epoch": 5.517202268431002, + "grad_norm": 1.8671875, + "learning_rate": 4.972721837510275e-06, + "loss": 0.5809, + "num_tokens": 39679846.0, + "step": 1826 + }, + { + "epoch": 5.520226843100189, + "grad_norm": 1.515625, + "learning_rate": 4.968348326415183e-06, + "loss": 0.7071, + "num_tokens": 39702858.0, + "step": 1827 + }, + { + "epoch": 5.523251417769377, + "grad_norm": 1.5703125, + "learning_rate": 4.96397532455343e-06, + "loss": 0.6975, + "num_tokens": 39719532.0, + "step": 1828 + }, + { + "epoch": 5.526275992438563, + "grad_norm": 1.8125, + "learning_rate": 4.9596028361136265e-06, + "loss": 0.5672, + "num_tokens": 39727176.0, + "step": 1829 + }, + { + "epoch": 5.52930056710775, + "grad_norm": 1.40625, + "learning_rate": 4.955230865283881e-06, + "loss": 0.7367, + "num_tokens": 39755000.0, + "step": 1830 + }, + { + "epoch": 5.532325141776938, + "grad_norm": 1.203125, + "learning_rate": 4.95085941625181e-06, + "loss": 0.8782, + "num_tokens": 39795892.0, + "step": 1831 + }, + { + "epoch": 5.535349716446125, + "grad_norm": 1.7890625, + "learning_rate": 4.946488493204528e-06, + "loss": 0.5593, + "num_tokens": 39805169.0, + "step": 1832 + }, + { + "epoch": 5.5383742911153115, + "grad_norm": 1.9453125, + "learning_rate": 4.942118100328651e-06, + "loss": 0.6032, + "num_tokens": 39813371.0, + "step": 1833 + }, + { + "epoch": 5.541398865784499, + "grad_norm": 1.3125, + "learning_rate": 4.937748241810286e-06, + "loss": 0.719, + "num_tokens": 39837358.0, + "step": 1834 + }, + { + "epoch": 5.544423440453686, + "grad_norm": 1.640625, + "learning_rate": 4.933378921835021e-06, + "loss": 0.7475, + "num_tokens": 39851967.0, + "step": 1835 + }, + { + "epoch": 5.547448015122873, + "grad_norm": 1.765625, + "learning_rate": 4.9290101445879375e-06, + "loss": 0.606, + "num_tokens": 39860122.0, + "step": 1836 + }, + { + "epoch": 5.550472589792061, + "grad_norm": 1.359375, + "learning_rate": 4.924641914253593e-06, + "loss": 0.9196, + "num_tokens": 39898470.0, + "step": 1837 + }, + { + "epoch": 5.553497164461247, + "grad_norm": 1.1953125, + "learning_rate": 4.920274235016016e-06, + "loss": 0.8496, + "num_tokens": 39938696.0, + "step": 1838 + }, + { + "epoch": 5.556521739130435, + "grad_norm": 1.5390625, + "learning_rate": 4.915907111058719e-06, + "loss": 0.8333, + "num_tokens": 39962283.0, + "step": 1839 + }, + { + "epoch": 5.559546313799622, + "grad_norm": 1.484375, + "learning_rate": 4.9115405465646735e-06, + "loss": 0.8292, + "num_tokens": 39985998.0, + "step": 1840 + }, + { + "epoch": 5.562570888468809, + "grad_norm": 1.515625, + "learning_rate": 4.907174545716317e-06, + "loss": 0.7458, + "num_tokens": 40009869.0, + "step": 1841 + }, + { + "epoch": 5.5655954631379965, + "grad_norm": 1.46875, + "learning_rate": 4.902809112695546e-06, + "loss": 0.7438, + "num_tokens": 40033655.0, + "step": 1842 + }, + { + "epoch": 5.568620037807183, + "grad_norm": 1.890625, + "learning_rate": 4.898444251683721e-06, + "loss": 0.5714, + "num_tokens": 40041273.0, + "step": 1843 + }, + { + "epoch": 5.571644612476371, + "grad_norm": 1.9609375, + "learning_rate": 4.8940799668616415e-06, + "loss": 0.5679, + "num_tokens": 40048384.0, + "step": 1844 + }, + { + "epoch": 5.574669187145558, + "grad_norm": 1.484375, + "learning_rate": 4.889716262409565e-06, + "loss": 0.7638, + "num_tokens": 40071238.0, + "step": 1845 + }, + { + "epoch": 5.577693761814745, + "grad_norm": 1.984375, + "learning_rate": 4.885353142507192e-06, + "loss": 0.6615, + "num_tokens": 40079828.0, + "step": 1846 + }, + { + "epoch": 5.5807183364839315, + "grad_norm": 1.5390625, + "learning_rate": 4.8809906113336584e-06, + "loss": 0.8974, + "num_tokens": 40103372.0, + "step": 1847 + }, + { + "epoch": 5.583742911153119, + "grad_norm": 1.9296875, + "learning_rate": 4.876628673067538e-06, + "loss": 0.5752, + "num_tokens": 40110459.0, + "step": 1848 + }, + { + "epoch": 5.586767485822306, + "grad_norm": 1.234375, + "learning_rate": 4.8722673318868444e-06, + "loss": 0.8476, + "num_tokens": 40147285.0, + "step": 1849 + }, + { + "epoch": 5.589792060491494, + "grad_norm": 1.21875, + "learning_rate": 4.867906591969006e-06, + "loss": 0.8825, + "num_tokens": 40192109.0, + "step": 1850 + }, + { + "epoch": 5.5928166351606805, + "grad_norm": 1.484375, + "learning_rate": 4.863546457490886e-06, + "loss": 0.7526, + "num_tokens": 40214909.0, + "step": 1851 + }, + { + "epoch": 5.595841209829867, + "grad_norm": 1.5703125, + "learning_rate": 4.859186932628758e-06, + "loss": 0.7549, + "num_tokens": 40237813.0, + "step": 1852 + }, + { + "epoch": 5.598865784499055, + "grad_norm": 1.453125, + "learning_rate": 4.854828021558323e-06, + "loss": 0.7619, + "num_tokens": 40261193.0, + "step": 1853 + }, + { + "epoch": 5.601890359168242, + "grad_norm": 1.7265625, + "learning_rate": 4.850469728454689e-06, + "loss": 0.7411, + "num_tokens": 40276310.0, + "step": 1854 + }, + { + "epoch": 5.60491493383743, + "grad_norm": 1.6640625, + "learning_rate": 4.846112057492366e-06, + "loss": 0.589, + "num_tokens": 40288042.0, + "step": 1855 + }, + { + "epoch": 5.607939508506616, + "grad_norm": 1.265625, + "learning_rate": 4.841755012845277e-06, + "loss": 0.8568, + "num_tokens": 40328099.0, + "step": 1856 + }, + { + "epoch": 5.610964083175803, + "grad_norm": 1.296875, + "learning_rate": 4.837398598686744e-06, + "loss": 0.8485, + "num_tokens": 40360959.0, + "step": 1857 + }, + { + "epoch": 5.613988657844991, + "grad_norm": 1.796875, + "learning_rate": 4.833042819189478e-06, + "loss": 0.5771, + "num_tokens": 40369248.0, + "step": 1858 + }, + { + "epoch": 5.617013232514178, + "grad_norm": 2.03125, + "learning_rate": 4.8286876785255895e-06, + "loss": 0.5722, + "num_tokens": 40376743.0, + "step": 1859 + }, + { + "epoch": 5.6200378071833645, + "grad_norm": 1.375, + "learning_rate": 4.824333180866575e-06, + "loss": 0.8409, + "num_tokens": 40406448.0, + "step": 1860 + }, + { + "epoch": 5.623062381852552, + "grad_norm": 1.5234375, + "learning_rate": 4.8199793303833135e-06, + "loss": 0.7765, + "num_tokens": 40429654.0, + "step": 1861 + }, + { + "epoch": 5.626086956521739, + "grad_norm": 1.1796875, + "learning_rate": 4.815626131246064e-06, + "loss": 0.8718, + "num_tokens": 40472215.0, + "step": 1862 + }, + { + "epoch": 5.629111531190926, + "grad_norm": 1.65625, + "learning_rate": 4.811273587624466e-06, + "loss": 0.8135, + "num_tokens": 40492551.0, + "step": 1863 + }, + { + "epoch": 5.632136105860114, + "grad_norm": 1.3046875, + "learning_rate": 4.806921703687526e-06, + "loss": 0.8104, + "num_tokens": 40525755.0, + "step": 1864 + }, + { + "epoch": 5.6351606805293, + "grad_norm": 1.03125, + "learning_rate": 4.80257048360362e-06, + "loss": 0.8546, + "num_tokens": 40583365.0, + "step": 1865 + }, + { + "epoch": 5.638185255198488, + "grad_norm": 1.7734375, + "learning_rate": 4.7982199315404905e-06, + "loss": 0.6434, + "num_tokens": 40593651.0, + "step": 1866 + }, + { + "epoch": 5.641209829867675, + "grad_norm": 1.59375, + "learning_rate": 4.793870051665239e-06, + "loss": 0.84, + "num_tokens": 40617010.0, + "step": 1867 + }, + { + "epoch": 5.644234404536862, + "grad_norm": 1.9296875, + "learning_rate": 4.789520848144319e-06, + "loss": 0.5875, + "num_tokens": 40624483.0, + "step": 1868 + }, + { + "epoch": 5.6472589792060495, + "grad_norm": 2.15625, + "learning_rate": 4.785172325143543e-06, + "loss": 0.6264, + "num_tokens": 40631667.0, + "step": 1869 + }, + { + "epoch": 5.650283553875236, + "grad_norm": 1.9453125, + "learning_rate": 4.780824486828069e-06, + "loss": 0.6538, + "num_tokens": 40639850.0, + "step": 1870 + }, + { + "epoch": 5.653308128544423, + "grad_norm": 1.234375, + "learning_rate": 4.776477337362394e-06, + "loss": 0.8632, + "num_tokens": 40675606.0, + "step": 1871 + }, + { + "epoch": 5.656332703213611, + "grad_norm": 1.8671875, + "learning_rate": 4.772130880910363e-06, + "loss": 0.6658, + "num_tokens": 40688430.0, + "step": 1872 + }, + { + "epoch": 5.659357277882798, + "grad_norm": 1.3046875, + "learning_rate": 4.7677851216351545e-06, + "loss": 0.8796, + "num_tokens": 40720474.0, + "step": 1873 + }, + { + "epoch": 5.6623818525519845, + "grad_norm": 1.484375, + "learning_rate": 4.763440063699278e-06, + "loss": 0.806, + "num_tokens": 40743384.0, + "step": 1874 + }, + { + "epoch": 5.665406427221172, + "grad_norm": 1.53125, + "learning_rate": 4.75909571126457e-06, + "loss": 0.8498, + "num_tokens": 40765732.0, + "step": 1875 + }, + { + "epoch": 5.668431001890359, + "grad_norm": 1.3125, + "learning_rate": 4.754752068492198e-06, + "loss": 0.8093, + "num_tokens": 40806378.0, + "step": 1876 + }, + { + "epoch": 5.671455576559547, + "grad_norm": 1.515625, + "learning_rate": 4.75040913954264e-06, + "loss": 0.7505, + "num_tokens": 40826469.0, + "step": 1877 + }, + { + "epoch": 5.6744801512287335, + "grad_norm": 1.734375, + "learning_rate": 4.746066928575697e-06, + "loss": 0.6329, + "num_tokens": 40837684.0, + "step": 1878 + }, + { + "epoch": 5.67750472589792, + "grad_norm": 1.7890625, + "learning_rate": 4.741725439750482e-06, + "loss": 0.4736, + "num_tokens": 40844996.0, + "step": 1879 + }, + { + "epoch": 5.680529300567108, + "grad_norm": 1.9765625, + "learning_rate": 4.737384677225413e-06, + "loss": 0.5613, + "num_tokens": 40852864.0, + "step": 1880 + }, + { + "epoch": 5.683553875236295, + "grad_norm": 1.28125, + "learning_rate": 4.733044645158214e-06, + "loss": 0.9068, + "num_tokens": 40892578.0, + "step": 1881 + }, + { + "epoch": 5.686578449905482, + "grad_norm": 1.9140625, + "learning_rate": 4.72870534770591e-06, + "loss": 0.6192, + "num_tokens": 40900130.0, + "step": 1882 + }, + { + "epoch": 5.689603024574669, + "grad_norm": 1.4765625, + "learning_rate": 4.724366789024822e-06, + "loss": 0.8954, + "num_tokens": 40923220.0, + "step": 1883 + }, + { + "epoch": 5.692627599243856, + "grad_norm": 1.4453125, + "learning_rate": 4.720028973270565e-06, + "loss": 0.8912, + "num_tokens": 40953795.0, + "step": 1884 + }, + { + "epoch": 5.695652173913043, + "grad_norm": 1.421875, + "learning_rate": 4.7156919045980365e-06, + "loss": 0.6721, + "num_tokens": 40973465.0, + "step": 1885 + }, + { + "epoch": 5.698676748582231, + "grad_norm": 1.6796875, + "learning_rate": 4.711355587161427e-06, + "loss": 0.795, + "num_tokens": 40988392.0, + "step": 1886 + }, + { + "epoch": 5.7017013232514175, + "grad_norm": 1.71875, + "learning_rate": 4.7070200251142005e-06, + "loss": 0.7359, + "num_tokens": 41002854.0, + "step": 1887 + }, + { + "epoch": 5.704725897920605, + "grad_norm": 1.5078125, + "learning_rate": 4.702685222609099e-06, + "loss": 0.8248, + "num_tokens": 41026046.0, + "step": 1888 + }, + { + "epoch": 5.707750472589792, + "grad_norm": 1.4140625, + "learning_rate": 4.698351183798141e-06, + "loss": 0.7578, + "num_tokens": 41049763.0, + "step": 1889 + }, + { + "epoch": 5.710775047258979, + "grad_norm": 2.234375, + "learning_rate": 4.6940179128326106e-06, + "loss": 0.5745, + "num_tokens": 41056038.0, + "step": 1890 + }, + { + "epoch": 5.713799621928167, + "grad_norm": 1.203125, + "learning_rate": 4.689685413863052e-06, + "loss": 0.8395, + "num_tokens": 41095678.0, + "step": 1891 + }, + { + "epoch": 5.716824196597353, + "grad_norm": 1.703125, + "learning_rate": 4.685353691039276e-06, + "loss": 0.7135, + "num_tokens": 41109384.0, + "step": 1892 + }, + { + "epoch": 5.719848771266541, + "grad_norm": 1.1953125, + "learning_rate": 4.68102274851035e-06, + "loss": 0.8079, + "num_tokens": 41151427.0, + "step": 1893 + }, + { + "epoch": 5.722873345935728, + "grad_norm": 1.8203125, + "learning_rate": 4.676692590424593e-06, + "loss": 0.5919, + "num_tokens": 41159877.0, + "step": 1894 + }, + { + "epoch": 5.725897920604915, + "grad_norm": 1.765625, + "learning_rate": 4.672363220929567e-06, + "loss": 0.7483, + "num_tokens": 41172295.0, + "step": 1895 + }, + { + "epoch": 5.7289224952741025, + "grad_norm": 1.3671875, + "learning_rate": 4.66803464417209e-06, + "loss": 0.82, + "num_tokens": 41203401.0, + "step": 1896 + }, + { + "epoch": 5.731947069943289, + "grad_norm": 1.671875, + "learning_rate": 4.663706864298211e-06, + "loss": 0.7147, + "num_tokens": 41217030.0, + "step": 1897 + }, + { + "epoch": 5.734971644612476, + "grad_norm": 1.84375, + "learning_rate": 4.659379885453217e-06, + "loss": 0.6181, + "num_tokens": 41224568.0, + "step": 1898 + }, + { + "epoch": 5.737996219281664, + "grad_norm": 2.015625, + "learning_rate": 4.655053711781632e-06, + "loss": 0.5701, + "num_tokens": 41231386.0, + "step": 1899 + }, + { + "epoch": 5.741020793950851, + "grad_norm": 1.5078125, + "learning_rate": 4.650728347427207e-06, + "loss": 0.7883, + "num_tokens": 41250373.0, + "step": 1900 + }, + { + "epoch": 5.7440453686200375, + "grad_norm": 1.40625, + "learning_rate": 4.646403796532916e-06, + "loss": 0.8393, + "num_tokens": 41280934.0, + "step": 1901 + }, + { + "epoch": 5.747069943289225, + "grad_norm": 1.984375, + "learning_rate": 4.642080063240955e-06, + "loss": 0.6176, + "num_tokens": 41288815.0, + "step": 1902 + }, + { + "epoch": 5.750094517958412, + "grad_norm": 2.09375, + "learning_rate": 4.637757151692737e-06, + "loss": 0.6454, + "num_tokens": 41295772.0, + "step": 1903 + }, + { + "epoch": 5.7531190926276, + "grad_norm": 1.25, + "learning_rate": 4.633435066028891e-06, + "loss": 0.8656, + "num_tokens": 41334678.0, + "step": 1904 + }, + { + "epoch": 5.7561436672967865, + "grad_norm": 1.109375, + "learning_rate": 4.6291138103892466e-06, + "loss": 0.8783, + "num_tokens": 41388764.0, + "step": 1905 + }, + { + "epoch": 5.759168241965973, + "grad_norm": 2.03125, + "learning_rate": 4.624793388912849e-06, + "loss": 0.6239, + "num_tokens": 41396999.0, + "step": 1906 + }, + { + "epoch": 5.762192816635161, + "grad_norm": 1.90625, + "learning_rate": 4.620473805737934e-06, + "loss": 0.6522, + "num_tokens": 41405401.0, + "step": 1907 + }, + { + "epoch": 5.765217391304348, + "grad_norm": 1.640625, + "learning_rate": 4.616155065001941e-06, + "loss": 0.7722, + "num_tokens": 41426209.0, + "step": 1908 + }, + { + "epoch": 5.768241965973535, + "grad_norm": 1.0078125, + "learning_rate": 4.611837170841505e-06, + "loss": 0.8431, + "num_tokens": 41479474.0, + "step": 1909 + }, + { + "epoch": 5.771266540642722, + "grad_norm": 1.2421875, + "learning_rate": 4.607520127392442e-06, + "loss": 0.8629, + "num_tokens": 41518990.0, + "step": 1910 + }, + { + "epoch": 5.774291115311909, + "grad_norm": 1.8125, + "learning_rate": 4.603203938789757e-06, + "loss": 0.5868, + "num_tokens": 41527880.0, + "step": 1911 + }, + { + "epoch": 5.777315689981096, + "grad_norm": 1.390625, + "learning_rate": 4.598888609167638e-06, + "loss": 0.8131, + "num_tokens": 41557993.0, + "step": 1912 + }, + { + "epoch": 5.780340264650284, + "grad_norm": 1.359375, + "learning_rate": 4.594574142659448e-06, + "loss": 0.8065, + "num_tokens": 41586361.0, + "step": 1913 + }, + { + "epoch": 5.7833648393194705, + "grad_norm": 2.171875, + "learning_rate": 4.590260543397726e-06, + "loss": 0.6313, + "num_tokens": 41593050.0, + "step": 1914 + }, + { + "epoch": 5.786389413988658, + "grad_norm": 1.4375, + "learning_rate": 4.585947815514174e-06, + "loss": 0.7752, + "num_tokens": 41615666.0, + "step": 1915 + }, + { + "epoch": 5.789413988657845, + "grad_norm": 1.203125, + "learning_rate": 4.581635963139668e-06, + "loss": 0.8669, + "num_tokens": 41660382.0, + "step": 1916 + }, + { + "epoch": 5.792438563327032, + "grad_norm": 1.234375, + "learning_rate": 4.577324990404239e-06, + "loss": 0.8788, + "num_tokens": 41700089.0, + "step": 1917 + }, + { + "epoch": 5.79546313799622, + "grad_norm": 1.3125, + "learning_rate": 4.573014901437077e-06, + "loss": 0.8796, + "num_tokens": 41739084.0, + "step": 1918 + }, + { + "epoch": 5.798487712665406, + "grad_norm": 1.15625, + "learning_rate": 4.568705700366527e-06, + "loss": 0.7756, + "num_tokens": 41774725.0, + "step": 1919 + }, + { + "epoch": 5.801512287334593, + "grad_norm": 1.4765625, + "learning_rate": 4.564397391320085e-06, + "loss": 0.8321, + "num_tokens": 41802043.0, + "step": 1920 + }, + { + "epoch": 5.804536862003781, + "grad_norm": 1.765625, + "learning_rate": 4.560089978424384e-06, + "loss": 0.719, + "num_tokens": 41814484.0, + "step": 1921 + }, + { + "epoch": 5.807561436672968, + "grad_norm": 1.6796875, + "learning_rate": 4.555783465805209e-06, + "loss": 0.8518, + "num_tokens": 41837478.0, + "step": 1922 + }, + { + "epoch": 5.810586011342155, + "grad_norm": 1.2578125, + "learning_rate": 4.55147785758748e-06, + "loss": 0.805, + "num_tokens": 41875459.0, + "step": 1923 + }, + { + "epoch": 5.813610586011342, + "grad_norm": 1.90625, + "learning_rate": 4.547173157895247e-06, + "loss": 0.6094, + "num_tokens": 41883665.0, + "step": 1924 + }, + { + "epoch": 5.816635160680529, + "grad_norm": 1.609375, + "learning_rate": 4.542869370851689e-06, + "loss": 0.7566, + "num_tokens": 41909556.0, + "step": 1925 + }, + { + "epoch": 5.819659735349717, + "grad_norm": 1.6640625, + "learning_rate": 4.538566500579119e-06, + "loss": 0.6952, + "num_tokens": 41923663.0, + "step": 1926 + }, + { + "epoch": 5.822684310018904, + "grad_norm": 0.9609375, + "learning_rate": 4.534264551198966e-06, + "loss": 0.8938, + "num_tokens": 41989193.0, + "step": 1927 + }, + { + "epoch": 5.8257088846880904, + "grad_norm": 1.4765625, + "learning_rate": 4.5299635268317735e-06, + "loss": 0.8375, + "num_tokens": 42010358.0, + "step": 1928 + }, + { + "epoch": 5.828733459357278, + "grad_norm": 1.625, + "learning_rate": 4.525663431597207e-06, + "loss": 0.6867, + "num_tokens": 42022140.0, + "step": 1929 + }, + { + "epoch": 5.831758034026465, + "grad_norm": 1.7734375, + "learning_rate": 4.521364269614038e-06, + "loss": 0.6482, + "num_tokens": 42032043.0, + "step": 1930 + }, + { + "epoch": 5.834782608695652, + "grad_norm": 1.0703125, + "learning_rate": 4.517066045000142e-06, + "loss": 0.9041, + "num_tokens": 42090276.0, + "step": 1931 + }, + { + "epoch": 5.8378071833648395, + "grad_norm": 1.4921875, + "learning_rate": 4.5127687618725006e-06, + "loss": 0.8658, + "num_tokens": 42114171.0, + "step": 1932 + }, + { + "epoch": 5.840831758034026, + "grad_norm": 1.7421875, + "learning_rate": 4.508472424347194e-06, + "loss": 0.7788, + "num_tokens": 42131144.0, + "step": 1933 + }, + { + "epoch": 5.843856332703213, + "grad_norm": 2.0625, + "learning_rate": 4.50417703653939e-06, + "loss": 0.5846, + "num_tokens": 42138451.0, + "step": 1934 + }, + { + "epoch": 5.846880907372401, + "grad_norm": 1.734375, + "learning_rate": 4.499882602563353e-06, + "loss": 0.7475, + "num_tokens": 42153124.0, + "step": 1935 + }, + { + "epoch": 5.849905482041588, + "grad_norm": 1.8515625, + "learning_rate": 4.495589126532436e-06, + "loss": 0.5609, + "num_tokens": 42160526.0, + "step": 1936 + }, + { + "epoch": 5.852930056710775, + "grad_norm": 1.8359375, + "learning_rate": 4.491296612559066e-06, + "loss": 0.7524, + "num_tokens": 42171639.0, + "step": 1937 + }, + { + "epoch": 5.855954631379962, + "grad_norm": 1.984375, + "learning_rate": 4.487005064754754e-06, + "loss": 0.6034, + "num_tokens": 42178775.0, + "step": 1938 + }, + { + "epoch": 5.858979206049149, + "grad_norm": 1.71875, + "learning_rate": 4.482714487230086e-06, + "loss": 0.6766, + "num_tokens": 42191766.0, + "step": 1939 + }, + { + "epoch": 5.862003780718337, + "grad_norm": 1.609375, + "learning_rate": 4.478424884094716e-06, + "loss": 0.7543, + "num_tokens": 42213927.0, + "step": 1940 + }, + { + "epoch": 5.8650283553875235, + "grad_norm": 1.7890625, + "learning_rate": 4.4741362594573686e-06, + "loss": 0.6308, + "num_tokens": 42223336.0, + "step": 1941 + }, + { + "epoch": 5.868052930056711, + "grad_norm": 1.75, + "learning_rate": 4.469848617425825e-06, + "loss": 0.614, + "num_tokens": 42232671.0, + "step": 1942 + }, + { + "epoch": 5.871077504725898, + "grad_norm": 1.7421875, + "learning_rate": 4.465561962106931e-06, + "loss": 0.6496, + "num_tokens": 42244077.0, + "step": 1943 + }, + { + "epoch": 5.874102079395085, + "grad_norm": 1.6484375, + "learning_rate": 4.461276297606585e-06, + "loss": 0.8566, + "num_tokens": 42267060.0, + "step": 1944 + }, + { + "epoch": 5.877126654064273, + "grad_norm": 1.5703125, + "learning_rate": 4.4569916280297355e-06, + "loss": 0.7343, + "num_tokens": 42284989.0, + "step": 1945 + }, + { + "epoch": 5.880151228733459, + "grad_norm": 2.109375, + "learning_rate": 4.45270795748038e-06, + "loss": 0.641, + "num_tokens": 42293459.0, + "step": 1946 + }, + { + "epoch": 5.883175803402646, + "grad_norm": 2.03125, + "learning_rate": 4.448425290061558e-06, + "loss": 0.6355, + "num_tokens": 42300928.0, + "step": 1947 + }, + { + "epoch": 5.886200378071834, + "grad_norm": 1.8984375, + "learning_rate": 4.444143629875346e-06, + "loss": 0.5618, + "num_tokens": 42308895.0, + "step": 1948 + }, + { + "epoch": 5.889224952741021, + "grad_norm": 2.046875, + "learning_rate": 4.439862981022862e-06, + "loss": 0.5586, + "num_tokens": 42315037.0, + "step": 1949 + }, + { + "epoch": 5.892249527410208, + "grad_norm": 2.09375, + "learning_rate": 4.43558334760425e-06, + "loss": 0.5808, + "num_tokens": 42321779.0, + "step": 1950 + }, + { + "epoch": 5.895274102079395, + "grad_norm": 1.5546875, + "learning_rate": 4.43130473371868e-06, + "loss": 0.8216, + "num_tokens": 42342282.0, + "step": 1951 + }, + { + "epoch": 5.898298676748582, + "grad_norm": 1.34375, + "learning_rate": 4.427027143464349e-06, + "loss": 0.7654, + "num_tokens": 42365868.0, + "step": 1952 + }, + { + "epoch": 5.90132325141777, + "grad_norm": 1.25, + "learning_rate": 4.422750580938477e-06, + "loss": 0.9011, + "num_tokens": 42404922.0, + "step": 1953 + }, + { + "epoch": 5.904347826086957, + "grad_norm": 1.5078125, + "learning_rate": 4.418475050237291e-06, + "loss": 0.6895, + "num_tokens": 42419640.0, + "step": 1954 + }, + { + "epoch": 5.9073724007561434, + "grad_norm": 1.71875, + "learning_rate": 4.4142005554560345e-06, + "loss": 0.7818, + "num_tokens": 42435954.0, + "step": 1955 + }, + { + "epoch": 5.910396975425331, + "grad_norm": 1.46875, + "learning_rate": 4.409927100688959e-06, + "loss": 0.7456, + "num_tokens": 42458194.0, + "step": 1956 + }, + { + "epoch": 5.913421550094518, + "grad_norm": 1.328125, + "learning_rate": 4.405654690029321e-06, + "loss": 0.8808, + "num_tokens": 42494482.0, + "step": 1957 + }, + { + "epoch": 5.916446124763705, + "grad_norm": 1.5546875, + "learning_rate": 4.40138332756937e-06, + "loss": 0.8315, + "num_tokens": 42517941.0, + "step": 1958 + }, + { + "epoch": 5.9194706994328925, + "grad_norm": 1.234375, + "learning_rate": 4.397113017400359e-06, + "loss": 0.7845, + "num_tokens": 42552364.0, + "step": 1959 + }, + { + "epoch": 5.922495274102079, + "grad_norm": 1.25, + "learning_rate": 4.392843763612535e-06, + "loss": 0.8312, + "num_tokens": 42591130.0, + "step": 1960 + }, + { + "epoch": 5.925519848771266, + "grad_norm": 1.515625, + "learning_rate": 4.388575570295123e-06, + "loss": 0.8299, + "num_tokens": 42617620.0, + "step": 1961 + }, + { + "epoch": 5.928544423440454, + "grad_norm": 1.7578125, + "learning_rate": 4.384308441536338e-06, + "loss": 0.5764, + "num_tokens": 42625382.0, + "step": 1962 + }, + { + "epoch": 5.931568998109641, + "grad_norm": 1.265625, + "learning_rate": 4.380042381423379e-06, + "loss": 0.8394, + "num_tokens": 42664078.0, + "step": 1963 + }, + { + "epoch": 5.934593572778828, + "grad_norm": 1.8671875, + "learning_rate": 4.3757773940424165e-06, + "loss": 0.6154, + "num_tokens": 42672441.0, + "step": 1964 + }, + { + "epoch": 5.937618147448015, + "grad_norm": 1.890625, + "learning_rate": 4.3715134834785925e-06, + "loss": 0.5578, + "num_tokens": 42680321.0, + "step": 1965 + }, + { + "epoch": 5.940642722117202, + "grad_norm": 1.5234375, + "learning_rate": 4.367250653816022e-06, + "loss": 0.706, + "num_tokens": 42699057.0, + "step": 1966 + }, + { + "epoch": 5.94366729678639, + "grad_norm": 1.21875, + "learning_rate": 4.362988909137783e-06, + "loss": 0.8064, + "num_tokens": 42741724.0, + "step": 1967 + }, + { + "epoch": 5.9466918714555765, + "grad_norm": 1.328125, + "learning_rate": 4.35872825352591e-06, + "loss": 0.8439, + "num_tokens": 42774197.0, + "step": 1968 + }, + { + "epoch": 5.949716446124763, + "grad_norm": 1.359375, + "learning_rate": 4.354468691061404e-06, + "loss": 0.8648, + "num_tokens": 42803493.0, + "step": 1969 + }, + { + "epoch": 5.952741020793951, + "grad_norm": 1.4921875, + "learning_rate": 4.350210225824209e-06, + "loss": 0.8001, + "num_tokens": 42826968.0, + "step": 1970 + }, + { + "epoch": 5.955765595463138, + "grad_norm": 1.546875, + "learning_rate": 4.345952861893225e-06, + "loss": 0.7298, + "num_tokens": 42845206.0, + "step": 1971 + }, + { + "epoch": 5.958790170132325, + "grad_norm": 1.453125, + "learning_rate": 4.341696603346289e-06, + "loss": 0.8037, + "num_tokens": 42868274.0, + "step": 1972 + }, + { + "epoch": 5.961814744801512, + "grad_norm": 1.15625, + "learning_rate": 4.33744145426019e-06, + "loss": 0.8713, + "num_tokens": 42915235.0, + "step": 1973 + }, + { + "epoch": 5.964839319470699, + "grad_norm": 1.375, + "learning_rate": 4.333187418710647e-06, + "loss": 0.722, + "num_tokens": 42937373.0, + "step": 1974 + }, + { + "epoch": 5.967863894139887, + "grad_norm": 1.15625, + "learning_rate": 4.3289345007723115e-06, + "loss": 0.8072, + "num_tokens": 42977136.0, + "step": 1975 + }, + { + "epoch": 5.970888468809074, + "grad_norm": 1.953125, + "learning_rate": 4.3246827045187715e-06, + "loss": 0.5833, + "num_tokens": 42985240.0, + "step": 1976 + }, + { + "epoch": 5.973913043478261, + "grad_norm": 1.40625, + "learning_rate": 4.320432034022535e-06, + "loss": 0.7359, + "num_tokens": 43007281.0, + "step": 1977 + }, + { + "epoch": 5.976937618147448, + "grad_norm": 1.875, + "learning_rate": 4.3161824933550315e-06, + "loss": 0.6806, + "num_tokens": 43016123.0, + "step": 1978 + }, + { + "epoch": 5.979962192816635, + "grad_norm": 1.2734375, + "learning_rate": 4.311934086586611e-06, + "loss": 0.8025, + "num_tokens": 43049435.0, + "step": 1979 + }, + { + "epoch": 5.982986767485822, + "grad_norm": 1.578125, + "learning_rate": 4.307686817786537e-06, + "loss": 0.8848, + "num_tokens": 43071215.0, + "step": 1980 + }, + { + "epoch": 5.98601134215501, + "grad_norm": 1.7265625, + "learning_rate": 4.303440691022982e-06, + "loss": 0.6358, + "num_tokens": 43082240.0, + "step": 1981 + }, + { + "epoch": 5.9890359168241964, + "grad_norm": 2.015625, + "learning_rate": 4.299195710363023e-06, + "loss": 0.6507, + "num_tokens": 43089353.0, + "step": 1982 + }, + { + "epoch": 5.992060491493383, + "grad_norm": 1.2578125, + "learning_rate": 4.294951879872644e-06, + "loss": 0.8406, + "num_tokens": 43128210.0, + "step": 1983 + }, + { + "epoch": 5.995085066162571, + "grad_norm": 1.7578125, + "learning_rate": 4.290709203616723e-06, + "loss": 0.9616, + "num_tokens": 43156719.0, + "step": 1984 + }, + { + "epoch": 5.998109640831758, + "grad_norm": 1.9453125, + "learning_rate": 4.286467685659034e-06, + "loss": 0.5465, + "num_tokens": 43163836.0, + "step": 1985 + }, + { + "epoch": 6.0, + "grad_norm": 2.28125, + "learning_rate": 4.282227330062241e-06, + "loss": 0.6411, + "num_tokens": 43171878.0, + "step": 1986 + }, + { + "epoch": 6.003024574669187, + "grad_norm": 1.5859375, + "learning_rate": 0.0, + "loss": 0.9473, + "num_tokens": 26171.0, + "step": 1987 + }, + { + "epoch": 6.0060491493383745, + "grad_norm": 1.6015625, + "learning_rate": 1.0000000000000001e-07, + "loss": 0.7196, + "num_tokens": 43292.0, + "step": 1988 + }, + { + "epoch": 6.009073724007561, + "grad_norm": 1.3203125, + "learning_rate": 2.0000000000000002e-07, + "loss": 0.8557, + "num_tokens": 73428.0, + "step": 1989 + }, + { + "epoch": 6.012098298676748, + "grad_norm": 1.9609375, + "learning_rate": 3.0000000000000004e-07, + "loss": 0.6327, + "num_tokens": 80817.0, + "step": 1990 + }, + { + "epoch": 6.015122873345936, + "grad_norm": 1.1171875, + "learning_rate": 4.0000000000000003e-07, + "loss": 0.8403, + "num_tokens": 130609.0, + "step": 1991 + }, + { + "epoch": 6.018147448015123, + "grad_norm": 1.109375, + "learning_rate": 5.000000000000001e-07, + "loss": 0.8032, + "num_tokens": 170887.0, + "step": 1992 + }, + { + "epoch": 6.02117202268431, + "grad_norm": 2.078125, + "learning_rate": 6.000000000000001e-07, + "loss": 0.6676, + "num_tokens": 177349.0, + "step": 1993 + }, + { + "epoch": 6.024196597353497, + "grad_norm": 1.234375, + "learning_rate": 7.000000000000001e-07, + "loss": 0.8217, + "num_tokens": 217813.0, + "step": 1994 + }, + { + "epoch": 6.027221172022684, + "grad_norm": 1.5390625, + "learning_rate": 8.000000000000001e-07, + "loss": 0.7377, + "num_tokens": 234634.0, + "step": 1995 + }, + { + "epoch": 6.030245746691872, + "grad_norm": 1.78125, + "learning_rate": 9.000000000000001e-07, + "loss": 0.915, + "num_tokens": 259074.0, + "step": 1996 + }, + { + "epoch": 6.033270321361059, + "grad_norm": 1.9375, + "learning_rate": 1.0000000000000002e-06, + "loss": 0.6272, + "num_tokens": 266062.0, + "step": 1997 + }, + { + "epoch": 6.036294896030245, + "grad_norm": 2.109375, + "learning_rate": 1.1e-06, + "loss": 0.5749, + "num_tokens": 272849.0, + "step": 1998 + }, + { + "epoch": 6.039319470699433, + "grad_norm": 1.1796875, + "learning_rate": 1.2000000000000002e-06, + "loss": 0.7953, + "num_tokens": 311236.0, + "step": 1999 + }, + { + "epoch": 6.04234404536862, + "grad_norm": 1.8671875, + "learning_rate": 1.3e-06, + "loss": 0.5414, + "num_tokens": 318629.0, + "step": 2000 + }, + { + "epoch": 6.045368620037808, + "grad_norm": 1.7265625, + "learning_rate": 1.4000000000000001e-06, + "loss": 0.6916, + "num_tokens": 331513.0, + "step": 2001 + }, + { + "epoch": 6.048393194706994, + "grad_norm": 1.3828125, + "learning_rate": 1.5e-06, + "loss": 0.7905, + "num_tokens": 357080.0, + "step": 2002 + }, + { + "epoch": 6.051417769376181, + "grad_norm": 1.9921875, + "learning_rate": 1.6000000000000001e-06, + "loss": 0.6234, + "num_tokens": 364230.0, + "step": 2003 + }, + { + "epoch": 6.054442344045369, + "grad_norm": 1.6484375, + "learning_rate": 1.7000000000000002e-06, + "loss": 0.8015, + "num_tokens": 381095.0, + "step": 2004 + }, + { + "epoch": 6.057466918714556, + "grad_norm": 1.3125, + "learning_rate": 1.8000000000000001e-06, + "loss": 0.8209, + "num_tokens": 412496.0, + "step": 2005 + }, + { + "epoch": 6.060491493383743, + "grad_norm": 1.4453125, + "learning_rate": 1.9000000000000002e-06, + "loss": 0.7627, + "num_tokens": 438017.0, + "step": 2006 + }, + { + "epoch": 6.06351606805293, + "grad_norm": 1.9296875, + "learning_rate": 2.0000000000000003e-06, + "loss": 0.6885, + "num_tokens": 445806.0, + "step": 2007 + }, + { + "epoch": 6.066540642722117, + "grad_norm": 2.125, + "learning_rate": 2.1000000000000002e-06, + "loss": 0.6331, + "num_tokens": 452789.0, + "step": 2008 + }, + { + "epoch": 6.069565217391304, + "grad_norm": 1.421875, + "learning_rate": 2.2e-06, + "loss": 0.8112, + "num_tokens": 475957.0, + "step": 2009 + }, + { + "epoch": 6.072589792060492, + "grad_norm": 1.21875, + "learning_rate": 2.3000000000000004e-06, + "loss": 0.9639, + "num_tokens": 531107.0, + "step": 2010 + }, + { + "epoch": 6.0756143667296785, + "grad_norm": 1.6953125, + "learning_rate": 2.4000000000000003e-06, + "loss": 0.7598, + "num_tokens": 546999.0, + "step": 2011 + }, + { + "epoch": 6.078638941398866, + "grad_norm": 1.9921875, + "learning_rate": 2.5e-06, + "loss": 0.5655, + "num_tokens": 553805.0, + "step": 2012 + }, + { + "epoch": 6.081663516068053, + "grad_norm": 2.0, + "learning_rate": 2.6e-06, + "loss": 0.5578, + "num_tokens": 562480.0, + "step": 2013 + }, + { + "epoch": 6.08468809073724, + "grad_norm": 1.7265625, + "learning_rate": 2.7000000000000004e-06, + "loss": 0.7453, + "num_tokens": 577012.0, + "step": 2014 + }, + { + "epoch": 6.0877126654064275, + "grad_norm": 1.9609375, + "learning_rate": 2.8000000000000003e-06, + "loss": 0.669, + "num_tokens": 584686.0, + "step": 2015 + }, + { + "epoch": 6.090737240075614, + "grad_norm": 1.9609375, + "learning_rate": 2.9e-06, + "loss": 0.6078, + "num_tokens": 592037.0, + "step": 2016 + }, + { + "epoch": 6.093761814744801, + "grad_norm": 1.9140625, + "learning_rate": 3e-06, + "loss": 0.587, + "num_tokens": 599510.0, + "step": 2017 + }, + { + "epoch": 6.096786389413989, + "grad_norm": 1.859375, + "learning_rate": 3.1000000000000004e-06, + "loss": 0.6324, + "num_tokens": 607392.0, + "step": 2018 + }, + { + "epoch": 6.099810964083176, + "grad_norm": 1.515625, + "learning_rate": 3.2000000000000003e-06, + "loss": 0.8082, + "num_tokens": 627673.0, + "step": 2019 + }, + { + "epoch": 6.1028355387523625, + "grad_norm": 2.03125, + "learning_rate": 3.3000000000000006e-06, + "loss": 0.5648, + "num_tokens": 634579.0, + "step": 2020 + }, + { + "epoch": 6.10586011342155, + "grad_norm": 1.8203125, + "learning_rate": 3.4000000000000005e-06, + "loss": 0.6982, + "num_tokens": 646714.0, + "step": 2021 + }, + { + "epoch": 6.108884688090737, + "grad_norm": 1.328125, + "learning_rate": 3.5e-06, + "loss": 0.9066, + "num_tokens": 683147.0, + "step": 2022 + }, + { + "epoch": 6.111909262759925, + "grad_norm": 1.5859375, + "learning_rate": 3.6000000000000003e-06, + "loss": 0.841, + "num_tokens": 706359.0, + "step": 2023 + }, + { + "epoch": 6.114933837429112, + "grad_norm": 1.1875, + "learning_rate": 3.7e-06, + "loss": 0.852, + "num_tokens": 744267.0, + "step": 2024 + }, + { + "epoch": 6.117958412098298, + "grad_norm": 1.2578125, + "learning_rate": 3.8000000000000005e-06, + "loss": 0.7558, + "num_tokens": 783672.0, + "step": 2025 + }, + { + "epoch": 6.120982986767486, + "grad_norm": 1.546875, + "learning_rate": 3.900000000000001e-06, + "loss": 0.8665, + "num_tokens": 806557.0, + "step": 2026 + }, + { + "epoch": 6.124007561436673, + "grad_norm": 2.203125, + "learning_rate": 4.000000000000001e-06, + "loss": 0.6493, + "num_tokens": 815148.0, + "step": 2027 + }, + { + "epoch": 6.12703213610586, + "grad_norm": 1.625, + "learning_rate": 4.1e-06, + "loss": 0.7393, + "num_tokens": 830649.0, + "step": 2028 + }, + { + "epoch": 6.130056710775047, + "grad_norm": 1.109375, + "learning_rate": 4.2000000000000004e-06, + "loss": 0.8818, + "num_tokens": 879927.0, + "step": 2029 + }, + { + "epoch": 6.133081285444234, + "grad_norm": 1.0078125, + "learning_rate": 4.3e-06, + "loss": 0.8816, + "num_tokens": 936454.0, + "step": 2030 + }, + { + "epoch": 6.136105860113422, + "grad_norm": 1.9375, + "learning_rate": 4.4e-06, + "loss": 0.6789, + "num_tokens": 944471.0, + "step": 2031 + }, + { + "epoch": 6.139130434782609, + "grad_norm": 1.28125, + "learning_rate": 4.5e-06, + "loss": 0.7715, + "num_tokens": 975945.0, + "step": 2032 + }, + { + "epoch": 6.142155009451796, + "grad_norm": 1.9140625, + "learning_rate": 4.600000000000001e-06, + "loss": 0.6395, + "num_tokens": 983554.0, + "step": 2033 + }, + { + "epoch": 6.145179584120983, + "grad_norm": 1.4296875, + "learning_rate": 4.7e-06, + "loss": 0.7842, + "num_tokens": 1006053.0, + "step": 2034 + }, + { + "epoch": 6.14820415879017, + "grad_norm": 1.5078125, + "learning_rate": 4.800000000000001e-06, + "loss": 0.7365, + "num_tokens": 1025663.0, + "step": 2035 + }, + { + "epoch": 6.151228733459357, + "grad_norm": 1.6171875, + "learning_rate": 4.9000000000000005e-06, + "loss": 0.6238, + "num_tokens": 1044966.0, + "step": 2036 + }, + { + "epoch": 6.154253308128545, + "grad_norm": 1.515625, + "learning_rate": 5e-06, + "loss": 0.8312, + "num_tokens": 1066273.0, + "step": 2037 + }, + { + "epoch": 6.1572778827977315, + "grad_norm": 0.9765625, + "learning_rate": 5.1e-06, + "loss": 0.8821, + "num_tokens": 1124087.0, + "step": 2038 + }, + { + "epoch": 6.160302457466918, + "grad_norm": 1.328125, + "learning_rate": 5.2e-06, + "loss": 0.8024, + "num_tokens": 1157347.0, + "step": 2039 + }, + { + "epoch": 6.163327032136106, + "grad_norm": 1.5625, + "learning_rate": 5.300000000000001e-06, + "loss": 0.8368, + "num_tokens": 1180527.0, + "step": 2040 + }, + { + "epoch": 6.166351606805293, + "grad_norm": 1.5703125, + "learning_rate": 5.400000000000001e-06, + "loss": 0.8339, + "num_tokens": 1203612.0, + "step": 2041 + }, + { + "epoch": 6.1693761814744805, + "grad_norm": 1.203125, + "learning_rate": 5.500000000000001e-06, + "loss": 0.8174, + "num_tokens": 1235345.0, + "step": 2042 + }, + { + "epoch": 6.172400756143667, + "grad_norm": 1.859375, + "learning_rate": 5.600000000000001e-06, + "loss": 0.5819, + "num_tokens": 1243100.0, + "step": 2043 + }, + { + "epoch": 6.175425330812854, + "grad_norm": 2.078125, + "learning_rate": 5.7e-06, + "loss": 0.5946, + "num_tokens": 1249540.0, + "step": 2044 + }, + { + "epoch": 6.178449905482042, + "grad_norm": 1.9140625, + "learning_rate": 5.8e-06, + "loss": 0.5924, + "num_tokens": 1257121.0, + "step": 2045 + }, + { + "epoch": 6.181474480151229, + "grad_norm": 1.28125, + "learning_rate": 5.9e-06, + "loss": 0.8012, + "num_tokens": 1290738.0, + "step": 2046 + }, + { + "epoch": 6.1844990548204155, + "grad_norm": 1.5234375, + "learning_rate": 6e-06, + "loss": 0.8265, + "num_tokens": 1314419.0, + "step": 2047 + }, + { + "epoch": 6.187523629489603, + "grad_norm": 1.546875, + "learning_rate": 6.1e-06, + "loss": 0.7946, + "num_tokens": 1337041.0, + "step": 2048 + }, + { + "epoch": 6.19054820415879, + "grad_norm": 1.9375, + "learning_rate": 6.200000000000001e-06, + "loss": 0.5768, + "num_tokens": 1344708.0, + "step": 2049 + }, + { + "epoch": 6.193572778827978, + "grad_norm": 1.4921875, + "learning_rate": 6.300000000000001e-06, + "loss": 0.8018, + "num_tokens": 1368280.0, + "step": 2050 + }, + { + "epoch": 6.196597353497165, + "grad_norm": 1.4296875, + "learning_rate": 6.4000000000000006e-06, + "loss": 0.7352, + "num_tokens": 1394066.0, + "step": 2051 + }, + { + "epoch": 6.199621928166351, + "grad_norm": 1.5390625, + "learning_rate": 6.5000000000000004e-06, + "loss": 0.7407, + "num_tokens": 1412223.0, + "step": 2052 + }, + { + "epoch": 6.202646502835539, + "grad_norm": 2.078125, + "learning_rate": 6.600000000000001e-06, + "loss": 0.6173, + "num_tokens": 1419945.0, + "step": 2053 + }, + { + "epoch": 6.205671077504726, + "grad_norm": 1.2578125, + "learning_rate": 6.700000000000001e-06, + "loss": 0.7699, + "num_tokens": 1450498.0, + "step": 2054 + }, + { + "epoch": 6.208695652173913, + "grad_norm": 0.9765625, + "learning_rate": 6.800000000000001e-06, + "loss": 0.8172, + "num_tokens": 1504044.0, + "step": 2055 + }, + { + "epoch": 6.2117202268431, + "grad_norm": 1.6171875, + "learning_rate": 6.9e-06, + "loss": 0.7083, + "num_tokens": 1519659.0, + "step": 2056 + }, + { + "epoch": 6.214744801512287, + "grad_norm": 1.546875, + "learning_rate": 7e-06, + "loss": 0.8425, + "num_tokens": 1543537.0, + "step": 2057 + }, + { + "epoch": 6.217769376181474, + "grad_norm": 1.9140625, + "learning_rate": 7.100000000000001e-06, + "loss": 0.5751, + "num_tokens": 1551196.0, + "step": 2058 + }, + { + "epoch": 6.220793950850662, + "grad_norm": 1.921875, + "learning_rate": 7.2000000000000005e-06, + "loss": 0.5492, + "num_tokens": 1558154.0, + "step": 2059 + }, + { + "epoch": 6.223818525519849, + "grad_norm": 1.5078125, + "learning_rate": 7.3e-06, + "loss": 0.779, + "num_tokens": 1581294.0, + "step": 2060 + }, + { + "epoch": 6.226843100189036, + "grad_norm": 1.6640625, + "learning_rate": 7.4e-06, + "loss": 0.555, + "num_tokens": 1590469.0, + "step": 2061 + }, + { + "epoch": 6.229867674858223, + "grad_norm": 1.9140625, + "learning_rate": 7.500000000000001e-06, + "loss": 0.6141, + "num_tokens": 1599388.0, + "step": 2062 + }, + { + "epoch": 6.23289224952741, + "grad_norm": 1.6796875, + "learning_rate": 7.600000000000001e-06, + "loss": 0.6919, + "num_tokens": 1612735.0, + "step": 2063 + }, + { + "epoch": 6.235916824196598, + "grad_norm": 1.8828125, + "learning_rate": 7.7e-06, + "loss": 0.616, + "num_tokens": 1620732.0, + "step": 2064 + }, + { + "epoch": 6.2389413988657845, + "grad_norm": 1.21875, + "learning_rate": 7.800000000000002e-06, + "loss": 0.7951, + "num_tokens": 1655263.0, + "step": 2065 + }, + { + "epoch": 6.241965973534971, + "grad_norm": 1.078125, + "learning_rate": 7.9e-06, + "loss": 0.9159, + "num_tokens": 1710453.0, + "step": 2066 + }, + { + "epoch": 6.244990548204159, + "grad_norm": 1.7734375, + "learning_rate": 8.000000000000001e-06, + "loss": 0.8181, + "num_tokens": 1725889.0, + "step": 2067 + }, + { + "epoch": 6.248015122873346, + "grad_norm": 1.28125, + "learning_rate": 8.1e-06, + "loss": 0.8007, + "num_tokens": 1759041.0, + "step": 2068 + }, + { + "epoch": 6.251039697542533, + "grad_norm": 1.8671875, + "learning_rate": 8.2e-06, + "loss": 0.4863, + "num_tokens": 1766475.0, + "step": 2069 + }, + { + "epoch": 6.25406427221172, + "grad_norm": 1.9296875, + "learning_rate": 8.3e-06, + "loss": 0.5592, + "num_tokens": 1774533.0, + "step": 2070 + }, + { + "epoch": 6.257088846880907, + "grad_norm": 1.875, + "learning_rate": 8.400000000000001e-06, + "loss": 0.5221, + "num_tokens": 1782976.0, + "step": 2071 + }, + { + "epoch": 6.260113421550095, + "grad_norm": 1.1328125, + "learning_rate": 8.5e-06, + "loss": 0.8639, + "num_tokens": 1828063.0, + "step": 2072 + }, + { + "epoch": 6.263137996219282, + "grad_norm": 1.625, + "learning_rate": 8.6e-06, + "loss": 0.6633, + "num_tokens": 1842489.0, + "step": 2073 + }, + { + "epoch": 6.2661625708884685, + "grad_norm": 1.65625, + "learning_rate": 8.700000000000001e-06, + "loss": 0.8896, + "num_tokens": 1864804.0, + "step": 2074 + }, + { + "epoch": 6.269187145557656, + "grad_norm": 1.8828125, + "learning_rate": 8.8e-06, + "loss": 0.6321, + "num_tokens": 1873458.0, + "step": 2075 + }, + { + "epoch": 6.272211720226843, + "grad_norm": 1.4765625, + "learning_rate": 8.900000000000001e-06, + "loss": 0.8868, + "num_tokens": 1905358.0, + "step": 2076 + }, + { + "epoch": 6.27523629489603, + "grad_norm": 1.7265625, + "learning_rate": 9e-06, + "loss": 0.6197, + "num_tokens": 1917188.0, + "step": 2077 + }, + { + "epoch": 6.278260869565218, + "grad_norm": 1.5703125, + "learning_rate": 9.100000000000001e-06, + "loss": 0.8192, + "num_tokens": 1941110.0, + "step": 2078 + }, + { + "epoch": 6.281285444234404, + "grad_norm": 1.5078125, + "learning_rate": 9.200000000000002e-06, + "loss": 0.7242, + "num_tokens": 1959239.0, + "step": 2079 + }, + { + "epoch": 6.284310018903592, + "grad_norm": 1.421875, + "learning_rate": 9.3e-06, + "loss": 0.8318, + "num_tokens": 1983088.0, + "step": 2080 + }, + { + "epoch": 6.287334593572779, + "grad_norm": 1.6328125, + "learning_rate": 9.4e-06, + "loss": 0.735, + "num_tokens": 2000918.0, + "step": 2081 + }, + { + "epoch": 6.290359168241966, + "grad_norm": 2.0, + "learning_rate": 9.5e-06, + "loss": 0.6341, + "num_tokens": 2008006.0, + "step": 2082 + }, + { + "epoch": 6.293383742911153, + "grad_norm": 1.3671875, + "learning_rate": 9.600000000000001e-06, + "loss": 0.6783, + "num_tokens": 2031615.0, + "step": 2083 + }, + { + "epoch": 6.29640831758034, + "grad_norm": 1.6796875, + "learning_rate": 9.7e-06, + "loss": 0.8817, + "num_tokens": 2054662.0, + "step": 2084 + }, + { + "epoch": 6.299432892249527, + "grad_norm": 1.265625, + "learning_rate": 9.800000000000001e-06, + "loss": 0.802, + "num_tokens": 2087002.0, + "step": 2085 + }, + { + "epoch": 6.302457466918715, + "grad_norm": 2.234375, + "learning_rate": 9.9e-06, + "loss": 0.6095, + "num_tokens": 2092815.0, + "step": 2086 + }, + { + "epoch": 6.305482041587902, + "grad_norm": 1.4375, + "learning_rate": 1e-05, + "loss": 0.8556, + "num_tokens": 2120077.0, + "step": 2087 + }, + { + "epoch": 6.308506616257089, + "grad_norm": 2.359375, + "learning_rate": 9.999997844876494e-06, + "loss": 0.5996, + "num_tokens": 2126749.0, + "step": 2088 + }, + { + "epoch": 6.311531190926276, + "grad_norm": 1.046875, + "learning_rate": 9.999991379508036e-06, + "loss": 0.8538, + "num_tokens": 2176527.0, + "step": 2089 + }, + { + "epoch": 6.314555765595463, + "grad_norm": 1.6875, + "learning_rate": 9.999980603900821e-06, + "loss": 0.8324, + "num_tokens": 2199466.0, + "step": 2090 + }, + { + "epoch": 6.317580340264651, + "grad_norm": 1.734375, + "learning_rate": 9.999965518065171e-06, + "loss": 0.5396, + "num_tokens": 2207583.0, + "step": 2091 + }, + { + "epoch": 6.3206049149338375, + "grad_norm": 1.796875, + "learning_rate": 9.999946122015535e-06, + "loss": 0.5719, + "num_tokens": 2215835.0, + "step": 2092 + }, + { + "epoch": 6.323629489603024, + "grad_norm": 1.4765625, + "learning_rate": 9.99992241577049e-06, + "loss": 0.7702, + "num_tokens": 2239209.0, + "step": 2093 + }, + { + "epoch": 6.326654064272212, + "grad_norm": 1.9375, + "learning_rate": 9.999894399352742e-06, + "loss": 0.5452, + "num_tokens": 2246722.0, + "step": 2094 + }, + { + "epoch": 6.329678638941399, + "grad_norm": 1.96875, + "learning_rate": 9.999862072789131e-06, + "loss": 0.6274, + "num_tokens": 2254512.0, + "step": 2095 + }, + { + "epoch": 6.332703213610586, + "grad_norm": 1.3828125, + "learning_rate": 9.999825436110612e-06, + "loss": 0.8496, + "num_tokens": 2283552.0, + "step": 2096 + }, + { + "epoch": 6.335727788279773, + "grad_norm": 1.1796875, + "learning_rate": 9.999784489352286e-06, + "loss": 0.9179, + "num_tokens": 2330201.0, + "step": 2097 + }, + { + "epoch": 6.33875236294896, + "grad_norm": 1.1796875, + "learning_rate": 9.999739232553367e-06, + "loss": 0.8304, + "num_tokens": 2370269.0, + "step": 2098 + }, + { + "epoch": 6.341776937618148, + "grad_norm": 1.7890625, + "learning_rate": 9.999689665757205e-06, + "loss": 0.6257, + "num_tokens": 2380458.0, + "step": 2099 + }, + { + "epoch": 6.344801512287335, + "grad_norm": 1.171875, + "learning_rate": 9.999635789011277e-06, + "loss": 0.8069, + "num_tokens": 2418658.0, + "step": 2100 + }, + { + "epoch": 6.3478260869565215, + "grad_norm": 1.5859375, + "learning_rate": 9.99957760236719e-06, + "loss": 0.6694, + "num_tokens": 2435480.0, + "step": 2101 + }, + { + "epoch": 6.350850661625709, + "grad_norm": 1.0625, + "learning_rate": 9.999515105880674e-06, + "loss": 0.8392, + "num_tokens": 2489710.0, + "step": 2102 + }, + { + "epoch": 6.353875236294896, + "grad_norm": 1.609375, + "learning_rate": 9.999448299611589e-06, + "loss": 0.7053, + "num_tokens": 2503406.0, + "step": 2103 + }, + { + "epoch": 6.356899810964083, + "grad_norm": 1.671875, + "learning_rate": 9.999377183623929e-06, + "loss": 0.8445, + "num_tokens": 2529625.0, + "step": 2104 + }, + { + "epoch": 6.359924385633271, + "grad_norm": 2.078125, + "learning_rate": 9.999301757985807e-06, + "loss": 0.6325, + "num_tokens": 2537048.0, + "step": 2105 + }, + { + "epoch": 6.362948960302457, + "grad_norm": 1.96875, + "learning_rate": 9.99922202276947e-06, + "loss": 0.6312, + "num_tokens": 2544121.0, + "step": 2106 + }, + { + "epoch": 6.365973534971644, + "grad_norm": 1.1953125, + "learning_rate": 9.999137978051292e-06, + "loss": 0.8025, + "num_tokens": 2580374.0, + "step": 2107 + }, + { + "epoch": 6.368998109640832, + "grad_norm": 1.3046875, + "learning_rate": 9.999049623911771e-06, + "loss": 0.7515, + "num_tokens": 2611057.0, + "step": 2108 + }, + { + "epoch": 6.372022684310019, + "grad_norm": 1.2890625, + "learning_rate": 9.998956960435538e-06, + "loss": 0.9219, + "num_tokens": 2650761.0, + "step": 2109 + }, + { + "epoch": 6.375047258979206, + "grad_norm": 1.265625, + "learning_rate": 9.998859987711347e-06, + "loss": 0.7917, + "num_tokens": 2686468.0, + "step": 2110 + }, + { + "epoch": 6.378071833648393, + "grad_norm": 1.3359375, + "learning_rate": 9.998758705832084e-06, + "loss": 0.7958, + "num_tokens": 2718113.0, + "step": 2111 + }, + { + "epoch": 6.38109640831758, + "grad_norm": 1.9375, + "learning_rate": 9.998653114894759e-06, + "loss": 0.6057, + "num_tokens": 2725357.0, + "step": 2112 + }, + { + "epoch": 6.384120982986768, + "grad_norm": 1.3359375, + "learning_rate": 9.998543215000508e-06, + "loss": 0.8008, + "num_tokens": 2755112.0, + "step": 2113 + }, + { + "epoch": 6.387145557655955, + "grad_norm": 1.359375, + "learning_rate": 9.998429006254601e-06, + "loss": 0.8676, + "num_tokens": 2795556.0, + "step": 2114 + }, + { + "epoch": 6.390170132325141, + "grad_norm": 1.625, + "learning_rate": 9.998310488766428e-06, + "loss": 0.8559, + "num_tokens": 2819118.0, + "step": 2115 + }, + { + "epoch": 6.393194706994329, + "grad_norm": 1.9453125, + "learning_rate": 9.99818766264951e-06, + "loss": 0.7051, + "num_tokens": 2830770.0, + "step": 2116 + }, + { + "epoch": 6.396219281663516, + "grad_norm": 1.6875, + "learning_rate": 9.998060528021493e-06, + "loss": 0.6765, + "num_tokens": 2844757.0, + "step": 2117 + }, + { + "epoch": 6.399243856332703, + "grad_norm": 1.2109375, + "learning_rate": 9.997929085004153e-06, + "loss": 0.87, + "num_tokens": 2884428.0, + "step": 2118 + }, + { + "epoch": 6.4022684310018905, + "grad_norm": 1.6953125, + "learning_rate": 9.997793333723389e-06, + "loss": 0.8177, + "num_tokens": 2903206.0, + "step": 2119 + }, + { + "epoch": 6.405293005671077, + "grad_norm": 1.5859375, + "learning_rate": 9.997653274309225e-06, + "loss": 0.8497, + "num_tokens": 2926834.0, + "step": 2120 + }, + { + "epoch": 6.408317580340265, + "grad_norm": 1.28125, + "learning_rate": 9.997508906895819e-06, + "loss": 0.8175, + "num_tokens": 2964344.0, + "step": 2121 + }, + { + "epoch": 6.411342155009452, + "grad_norm": 1.6015625, + "learning_rate": 9.997360231621448e-06, + "loss": 0.8293, + "num_tokens": 2988066.0, + "step": 2122 + }, + { + "epoch": 6.414366729678639, + "grad_norm": 1.5859375, + "learning_rate": 9.99720724862852e-06, + "loss": 0.8082, + "num_tokens": 3010216.0, + "step": 2123 + }, + { + "epoch": 6.417391304347826, + "grad_norm": 1.7890625, + "learning_rate": 9.997049958063568e-06, + "loss": 0.7391, + "num_tokens": 3025524.0, + "step": 2124 + }, + { + "epoch": 6.420415879017013, + "grad_norm": 1.609375, + "learning_rate": 9.996888360077244e-06, + "loss": 0.8104, + "num_tokens": 3049106.0, + "step": 2125 + }, + { + "epoch": 6.4234404536862, + "grad_norm": 1.8203125, + "learning_rate": 9.99672245482434e-06, + "loss": 0.6793, + "num_tokens": 3058681.0, + "step": 2126 + }, + { + "epoch": 6.426465028355388, + "grad_norm": 1.25, + "learning_rate": 9.99655224246376e-06, + "loss": 0.7664, + "num_tokens": 3095337.0, + "step": 2127 + }, + { + "epoch": 6.4294896030245745, + "grad_norm": 1.0625, + "learning_rate": 9.99637772315854e-06, + "loss": 0.7948, + "num_tokens": 3146506.0, + "step": 2128 + }, + { + "epoch": 6.432514177693762, + "grad_norm": 1.0859375, + "learning_rate": 9.996198897075842e-06, + "loss": 0.8593, + "num_tokens": 3196683.0, + "step": 2129 + }, + { + "epoch": 6.435538752362949, + "grad_norm": 1.53125, + "learning_rate": 9.99601576438695e-06, + "loss": 0.7018, + "num_tokens": 3213378.0, + "step": 2130 + }, + { + "epoch": 6.438563327032136, + "grad_norm": 1.6171875, + "learning_rate": 9.995828325267276e-06, + "loss": 0.8038, + "num_tokens": 3237350.0, + "step": 2131 + }, + { + "epoch": 6.441587901701324, + "grad_norm": 1.6484375, + "learning_rate": 9.99563657989635e-06, + "loss": 0.8996, + "num_tokens": 3260757.0, + "step": 2132 + }, + { + "epoch": 6.44461247637051, + "grad_norm": 1.25, + "learning_rate": 9.995440528457839e-06, + "loss": 0.8661, + "num_tokens": 3300092.0, + "step": 2133 + }, + { + "epoch": 6.447637051039697, + "grad_norm": 1.8984375, + "learning_rate": 9.995240171139524e-06, + "loss": 0.7713, + "num_tokens": 3313269.0, + "step": 2134 + }, + { + "epoch": 6.450661625708885, + "grad_norm": 1.921875, + "learning_rate": 9.995035508133316e-06, + "loss": 0.5975, + "num_tokens": 3321508.0, + "step": 2135 + }, + { + "epoch": 6.453686200378072, + "grad_norm": 1.5859375, + "learning_rate": 9.994826539635244e-06, + "loss": 0.8633, + "num_tokens": 3345100.0, + "step": 2136 + }, + { + "epoch": 6.456710775047259, + "grad_norm": 2.125, + "learning_rate": 9.994613265845468e-06, + "loss": 0.6122, + "num_tokens": 3352061.0, + "step": 2137 + }, + { + "epoch": 6.459735349716446, + "grad_norm": 1.890625, + "learning_rate": 9.994395686968267e-06, + "loss": 0.6695, + "num_tokens": 3363905.0, + "step": 2138 + }, + { + "epoch": 6.462759924385633, + "grad_norm": 1.4765625, + "learning_rate": 9.994173803212046e-06, + "loss": 0.8455, + "num_tokens": 3388879.0, + "step": 2139 + }, + { + "epoch": 6.465784499054821, + "grad_norm": 1.5078125, + "learning_rate": 9.993947614789332e-06, + "loss": 0.7222, + "num_tokens": 3408869.0, + "step": 2140 + }, + { + "epoch": 6.468809073724008, + "grad_norm": 1.359375, + "learning_rate": 9.993717121916778e-06, + "loss": 0.7898, + "num_tokens": 3435478.0, + "step": 2141 + }, + { + "epoch": 6.471833648393194, + "grad_norm": 1.21875, + "learning_rate": 9.993482324815152e-06, + "loss": 0.842, + "num_tokens": 3480410.0, + "step": 2142 + }, + { + "epoch": 6.474858223062382, + "grad_norm": 1.9140625, + "learning_rate": 9.993243223709357e-06, + "loss": 0.6338, + "num_tokens": 3488326.0, + "step": 2143 + }, + { + "epoch": 6.477882797731569, + "grad_norm": 1.84375, + "learning_rate": 9.992999818828408e-06, + "loss": 0.6177, + "num_tokens": 3496338.0, + "step": 2144 + }, + { + "epoch": 6.480907372400756, + "grad_norm": 1.453125, + "learning_rate": 9.992752110405448e-06, + "loss": 0.8439, + "num_tokens": 3525094.0, + "step": 2145 + }, + { + "epoch": 6.4839319470699435, + "grad_norm": 1.421875, + "learning_rate": 9.992500098677737e-06, + "loss": 0.7905, + "num_tokens": 3554776.0, + "step": 2146 + }, + { + "epoch": 6.48695652173913, + "grad_norm": 1.765625, + "learning_rate": 9.992243783886663e-06, + "loss": 0.6378, + "num_tokens": 3570213.0, + "step": 2147 + }, + { + "epoch": 6.489981096408318, + "grad_norm": 1.8359375, + "learning_rate": 9.991983166277731e-06, + "loss": 0.5718, + "num_tokens": 3578593.0, + "step": 2148 + }, + { + "epoch": 6.493005671077505, + "grad_norm": 1.8828125, + "learning_rate": 9.991718246100573e-06, + "loss": 0.719, + "num_tokens": 3592005.0, + "step": 2149 + }, + { + "epoch": 6.496030245746692, + "grad_norm": 1.0078125, + "learning_rate": 9.991449023608932e-06, + "loss": 0.837, + "num_tokens": 3659871.0, + "step": 2150 + }, + { + "epoch": 6.499054820415879, + "grad_norm": 1.6484375, + "learning_rate": 9.991175499060684e-06, + "loss": 0.8505, + "num_tokens": 3683703.0, + "step": 2151 + }, + { + "epoch": 6.502079395085066, + "grad_norm": 1.578125, + "learning_rate": 9.990897672717813e-06, + "loss": 0.8208, + "num_tokens": 3708272.0, + "step": 2152 + }, + { + "epoch": 6.505103969754253, + "grad_norm": 1.3671875, + "learning_rate": 9.990615544846439e-06, + "loss": 0.8185, + "num_tokens": 3736118.0, + "step": 2153 + }, + { + "epoch": 6.508128544423441, + "grad_norm": 1.8671875, + "learning_rate": 9.990329115716785e-06, + "loss": 0.5984, + "num_tokens": 3743989.0, + "step": 2154 + }, + { + "epoch": 6.5111531190926275, + "grad_norm": 1.5234375, + "learning_rate": 9.990038385603208e-06, + "loss": 0.7613, + "num_tokens": 3767475.0, + "step": 2155 + }, + { + "epoch": 6.514177693761814, + "grad_norm": 1.875, + "learning_rate": 9.989743354784174e-06, + "loss": 0.7801, + "num_tokens": 3782704.0, + "step": 2156 + }, + { + "epoch": 6.517202268431002, + "grad_norm": 1.71875, + "learning_rate": 9.989444023542278e-06, + "loss": 0.6558, + "num_tokens": 3797306.0, + "step": 2157 + }, + { + "epoch": 6.520226843100189, + "grad_norm": 1.5078125, + "learning_rate": 9.989140392164225e-06, + "loss": 0.689, + "num_tokens": 3816867.0, + "step": 2158 + }, + { + "epoch": 6.523251417769377, + "grad_norm": 1.640625, + "learning_rate": 9.988832460940846e-06, + "loss": 0.8048, + "num_tokens": 3837774.0, + "step": 2159 + }, + { + "epoch": 6.526275992438563, + "grad_norm": 1.5078125, + "learning_rate": 9.988520230167084e-06, + "loss": 0.7108, + "num_tokens": 3856412.0, + "step": 2160 + }, + { + "epoch": 6.52930056710775, + "grad_norm": 1.2578125, + "learning_rate": 9.98820370014201e-06, + "loss": 0.9037, + "num_tokens": 3899474.0, + "step": 2161 + }, + { + "epoch": 6.532325141776938, + "grad_norm": 1.7734375, + "learning_rate": 9.987882871168801e-06, + "loss": 0.5438, + "num_tokens": 3908797.0, + "step": 2162 + }, + { + "epoch": 6.535349716446125, + "grad_norm": 1.40625, + "learning_rate": 9.98755774355476e-06, + "loss": 0.7373, + "num_tokens": 3929968.0, + "step": 2163 + }, + { + "epoch": 6.5383742911153115, + "grad_norm": 1.859375, + "learning_rate": 9.987228317611305e-06, + "loss": 0.5787, + "num_tokens": 3937946.0, + "step": 2164 + }, + { + "epoch": 6.541398865784499, + "grad_norm": 1.5, + "learning_rate": 9.986894593653969e-06, + "loss": 0.8172, + "num_tokens": 3960720.0, + "step": 2165 + }, + { + "epoch": 6.544423440453686, + "grad_norm": 1.4765625, + "learning_rate": 9.986556572002408e-06, + "loss": 0.8596, + "num_tokens": 3984511.0, + "step": 2166 + }, + { + "epoch": 6.547448015122873, + "grad_norm": 1.7890625, + "learning_rate": 9.986214252980384e-06, + "loss": 0.5865, + "num_tokens": 3992291.0, + "step": 2167 + }, + { + "epoch": 6.550472589792061, + "grad_norm": 1.5078125, + "learning_rate": 9.985867636915784e-06, + "loss": 0.7017, + "num_tokens": 4011249.0, + "step": 2168 + }, + { + "epoch": 6.553497164461247, + "grad_norm": 1.2109375, + "learning_rate": 9.985516724140613e-06, + "loss": 0.862, + "num_tokens": 4050037.0, + "step": 2169 + }, + { + "epoch": 6.556521739130435, + "grad_norm": 1.859375, + "learning_rate": 9.985161514990979e-06, + "loss": 0.5538, + "num_tokens": 4057585.0, + "step": 2170 + }, + { + "epoch": 6.559546313799622, + "grad_norm": 1.53125, + "learning_rate": 9.984802009807117e-06, + "loss": 0.8257, + "num_tokens": 4081197.0, + "step": 2171 + }, + { + "epoch": 6.562570888468809, + "grad_norm": 1.4921875, + "learning_rate": 9.984438208933371e-06, + "loss": 0.897, + "num_tokens": 4105160.0, + "step": 2172 + }, + { + "epoch": 6.5655954631379965, + "grad_norm": 2.125, + "learning_rate": 9.984070112718204e-06, + "loss": 0.5904, + "num_tokens": 4112426.0, + "step": 2173 + }, + { + "epoch": 6.568620037807183, + "grad_norm": 1.0703125, + "learning_rate": 9.98369772151419e-06, + "loss": 0.8625, + "num_tokens": 4170371.0, + "step": 2174 + }, + { + "epoch": 6.571644612476371, + "grad_norm": 1.2421875, + "learning_rate": 9.983321035678016e-06, + "loss": 0.8093, + "num_tokens": 4209834.0, + "step": 2175 + }, + { + "epoch": 6.574669187145558, + "grad_norm": 1.296875, + "learning_rate": 9.982940055570484e-06, + "loss": 0.9044, + "num_tokens": 4249042.0, + "step": 2176 + }, + { + "epoch": 6.577693761814745, + "grad_norm": 1.5546875, + "learning_rate": 9.982554781556512e-06, + "loss": 0.8147, + "num_tokens": 4273516.0, + "step": 2177 + }, + { + "epoch": 6.5807183364839315, + "grad_norm": 1.5703125, + "learning_rate": 9.982165214005124e-06, + "loss": 0.7841, + "num_tokens": 4292398.0, + "step": 2178 + }, + { + "epoch": 6.583742911153119, + "grad_norm": 1.6484375, + "learning_rate": 9.981771353289464e-06, + "loss": 0.7159, + "num_tokens": 4307301.0, + "step": 2179 + }, + { + "epoch": 6.586767485822306, + "grad_norm": 1.90625, + "learning_rate": 9.981373199786782e-06, + "loss": 0.6828, + "num_tokens": 4317915.0, + "step": 2180 + }, + { + "epoch": 6.589792060491494, + "grad_norm": 1.8515625, + "learning_rate": 9.980970753878443e-06, + "loss": 0.5672, + "num_tokens": 4326443.0, + "step": 2181 + }, + { + "epoch": 6.5928166351606805, + "grad_norm": 1.9296875, + "learning_rate": 9.980564015949924e-06, + "loss": 0.7393, + "num_tokens": 4341752.0, + "step": 2182 + }, + { + "epoch": 6.595841209829867, + "grad_norm": 1.4140625, + "learning_rate": 9.98015298639081e-06, + "loss": 0.6769, + "num_tokens": 4361811.0, + "step": 2183 + }, + { + "epoch": 6.598865784499055, + "grad_norm": 2.203125, + "learning_rate": 9.9797376655948e-06, + "loss": 0.6081, + "num_tokens": 4367930.0, + "step": 2184 + }, + { + "epoch": 6.601890359168242, + "grad_norm": 1.53125, + "learning_rate": 9.979318053959704e-06, + "loss": 0.8041, + "num_tokens": 4397769.0, + "step": 2185 + }, + { + "epoch": 6.60491493383743, + "grad_norm": 1.40625, + "learning_rate": 9.978894151887435e-06, + "loss": 0.832, + "num_tokens": 4429416.0, + "step": 2186 + }, + { + "epoch": 6.607939508506616, + "grad_norm": 1.7109375, + "learning_rate": 9.978465959784022e-06, + "loss": 0.7147, + "num_tokens": 4449211.0, + "step": 2187 + }, + { + "epoch": 6.610964083175803, + "grad_norm": 1.546875, + "learning_rate": 9.978033478059604e-06, + "loss": 0.8018, + "num_tokens": 4471998.0, + "step": 2188 + }, + { + "epoch": 6.613988657844991, + "grad_norm": 1.2734375, + "learning_rate": 9.977596707128424e-06, + "loss": 0.8997, + "num_tokens": 4510731.0, + "step": 2189 + }, + { + "epoch": 6.617013232514178, + "grad_norm": 1.421875, + "learning_rate": 9.977155647408834e-06, + "loss": 0.8536, + "num_tokens": 4537821.0, + "step": 2190 + }, + { + "epoch": 6.6200378071833645, + "grad_norm": 1.1953125, + "learning_rate": 9.976710299323296e-06, + "loss": 0.9149, + "num_tokens": 4580743.0, + "step": 2191 + }, + { + "epoch": 6.623062381852552, + "grad_norm": 1.9609375, + "learning_rate": 9.976260663298384e-06, + "loss": 0.6002, + "num_tokens": 4588710.0, + "step": 2192 + }, + { + "epoch": 6.626086956521739, + "grad_norm": 1.8828125, + "learning_rate": 9.975806739764766e-06, + "loss": 0.5578, + "num_tokens": 4596278.0, + "step": 2193 + }, + { + "epoch": 6.629111531190926, + "grad_norm": 1.390625, + "learning_rate": 9.97534852915723e-06, + "loss": 0.7679, + "num_tokens": 4622060.0, + "step": 2194 + }, + { + "epoch": 6.632136105860114, + "grad_norm": 1.3203125, + "learning_rate": 9.974886031914665e-06, + "loss": 0.7711, + "num_tokens": 4661463.0, + "step": 2195 + }, + { + "epoch": 6.6351606805293, + "grad_norm": 1.453125, + "learning_rate": 9.974419248480064e-06, + "loss": 0.7081, + "num_tokens": 4682317.0, + "step": 2196 + }, + { + "epoch": 6.638185255198488, + "grad_norm": 1.984375, + "learning_rate": 9.973948179300528e-06, + "loss": 0.6695, + "num_tokens": 4689734.0, + "step": 2197 + }, + { + "epoch": 6.641209829867675, + "grad_norm": 1.875, + "learning_rate": 9.973472824827262e-06, + "loss": 0.5639, + "num_tokens": 4697464.0, + "step": 2198 + }, + { + "epoch": 6.644234404536862, + "grad_norm": 1.8984375, + "learning_rate": 9.972993185515579e-06, + "loss": 0.6253, + "num_tokens": 4706268.0, + "step": 2199 + }, + { + "epoch": 6.6472589792060495, + "grad_norm": 1.5859375, + "learning_rate": 9.97250926182489e-06, + "loss": 0.8141, + "num_tokens": 4730202.0, + "step": 2200 + }, + { + "epoch": 6.650283553875236, + "grad_norm": 2.0625, + "learning_rate": 9.972021054218712e-06, + "loss": 0.6068, + "num_tokens": 4737891.0, + "step": 2201 + }, + { + "epoch": 6.653308128544423, + "grad_norm": 1.1796875, + "learning_rate": 9.971528563164669e-06, + "loss": 0.8376, + "num_tokens": 4773707.0, + "step": 2202 + }, + { + "epoch": 6.656332703213611, + "grad_norm": 1.7890625, + "learning_rate": 9.971031789134486e-06, + "loss": 0.7081, + "num_tokens": 4787114.0, + "step": 2203 + }, + { + "epoch": 6.659357277882798, + "grad_norm": 1.6796875, + "learning_rate": 9.970530732603984e-06, + "loss": 0.824, + "num_tokens": 4810574.0, + "step": 2204 + }, + { + "epoch": 6.6623818525519845, + "grad_norm": 2.03125, + "learning_rate": 9.9700253940531e-06, + "loss": 0.6395, + "num_tokens": 4818191.0, + "step": 2205 + }, + { + "epoch": 6.665406427221172, + "grad_norm": 1.3515625, + "learning_rate": 9.969515773965855e-06, + "loss": 0.8685, + "num_tokens": 4857894.0, + "step": 2206 + }, + { + "epoch": 6.668431001890359, + "grad_norm": 1.1953125, + "learning_rate": 9.969001872830383e-06, + "loss": 0.9049, + "num_tokens": 4902266.0, + "step": 2207 + }, + { + "epoch": 6.671455576559547, + "grad_norm": 1.765625, + "learning_rate": 9.968483691138917e-06, + "loss": 0.6992, + "num_tokens": 4917712.0, + "step": 2208 + }, + { + "epoch": 6.6744801512287335, + "grad_norm": 1.578125, + "learning_rate": 9.967961229387784e-06, + "loss": 0.8521, + "num_tokens": 4940087.0, + "step": 2209 + }, + { + "epoch": 6.67750472589792, + "grad_norm": 1.296875, + "learning_rate": 9.967434488077422e-06, + "loss": 0.8427, + "num_tokens": 4969595.0, + "step": 2210 + }, + { + "epoch": 6.680529300567108, + "grad_norm": 1.5078125, + "learning_rate": 9.966903467712355e-06, + "loss": 0.7501, + "num_tokens": 4992842.0, + "step": 2211 + }, + { + "epoch": 6.683553875236295, + "grad_norm": 1.578125, + "learning_rate": 9.966368168801214e-06, + "loss": 0.8591, + "num_tokens": 5017528.0, + "step": 2212 + }, + { + "epoch": 6.686578449905482, + "grad_norm": 1.4375, + "learning_rate": 9.965828591856725e-06, + "loss": 0.8418, + "num_tokens": 5042432.0, + "step": 2213 + }, + { + "epoch": 6.689603024574669, + "grad_norm": 1.640625, + "learning_rate": 9.965284737395714e-06, + "loss": 0.8446, + "num_tokens": 5065163.0, + "step": 2214 + }, + { + "epoch": 6.692627599243856, + "grad_norm": 1.8984375, + "learning_rate": 9.964736605939098e-06, + "loss": 0.647, + "num_tokens": 5079655.0, + "step": 2215 + }, + { + "epoch": 6.695652173913043, + "grad_norm": 1.453125, + "learning_rate": 9.964184198011903e-06, + "loss": 0.7973, + "num_tokens": 5105258.0, + "step": 2216 + }, + { + "epoch": 6.698676748582231, + "grad_norm": 1.5390625, + "learning_rate": 9.963627514143235e-06, + "loss": 0.8141, + "num_tokens": 5127948.0, + "step": 2217 + }, + { + "epoch": 6.7017013232514175, + "grad_norm": 2.0625, + "learning_rate": 9.963066554866311e-06, + "loss": 0.6055, + "num_tokens": 5135092.0, + "step": 2218 + }, + { + "epoch": 6.704725897920605, + "grad_norm": 1.2265625, + "learning_rate": 9.962501320718432e-06, + "loss": 0.8316, + "num_tokens": 5174405.0, + "step": 2219 + }, + { + "epoch": 6.707750472589792, + "grad_norm": 1.890625, + "learning_rate": 9.961931812240998e-06, + "loss": 0.5795, + "num_tokens": 5181549.0, + "step": 2220 + }, + { + "epoch": 6.710775047258979, + "grad_norm": 1.7109375, + "learning_rate": 9.961358029979504e-06, + "loss": 0.8046, + "num_tokens": 5200158.0, + "step": 2221 + }, + { + "epoch": 6.713799621928167, + "grad_norm": 1.4921875, + "learning_rate": 9.960779974483537e-06, + "loss": 0.8223, + "num_tokens": 5220403.0, + "step": 2222 + }, + { + "epoch": 6.716824196597353, + "grad_norm": 1.5546875, + "learning_rate": 9.960197646306778e-06, + "loss": 0.8018, + "num_tokens": 5240150.0, + "step": 2223 + }, + { + "epoch": 6.719848771266541, + "grad_norm": 1.5546875, + "learning_rate": 9.959611046007e-06, + "loss": 0.7519, + "num_tokens": 5263503.0, + "step": 2224 + }, + { + "epoch": 6.722873345935728, + "grad_norm": 1.9453125, + "learning_rate": 9.959020174146066e-06, + "loss": 0.5956, + "num_tokens": 5270973.0, + "step": 2225 + }, + { + "epoch": 6.725897920604915, + "grad_norm": 1.6640625, + "learning_rate": 9.958425031289935e-06, + "loss": 0.7478, + "num_tokens": 5287245.0, + "step": 2226 + }, + { + "epoch": 6.7289224952741025, + "grad_norm": 1.1953125, + "learning_rate": 9.957825618008653e-06, + "loss": 0.7624, + "num_tokens": 5321128.0, + "step": 2227 + }, + { + "epoch": 6.731947069943289, + "grad_norm": 1.1875, + "learning_rate": 9.957221934876355e-06, + "loss": 0.8307, + "num_tokens": 5364275.0, + "step": 2228 + }, + { + "epoch": 6.734971644612476, + "grad_norm": 1.40625, + "learning_rate": 9.956613982471274e-06, + "loss": 0.7754, + "num_tokens": 5388824.0, + "step": 2229 + }, + { + "epoch": 6.737996219281664, + "grad_norm": 2.125, + "learning_rate": 9.956001761375723e-06, + "loss": 0.6457, + "num_tokens": 5395926.0, + "step": 2230 + }, + { + "epoch": 6.741020793950851, + "grad_norm": 2.015625, + "learning_rate": 9.955385272176108e-06, + "loss": 0.6578, + "num_tokens": 5403295.0, + "step": 2231 + }, + { + "epoch": 6.7440453686200375, + "grad_norm": 1.1875, + "learning_rate": 9.954764515462921e-06, + "loss": 0.8436, + "num_tokens": 5441628.0, + "step": 2232 + }, + { + "epoch": 6.747069943289225, + "grad_norm": 1.9375, + "learning_rate": 9.954139491830748e-06, + "loss": 0.5819, + "num_tokens": 5449287.0, + "step": 2233 + }, + { + "epoch": 6.750094517958412, + "grad_norm": 1.984375, + "learning_rate": 9.953510201878251e-06, + "loss": 0.7169, + "num_tokens": 5461195.0, + "step": 2234 + }, + { + "epoch": 6.7531190926276, + "grad_norm": 1.9921875, + "learning_rate": 9.952876646208187e-06, + "loss": 0.6227, + "num_tokens": 5468592.0, + "step": 2235 + }, + { + "epoch": 6.7561436672967865, + "grad_norm": 1.828125, + "learning_rate": 9.952238825427398e-06, + "loss": 0.8306, + "num_tokens": 5483415.0, + "step": 2236 + }, + { + "epoch": 6.759168241965973, + "grad_norm": 1.6171875, + "learning_rate": 9.951596740146809e-06, + "loss": 0.82, + "num_tokens": 5506092.0, + "step": 2237 + }, + { + "epoch": 6.762192816635161, + "grad_norm": 1.78125, + "learning_rate": 9.950950390981427e-06, + "loss": 0.7946, + "num_tokens": 5528929.0, + "step": 2238 + }, + { + "epoch": 6.765217391304348, + "grad_norm": 1.5234375, + "learning_rate": 9.95029977855035e-06, + "loss": 0.8215, + "num_tokens": 5553395.0, + "step": 2239 + }, + { + "epoch": 6.768241965973535, + "grad_norm": 1.8515625, + "learning_rate": 9.949644903476752e-06, + "loss": 0.5404, + "num_tokens": 5561182.0, + "step": 2240 + }, + { + "epoch": 6.771266540642722, + "grad_norm": 1.453125, + "learning_rate": 9.948985766387899e-06, + "loss": 0.834, + "num_tokens": 5591453.0, + "step": 2241 + }, + { + "epoch": 6.774291115311909, + "grad_norm": 1.09375, + "learning_rate": 9.948322367915131e-06, + "loss": 0.787, + "num_tokens": 5640989.0, + "step": 2242 + }, + { + "epoch": 6.777315689981096, + "grad_norm": 1.3359375, + "learning_rate": 9.947654708693872e-06, + "loss": 0.8161, + "num_tokens": 5670247.0, + "step": 2243 + }, + { + "epoch": 6.780340264650284, + "grad_norm": 1.4453125, + "learning_rate": 9.946982789363631e-06, + "loss": 0.7668, + "num_tokens": 5693017.0, + "step": 2244 + }, + { + "epoch": 6.7833648393194705, + "grad_norm": 2.0, + "learning_rate": 9.946306610567989e-06, + "loss": 0.6101, + "num_tokens": 5700736.0, + "step": 2245 + }, + { + "epoch": 6.786389413988658, + "grad_norm": 1.484375, + "learning_rate": 9.945626172954617e-06, + "loss": 0.7944, + "num_tokens": 5724797.0, + "step": 2246 + }, + { + "epoch": 6.789413988657845, + "grad_norm": 1.9921875, + "learning_rate": 9.944941477175258e-06, + "loss": 0.6763, + "num_tokens": 5733912.0, + "step": 2247 + }, + { + "epoch": 6.792438563327032, + "grad_norm": 1.4375, + "learning_rate": 9.944252523885739e-06, + "loss": 0.7616, + "num_tokens": 5757686.0, + "step": 2248 + }, + { + "epoch": 6.79546313799622, + "grad_norm": 1.5234375, + "learning_rate": 9.943559313745957e-06, + "loss": 0.7847, + "num_tokens": 5785190.0, + "step": 2249 + }, + { + "epoch": 6.798487712665406, + "grad_norm": 1.5625, + "learning_rate": 9.942861847419895e-06, + "loss": 0.8219, + "num_tokens": 5808457.0, + "step": 2250 + }, + { + "epoch": 6.801512287334593, + "grad_norm": 1.859375, + "learning_rate": 9.942160125575608e-06, + "loss": 0.5506, + "num_tokens": 5816452.0, + "step": 2251 + }, + { + "epoch": 6.804536862003781, + "grad_norm": 1.6796875, + "learning_rate": 9.941454148885226e-06, + "loss": 0.84, + "num_tokens": 5838631.0, + "step": 2252 + }, + { + "epoch": 6.807561436672968, + "grad_norm": 1.90625, + "learning_rate": 9.940743918024961e-06, + "loss": 0.5731, + "num_tokens": 5846117.0, + "step": 2253 + }, + { + "epoch": 6.810586011342155, + "grad_norm": 1.9140625, + "learning_rate": 9.940029433675093e-06, + "loss": 0.6481, + "num_tokens": 5853930.0, + "step": 2254 + }, + { + "epoch": 6.813610586011342, + "grad_norm": 1.7265625, + "learning_rate": 9.939310696519977e-06, + "loss": 0.7701, + "num_tokens": 5874065.0, + "step": 2255 + }, + { + "epoch": 6.816635160680529, + "grad_norm": 1.71875, + "learning_rate": 9.938587707248045e-06, + "loss": 0.9082, + "num_tokens": 5896400.0, + "step": 2256 + }, + { + "epoch": 6.819659735349717, + "grad_norm": 1.5234375, + "learning_rate": 9.9378604665518e-06, + "loss": 0.7433, + "num_tokens": 5917871.0, + "step": 2257 + }, + { + "epoch": 6.822684310018904, + "grad_norm": 1.28125, + "learning_rate": 9.937128975127814e-06, + "loss": 0.813, + "num_tokens": 5950470.0, + "step": 2258 + }, + { + "epoch": 6.8257088846880904, + "grad_norm": 1.5, + "learning_rate": 9.936393233676737e-06, + "loss": 0.8163, + "num_tokens": 5973952.0, + "step": 2259 + }, + { + "epoch": 6.828733459357278, + "grad_norm": 2.03125, + "learning_rate": 9.935653242903283e-06, + "loss": 0.59, + "num_tokens": 5981578.0, + "step": 2260 + }, + { + "epoch": 6.831758034026465, + "grad_norm": 1.6796875, + "learning_rate": 9.93490900351624e-06, + "loss": 0.7709, + "num_tokens": 6000081.0, + "step": 2261 + }, + { + "epoch": 6.834782608695652, + "grad_norm": 2.0, + "learning_rate": 9.934160516228468e-06, + "loss": 0.5705, + "num_tokens": 6006888.0, + "step": 2262 + }, + { + "epoch": 6.8378071833648395, + "grad_norm": 1.8359375, + "learning_rate": 9.933407781756888e-06, + "loss": 0.896, + "num_tokens": 6029458.0, + "step": 2263 + }, + { + "epoch": 6.840831758034026, + "grad_norm": 1.359375, + "learning_rate": 9.9326508008225e-06, + "loss": 0.8193, + "num_tokens": 6064511.0, + "step": 2264 + }, + { + "epoch": 6.843856332703213, + "grad_norm": 1.96875, + "learning_rate": 9.931889574150357e-06, + "loss": 0.5639, + "num_tokens": 6071692.0, + "step": 2265 + }, + { + "epoch": 6.846880907372401, + "grad_norm": 1.84375, + "learning_rate": 9.931124102469592e-06, + "loss": 0.5339, + "num_tokens": 6079758.0, + "step": 2266 + }, + { + "epoch": 6.849905482041588, + "grad_norm": 1.4296875, + "learning_rate": 9.930354386513399e-06, + "loss": 0.7791, + "num_tokens": 6103184.0, + "step": 2267 + }, + { + "epoch": 6.852930056710775, + "grad_norm": 1.5, + "learning_rate": 9.929580427019034e-06, + "loss": 0.7178, + "num_tokens": 6123743.0, + "step": 2268 + }, + { + "epoch": 6.855954631379962, + "grad_norm": 1.7421875, + "learning_rate": 9.928802224727822e-06, + "loss": 0.6997, + "num_tokens": 6144695.0, + "step": 2269 + }, + { + "epoch": 6.858979206049149, + "grad_norm": 1.828125, + "learning_rate": 9.928019780385152e-06, + "loss": 0.6977, + "num_tokens": 6157787.0, + "step": 2270 + }, + { + "epoch": 6.862003780718337, + "grad_norm": 1.96875, + "learning_rate": 9.927233094740473e-06, + "loss": 0.5817, + "num_tokens": 6164874.0, + "step": 2271 + }, + { + "epoch": 6.8650283553875235, + "grad_norm": 1.265625, + "learning_rate": 9.926442168547299e-06, + "loss": 0.9005, + "num_tokens": 6204564.0, + "step": 2272 + }, + { + "epoch": 6.868052930056711, + "grad_norm": 1.5, + "learning_rate": 9.925647002563205e-06, + "loss": 0.7321, + "num_tokens": 6228989.0, + "step": 2273 + }, + { + "epoch": 6.871077504725898, + "grad_norm": 1.9609375, + "learning_rate": 9.924847597549825e-06, + "loss": 0.6865, + "num_tokens": 6236880.0, + "step": 2274 + }, + { + "epoch": 6.874102079395085, + "grad_norm": 1.8828125, + "learning_rate": 9.924043954272859e-06, + "loss": 0.6359, + "num_tokens": 6244684.0, + "step": 2275 + }, + { + "epoch": 6.877126654064273, + "grad_norm": 1.4296875, + "learning_rate": 9.92323607350206e-06, + "loss": 0.7113, + "num_tokens": 6268309.0, + "step": 2276 + }, + { + "epoch": 6.880151228733459, + "grad_norm": 1.5390625, + "learning_rate": 9.922423956011245e-06, + "loss": 0.7843, + "num_tokens": 6292444.0, + "step": 2277 + }, + { + "epoch": 6.883175803402646, + "grad_norm": 1.5625, + "learning_rate": 9.921607602578284e-06, + "loss": 0.7758, + "num_tokens": 6315628.0, + "step": 2278 + }, + { + "epoch": 6.886200378071834, + "grad_norm": 1.96875, + "learning_rate": 9.920787013985106e-06, + "loss": 0.5706, + "num_tokens": 6322509.0, + "step": 2279 + }, + { + "epoch": 6.889224952741021, + "grad_norm": 1.390625, + "learning_rate": 9.919962191017703e-06, + "loss": 0.7833, + "num_tokens": 6345053.0, + "step": 2280 + }, + { + "epoch": 6.892249527410208, + "grad_norm": 1.8125, + "learning_rate": 9.919133134466111e-06, + "loss": 0.5662, + "num_tokens": 6352752.0, + "step": 2281 + }, + { + "epoch": 6.895274102079395, + "grad_norm": 1.7578125, + "learning_rate": 9.918299845124433e-06, + "loss": 0.6277, + "num_tokens": 6364194.0, + "step": 2282 + }, + { + "epoch": 6.898298676748582, + "grad_norm": 1.1875, + "learning_rate": 9.917462323790818e-06, + "loss": 0.8677, + "num_tokens": 6406128.0, + "step": 2283 + }, + { + "epoch": 6.90132325141777, + "grad_norm": 1.9609375, + "learning_rate": 9.916620571267469e-06, + "loss": 0.6108, + "num_tokens": 6414481.0, + "step": 2284 + }, + { + "epoch": 6.904347826086957, + "grad_norm": 1.0859375, + "learning_rate": 9.915774588360649e-06, + "loss": 0.7964, + "num_tokens": 6467771.0, + "step": 2285 + }, + { + "epoch": 6.9073724007561434, + "grad_norm": 1.8984375, + "learning_rate": 9.914924375880663e-06, + "loss": 0.5836, + "num_tokens": 6476285.0, + "step": 2286 + }, + { + "epoch": 6.910396975425331, + "grad_norm": 1.3515625, + "learning_rate": 9.914069934641877e-06, + "loss": 0.7732, + "num_tokens": 6499683.0, + "step": 2287 + }, + { + "epoch": 6.913421550094518, + "grad_norm": 1.5, + "learning_rate": 9.9132112654627e-06, + "loss": 0.7408, + "num_tokens": 6522649.0, + "step": 2288 + }, + { + "epoch": 6.916446124763705, + "grad_norm": 1.9296875, + "learning_rate": 9.912348369165594e-06, + "loss": 0.6789, + "num_tokens": 6531920.0, + "step": 2289 + }, + { + "epoch": 6.9194706994328925, + "grad_norm": 1.1640625, + "learning_rate": 9.91148124657707e-06, + "loss": 0.8402, + "num_tokens": 6570722.0, + "step": 2290 + }, + { + "epoch": 6.922495274102079, + "grad_norm": 1.796875, + "learning_rate": 9.910609898527686e-06, + "loss": 0.9596, + "num_tokens": 6593382.0, + "step": 2291 + }, + { + "epoch": 6.925519848771266, + "grad_norm": 2.015625, + "learning_rate": 9.909734325852046e-06, + "loss": 0.586, + "num_tokens": 6600874.0, + "step": 2292 + }, + { + "epoch": 6.928544423440454, + "grad_norm": 1.8984375, + "learning_rate": 9.908854529388802e-06, + "loss": 0.6205, + "num_tokens": 6610400.0, + "step": 2293 + }, + { + "epoch": 6.931568998109641, + "grad_norm": 1.25, + "learning_rate": 9.907970509980657e-06, + "loss": 0.853, + "num_tokens": 6650349.0, + "step": 2294 + }, + { + "epoch": 6.934593572778828, + "grad_norm": 1.875, + "learning_rate": 9.907082268474348e-06, + "loss": 0.5324, + "num_tokens": 6657967.0, + "step": 2295 + }, + { + "epoch": 6.937618147448015, + "grad_norm": 1.2734375, + "learning_rate": 9.906189805720664e-06, + "loss": 0.8822, + "num_tokens": 6698236.0, + "step": 2296 + }, + { + "epoch": 6.940642722117202, + "grad_norm": 1.2734375, + "learning_rate": 9.905293122574433e-06, + "loss": 0.8409, + "num_tokens": 6737527.0, + "step": 2297 + }, + { + "epoch": 6.94366729678639, + "grad_norm": 1.59375, + "learning_rate": 9.90439221989453e-06, + "loss": 0.8063, + "num_tokens": 6757784.0, + "step": 2298 + }, + { + "epoch": 6.9466918714555765, + "grad_norm": 1.296875, + "learning_rate": 9.90348709854387e-06, + "loss": 0.9276, + "num_tokens": 6798559.0, + "step": 2299 + }, + { + "epoch": 6.949716446124763, + "grad_norm": 1.8046875, + "learning_rate": 9.902577759389402e-06, + "loss": 0.6773, + "num_tokens": 6809176.0, + "step": 2300 + }, + { + "epoch": 6.952741020793951, + "grad_norm": 1.65625, + "learning_rate": 9.901664203302126e-06, + "loss": 0.8485, + "num_tokens": 6832534.0, + "step": 2301 + }, + { + "epoch": 6.955765595463138, + "grad_norm": 1.4375, + "learning_rate": 9.900746431157074e-06, + "loss": 0.6388, + "num_tokens": 6849616.0, + "step": 2302 + }, + { + "epoch": 6.958790170132325, + "grad_norm": 2.015625, + "learning_rate": 9.89982444383332e-06, + "loss": 0.5613, + "num_tokens": 6856863.0, + "step": 2303 + }, + { + "epoch": 6.961814744801512, + "grad_norm": 2.140625, + "learning_rate": 9.898898242213971e-06, + "loss": 0.7262, + "num_tokens": 6864545.0, + "step": 2304 + }, + { + "epoch": 6.964839319470699, + "grad_norm": 1.796875, + "learning_rate": 9.897967827186175e-06, + "loss": 0.6509, + "num_tokens": 6875124.0, + "step": 2305 + }, + { + "epoch": 6.967863894139887, + "grad_norm": 1.515625, + "learning_rate": 9.897033199641114e-06, + "loss": 0.7282, + "num_tokens": 6890571.0, + "step": 2306 + }, + { + "epoch": 6.970888468809074, + "grad_norm": 1.265625, + "learning_rate": 9.896094360474e-06, + "loss": 0.8419, + "num_tokens": 6930173.0, + "step": 2307 + }, + { + "epoch": 6.973913043478261, + "grad_norm": 1.578125, + "learning_rate": 9.895151310584092e-06, + "loss": 0.8317, + "num_tokens": 6953811.0, + "step": 2308 + }, + { + "epoch": 6.976937618147448, + "grad_norm": 1.546875, + "learning_rate": 9.89420405087467e-06, + "loss": 0.8453, + "num_tokens": 6977629.0, + "step": 2309 + }, + { + "epoch": 6.979962192816635, + "grad_norm": 1.6015625, + "learning_rate": 9.893252582253048e-06, + "loss": 0.8743, + "num_tokens": 7000801.0, + "step": 2310 + }, + { + "epoch": 6.982986767485822, + "grad_norm": 1.8671875, + "learning_rate": 9.892296905630576e-06, + "loss": 0.5638, + "num_tokens": 7008330.0, + "step": 2311 + }, + { + "epoch": 6.98601134215501, + "grad_norm": 1.1953125, + "learning_rate": 9.891337021922633e-06, + "loss": 0.8862, + "num_tokens": 7051752.0, + "step": 2312 + }, + { + "epoch": 6.9890359168241964, + "grad_norm": 1.921875, + "learning_rate": 9.890372932048627e-06, + "loss": 0.5723, + "num_tokens": 7061560.0, + "step": 2313 + }, + { + "epoch": 6.992060491493383, + "grad_norm": 1.2890625, + "learning_rate": 9.889404636931991e-06, + "loss": 0.8112, + "num_tokens": 7099810.0, + "step": 2314 + }, + { + "epoch": 6.995085066162571, + "grad_norm": 1.640625, + "learning_rate": 9.888432137500194e-06, + "loss": 0.9016, + "num_tokens": 7122661.0, + "step": 2315 + }, + { + "epoch": 6.998109640831758, + "grad_norm": 1.0625, + "learning_rate": 9.887455434684728e-06, + "loss": 0.8399, + "num_tokens": 7174797.0, + "step": 2316 + }, + { + "epoch": 7.0, + "grad_norm": 1.5625, + "learning_rate": 9.886474529421108e-06, + "loss": 0.7552, + "num_tokens": 7195313.0, + "step": 2317 + }, + { + "epoch": 7.003024574669187, + "grad_norm": 1.078125, + "learning_rate": 9.885489422648878e-06, + "loss": 0.7948, + "num_tokens": 7235142.0, + "step": 2318 + }, + { + "epoch": 7.0060491493383745, + "grad_norm": 1.5859375, + "learning_rate": 9.884500115311606e-06, + "loss": 0.8025, + "num_tokens": 7256628.0, + "step": 2319 + }, + { + "epoch": 7.009073724007561, + "grad_norm": 1.5546875, + "learning_rate": 9.883506608356884e-06, + "loss": 0.8333, + "num_tokens": 7282117.0, + "step": 2320 + }, + { + "epoch": 7.012098298676748, + "grad_norm": 1.390625, + "learning_rate": 9.88250890273632e-06, + "loss": 0.7523, + "num_tokens": 7305598.0, + "step": 2321 + }, + { + "epoch": 7.015122873345936, + "grad_norm": 1.015625, + "learning_rate": 9.881506999405557e-06, + "loss": 0.8909, + "num_tokens": 7360974.0, + "step": 2322 + }, + { + "epoch": 7.018147448015123, + "grad_norm": 1.46875, + "learning_rate": 9.880500899324246e-06, + "loss": 0.789, + "num_tokens": 7383642.0, + "step": 2323 + }, + { + "epoch": 7.02117202268431, + "grad_norm": 1.8515625, + "learning_rate": 9.879490603456062e-06, + "loss": 0.5975, + "num_tokens": 7392354.0, + "step": 2324 + }, + { + "epoch": 7.024196597353497, + "grad_norm": 1.28125, + "learning_rate": 9.878476112768701e-06, + "loss": 0.7619, + "num_tokens": 7419960.0, + "step": 2325 + }, + { + "epoch": 7.027221172022684, + "grad_norm": 1.265625, + "learning_rate": 9.877457428233877e-06, + "loss": 0.8125, + "num_tokens": 7456419.0, + "step": 2326 + }, + { + "epoch": 7.030245746691872, + "grad_norm": 1.578125, + "learning_rate": 9.876434550827315e-06, + "loss": 0.7064, + "num_tokens": 7470600.0, + "step": 2327 + }, + { + "epoch": 7.033270321361059, + "grad_norm": 1.1875, + "learning_rate": 9.875407481528765e-06, + "loss": 0.8261, + "num_tokens": 7505320.0, + "step": 2328 + }, + { + "epoch": 7.036294896030245, + "grad_norm": 1.390625, + "learning_rate": 9.874376221321983e-06, + "loss": 0.7718, + "num_tokens": 7528900.0, + "step": 2329 + }, + { + "epoch": 7.039319470699433, + "grad_norm": 1.90625, + "learning_rate": 9.873340771194749e-06, + "loss": 0.6611, + "num_tokens": 7539930.0, + "step": 2330 + }, + { + "epoch": 7.04234404536862, + "grad_norm": 1.6796875, + "learning_rate": 9.872301132138845e-06, + "loss": 0.6477, + "num_tokens": 7557906.0, + "step": 2331 + }, + { + "epoch": 7.045368620037808, + "grad_norm": 1.703125, + "learning_rate": 9.871257305150076e-06, + "loss": 0.7168, + "num_tokens": 7574385.0, + "step": 2332 + }, + { + "epoch": 7.048393194706994, + "grad_norm": 1.4453125, + "learning_rate": 9.87020929122825e-06, + "loss": 0.8839, + "num_tokens": 7604982.0, + "step": 2333 + }, + { + "epoch": 7.051417769376181, + "grad_norm": 1.578125, + "learning_rate": 9.869157091377192e-06, + "loss": 0.8429, + "num_tokens": 7628504.0, + "step": 2334 + }, + { + "epoch": 7.054442344045369, + "grad_norm": 1.6015625, + "learning_rate": 9.868100706604733e-06, + "loss": 0.7365, + "num_tokens": 7649161.0, + "step": 2335 + }, + { + "epoch": 7.057466918714556, + "grad_norm": 1.703125, + "learning_rate": 9.867040137922712e-06, + "loss": 0.9001, + "num_tokens": 7675772.0, + "step": 2336 + }, + { + "epoch": 7.060491493383743, + "grad_norm": 1.59375, + "learning_rate": 9.865975386346977e-06, + "loss": 0.844, + "num_tokens": 7699305.0, + "step": 2337 + }, + { + "epoch": 7.06351606805293, + "grad_norm": 2.078125, + "learning_rate": 9.86490645289738e-06, + "loss": 0.6297, + "num_tokens": 7706486.0, + "step": 2338 + }, + { + "epoch": 7.066540642722117, + "grad_norm": 2.015625, + "learning_rate": 9.86383333859778e-06, + "loss": 0.5962, + "num_tokens": 7713825.0, + "step": 2339 + }, + { + "epoch": 7.069565217391304, + "grad_norm": 2.0625, + "learning_rate": 9.862756044476044e-06, + "loss": 0.5888, + "num_tokens": 7721721.0, + "step": 2340 + }, + { + "epoch": 7.072589792060492, + "grad_norm": 2.265625, + "learning_rate": 9.86167457156404e-06, + "loss": 0.579, + "num_tokens": 7728309.0, + "step": 2341 + }, + { + "epoch": 7.0756143667296785, + "grad_norm": 1.3515625, + "learning_rate": 9.860588920897633e-06, + "loss": 0.7577, + "num_tokens": 7756475.0, + "step": 2342 + }, + { + "epoch": 7.078638941398866, + "grad_norm": 1.859375, + "learning_rate": 9.8594990935167e-06, + "loss": 0.5872, + "num_tokens": 7765155.0, + "step": 2343 + }, + { + "epoch": 7.081663516068053, + "grad_norm": 1.390625, + "learning_rate": 9.85840509046511e-06, + "loss": 0.7114, + "num_tokens": 7788676.0, + "step": 2344 + }, + { + "epoch": 7.08468809073724, + "grad_norm": 1.609375, + "learning_rate": 9.857306912790737e-06, + "loss": 0.7668, + "num_tokens": 7810340.0, + "step": 2345 + }, + { + "epoch": 7.0877126654064275, + "grad_norm": 1.4375, + "learning_rate": 9.856204561545449e-06, + "loss": 0.7648, + "num_tokens": 7834560.0, + "step": 2346 + }, + { + "epoch": 7.090737240075614, + "grad_norm": 1.5390625, + "learning_rate": 9.855098037785118e-06, + "loss": 0.8282, + "num_tokens": 7857763.0, + "step": 2347 + }, + { + "epoch": 7.093761814744801, + "grad_norm": 1.9765625, + "learning_rate": 9.853987342569604e-06, + "loss": 0.5915, + "num_tokens": 7865395.0, + "step": 2348 + }, + { + "epoch": 7.096786389413989, + "grad_norm": 1.640625, + "learning_rate": 9.852872476962769e-06, + "loss": 0.8391, + "num_tokens": 7888191.0, + "step": 2349 + }, + { + "epoch": 7.099810964083176, + "grad_norm": 1.3828125, + "learning_rate": 9.851753442032467e-06, + "loss": 0.841, + "num_tokens": 7923125.0, + "step": 2350 + }, + { + "epoch": 7.1028355387523625, + "grad_norm": 2.0, + "learning_rate": 9.850630238850549e-06, + "loss": 0.6141, + "num_tokens": 7930110.0, + "step": 2351 + }, + { + "epoch": 7.10586011342155, + "grad_norm": 1.3125, + "learning_rate": 9.849502868492852e-06, + "loss": 0.8303, + "num_tokens": 7969542.0, + "step": 2352 + }, + { + "epoch": 7.108884688090737, + "grad_norm": 1.6875, + "learning_rate": 9.848371332039214e-06, + "loss": 0.7717, + "num_tokens": 7986228.0, + "step": 2353 + }, + { + "epoch": 7.111909262759925, + "grad_norm": 1.7734375, + "learning_rate": 9.84723563057345e-06, + "loss": 0.5486, + "num_tokens": 7993641.0, + "step": 2354 + }, + { + "epoch": 7.114933837429112, + "grad_norm": 1.46875, + "learning_rate": 9.846095765183376e-06, + "loss": 0.7253, + "num_tokens": 8013324.0, + "step": 2355 + }, + { + "epoch": 7.117958412098298, + "grad_norm": 1.421875, + "learning_rate": 9.844951736960795e-06, + "loss": 0.7702, + "num_tokens": 8037673.0, + "step": 2356 + }, + { + "epoch": 7.120982986767486, + "grad_norm": 1.5234375, + "learning_rate": 9.843803547001487e-06, + "loss": 0.8014, + "num_tokens": 8060361.0, + "step": 2357 + }, + { + "epoch": 7.124007561436673, + "grad_norm": 1.8203125, + "learning_rate": 9.842651196405232e-06, + "loss": 0.6129, + "num_tokens": 8071458.0, + "step": 2358 + }, + { + "epoch": 7.12703213610586, + "grad_norm": 1.9921875, + "learning_rate": 9.841494686275786e-06, + "loss": 0.6547, + "num_tokens": 8082464.0, + "step": 2359 + }, + { + "epoch": 7.130056710775047, + "grad_norm": 1.203125, + "learning_rate": 9.840334017720896e-06, + "loss": 0.8648, + "num_tokens": 8120677.0, + "step": 2360 + }, + { + "epoch": 7.133081285444234, + "grad_norm": 1.4296875, + "learning_rate": 9.839169191852286e-06, + "loss": 0.8501, + "num_tokens": 8144694.0, + "step": 2361 + }, + { + "epoch": 7.136105860113422, + "grad_norm": 1.84375, + "learning_rate": 9.838000209785663e-06, + "loss": 0.7781, + "num_tokens": 8159508.0, + "step": 2362 + }, + { + "epoch": 7.139130434782609, + "grad_norm": 1.6953125, + "learning_rate": 9.836827072640716e-06, + "loss": 0.6239, + "num_tokens": 8172019.0, + "step": 2363 + }, + { + "epoch": 7.142155009451796, + "grad_norm": 1.4140625, + "learning_rate": 9.835649781541117e-06, + "loss": 0.6687, + "num_tokens": 8191987.0, + "step": 2364 + }, + { + "epoch": 7.145179584120983, + "grad_norm": 2.015625, + "learning_rate": 9.834468337614512e-06, + "loss": 0.6038, + "num_tokens": 8199823.0, + "step": 2365 + }, + { + "epoch": 7.14820415879017, + "grad_norm": 1.3359375, + "learning_rate": 9.833282741992526e-06, + "loss": 0.6928, + "num_tokens": 8223031.0, + "step": 2366 + }, + { + "epoch": 7.151228733459357, + "grad_norm": 1.4921875, + "learning_rate": 9.832092995810763e-06, + "loss": 0.7524, + "num_tokens": 8244058.0, + "step": 2367 + }, + { + "epoch": 7.154253308128545, + "grad_norm": 2.09375, + "learning_rate": 9.830899100208799e-06, + "loss": 0.5916, + "num_tokens": 8251229.0, + "step": 2368 + }, + { + "epoch": 7.1572778827977315, + "grad_norm": 1.3828125, + "learning_rate": 9.829701056330188e-06, + "loss": 0.8045, + "num_tokens": 8278286.0, + "step": 2369 + }, + { + "epoch": 7.160302457466918, + "grad_norm": 1.625, + "learning_rate": 9.828498865322454e-06, + "loss": 0.7911, + "num_tokens": 8301538.0, + "step": 2370 + }, + { + "epoch": 7.163327032136106, + "grad_norm": 1.59375, + "learning_rate": 9.827292528337095e-06, + "loss": 0.8772, + "num_tokens": 8324649.0, + "step": 2371 + }, + { + "epoch": 7.166351606805293, + "grad_norm": 1.09375, + "learning_rate": 9.826082046529581e-06, + "loss": 0.8245, + "num_tokens": 8372251.0, + "step": 2372 + }, + { + "epoch": 7.1693761814744805, + "grad_norm": 1.3359375, + "learning_rate": 9.82486742105935e-06, + "loss": 0.7767, + "num_tokens": 8399831.0, + "step": 2373 + }, + { + "epoch": 7.172400756143667, + "grad_norm": 2.0, + "learning_rate": 9.82364865308981e-06, + "loss": 0.6129, + "num_tokens": 8407400.0, + "step": 2374 + }, + { + "epoch": 7.175425330812854, + "grad_norm": 1.4921875, + "learning_rate": 9.82242574378834e-06, + "loss": 0.7248, + "num_tokens": 8430530.0, + "step": 2375 + }, + { + "epoch": 7.178449905482042, + "grad_norm": 1.53125, + "learning_rate": 9.821198694326278e-06, + "loss": 0.7177, + "num_tokens": 8450259.0, + "step": 2376 + }, + { + "epoch": 7.181474480151229, + "grad_norm": 1.8046875, + "learning_rate": 9.819967505878932e-06, + "loss": 0.6404, + "num_tokens": 8460897.0, + "step": 2377 + }, + { + "epoch": 7.1844990548204155, + "grad_norm": 1.9921875, + "learning_rate": 9.818732179625578e-06, + "loss": 0.621, + "num_tokens": 8468243.0, + "step": 2378 + }, + { + "epoch": 7.187523629489603, + "grad_norm": 1.140625, + "learning_rate": 9.81749271674945e-06, + "loss": 0.8491, + "num_tokens": 8524150.0, + "step": 2379 + }, + { + "epoch": 7.19054820415879, + "grad_norm": 1.3984375, + "learning_rate": 9.816249118437746e-06, + "loss": 0.7751, + "num_tokens": 8551813.0, + "step": 2380 + }, + { + "epoch": 7.193572778827978, + "grad_norm": 1.25, + "learning_rate": 9.815001385881624e-06, + "loss": 0.8804, + "num_tokens": 8590618.0, + "step": 2381 + }, + { + "epoch": 7.196597353497165, + "grad_norm": 1.2421875, + "learning_rate": 9.813749520276205e-06, + "loss": 0.7601, + "num_tokens": 8621147.0, + "step": 2382 + }, + { + "epoch": 7.199621928166351, + "grad_norm": 2.078125, + "learning_rate": 9.812493522820567e-06, + "loss": 0.5931, + "num_tokens": 8628176.0, + "step": 2383 + }, + { + "epoch": 7.202646502835539, + "grad_norm": 1.59375, + "learning_rate": 9.811233394717742e-06, + "loss": 0.8154, + "num_tokens": 8652548.0, + "step": 2384 + }, + { + "epoch": 7.205671077504726, + "grad_norm": 1.171875, + "learning_rate": 9.809969137174725e-06, + "loss": 0.8344, + "num_tokens": 8687653.0, + "step": 2385 + }, + { + "epoch": 7.208695652173913, + "grad_norm": 1.4921875, + "learning_rate": 9.808700751402465e-06, + "loss": 0.8483, + "num_tokens": 8710226.0, + "step": 2386 + }, + { + "epoch": 7.2117202268431, + "grad_norm": 2.0625, + "learning_rate": 9.807428238615858e-06, + "loss": 0.6416, + "num_tokens": 8717604.0, + "step": 2387 + }, + { + "epoch": 7.214744801512287, + "grad_norm": 1.546875, + "learning_rate": 9.806151600033761e-06, + "loss": 0.7736, + "num_tokens": 8734425.0, + "step": 2388 + }, + { + "epoch": 7.217769376181474, + "grad_norm": 1.453125, + "learning_rate": 9.804870836878982e-06, + "loss": 0.8106, + "num_tokens": 8757936.0, + "step": 2389 + }, + { + "epoch": 7.220793950850662, + "grad_norm": 1.6328125, + "learning_rate": 9.803585950378274e-06, + "loss": 0.816, + "num_tokens": 8781024.0, + "step": 2390 + }, + { + "epoch": 7.223818525519849, + "grad_norm": 1.640625, + "learning_rate": 9.802296941762347e-06, + "loss": 0.7642, + "num_tokens": 8799342.0, + "step": 2391 + }, + { + "epoch": 7.226843100189036, + "grad_norm": 1.2578125, + "learning_rate": 9.801003812265853e-06, + "loss": 0.8629, + "num_tokens": 8840282.0, + "step": 2392 + }, + { + "epoch": 7.229867674858223, + "grad_norm": 1.21875, + "learning_rate": 9.799706563127395e-06, + "loss": 0.8546, + "num_tokens": 8882521.0, + "step": 2393 + }, + { + "epoch": 7.23289224952741, + "grad_norm": 2.0625, + "learning_rate": 9.79840519558952e-06, + "loss": 0.5877, + "num_tokens": 8889655.0, + "step": 2394 + }, + { + "epoch": 7.235916824196598, + "grad_norm": 1.828125, + "learning_rate": 9.79709971089872e-06, + "loss": 0.6075, + "num_tokens": 8897803.0, + "step": 2395 + }, + { + "epoch": 7.2389413988657845, + "grad_norm": 1.8828125, + "learning_rate": 9.795790110305431e-06, + "loss": 0.5711, + "num_tokens": 8905332.0, + "step": 2396 + }, + { + "epoch": 7.241965973534971, + "grad_norm": 1.2578125, + "learning_rate": 9.794476395064032e-06, + "loss": 0.8265, + "num_tokens": 8943981.0, + "step": 2397 + }, + { + "epoch": 7.244990548204159, + "grad_norm": 1.3125, + "learning_rate": 9.793158566432843e-06, + "loss": 0.8627, + "num_tokens": 8983236.0, + "step": 2398 + }, + { + "epoch": 7.248015122873346, + "grad_norm": 1.4296875, + "learning_rate": 9.79183662567412e-06, + "loss": 0.6819, + "num_tokens": 9007313.0, + "step": 2399 + }, + { + "epoch": 7.251039697542533, + "grad_norm": 1.2734375, + "learning_rate": 9.790510574054065e-06, + "loss": 0.9029, + "num_tokens": 9049423.0, + "step": 2400 + }, + { + "epoch": 7.25406427221172, + "grad_norm": 1.5078125, + "learning_rate": 9.789180412842813e-06, + "loss": 0.7936, + "num_tokens": 9072714.0, + "step": 2401 + }, + { + "epoch": 7.257088846880907, + "grad_norm": 1.921875, + "learning_rate": 9.787846143314433e-06, + "loss": 0.5512, + "num_tokens": 9080472.0, + "step": 2402 + }, + { + "epoch": 7.260113421550095, + "grad_norm": 1.8515625, + "learning_rate": 9.786507766746932e-06, + "loss": 0.5984, + "num_tokens": 9088375.0, + "step": 2403 + }, + { + "epoch": 7.263137996219282, + "grad_norm": 1.7734375, + "learning_rate": 9.785165284422256e-06, + "loss": 0.5797, + "num_tokens": 9097472.0, + "step": 2404 + }, + { + "epoch": 7.2661625708884685, + "grad_norm": 1.4453125, + "learning_rate": 9.783818697626273e-06, + "loss": 0.7052, + "num_tokens": 9119431.0, + "step": 2405 + }, + { + "epoch": 7.269187145557656, + "grad_norm": 1.515625, + "learning_rate": 9.78246800764879e-06, + "loss": 0.7256, + "num_tokens": 9138796.0, + "step": 2406 + }, + { + "epoch": 7.272211720226843, + "grad_norm": 1.65625, + "learning_rate": 9.781113215783541e-06, + "loss": 0.8119, + "num_tokens": 9162163.0, + "step": 2407 + }, + { + "epoch": 7.27523629489603, + "grad_norm": 1.53125, + "learning_rate": 9.779754323328192e-06, + "loss": 0.7925, + "num_tokens": 9183411.0, + "step": 2408 + }, + { + "epoch": 7.278260869565218, + "grad_norm": 1.7578125, + "learning_rate": 9.778391331584332e-06, + "loss": 0.542, + "num_tokens": 9191678.0, + "step": 2409 + }, + { + "epoch": 7.281285444234404, + "grad_norm": 1.7109375, + "learning_rate": 9.777024241857482e-06, + "loss": 0.8387, + "num_tokens": 9214400.0, + "step": 2410 + }, + { + "epoch": 7.284310018903592, + "grad_norm": 1.3359375, + "learning_rate": 9.775653055457082e-06, + "loss": 0.7589, + "num_tokens": 9237665.0, + "step": 2411 + }, + { + "epoch": 7.287334593572779, + "grad_norm": 2.0, + "learning_rate": 9.7742777736965e-06, + "loss": 0.6213, + "num_tokens": 9244569.0, + "step": 2412 + }, + { + "epoch": 7.290359168241966, + "grad_norm": 1.1875, + "learning_rate": 9.772898397893028e-06, + "loss": 0.8051, + "num_tokens": 9284687.0, + "step": 2413 + }, + { + "epoch": 7.293383742911153, + "grad_norm": 1.5390625, + "learning_rate": 9.771514929367875e-06, + "loss": 0.9174, + "num_tokens": 9308603.0, + "step": 2414 + }, + { + "epoch": 7.29640831758034, + "grad_norm": 1.734375, + "learning_rate": 9.770127369446173e-06, + "loss": 0.584, + "num_tokens": 9317380.0, + "step": 2415 + }, + { + "epoch": 7.299432892249527, + "grad_norm": 1.953125, + "learning_rate": 9.76873571945697e-06, + "loss": 0.6002, + "num_tokens": 9324855.0, + "step": 2416 + }, + { + "epoch": 7.302457466918715, + "grad_norm": 0.9921875, + "learning_rate": 9.76733998073324e-06, + "loss": 0.7899, + "num_tokens": 9380519.0, + "step": 2417 + }, + { + "epoch": 7.305482041587902, + "grad_norm": 1.953125, + "learning_rate": 9.76594015461186e-06, + "loss": 0.6067, + "num_tokens": 9390894.0, + "step": 2418 + }, + { + "epoch": 7.308506616257089, + "grad_norm": 1.4453125, + "learning_rate": 9.764536242433635e-06, + "loss": 0.6769, + "num_tokens": 9409879.0, + "step": 2419 + }, + { + "epoch": 7.311531190926276, + "grad_norm": 1.0546875, + "learning_rate": 9.763128245543272e-06, + "loss": 0.7997, + "num_tokens": 9461144.0, + "step": 2420 + }, + { + "epoch": 7.314555765595463, + "grad_norm": 1.3203125, + "learning_rate": 9.761716165289402e-06, + "loss": 0.8422, + "num_tokens": 9494754.0, + "step": 2421 + }, + { + "epoch": 7.317580340264651, + "grad_norm": 1.6875, + "learning_rate": 9.76030000302456e-06, + "loss": 0.7706, + "num_tokens": 9510280.0, + "step": 2422 + }, + { + "epoch": 7.3206049149338375, + "grad_norm": 2.015625, + "learning_rate": 9.75887976010519e-06, + "loss": 0.5455, + "num_tokens": 9517201.0, + "step": 2423 + }, + { + "epoch": 7.323629489603024, + "grad_norm": 1.015625, + "learning_rate": 9.757455437891649e-06, + "loss": 0.8373, + "num_tokens": 9571499.0, + "step": 2424 + }, + { + "epoch": 7.326654064272212, + "grad_norm": 1.8671875, + "learning_rate": 9.7560270377482e-06, + "loss": 0.6647, + "num_tokens": 9581999.0, + "step": 2425 + }, + { + "epoch": 7.329678638941399, + "grad_norm": 1.1640625, + "learning_rate": 9.75459456104301e-06, + "loss": 0.9357, + "num_tokens": 9636923.0, + "step": 2426 + }, + { + "epoch": 7.332703213610586, + "grad_norm": 1.3125, + "learning_rate": 9.75315800914815e-06, + "loss": 0.8575, + "num_tokens": 9675033.0, + "step": 2427 + }, + { + "epoch": 7.335727788279773, + "grad_norm": 1.328125, + "learning_rate": 9.751717383439604e-06, + "loss": 0.7397, + "num_tokens": 9699157.0, + "step": 2428 + }, + { + "epoch": 7.33875236294896, + "grad_norm": 1.5703125, + "learning_rate": 9.750272685297241e-06, + "loss": 0.8575, + "num_tokens": 9721827.0, + "step": 2429 + }, + { + "epoch": 7.341776937618148, + "grad_norm": 1.125, + "learning_rate": 9.748823916104847e-06, + "loss": 0.8441, + "num_tokens": 9769867.0, + "step": 2430 + }, + { + "epoch": 7.344801512287335, + "grad_norm": 1.421875, + "learning_rate": 9.747371077250096e-06, + "loss": 0.6753, + "num_tokens": 9791334.0, + "step": 2431 + }, + { + "epoch": 7.3478260869565215, + "grad_norm": 1.734375, + "learning_rate": 9.745914170124568e-06, + "loss": 0.7164, + "num_tokens": 9804906.0, + "step": 2432 + }, + { + "epoch": 7.350850661625709, + "grad_norm": 1.484375, + "learning_rate": 9.744453196123733e-06, + "loss": 0.7784, + "num_tokens": 9828690.0, + "step": 2433 + }, + { + "epoch": 7.353875236294896, + "grad_norm": 1.9921875, + "learning_rate": 9.742988156646962e-06, + "loss": 0.6592, + "num_tokens": 9836390.0, + "step": 2434 + }, + { + "epoch": 7.356899810964083, + "grad_norm": 1.4140625, + "learning_rate": 9.741519053097516e-06, + "loss": 0.8113, + "num_tokens": 9863588.0, + "step": 2435 + }, + { + "epoch": 7.359924385633271, + "grad_norm": 1.6875, + "learning_rate": 9.740045886882552e-06, + "loss": 0.8177, + "num_tokens": 9886656.0, + "step": 2436 + }, + { + "epoch": 7.362948960302457, + "grad_norm": 1.9453125, + "learning_rate": 9.738568659413116e-06, + "loss": 0.5565, + "num_tokens": 9894405.0, + "step": 2437 + }, + { + "epoch": 7.365973534971644, + "grad_norm": 1.734375, + "learning_rate": 9.737087372104143e-06, + "loss": 0.6533, + "num_tokens": 9906895.0, + "step": 2438 + }, + { + "epoch": 7.368998109640832, + "grad_norm": 2.140625, + "learning_rate": 9.73560202637446e-06, + "loss": 0.568, + "num_tokens": 9913453.0, + "step": 2439 + }, + { + "epoch": 7.372022684310019, + "grad_norm": 1.6484375, + "learning_rate": 9.734112623646783e-06, + "loss": 0.7314, + "num_tokens": 9931294.0, + "step": 2440 + }, + { + "epoch": 7.375047258979206, + "grad_norm": 1.7578125, + "learning_rate": 9.732619165347705e-06, + "loss": 0.7375, + "num_tokens": 9946025.0, + "step": 2441 + }, + { + "epoch": 7.378071833648393, + "grad_norm": 1.8203125, + "learning_rate": 9.731121652907712e-06, + "loss": 0.7087, + "num_tokens": 9960709.0, + "step": 2442 + }, + { + "epoch": 7.38109640831758, + "grad_norm": 1.4140625, + "learning_rate": 9.729620087761167e-06, + "loss": 0.7571, + "num_tokens": 9984858.0, + "step": 2443 + }, + { + "epoch": 7.384120982986768, + "grad_norm": 1.890625, + "learning_rate": 9.728114471346324e-06, + "loss": 0.5809, + "num_tokens": 9992839.0, + "step": 2444 + }, + { + "epoch": 7.387145557655955, + "grad_norm": 1.40625, + "learning_rate": 9.72660480510531e-06, + "loss": 0.7944, + "num_tokens": 10017529.0, + "step": 2445 + }, + { + "epoch": 7.390170132325141, + "grad_norm": 0.9609375, + "learning_rate": 9.725091090484128e-06, + "loss": 0.8604, + "num_tokens": 10089247.0, + "step": 2446 + }, + { + "epoch": 7.393194706994329, + "grad_norm": 2.015625, + "learning_rate": 9.723573328932669e-06, + "loss": 0.6285, + "num_tokens": 10096575.0, + "step": 2447 + }, + { + "epoch": 7.396219281663516, + "grad_norm": 1.171875, + "learning_rate": 9.722051521904689e-06, + "loss": 0.8314, + "num_tokens": 10136047.0, + "step": 2448 + }, + { + "epoch": 7.399243856332703, + "grad_norm": 1.71875, + "learning_rate": 9.72052567085783e-06, + "loss": 0.5955, + "num_tokens": 10146886.0, + "step": 2449 + }, + { + "epoch": 7.4022684310018905, + "grad_norm": 1.2421875, + "learning_rate": 9.718995777253598e-06, + "loss": 0.8022, + "num_tokens": 10182657.0, + "step": 2450 + }, + { + "epoch": 7.405293005671077, + "grad_norm": 1.796875, + "learning_rate": 9.717461842557377e-06, + "loss": 0.6858, + "num_tokens": 10195852.0, + "step": 2451 + }, + { + "epoch": 7.408317580340265, + "grad_norm": 1.890625, + "learning_rate": 9.71592386823842e-06, + "loss": 0.6049, + "num_tokens": 10205433.0, + "step": 2452 + }, + { + "epoch": 7.411342155009452, + "grad_norm": 1.4140625, + "learning_rate": 9.71438185576985e-06, + "loss": 0.8397, + "num_tokens": 10232031.0, + "step": 2453 + }, + { + "epoch": 7.414366729678639, + "grad_norm": 1.2734375, + "learning_rate": 9.712835806628653e-06, + "loss": 0.836, + "num_tokens": 10267858.0, + "step": 2454 + }, + { + "epoch": 7.417391304347826, + "grad_norm": 1.4296875, + "learning_rate": 9.711285722295688e-06, + "loss": 0.7327, + "num_tokens": 10291532.0, + "step": 2455 + }, + { + "epoch": 7.420415879017013, + "grad_norm": 1.734375, + "learning_rate": 9.709731604255675e-06, + "loss": 0.7563, + "num_tokens": 10306458.0, + "step": 2456 + }, + { + "epoch": 7.4234404536862, + "grad_norm": 1.4765625, + "learning_rate": 9.708173453997202e-06, + "loss": 0.9072, + "num_tokens": 10336264.0, + "step": 2457 + }, + { + "epoch": 7.426465028355388, + "grad_norm": 1.234375, + "learning_rate": 9.706611273012715e-06, + "loss": 0.7462, + "num_tokens": 10372103.0, + "step": 2458 + }, + { + "epoch": 7.4294896030245745, + "grad_norm": 1.765625, + "learning_rate": 9.705045062798519e-06, + "loss": 0.5661, + "num_tokens": 10380870.0, + "step": 2459 + }, + { + "epoch": 7.432514177693762, + "grad_norm": 1.3203125, + "learning_rate": 9.703474824854785e-06, + "loss": 0.815, + "num_tokens": 10413276.0, + "step": 2460 + }, + { + "epoch": 7.435538752362949, + "grad_norm": 1.3125, + "learning_rate": 9.701900560685539e-06, + "loss": 0.8973, + "num_tokens": 10449336.0, + "step": 2461 + }, + { + "epoch": 7.438563327032136, + "grad_norm": 1.59375, + "learning_rate": 9.700322271798657e-06, + "loss": 0.8299, + "num_tokens": 10471971.0, + "step": 2462 + }, + { + "epoch": 7.441587901701324, + "grad_norm": 1.109375, + "learning_rate": 9.698739959705882e-06, + "loss": 0.8059, + "num_tokens": 10519935.0, + "step": 2463 + }, + { + "epoch": 7.44461247637051, + "grad_norm": 2.109375, + "learning_rate": 9.6971536259228e-06, + "loss": 0.5321, + "num_tokens": 10526716.0, + "step": 2464 + }, + { + "epoch": 7.447637051039697, + "grad_norm": 1.3203125, + "learning_rate": 9.695563271968853e-06, + "loss": 0.8126, + "num_tokens": 10558453.0, + "step": 2465 + }, + { + "epoch": 7.450661625708885, + "grad_norm": 1.484375, + "learning_rate": 9.69396889936734e-06, + "loss": 0.827, + "num_tokens": 10583971.0, + "step": 2466 + }, + { + "epoch": 7.453686200378072, + "grad_norm": 1.9765625, + "learning_rate": 9.692370509645396e-06, + "loss": 0.5703, + "num_tokens": 10590939.0, + "step": 2467 + }, + { + "epoch": 7.456710775047259, + "grad_norm": 1.9296875, + "learning_rate": 9.690768104334015e-06, + "loss": 0.5982, + "num_tokens": 10599811.0, + "step": 2468 + }, + { + "epoch": 7.459735349716446, + "grad_norm": 1.515625, + "learning_rate": 9.68916168496803e-06, + "loss": 0.7177, + "num_tokens": 10617884.0, + "step": 2469 + }, + { + "epoch": 7.462759924385633, + "grad_norm": 1.734375, + "learning_rate": 9.687551253086124e-06, + "loss": 0.6127, + "num_tokens": 10636179.0, + "step": 2470 + }, + { + "epoch": 7.465784499054821, + "grad_norm": 1.6171875, + "learning_rate": 9.685936810230824e-06, + "loss": 0.7347, + "num_tokens": 10659795.0, + "step": 2471 + }, + { + "epoch": 7.468809073724008, + "grad_norm": 2.171875, + "learning_rate": 9.684318357948492e-06, + "loss": 0.6635, + "num_tokens": 10666468.0, + "step": 2472 + }, + { + "epoch": 7.471833648393194, + "grad_norm": 2.1875, + "learning_rate": 9.682695897789334e-06, + "loss": 0.5481, + "num_tokens": 10673488.0, + "step": 2473 + }, + { + "epoch": 7.474858223062382, + "grad_norm": 1.421875, + "learning_rate": 9.6810694313074e-06, + "loss": 0.8311, + "num_tokens": 10696557.0, + "step": 2474 + }, + { + "epoch": 7.477882797731569, + "grad_norm": 1.3359375, + "learning_rate": 9.67943896006057e-06, + "loss": 0.8451, + "num_tokens": 10735636.0, + "step": 2475 + }, + { + "epoch": 7.480907372400756, + "grad_norm": 1.25, + "learning_rate": 9.677804485610558e-06, + "loss": 0.8054, + "num_tokens": 10770650.0, + "step": 2476 + }, + { + "epoch": 7.4839319470699435, + "grad_norm": 1.578125, + "learning_rate": 9.676166009522925e-06, + "loss": 0.8661, + "num_tokens": 10794079.0, + "step": 2477 + }, + { + "epoch": 7.48695652173913, + "grad_norm": 1.4921875, + "learning_rate": 9.674523533367052e-06, + "loss": 0.7705, + "num_tokens": 10814885.0, + "step": 2478 + }, + { + "epoch": 7.489981096408318, + "grad_norm": 1.1953125, + "learning_rate": 9.672877058716158e-06, + "loss": 0.7873, + "num_tokens": 10853484.0, + "step": 2479 + }, + { + "epoch": 7.493005671077505, + "grad_norm": 1.4921875, + "learning_rate": 9.67122658714729e-06, + "loss": 0.7159, + "num_tokens": 10876679.0, + "step": 2480 + }, + { + "epoch": 7.496030245746692, + "grad_norm": 1.3125, + "learning_rate": 9.669572120241326e-06, + "loss": 0.7998, + "num_tokens": 10914665.0, + "step": 2481 + }, + { + "epoch": 7.499054820415879, + "grad_norm": 1.1875, + "learning_rate": 9.667913659582962e-06, + "loss": 0.872, + "num_tokens": 10960959.0, + "step": 2482 + }, + { + "epoch": 7.502079395085066, + "grad_norm": 1.6328125, + "learning_rate": 9.666251206760732e-06, + "loss": 0.8469, + "num_tokens": 10983854.0, + "step": 2483 + }, + { + "epoch": 7.505103969754253, + "grad_norm": 1.3203125, + "learning_rate": 9.664584763366983e-06, + "loss": 0.8049, + "num_tokens": 11018702.0, + "step": 2484 + }, + { + "epoch": 7.508128544423441, + "grad_norm": 1.921875, + "learning_rate": 9.662914330997894e-06, + "loss": 0.5646, + "num_tokens": 11026575.0, + "step": 2485 + }, + { + "epoch": 7.5111531190926275, + "grad_norm": 1.734375, + "learning_rate": 9.661239911253457e-06, + "loss": 0.8709, + "num_tokens": 11048730.0, + "step": 2486 + }, + { + "epoch": 7.514177693761814, + "grad_norm": 2.0625, + "learning_rate": 9.659561505737486e-06, + "loss": 0.5864, + "num_tokens": 11055394.0, + "step": 2487 + }, + { + "epoch": 7.517202268431002, + "grad_norm": 1.5, + "learning_rate": 9.657879116057612e-06, + "loss": 0.8423, + "num_tokens": 11078604.0, + "step": 2488 + }, + { + "epoch": 7.520226843100189, + "grad_norm": 1.703125, + "learning_rate": 9.656192743825283e-06, + "loss": 0.81, + "num_tokens": 11096010.0, + "step": 2489 + }, + { + "epoch": 7.523251417769377, + "grad_norm": 1.9921875, + "learning_rate": 9.654502390655762e-06, + "loss": 0.7876, + "num_tokens": 11107875.0, + "step": 2490 + }, + { + "epoch": 7.526275992438563, + "grad_norm": 1.3828125, + "learning_rate": 9.652808058168125e-06, + "loss": 0.8513, + "num_tokens": 11147340.0, + "step": 2491 + }, + { + "epoch": 7.52930056710775, + "grad_norm": 1.453125, + "learning_rate": 9.651109747985257e-06, + "loss": 0.816, + "num_tokens": 11171663.0, + "step": 2492 + }, + { + "epoch": 7.532325141776938, + "grad_norm": 1.8828125, + "learning_rate": 9.649407461733857e-06, + "loss": 0.5414, + "num_tokens": 11178772.0, + "step": 2493 + }, + { + "epoch": 7.535349716446125, + "grad_norm": 1.078125, + "learning_rate": 9.647701201044429e-06, + "loss": 0.8273, + "num_tokens": 11233253.0, + "step": 2494 + }, + { + "epoch": 7.5383742911153115, + "grad_norm": 1.9453125, + "learning_rate": 9.645990967551287e-06, + "loss": 0.6069, + "num_tokens": 11245127.0, + "step": 2495 + }, + { + "epoch": 7.541398865784499, + "grad_norm": 1.75, + "learning_rate": 9.644276762892548e-06, + "loss": 0.5653, + "num_tokens": 11253135.0, + "step": 2496 + }, + { + "epoch": 7.544423440453686, + "grad_norm": 1.5, + "learning_rate": 9.64255858871013e-06, + "loss": 0.7481, + "num_tokens": 11276571.0, + "step": 2497 + }, + { + "epoch": 7.547448015122873, + "grad_norm": 2.4375, + "learning_rate": 9.640836446649761e-06, + "loss": 0.7222, + "num_tokens": 11283627.0, + "step": 2498 + }, + { + "epoch": 7.550472589792061, + "grad_norm": 2.0625, + "learning_rate": 9.639110338360963e-06, + "loss": 0.568, + "num_tokens": 11290459.0, + "step": 2499 + }, + { + "epoch": 7.553497164461247, + "grad_norm": 1.3046875, + "learning_rate": 9.637380265497057e-06, + "loss": 0.8822, + "num_tokens": 11330528.0, + "step": 2500 + }, + { + "epoch": 7.556521739130435, + "grad_norm": 2.015625, + "learning_rate": 9.635646229715168e-06, + "loss": 0.5939, + "num_tokens": 11337618.0, + "step": 2501 + }, + { + "epoch": 7.559546313799622, + "grad_norm": 1.515625, + "learning_rate": 9.633908232676206e-06, + "loss": 0.8133, + "num_tokens": 11359058.0, + "step": 2502 + }, + { + "epoch": 7.562570888468809, + "grad_norm": 1.921875, + "learning_rate": 9.632166276044888e-06, + "loss": 0.7711, + "num_tokens": 11372075.0, + "step": 2503 + }, + { + "epoch": 7.5655954631379965, + "grad_norm": 1.953125, + "learning_rate": 9.630420361489711e-06, + "loss": 0.5532, + "num_tokens": 11379510.0, + "step": 2504 + }, + { + "epoch": 7.568620037807183, + "grad_norm": 1.2890625, + "learning_rate": 9.628670490682974e-06, + "loss": 0.7722, + "num_tokens": 11410990.0, + "step": 2505 + }, + { + "epoch": 7.571644612476371, + "grad_norm": 2.078125, + "learning_rate": 9.626916665300756e-06, + "loss": 0.5933, + "num_tokens": 11418165.0, + "step": 2506 + }, + { + "epoch": 7.574669187145558, + "grad_norm": 1.2578125, + "learning_rate": 9.62515888702293e-06, + "loss": 0.7395, + "num_tokens": 11447461.0, + "step": 2507 + }, + { + "epoch": 7.577693761814745, + "grad_norm": 1.2890625, + "learning_rate": 9.623397157533155e-06, + "loss": 0.7925, + "num_tokens": 11480453.0, + "step": 2508 + }, + { + "epoch": 7.5807183364839315, + "grad_norm": 1.9453125, + "learning_rate": 9.62163147851887e-06, + "loss": 0.5835, + "num_tokens": 11488403.0, + "step": 2509 + }, + { + "epoch": 7.583742911153119, + "grad_norm": 1.3203125, + "learning_rate": 9.619861851671306e-06, + "loss": 0.835, + "num_tokens": 11520077.0, + "step": 2510 + }, + { + "epoch": 7.586767485822306, + "grad_norm": 1.625, + "learning_rate": 9.618088278685462e-06, + "loss": 0.8002, + "num_tokens": 11540686.0, + "step": 2511 + }, + { + "epoch": 7.589792060491494, + "grad_norm": 1.59375, + "learning_rate": 9.616310761260128e-06, + "loss": 0.7151, + "num_tokens": 11556994.0, + "step": 2512 + }, + { + "epoch": 7.5928166351606805, + "grad_norm": 1.3046875, + "learning_rate": 9.614529301097867e-06, + "loss": 0.7934, + "num_tokens": 11588070.0, + "step": 2513 + }, + { + "epoch": 7.595841209829867, + "grad_norm": 1.953125, + "learning_rate": 9.612743899905021e-06, + "loss": 0.6699, + "num_tokens": 11596608.0, + "step": 2514 + }, + { + "epoch": 7.598865784499055, + "grad_norm": 1.921875, + "learning_rate": 9.610954559391704e-06, + "loss": 0.5806, + "num_tokens": 11603822.0, + "step": 2515 + }, + { + "epoch": 7.601890359168242, + "grad_norm": 2.015625, + "learning_rate": 9.609161281271808e-06, + "loss": 0.5723, + "num_tokens": 11610628.0, + "step": 2516 + }, + { + "epoch": 7.60491493383743, + "grad_norm": 1.71875, + "learning_rate": 9.607364067262988e-06, + "loss": 0.9086, + "num_tokens": 11634438.0, + "step": 2517 + }, + { + "epoch": 7.607939508506616, + "grad_norm": 0.90625, + "learning_rate": 9.60556291908668e-06, + "loss": 0.8266, + "num_tokens": 11710145.0, + "step": 2518 + }, + { + "epoch": 7.610964083175803, + "grad_norm": 2.046875, + "learning_rate": 9.603757838468079e-06, + "loss": 0.5748, + "num_tokens": 11716830.0, + "step": 2519 + }, + { + "epoch": 7.613988657844991, + "grad_norm": 2.015625, + "learning_rate": 9.601948827136153e-06, + "loss": 0.6232, + "num_tokens": 11724598.0, + "step": 2520 + }, + { + "epoch": 7.617013232514178, + "grad_norm": 1.28125, + "learning_rate": 9.600135886823627e-06, + "loss": 0.8799, + "num_tokens": 11758091.0, + "step": 2521 + }, + { + "epoch": 7.6200378071833645, + "grad_norm": 1.578125, + "learning_rate": 9.598319019267e-06, + "loss": 0.6711, + "num_tokens": 11773565.0, + "step": 2522 + }, + { + "epoch": 7.623062381852552, + "grad_norm": 1.3984375, + "learning_rate": 9.596498226206525e-06, + "loss": 0.7901, + "num_tokens": 11799872.0, + "step": 2523 + }, + { + "epoch": 7.626086956521739, + "grad_norm": 1.953125, + "learning_rate": 9.594673509386217e-06, + "loss": 0.4983, + "num_tokens": 11806763.0, + "step": 2524 + }, + { + "epoch": 7.629111531190926, + "grad_norm": 1.7265625, + "learning_rate": 9.592844870553849e-06, + "loss": 0.7493, + "num_tokens": 11822452.0, + "step": 2525 + }, + { + "epoch": 7.632136105860114, + "grad_norm": 1.671875, + "learning_rate": 9.591012311460952e-06, + "loss": 0.8732, + "num_tokens": 11845643.0, + "step": 2526 + }, + { + "epoch": 7.6351606805293, + "grad_norm": 1.4921875, + "learning_rate": 9.58917583386281e-06, + "loss": 0.7835, + "num_tokens": 11869025.0, + "step": 2527 + }, + { + "epoch": 7.638185255198488, + "grad_norm": 1.84375, + "learning_rate": 9.58733543951846e-06, + "loss": 0.7677, + "num_tokens": 11883831.0, + "step": 2528 + }, + { + "epoch": 7.641209829867675, + "grad_norm": 1.4296875, + "learning_rate": 9.585491130190696e-06, + "loss": 0.8229, + "num_tokens": 11912891.0, + "step": 2529 + }, + { + "epoch": 7.644234404536862, + "grad_norm": 1.75, + "learning_rate": 9.583642907646053e-06, + "loss": 0.6099, + "num_tokens": 11922244.0, + "step": 2530 + }, + { + "epoch": 7.6472589792060495, + "grad_norm": 1.5703125, + "learning_rate": 9.581790773654821e-06, + "loss": 0.8255, + "num_tokens": 11945796.0, + "step": 2531 + }, + { + "epoch": 7.650283553875236, + "grad_norm": 1.96875, + "learning_rate": 9.579934729991037e-06, + "loss": 0.6686, + "num_tokens": 11953602.0, + "step": 2532 + }, + { + "epoch": 7.653308128544423, + "grad_norm": 1.3046875, + "learning_rate": 9.578074778432475e-06, + "loss": 0.717, + "num_tokens": 11977736.0, + "step": 2533 + }, + { + "epoch": 7.656332703213611, + "grad_norm": 1.7578125, + "learning_rate": 9.576210920760662e-06, + "loss": 0.6605, + "num_tokens": 11991591.0, + "step": 2534 + }, + { + "epoch": 7.659357277882798, + "grad_norm": 1.1796875, + "learning_rate": 9.57434315876086e-06, + "loss": 0.7898, + "num_tokens": 12029891.0, + "step": 2535 + }, + { + "epoch": 7.6623818525519845, + "grad_norm": 2.015625, + "learning_rate": 9.572471494222072e-06, + "loss": 0.59, + "num_tokens": 12037013.0, + "step": 2536 + }, + { + "epoch": 7.665406427221172, + "grad_norm": 1.4609375, + "learning_rate": 9.57059592893704e-06, + "loss": 0.8056, + "num_tokens": 12059588.0, + "step": 2537 + }, + { + "epoch": 7.668431001890359, + "grad_norm": 1.3203125, + "learning_rate": 9.56871646470224e-06, + "loss": 0.8253, + "num_tokens": 12093041.0, + "step": 2538 + }, + { + "epoch": 7.671455576559547, + "grad_norm": 1.78125, + "learning_rate": 9.566833103317887e-06, + "loss": 0.7782, + "num_tokens": 12107189.0, + "step": 2539 + }, + { + "epoch": 7.6744801512287335, + "grad_norm": 1.28125, + "learning_rate": 9.564945846587925e-06, + "loss": 0.8834, + "num_tokens": 12146097.0, + "step": 2540 + }, + { + "epoch": 7.67750472589792, + "grad_norm": 1.8203125, + "learning_rate": 9.56305469632003e-06, + "loss": 0.5709, + "num_tokens": 12154102.0, + "step": 2541 + }, + { + "epoch": 7.680529300567108, + "grad_norm": 2.15625, + "learning_rate": 9.561159654325607e-06, + "loss": 0.6418, + "num_tokens": 12161380.0, + "step": 2542 + }, + { + "epoch": 7.683553875236295, + "grad_norm": 1.3203125, + "learning_rate": 9.55926072241979e-06, + "loss": 0.7247, + "num_tokens": 12187779.0, + "step": 2543 + }, + { + "epoch": 7.686578449905482, + "grad_norm": 1.7734375, + "learning_rate": 9.557357902421438e-06, + "loss": 0.513, + "num_tokens": 12196548.0, + "step": 2544 + }, + { + "epoch": 7.689603024574669, + "grad_norm": 1.828125, + "learning_rate": 9.555451196153133e-06, + "loss": 0.608, + "num_tokens": 12205893.0, + "step": 2545 + }, + { + "epoch": 7.692627599243856, + "grad_norm": 1.640625, + "learning_rate": 9.553540605441182e-06, + "loss": 0.8912, + "num_tokens": 12228373.0, + "step": 2546 + }, + { + "epoch": 7.695652173913043, + "grad_norm": 1.2890625, + "learning_rate": 9.55162613211561e-06, + "loss": 0.7848, + "num_tokens": 12268286.0, + "step": 2547 + }, + { + "epoch": 7.698676748582231, + "grad_norm": 1.8671875, + "learning_rate": 9.549707778010164e-06, + "loss": 0.57, + "num_tokens": 12275779.0, + "step": 2548 + }, + { + "epoch": 7.7017013232514175, + "grad_norm": 1.453125, + "learning_rate": 9.547785544962303e-06, + "loss": 0.7105, + "num_tokens": 12298908.0, + "step": 2549 + }, + { + "epoch": 7.704725897920605, + "grad_norm": 1.84375, + "learning_rate": 9.545859434813209e-06, + "loss": 0.7211, + "num_tokens": 12311704.0, + "step": 2550 + }, + { + "epoch": 7.707750472589792, + "grad_norm": 1.5078125, + "learning_rate": 9.543929449407768e-06, + "loss": 0.8415, + "num_tokens": 12335688.0, + "step": 2551 + }, + { + "epoch": 7.710775047258979, + "grad_norm": 1.3671875, + "learning_rate": 9.541995590594589e-06, + "loss": 0.7855, + "num_tokens": 12364818.0, + "step": 2552 + }, + { + "epoch": 7.713799621928167, + "grad_norm": 1.7578125, + "learning_rate": 9.54005786022598e-06, + "loss": 0.6996, + "num_tokens": 12377792.0, + "step": 2553 + }, + { + "epoch": 7.716824196597353, + "grad_norm": 1.8125, + "learning_rate": 9.538116260157965e-06, + "loss": 0.6424, + "num_tokens": 12386479.0, + "step": 2554 + }, + { + "epoch": 7.719848771266541, + "grad_norm": 1.3515625, + "learning_rate": 9.53617079225027e-06, + "loss": 0.8712, + "num_tokens": 12422463.0, + "step": 2555 + }, + { + "epoch": 7.722873345935728, + "grad_norm": 1.96875, + "learning_rate": 9.534221458366331e-06, + "loss": 0.5679, + "num_tokens": 12430206.0, + "step": 2556 + }, + { + "epoch": 7.725897920604915, + "grad_norm": 2.09375, + "learning_rate": 9.53226826037328e-06, + "loss": 0.632, + "num_tokens": 12437623.0, + "step": 2557 + }, + { + "epoch": 7.7289224952741025, + "grad_norm": 1.7890625, + "learning_rate": 9.530311200141957e-06, + "loss": 0.6874, + "num_tokens": 12450892.0, + "step": 2558 + }, + { + "epoch": 7.731947069943289, + "grad_norm": 2.078125, + "learning_rate": 9.528350279546897e-06, + "loss": 0.6758, + "num_tokens": 12458138.0, + "step": 2559 + }, + { + "epoch": 7.734971644612476, + "grad_norm": 1.4765625, + "learning_rate": 9.526385500466332e-06, + "loss": 0.8208, + "num_tokens": 12481528.0, + "step": 2560 + }, + { + "epoch": 7.737996219281664, + "grad_norm": 1.171875, + "learning_rate": 9.524416864782196e-06, + "loss": 0.8453, + "num_tokens": 12520609.0, + "step": 2561 + }, + { + "epoch": 7.741020793950851, + "grad_norm": 1.953125, + "learning_rate": 9.522444374380107e-06, + "loss": 0.612, + "num_tokens": 12528398.0, + "step": 2562 + }, + { + "epoch": 7.7440453686200375, + "grad_norm": 1.6484375, + "learning_rate": 9.520468031149385e-06, + "loss": 0.837, + "num_tokens": 12556208.0, + "step": 2563 + }, + { + "epoch": 7.747069943289225, + "grad_norm": 1.265625, + "learning_rate": 9.518487836983035e-06, + "loss": 0.8255, + "num_tokens": 12589648.0, + "step": 2564 + }, + { + "epoch": 7.750094517958412, + "grad_norm": 1.4921875, + "learning_rate": 9.51650379377775e-06, + "loss": 0.7949, + "num_tokens": 12613689.0, + "step": 2565 + }, + { + "epoch": 7.7531190926276, + "grad_norm": 1.6484375, + "learning_rate": 9.514515903433915e-06, + "loss": 0.7312, + "num_tokens": 12631318.0, + "step": 2566 + }, + { + "epoch": 7.7561436672967865, + "grad_norm": 1.2578125, + "learning_rate": 9.51252416785559e-06, + "loss": 0.8814, + "num_tokens": 12668247.0, + "step": 2567 + }, + { + "epoch": 7.759168241965973, + "grad_norm": 2.046875, + "learning_rate": 9.510528588950533e-06, + "loss": 0.6308, + "num_tokens": 12675639.0, + "step": 2568 + }, + { + "epoch": 7.762192816635161, + "grad_norm": 1.125, + "learning_rate": 9.50852916863017e-06, + "loss": 0.8441, + "num_tokens": 12718616.0, + "step": 2569 + }, + { + "epoch": 7.765217391304348, + "grad_norm": 1.6875, + "learning_rate": 9.50652590880961e-06, + "loss": 0.8919, + "num_tokens": 12742179.0, + "step": 2570 + }, + { + "epoch": 7.768241965973535, + "grad_norm": 2.046875, + "learning_rate": 9.50451881140764e-06, + "loss": 0.5948, + "num_tokens": 12749382.0, + "step": 2571 + }, + { + "epoch": 7.771266540642722, + "grad_norm": 1.7734375, + "learning_rate": 9.502507878346729e-06, + "loss": 0.7124, + "num_tokens": 12763167.0, + "step": 2572 + }, + { + "epoch": 7.774291115311909, + "grad_norm": 1.5703125, + "learning_rate": 9.500493111553007e-06, + "loss": 0.8372, + "num_tokens": 12786769.0, + "step": 2573 + }, + { + "epoch": 7.777315689981096, + "grad_norm": 2.03125, + "learning_rate": 9.498474512956291e-06, + "loss": 0.6063, + "num_tokens": 12793640.0, + "step": 2574 + }, + { + "epoch": 7.780340264650284, + "grad_norm": 1.4453125, + "learning_rate": 9.496452084490054e-06, + "loss": 0.7044, + "num_tokens": 12811977.0, + "step": 2575 + }, + { + "epoch": 7.7833648393194705, + "grad_norm": 1.2109375, + "learning_rate": 9.49442582809145e-06, + "loss": 0.8183, + "num_tokens": 12849345.0, + "step": 2576 + }, + { + "epoch": 7.786389413988658, + "grad_norm": 1.1015625, + "learning_rate": 9.492395745701287e-06, + "loss": 0.8445, + "num_tokens": 12898976.0, + "step": 2577 + }, + { + "epoch": 7.789413988657845, + "grad_norm": 1.90625, + "learning_rate": 9.49036183926405e-06, + "loss": 0.5912, + "num_tokens": 12908453.0, + "step": 2578 + }, + { + "epoch": 7.792438563327032, + "grad_norm": 1.59375, + "learning_rate": 9.488324110727878e-06, + "loss": 0.726, + "num_tokens": 12923867.0, + "step": 2579 + }, + { + "epoch": 7.79546313799622, + "grad_norm": 2.109375, + "learning_rate": 9.486282562044577e-06, + "loss": 0.5574, + "num_tokens": 12931212.0, + "step": 2580 + }, + { + "epoch": 7.798487712665406, + "grad_norm": 1.8203125, + "learning_rate": 9.484237195169606e-06, + "loss": 0.5198, + "num_tokens": 12938958.0, + "step": 2581 + }, + { + "epoch": 7.801512287334593, + "grad_norm": 1.28125, + "learning_rate": 9.482188012062084e-06, + "loss": 0.879, + "num_tokens": 12975615.0, + "step": 2582 + }, + { + "epoch": 7.804536862003781, + "grad_norm": 1.7578125, + "learning_rate": 9.480135014684787e-06, + "loss": 0.7911, + "num_tokens": 12994200.0, + "step": 2583 + }, + { + "epoch": 7.807561436672968, + "grad_norm": 2.015625, + "learning_rate": 9.478078205004142e-06, + "loss": 0.6207, + "num_tokens": 13001453.0, + "step": 2584 + }, + { + "epoch": 7.810586011342155, + "grad_norm": 1.21875, + "learning_rate": 9.476017584990229e-06, + "loss": 0.8783, + "num_tokens": 13050024.0, + "step": 2585 + }, + { + "epoch": 7.813610586011342, + "grad_norm": 1.5234375, + "learning_rate": 9.473953156616777e-06, + "loss": 0.825, + "num_tokens": 13072697.0, + "step": 2586 + }, + { + "epoch": 7.816635160680529, + "grad_norm": 1.4609375, + "learning_rate": 9.471884921861165e-06, + "loss": 0.8175, + "num_tokens": 13102138.0, + "step": 2587 + }, + { + "epoch": 7.819659735349717, + "grad_norm": 1.2890625, + "learning_rate": 9.469812882704413e-06, + "loss": 0.8877, + "num_tokens": 13138513.0, + "step": 2588 + }, + { + "epoch": 7.822684310018904, + "grad_norm": 1.84375, + "learning_rate": 9.467737041131188e-06, + "loss": 0.8024, + "num_tokens": 13155864.0, + "step": 2589 + }, + { + "epoch": 7.8257088846880904, + "grad_norm": 1.6171875, + "learning_rate": 9.465657399129801e-06, + "loss": 0.7617, + "num_tokens": 13179599.0, + "step": 2590 + }, + { + "epoch": 7.828733459357278, + "grad_norm": 1.1953125, + "learning_rate": 9.4635739586922e-06, + "loss": 0.8052, + "num_tokens": 13220464.0, + "step": 2591 + }, + { + "epoch": 7.831758034026465, + "grad_norm": 1.3359375, + "learning_rate": 9.461486721813972e-06, + "loss": 0.8114, + "num_tokens": 13252911.0, + "step": 2592 + }, + { + "epoch": 7.834782608695652, + "grad_norm": 1.625, + "learning_rate": 9.45939569049434e-06, + "loss": 0.7572, + "num_tokens": 13270500.0, + "step": 2593 + }, + { + "epoch": 7.8378071833648395, + "grad_norm": 1.65625, + "learning_rate": 9.45730086673616e-06, + "loss": 0.7629, + "num_tokens": 13292697.0, + "step": 2594 + }, + { + "epoch": 7.840831758034026, + "grad_norm": 1.7421875, + "learning_rate": 9.45520225254593e-06, + "loss": 0.6323, + "num_tokens": 13303350.0, + "step": 2595 + }, + { + "epoch": 7.843856332703213, + "grad_norm": 1.7734375, + "learning_rate": 9.453099849933766e-06, + "loss": 0.6784, + "num_tokens": 13317565.0, + "step": 2596 + }, + { + "epoch": 7.846880907372401, + "grad_norm": 1.421875, + "learning_rate": 9.450993660913418e-06, + "loss": 0.8253, + "num_tokens": 13343245.0, + "step": 2597 + }, + { + "epoch": 7.849905482041588, + "grad_norm": 1.3828125, + "learning_rate": 9.448883687502261e-06, + "loss": 0.7696, + "num_tokens": 13365774.0, + "step": 2598 + }, + { + "epoch": 7.852930056710775, + "grad_norm": 1.9765625, + "learning_rate": 9.446769931721301e-06, + "loss": 0.5672, + "num_tokens": 13373305.0, + "step": 2599 + }, + { + "epoch": 7.855954631379962, + "grad_norm": 1.1875, + "learning_rate": 9.444652395595159e-06, + "loss": 0.8485, + "num_tokens": 13413240.0, + "step": 2600 + }, + { + "epoch": 7.858979206049149, + "grad_norm": 1.515625, + "learning_rate": 9.442531081152082e-06, + "loss": 0.7498, + "num_tokens": 13436221.0, + "step": 2601 + }, + { + "epoch": 7.862003780718337, + "grad_norm": 1.1171875, + "learning_rate": 9.440405990423933e-06, + "loss": 0.711, + "num_tokens": 13472905.0, + "step": 2602 + }, + { + "epoch": 7.8650283553875235, + "grad_norm": 1.3046875, + "learning_rate": 9.438277125446194e-06, + "loss": 0.9296, + "num_tokens": 13512602.0, + "step": 2603 + }, + { + "epoch": 7.868052930056711, + "grad_norm": 2.09375, + "learning_rate": 9.436144488257963e-06, + "loss": 0.9484, + "num_tokens": 13536182.0, + "step": 2604 + }, + { + "epoch": 7.871077504725898, + "grad_norm": 1.4453125, + "learning_rate": 9.434008080901946e-06, + "loss": 0.7644, + "num_tokens": 13559289.0, + "step": 2605 + }, + { + "epoch": 7.874102079395085, + "grad_norm": 2.015625, + "learning_rate": 9.431867905424466e-06, + "loss": 0.6846, + "num_tokens": 13569696.0, + "step": 2606 + }, + { + "epoch": 7.877126654064273, + "grad_norm": 1.6953125, + "learning_rate": 9.429723963875454e-06, + "loss": 0.6708, + "num_tokens": 13583713.0, + "step": 2607 + }, + { + "epoch": 7.880151228733459, + "grad_norm": 1.40625, + "learning_rate": 9.427576258308446e-06, + "loss": 0.7661, + "num_tokens": 13606861.0, + "step": 2608 + }, + { + "epoch": 7.883175803402646, + "grad_norm": 1.1953125, + "learning_rate": 9.425424790780581e-06, + "loss": 0.7525, + "num_tokens": 13650484.0, + "step": 2609 + }, + { + "epoch": 7.886200378071834, + "grad_norm": 1.9453125, + "learning_rate": 9.423269563352611e-06, + "loss": 0.5608, + "num_tokens": 13658676.0, + "step": 2610 + }, + { + "epoch": 7.889224952741021, + "grad_norm": 1.609375, + "learning_rate": 9.421110578088883e-06, + "loss": 0.6586, + "num_tokens": 13671031.0, + "step": 2611 + }, + { + "epoch": 7.892249527410208, + "grad_norm": 1.5234375, + "learning_rate": 9.418947837057338e-06, + "loss": 0.7976, + "num_tokens": 13691323.0, + "step": 2612 + }, + { + "epoch": 7.895274102079395, + "grad_norm": 1.6484375, + "learning_rate": 9.416781342329523e-06, + "loss": 0.8826, + "num_tokens": 13716242.0, + "step": 2613 + }, + { + "epoch": 7.898298676748582, + "grad_norm": 1.3984375, + "learning_rate": 9.41461109598058e-06, + "loss": 0.8194, + "num_tokens": 13747669.0, + "step": 2614 + }, + { + "epoch": 7.90132325141777, + "grad_norm": 1.8359375, + "learning_rate": 9.412437100089236e-06, + "loss": 0.6863, + "num_tokens": 13758445.0, + "step": 2615 + }, + { + "epoch": 7.904347826086957, + "grad_norm": 1.5859375, + "learning_rate": 9.410259356737822e-06, + "loss": 0.8975, + "num_tokens": 13781586.0, + "step": 2616 + }, + { + "epoch": 7.9073724007561434, + "grad_norm": 1.1328125, + "learning_rate": 9.408077868012248e-06, + "loss": 0.7666, + "num_tokens": 13816207.0, + "step": 2617 + }, + { + "epoch": 7.910396975425331, + "grad_norm": 1.6484375, + "learning_rate": 9.405892636002016e-06, + "loss": 0.722, + "num_tokens": 13830643.0, + "step": 2618 + }, + { + "epoch": 7.913421550094518, + "grad_norm": 1.328125, + "learning_rate": 9.403703662800213e-06, + "loss": 0.8211, + "num_tokens": 13864595.0, + "step": 2619 + }, + { + "epoch": 7.916446124763705, + "grad_norm": 1.6171875, + "learning_rate": 9.401510950503506e-06, + "loss": 0.7643, + "num_tokens": 13884766.0, + "step": 2620 + }, + { + "epoch": 7.9194706994328925, + "grad_norm": 1.7578125, + "learning_rate": 9.39931450121215e-06, + "loss": 0.5511, + "num_tokens": 13893072.0, + "step": 2621 + }, + { + "epoch": 7.922495274102079, + "grad_norm": 1.5859375, + "learning_rate": 9.397114317029975e-06, + "loss": 0.7429, + "num_tokens": 13912373.0, + "step": 2622 + }, + { + "epoch": 7.925519848771266, + "grad_norm": 1.6328125, + "learning_rate": 9.394910400064387e-06, + "loss": 0.7678, + "num_tokens": 13937172.0, + "step": 2623 + }, + { + "epoch": 7.928544423440454, + "grad_norm": 1.2734375, + "learning_rate": 9.392702752426377e-06, + "loss": 0.8214, + "num_tokens": 13977949.0, + "step": 2624 + }, + { + "epoch": 7.931568998109641, + "grad_norm": 2.015625, + "learning_rate": 9.390491376230491e-06, + "loss": 0.6348, + "num_tokens": 13984870.0, + "step": 2625 + }, + { + "epoch": 7.934593572778828, + "grad_norm": 1.46875, + "learning_rate": 9.388276273594865e-06, + "loss": 0.7986, + "num_tokens": 14008252.0, + "step": 2626 + }, + { + "epoch": 7.937618147448015, + "grad_norm": 1.5546875, + "learning_rate": 9.386057446641195e-06, + "loss": 0.7715, + "num_tokens": 14032041.0, + "step": 2627 + }, + { + "epoch": 7.940642722117202, + "grad_norm": 1.9921875, + "learning_rate": 9.383834897494747e-06, + "loss": 0.5694, + "num_tokens": 14038843.0, + "step": 2628 + }, + { + "epoch": 7.94366729678639, + "grad_norm": 1.8046875, + "learning_rate": 9.38160862828435e-06, + "loss": 0.5206, + "num_tokens": 14047426.0, + "step": 2629 + }, + { + "epoch": 7.9466918714555765, + "grad_norm": 1.671875, + "learning_rate": 9.379378641142394e-06, + "loss": 0.7987, + "num_tokens": 14071378.0, + "step": 2630 + }, + { + "epoch": 7.949716446124763, + "grad_norm": 1.71875, + "learning_rate": 9.37714493820484e-06, + "loss": 0.7111, + "num_tokens": 14084323.0, + "step": 2631 + }, + { + "epoch": 7.952741020793951, + "grad_norm": 1.9921875, + "learning_rate": 9.374907521611198e-06, + "loss": 0.5739, + "num_tokens": 14091256.0, + "step": 2632 + }, + { + "epoch": 7.955765595463138, + "grad_norm": 1.96875, + "learning_rate": 9.372666393504537e-06, + "loss": 0.5351, + "num_tokens": 14098176.0, + "step": 2633 + }, + { + "epoch": 7.958790170132325, + "grad_norm": 1.0, + "learning_rate": 9.370421556031487e-06, + "loss": 0.8273, + "num_tokens": 14159439.0, + "step": 2634 + }, + { + "epoch": 7.961814744801512, + "grad_norm": 1.859375, + "learning_rate": 9.368173011342224e-06, + "loss": 0.5863, + "num_tokens": 14167526.0, + "step": 2635 + }, + { + "epoch": 7.964839319470699, + "grad_norm": 1.59375, + "learning_rate": 9.365920761590478e-06, + "loss": 0.7807, + "num_tokens": 14190445.0, + "step": 2636 + }, + { + "epoch": 7.967863894139887, + "grad_norm": 1.796875, + "learning_rate": 9.363664808933529e-06, + "loss": 0.6457, + "num_tokens": 14201019.0, + "step": 2637 + }, + { + "epoch": 7.970888468809074, + "grad_norm": 1.9296875, + "learning_rate": 9.3614051555322e-06, + "loss": 0.6876, + "num_tokens": 14211884.0, + "step": 2638 + }, + { + "epoch": 7.973913043478261, + "grad_norm": 1.140625, + "learning_rate": 9.35914180355086e-06, + "loss": 0.7337, + "num_tokens": 14248424.0, + "step": 2639 + }, + { + "epoch": 7.976937618147448, + "grad_norm": 1.828125, + "learning_rate": 9.356874755157424e-06, + "loss": 0.6475, + "num_tokens": 14257972.0, + "step": 2640 + }, + { + "epoch": 7.979962192816635, + "grad_norm": 1.8828125, + "learning_rate": 9.354604012523344e-06, + "loss": 0.5655, + "num_tokens": 14266753.0, + "step": 2641 + }, + { + "epoch": 7.982986767485822, + "grad_norm": 1.9296875, + "learning_rate": 9.352329577823613e-06, + "loss": 0.6707, + "num_tokens": 14280177.0, + "step": 2642 + }, + { + "epoch": 7.98601134215501, + "grad_norm": 2.0625, + "learning_rate": 9.350051453236755e-06, + "loss": 0.5651, + "num_tokens": 14287621.0, + "step": 2643 + }, + { + "epoch": 7.9890359168241964, + "grad_norm": 1.5625, + "learning_rate": 9.347769640944834e-06, + "loss": 0.8267, + "num_tokens": 14311097.0, + "step": 2644 + }, + { + "epoch": 7.992060491493383, + "grad_norm": 1.8828125, + "learning_rate": 9.345484143133447e-06, + "loss": 0.5922, + "num_tokens": 14318601.0, + "step": 2645 + }, + { + "epoch": 7.995085066162571, + "grad_norm": 1.5078125, + "learning_rate": 9.343194961991716e-06, + "loss": 0.8344, + "num_tokens": 14346184.0, + "step": 2646 + }, + { + "epoch": 7.998109640831758, + "grad_norm": 1.6171875, + "learning_rate": 9.340902099712294e-06, + "loss": 0.8498, + "num_tokens": 14370140.0, + "step": 2647 + }, + { + "epoch": 8.0, + "grad_norm": 1.6796875, + "learning_rate": 9.33860555849136e-06, + "loss": 0.8563, + "num_tokens": 14390626.0, + "step": 2648 + }, + { + "epoch": 8.003024574669187, + "grad_norm": 1.9609375, + "learning_rate": 9.336305340528613e-06, + "loss": 0.6143, + "num_tokens": 14398353.0, + "step": 2649 + }, + { + "epoch": 8.006049149338374, + "grad_norm": 1.5, + "learning_rate": 9.33400144802728e-06, + "loss": 0.8171, + "num_tokens": 14421027.0, + "step": 2650 + }, + { + "epoch": 8.009073724007562, + "grad_norm": 1.28125, + "learning_rate": 9.331693883194105e-06, + "loss": 0.7708, + "num_tokens": 14450806.0, + "step": 2651 + }, + { + "epoch": 8.012098298676749, + "grad_norm": 1.3671875, + "learning_rate": 9.329382648239348e-06, + "loss": 0.6471, + "num_tokens": 14470321.0, + "step": 2652 + }, + { + "epoch": 8.015122873345936, + "grad_norm": 1.8984375, + "learning_rate": 9.327067745376785e-06, + "loss": 0.5302, + "num_tokens": 14477823.0, + "step": 2653 + }, + { + "epoch": 8.018147448015123, + "grad_norm": 0.984375, + "learning_rate": 9.324749176823704e-06, + "loss": 0.8646, + "num_tokens": 14530573.0, + "step": 2654 + }, + { + "epoch": 8.02117202268431, + "grad_norm": 1.4765625, + "learning_rate": 9.322426944800909e-06, + "loss": 0.6906, + "num_tokens": 14548923.0, + "step": 2655 + }, + { + "epoch": 8.024196597353496, + "grad_norm": 1.875, + "learning_rate": 9.320101051532709e-06, + "loss": 0.5792, + "num_tokens": 14557646.0, + "step": 2656 + }, + { + "epoch": 8.027221172022685, + "grad_norm": 1.1328125, + "learning_rate": 9.317771499246918e-06, + "loss": 0.8656, + "num_tokens": 14600482.0, + "step": 2657 + }, + { + "epoch": 8.030245746691872, + "grad_norm": 1.90625, + "learning_rate": 9.31543829017486e-06, + "loss": 0.5299, + "num_tokens": 14607864.0, + "step": 2658 + }, + { + "epoch": 8.033270321361059, + "grad_norm": 2.015625, + "learning_rate": 9.313101426551357e-06, + "loss": 0.5482, + "num_tokens": 14615583.0, + "step": 2659 + }, + { + "epoch": 8.036294896030245, + "grad_norm": 2.09375, + "learning_rate": 9.310760910614736e-06, + "loss": 0.5659, + "num_tokens": 14622538.0, + "step": 2660 + }, + { + "epoch": 8.039319470699432, + "grad_norm": 1.9296875, + "learning_rate": 9.308416744606816e-06, + "loss": 0.5547, + "num_tokens": 14629197.0, + "step": 2661 + }, + { + "epoch": 8.04234404536862, + "grad_norm": 1.6015625, + "learning_rate": 9.306068930772918e-06, + "loss": 0.6305, + "num_tokens": 14642404.0, + "step": 2662 + }, + { + "epoch": 8.045368620037808, + "grad_norm": 1.6171875, + "learning_rate": 9.303717471361855e-06, + "loss": 0.7267, + "num_tokens": 14659009.0, + "step": 2663 + }, + { + "epoch": 8.048393194706994, + "grad_norm": 1.5234375, + "learning_rate": 9.301362368625931e-06, + "loss": 0.7381, + "num_tokens": 14677805.0, + "step": 2664 + }, + { + "epoch": 8.051417769376181, + "grad_norm": 1.1484375, + "learning_rate": 9.29900362482094e-06, + "loss": 0.7376, + "num_tokens": 14714074.0, + "step": 2665 + }, + { + "epoch": 8.054442344045368, + "grad_norm": 1.421875, + "learning_rate": 9.296641242206165e-06, + "loss": 0.7486, + "num_tokens": 14738129.0, + "step": 2666 + }, + { + "epoch": 8.057466918714555, + "grad_norm": 1.6328125, + "learning_rate": 9.294275223044372e-06, + "loss": 0.8108, + "num_tokens": 14759501.0, + "step": 2667 + }, + { + "epoch": 8.060491493383743, + "grad_norm": 2.171875, + "learning_rate": 9.291905569601812e-06, + "loss": 0.6583, + "num_tokens": 14766216.0, + "step": 2668 + }, + { + "epoch": 8.06351606805293, + "grad_norm": 1.953125, + "learning_rate": 9.289532284148218e-06, + "loss": 0.5753, + "num_tokens": 14773486.0, + "step": 2669 + }, + { + "epoch": 8.066540642722117, + "grad_norm": 1.765625, + "learning_rate": 9.287155368956798e-06, + "loss": 0.8899, + "num_tokens": 14797158.0, + "step": 2670 + }, + { + "epoch": 8.069565217391304, + "grad_norm": 1.390625, + "learning_rate": 9.284774826304239e-06, + "loss": 0.8299, + "num_tokens": 14826215.0, + "step": 2671 + }, + { + "epoch": 8.07258979206049, + "grad_norm": 1.890625, + "learning_rate": 9.282390658470703e-06, + "loss": 0.5566, + "num_tokens": 14833771.0, + "step": 2672 + }, + { + "epoch": 8.07561436672968, + "grad_norm": 2.078125, + "learning_rate": 9.280002867739826e-06, + "loss": 0.635, + "num_tokens": 14841423.0, + "step": 2673 + }, + { + "epoch": 8.078638941398866, + "grad_norm": 1.4375, + "learning_rate": 9.277611456398709e-06, + "loss": 0.7607, + "num_tokens": 14865686.0, + "step": 2674 + }, + { + "epoch": 8.081663516068053, + "grad_norm": 1.390625, + "learning_rate": 9.275216426737924e-06, + "loss": 0.8023, + "num_tokens": 14892934.0, + "step": 2675 + }, + { + "epoch": 8.08468809073724, + "grad_norm": 1.8046875, + "learning_rate": 9.272817781051512e-06, + "loss": 0.6391, + "num_tokens": 14903447.0, + "step": 2676 + }, + { + "epoch": 8.087712665406427, + "grad_norm": 0.93359375, + "learning_rate": 9.270415521636971e-06, + "loss": 0.8693, + "num_tokens": 14972759.0, + "step": 2677 + }, + { + "epoch": 8.090737240075615, + "grad_norm": 1.953125, + "learning_rate": 9.268009650795264e-06, + "loss": 0.5985, + "num_tokens": 14980187.0, + "step": 2678 + }, + { + "epoch": 8.093761814744802, + "grad_norm": 1.53125, + "learning_rate": 9.265600170830813e-06, + "loss": 0.7789, + "num_tokens": 15003669.0, + "step": 2679 + }, + { + "epoch": 8.096786389413989, + "grad_norm": 1.0390625, + "learning_rate": 9.263187084051499e-06, + "loss": 0.8271, + "num_tokens": 15058807.0, + "step": 2680 + }, + { + "epoch": 8.099810964083176, + "grad_norm": 1.5234375, + "learning_rate": 9.260770392768652e-06, + "loss": 0.8779, + "num_tokens": 15088530.0, + "step": 2681 + }, + { + "epoch": 8.102835538752363, + "grad_norm": 1.46875, + "learning_rate": 9.258350099297057e-06, + "loss": 0.7995, + "num_tokens": 15112835.0, + "step": 2682 + }, + { + "epoch": 8.10586011342155, + "grad_norm": 1.9921875, + "learning_rate": 9.255926205954953e-06, + "loss": 0.5766, + "num_tokens": 15119471.0, + "step": 2683 + }, + { + "epoch": 8.108884688090738, + "grad_norm": 0.99609375, + "learning_rate": 9.253498715064025e-06, + "loss": 0.8573, + "num_tokens": 15173737.0, + "step": 2684 + }, + { + "epoch": 8.111909262759925, + "grad_norm": 2.09375, + "learning_rate": 9.251067628949402e-06, + "loss": 0.6799, + "num_tokens": 15181334.0, + "step": 2685 + }, + { + "epoch": 8.114933837429112, + "grad_norm": 1.8515625, + "learning_rate": 9.248632949939655e-06, + "loss": 0.5361, + "num_tokens": 15189030.0, + "step": 2686 + }, + { + "epoch": 8.117958412098298, + "grad_norm": 1.765625, + "learning_rate": 9.246194680366802e-06, + "loss": 0.6575, + "num_tokens": 15199101.0, + "step": 2687 + }, + { + "epoch": 8.120982986767485, + "grad_norm": 1.671875, + "learning_rate": 9.243752822566297e-06, + "loss": 0.729, + "num_tokens": 15213949.0, + "step": 2688 + }, + { + "epoch": 8.124007561436674, + "grad_norm": 1.4296875, + "learning_rate": 9.24130737887703e-06, + "loss": 0.7832, + "num_tokens": 15238725.0, + "step": 2689 + }, + { + "epoch": 8.12703213610586, + "grad_norm": 1.1796875, + "learning_rate": 9.23885835164133e-06, + "loss": 0.8575, + "num_tokens": 15280621.0, + "step": 2690 + }, + { + "epoch": 8.130056710775047, + "grad_norm": 2.015625, + "learning_rate": 9.23640574320495e-06, + "loss": 0.5736, + "num_tokens": 15287119.0, + "step": 2691 + }, + { + "epoch": 8.133081285444234, + "grad_norm": 1.5078125, + "learning_rate": 9.233949555917084e-06, + "loss": 0.8523, + "num_tokens": 15310780.0, + "step": 2692 + }, + { + "epoch": 8.136105860113421, + "grad_norm": 2.078125, + "learning_rate": 9.231489792130343e-06, + "loss": 0.5866, + "num_tokens": 15317224.0, + "step": 2693 + }, + { + "epoch": 8.139130434782608, + "grad_norm": 1.625, + "learning_rate": 9.229026454200773e-06, + "loss": 0.6447, + "num_tokens": 15335685.0, + "step": 2694 + }, + { + "epoch": 8.142155009451796, + "grad_norm": 1.2890625, + "learning_rate": 9.226559544487839e-06, + "loss": 0.8048, + "num_tokens": 15366541.0, + "step": 2695 + }, + { + "epoch": 8.145179584120983, + "grad_norm": 1.921875, + "learning_rate": 9.224089065354428e-06, + "loss": 0.7384, + "num_tokens": 15380319.0, + "step": 2696 + }, + { + "epoch": 8.14820415879017, + "grad_norm": 1.6328125, + "learning_rate": 9.221615019166844e-06, + "loss": 0.5803, + "num_tokens": 15391943.0, + "step": 2697 + }, + { + "epoch": 8.151228733459357, + "grad_norm": 1.484375, + "learning_rate": 9.219137408294808e-06, + "loss": 0.7655, + "num_tokens": 15414384.0, + "step": 2698 + }, + { + "epoch": 8.154253308128544, + "grad_norm": 1.9765625, + "learning_rate": 9.216656235111463e-06, + "loss": 0.5647, + "num_tokens": 15421341.0, + "step": 2699 + }, + { + "epoch": 8.157277882797732, + "grad_norm": 1.21875, + "learning_rate": 9.214171501993349e-06, + "loss": 0.7292, + "num_tokens": 15458032.0, + "step": 2700 + }, + { + "epoch": 8.16030245746692, + "grad_norm": 1.5859375, + "learning_rate": 9.211683211320431e-06, + "loss": 0.8268, + "num_tokens": 15481916.0, + "step": 2701 + }, + { + "epoch": 8.163327032136106, + "grad_norm": 1.09375, + "learning_rate": 9.209191365476074e-06, + "loss": 0.836, + "num_tokens": 15526137.0, + "step": 2702 + }, + { + "epoch": 8.166351606805293, + "grad_norm": 1.2109375, + "learning_rate": 9.206695966847048e-06, + "loss": 0.8414, + "num_tokens": 15563163.0, + "step": 2703 + }, + { + "epoch": 8.16937618147448, + "grad_norm": 1.421875, + "learning_rate": 9.204197017823526e-06, + "loss": 0.7574, + "num_tokens": 15584631.0, + "step": 2704 + }, + { + "epoch": 8.172400756143666, + "grad_norm": 1.8671875, + "learning_rate": 9.201694520799086e-06, + "loss": 0.6174, + "num_tokens": 15592886.0, + "step": 2705 + }, + { + "epoch": 8.175425330812855, + "grad_norm": 1.546875, + "learning_rate": 9.199188478170699e-06, + "loss": 0.8103, + "num_tokens": 15613975.0, + "step": 2706 + }, + { + "epoch": 8.178449905482042, + "grad_norm": 1.2109375, + "learning_rate": 9.19667889233874e-06, + "loss": 0.8074, + "num_tokens": 15653807.0, + "step": 2707 + }, + { + "epoch": 8.181474480151229, + "grad_norm": 1.5625, + "learning_rate": 9.194165765706963e-06, + "loss": 0.8244, + "num_tokens": 15676482.0, + "step": 2708 + }, + { + "epoch": 8.184499054820416, + "grad_norm": 1.3984375, + "learning_rate": 9.19164910068253e-06, + "loss": 0.8509, + "num_tokens": 15708659.0, + "step": 2709 + }, + { + "epoch": 8.187523629489602, + "grad_norm": 1.6328125, + "learning_rate": 9.189128899675982e-06, + "loss": 0.6798, + "num_tokens": 15723555.0, + "step": 2710 + }, + { + "epoch": 8.190548204158791, + "grad_norm": 1.59375, + "learning_rate": 9.186605165101253e-06, + "loss": 0.7663, + "num_tokens": 15745989.0, + "step": 2711 + }, + { + "epoch": 8.193572778827978, + "grad_norm": 1.4765625, + "learning_rate": 9.184077899375653e-06, + "loss": 0.7966, + "num_tokens": 15769301.0, + "step": 2712 + }, + { + "epoch": 8.196597353497165, + "grad_norm": 1.5625, + "learning_rate": 9.181547104919884e-06, + "loss": 0.8872, + "num_tokens": 15795386.0, + "step": 2713 + }, + { + "epoch": 8.199621928166351, + "grad_norm": 1.5546875, + "learning_rate": 9.179012784158023e-06, + "loss": 0.8099, + "num_tokens": 15819246.0, + "step": 2714 + }, + { + "epoch": 8.202646502835538, + "grad_norm": 1.953125, + "learning_rate": 9.176474939517524e-06, + "loss": 0.6152, + "num_tokens": 15826276.0, + "step": 2715 + }, + { + "epoch": 8.205671077504725, + "grad_norm": 1.609375, + "learning_rate": 9.173933573429217e-06, + "loss": 0.7715, + "num_tokens": 15847693.0, + "step": 2716 + }, + { + "epoch": 8.208695652173914, + "grad_norm": 1.3125, + "learning_rate": 9.171388688327307e-06, + "loss": 0.7458, + "num_tokens": 15873374.0, + "step": 2717 + }, + { + "epoch": 8.2117202268431, + "grad_norm": 2.0, + "learning_rate": 9.16884028664937e-06, + "loss": 0.5977, + "num_tokens": 15880425.0, + "step": 2718 + }, + { + "epoch": 8.214744801512287, + "grad_norm": 1.421875, + "learning_rate": 9.166288370836343e-06, + "loss": 0.6909, + "num_tokens": 15901041.0, + "step": 2719 + }, + { + "epoch": 8.217769376181474, + "grad_norm": 1.640625, + "learning_rate": 9.163732943332539e-06, + "loss": 0.7118, + "num_tokens": 15918247.0, + "step": 2720 + }, + { + "epoch": 8.220793950850661, + "grad_norm": 1.5546875, + "learning_rate": 9.161174006585629e-06, + "loss": 0.8306, + "num_tokens": 15941126.0, + "step": 2721 + }, + { + "epoch": 8.22381852551985, + "grad_norm": 2.125, + "learning_rate": 9.158611563046645e-06, + "loss": 0.6548, + "num_tokens": 15948331.0, + "step": 2722 + }, + { + "epoch": 8.226843100189036, + "grad_norm": 1.078125, + "learning_rate": 9.156045615169978e-06, + "loss": 0.8173, + "num_tokens": 15992113.0, + "step": 2723 + }, + { + "epoch": 8.229867674858223, + "grad_norm": 1.296875, + "learning_rate": 9.153476165413379e-06, + "loss": 0.8644, + "num_tokens": 16031299.0, + "step": 2724 + }, + { + "epoch": 8.23289224952741, + "grad_norm": 1.1640625, + "learning_rate": 9.15090321623795e-06, + "loss": 0.8339, + "num_tokens": 16075755.0, + "step": 2725 + }, + { + "epoch": 8.235916824196597, + "grad_norm": 1.2109375, + "learning_rate": 9.148326770108147e-06, + "loss": 0.773, + "num_tokens": 16114075.0, + "step": 2726 + }, + { + "epoch": 8.238941398865785, + "grad_norm": 1.921875, + "learning_rate": 9.145746829491775e-06, + "loss": 0.5658, + "num_tokens": 16123267.0, + "step": 2727 + }, + { + "epoch": 8.241965973534972, + "grad_norm": 1.28125, + "learning_rate": 9.14316339685998e-06, + "loss": 0.7523, + "num_tokens": 16152054.0, + "step": 2728 + }, + { + "epoch": 8.244990548204159, + "grad_norm": 1.8515625, + "learning_rate": 9.140576474687263e-06, + "loss": 0.5496, + "num_tokens": 16159980.0, + "step": 2729 + }, + { + "epoch": 8.248015122873346, + "grad_norm": 2.046875, + "learning_rate": 9.137986065451464e-06, + "loss": 0.6052, + "num_tokens": 16167084.0, + "step": 2730 + }, + { + "epoch": 8.251039697542533, + "grad_norm": 1.25, + "learning_rate": 9.135392171633756e-06, + "loss": 0.7725, + "num_tokens": 16199443.0, + "step": 2731 + }, + { + "epoch": 8.25406427221172, + "grad_norm": 1.2421875, + "learning_rate": 9.132794795718662e-06, + "loss": 0.7954, + "num_tokens": 16232555.0, + "step": 2732 + }, + { + "epoch": 8.257088846880908, + "grad_norm": 1.7890625, + "learning_rate": 9.130193940194027e-06, + "loss": 0.5177, + "num_tokens": 16240121.0, + "step": 2733 + }, + { + "epoch": 8.260113421550095, + "grad_norm": 1.5546875, + "learning_rate": 9.127589607551038e-06, + "loss": 0.8608, + "num_tokens": 16262180.0, + "step": 2734 + }, + { + "epoch": 8.263137996219282, + "grad_norm": 1.578125, + "learning_rate": 9.12498180028421e-06, + "loss": 0.7521, + "num_tokens": 16279434.0, + "step": 2735 + }, + { + "epoch": 8.266162570888469, + "grad_norm": 1.5546875, + "learning_rate": 9.122370520891387e-06, + "loss": 0.838, + "num_tokens": 16303120.0, + "step": 2736 + }, + { + "epoch": 8.269187145557655, + "grad_norm": 1.890625, + "learning_rate": 9.119755771873737e-06, + "loss": 0.5912, + "num_tokens": 16311900.0, + "step": 2737 + }, + { + "epoch": 8.272211720226844, + "grad_norm": 1.328125, + "learning_rate": 9.117137555735753e-06, + "loss": 0.7837, + "num_tokens": 16339986.0, + "step": 2738 + }, + { + "epoch": 8.27523629489603, + "grad_norm": 1.2421875, + "learning_rate": 9.114515874985245e-06, + "loss": 0.8614, + "num_tokens": 16380811.0, + "step": 2739 + }, + { + "epoch": 8.278260869565218, + "grad_norm": 1.5234375, + "learning_rate": 9.111890732133346e-06, + "loss": 0.6931, + "num_tokens": 16401775.0, + "step": 2740 + }, + { + "epoch": 8.281285444234404, + "grad_norm": 1.5703125, + "learning_rate": 9.109262129694506e-06, + "loss": 0.8532, + "num_tokens": 16425096.0, + "step": 2741 + }, + { + "epoch": 8.284310018903591, + "grad_norm": 1.6953125, + "learning_rate": 9.106630070186482e-06, + "loss": 0.6611, + "num_tokens": 16438176.0, + "step": 2742 + }, + { + "epoch": 8.287334593572778, + "grad_norm": 1.4921875, + "learning_rate": 9.103994556130349e-06, + "loss": 0.9048, + "num_tokens": 16461604.0, + "step": 2743 + }, + { + "epoch": 8.290359168241967, + "grad_norm": 1.34375, + "learning_rate": 9.101355590050489e-06, + "loss": 0.7978, + "num_tokens": 16491379.0, + "step": 2744 + }, + { + "epoch": 8.293383742911153, + "grad_norm": 1.5078125, + "learning_rate": 9.098713174474588e-06, + "loss": 0.787, + "num_tokens": 16514038.0, + "step": 2745 + }, + { + "epoch": 8.29640831758034, + "grad_norm": 1.84375, + "learning_rate": 9.096067311933641e-06, + "loss": 0.7308, + "num_tokens": 16525933.0, + "step": 2746 + }, + { + "epoch": 8.299432892249527, + "grad_norm": 1.2109375, + "learning_rate": 9.093418004961939e-06, + "loss": 0.8219, + "num_tokens": 16563057.0, + "step": 2747 + }, + { + "epoch": 8.302457466918714, + "grad_norm": 1.640625, + "learning_rate": 9.090765256097075e-06, + "loss": 0.8115, + "num_tokens": 16587732.0, + "step": 2748 + }, + { + "epoch": 8.305482041587902, + "grad_norm": 1.28125, + "learning_rate": 9.08810906787994e-06, + "loss": 0.8593, + "num_tokens": 16626606.0, + "step": 2749 + }, + { + "epoch": 8.30850661625709, + "grad_norm": 2.125, + "learning_rate": 9.085449442854715e-06, + "loss": 0.6025, + "num_tokens": 16633688.0, + "step": 2750 + }, + { + "epoch": 8.311531190926276, + "grad_norm": 1.9609375, + "learning_rate": 9.082786383568877e-06, + "loss": 0.6733, + "num_tokens": 16644365.0, + "step": 2751 + }, + { + "epoch": 8.314555765595463, + "grad_norm": 1.5390625, + "learning_rate": 9.080119892573194e-06, + "loss": 0.7012, + "num_tokens": 16663691.0, + "step": 2752 + }, + { + "epoch": 8.31758034026465, + "grad_norm": 1.5625, + "learning_rate": 9.077449972421716e-06, + "loss": 0.7575, + "num_tokens": 16685076.0, + "step": 2753 + }, + { + "epoch": 8.320604914933837, + "grad_norm": 1.3828125, + "learning_rate": 9.074776625671777e-06, + "loss": 0.8217, + "num_tokens": 16713578.0, + "step": 2754 + }, + { + "epoch": 8.323629489603025, + "grad_norm": 1.296875, + "learning_rate": 9.072099854884002e-06, + "loss": 0.8275, + "num_tokens": 16745845.0, + "step": 2755 + }, + { + "epoch": 8.326654064272212, + "grad_norm": 1.90625, + "learning_rate": 9.069419662622284e-06, + "loss": 0.6198, + "num_tokens": 16754750.0, + "step": 2756 + }, + { + "epoch": 8.329678638941399, + "grad_norm": 1.1640625, + "learning_rate": 9.066736051453802e-06, + "loss": 0.7495, + "num_tokens": 16794084.0, + "step": 2757 + }, + { + "epoch": 8.332703213610586, + "grad_norm": 1.25, + "learning_rate": 9.064049023949003e-06, + "loss": 0.7934, + "num_tokens": 16825645.0, + "step": 2758 + }, + { + "epoch": 8.335727788279772, + "grad_norm": 2.09375, + "learning_rate": 9.061358582681614e-06, + "loss": 0.6242, + "num_tokens": 16832798.0, + "step": 2759 + }, + { + "epoch": 8.338752362948961, + "grad_norm": 1.3203125, + "learning_rate": 9.058664730228625e-06, + "loss": 0.8424, + "num_tokens": 16871988.0, + "step": 2760 + }, + { + "epoch": 8.341776937618148, + "grad_norm": 1.2734375, + "learning_rate": 9.055967469170297e-06, + "loss": 0.913, + "num_tokens": 16911045.0, + "step": 2761 + }, + { + "epoch": 8.344801512287335, + "grad_norm": 1.5234375, + "learning_rate": 9.053266802090152e-06, + "loss": 0.7428, + "num_tokens": 16930996.0, + "step": 2762 + }, + { + "epoch": 8.347826086956522, + "grad_norm": 1.2578125, + "learning_rate": 9.050562731574983e-06, + "loss": 0.7588, + "num_tokens": 16958152.0, + "step": 2763 + }, + { + "epoch": 8.350850661625708, + "grad_norm": 1.984375, + "learning_rate": 9.047855260214829e-06, + "loss": 0.5656, + "num_tokens": 16965197.0, + "step": 2764 + }, + { + "epoch": 8.353875236294897, + "grad_norm": 1.375, + "learning_rate": 9.045144390603e-06, + "loss": 0.8133, + "num_tokens": 16990374.0, + "step": 2765 + }, + { + "epoch": 8.356899810964084, + "grad_norm": 1.609375, + "learning_rate": 9.042430125336057e-06, + "loss": 0.6412, + "num_tokens": 17005277.0, + "step": 2766 + }, + { + "epoch": 8.35992438563327, + "grad_norm": 1.4609375, + "learning_rate": 9.039712467013808e-06, + "loss": 0.7801, + "num_tokens": 17028098.0, + "step": 2767 + }, + { + "epoch": 8.362948960302457, + "grad_norm": 1.0703125, + "learning_rate": 9.036991418239316e-06, + "loss": 0.812, + "num_tokens": 17073216.0, + "step": 2768 + }, + { + "epoch": 8.365973534971644, + "grad_norm": 1.453125, + "learning_rate": 9.03426698161889e-06, + "loss": 0.9227, + "num_tokens": 17107961.0, + "step": 2769 + }, + { + "epoch": 8.368998109640831, + "grad_norm": 1.3125, + "learning_rate": 9.031539159762086e-06, + "loss": 0.855, + "num_tokens": 17140051.0, + "step": 2770 + }, + { + "epoch": 8.37202268431002, + "grad_norm": 1.5, + "learning_rate": 9.028807955281701e-06, + "loss": 0.7508, + "num_tokens": 17170250.0, + "step": 2771 + }, + { + "epoch": 8.375047258979206, + "grad_norm": 1.828125, + "learning_rate": 9.02607337079377e-06, + "loss": 0.5502, + "num_tokens": 17178311.0, + "step": 2772 + }, + { + "epoch": 8.378071833648393, + "grad_norm": 1.25, + "learning_rate": 9.023335408917566e-06, + "loss": 0.7541, + "num_tokens": 17218315.0, + "step": 2773 + }, + { + "epoch": 8.38109640831758, + "grad_norm": 1.796875, + "learning_rate": 9.020594072275602e-06, + "loss": 0.6674, + "num_tokens": 17229920.0, + "step": 2774 + }, + { + "epoch": 8.384120982986767, + "grad_norm": 2.09375, + "learning_rate": 9.01784936349362e-06, + "loss": 0.6043, + "num_tokens": 17237021.0, + "step": 2775 + }, + { + "epoch": 8.387145557655955, + "grad_norm": 1.4609375, + "learning_rate": 9.015101285200586e-06, + "loss": 0.6939, + "num_tokens": 17258870.0, + "step": 2776 + }, + { + "epoch": 8.390170132325142, + "grad_norm": 1.390625, + "learning_rate": 9.012349840028707e-06, + "loss": 0.7822, + "num_tokens": 17282380.0, + "step": 2777 + }, + { + "epoch": 8.39319470699433, + "grad_norm": 1.796875, + "learning_rate": 9.009595030613403e-06, + "loss": 0.515, + "num_tokens": 17290488.0, + "step": 2778 + }, + { + "epoch": 8.396219281663516, + "grad_norm": 1.4140625, + "learning_rate": 9.00683685959332e-06, + "loss": 0.729, + "num_tokens": 17313781.0, + "step": 2779 + }, + { + "epoch": 8.399243856332703, + "grad_norm": 1.4921875, + "learning_rate": 9.004075329610324e-06, + "loss": 0.7305, + "num_tokens": 17336399.0, + "step": 2780 + }, + { + "epoch": 8.40226843100189, + "grad_norm": 1.3046875, + "learning_rate": 9.001310443309503e-06, + "loss": 0.8561, + "num_tokens": 17376661.0, + "step": 2781 + }, + { + "epoch": 8.405293005671078, + "grad_norm": 1.8671875, + "learning_rate": 8.998542203339148e-06, + "loss": 0.5498, + "num_tokens": 17383904.0, + "step": 2782 + }, + { + "epoch": 8.408317580340265, + "grad_norm": 1.5859375, + "learning_rate": 8.995770612350778e-06, + "loss": 0.7055, + "num_tokens": 17399193.0, + "step": 2783 + }, + { + "epoch": 8.411342155009452, + "grad_norm": 1.4609375, + "learning_rate": 8.992995672999105e-06, + "loss": 0.8684, + "num_tokens": 17428484.0, + "step": 2784 + }, + { + "epoch": 8.414366729678639, + "grad_norm": 1.9765625, + "learning_rate": 8.990217387942061e-06, + "loss": 0.5453, + "num_tokens": 17435238.0, + "step": 2785 + }, + { + "epoch": 8.417391304347825, + "grad_norm": 1.2109375, + "learning_rate": 8.987435759840782e-06, + "loss": 0.8031, + "num_tokens": 17472613.0, + "step": 2786 + }, + { + "epoch": 8.420415879017014, + "grad_norm": 1.40625, + "learning_rate": 8.984650791359596e-06, + "loss": 0.7442, + "num_tokens": 17496295.0, + "step": 2787 + }, + { + "epoch": 8.4234404536862, + "grad_norm": 1.9453125, + "learning_rate": 8.98186248516604e-06, + "loss": 0.5561, + "num_tokens": 17503185.0, + "step": 2788 + }, + { + "epoch": 8.426465028355388, + "grad_norm": 1.2109375, + "learning_rate": 8.979070843930841e-06, + "loss": 0.7304, + "num_tokens": 17534597.0, + "step": 2789 + }, + { + "epoch": 8.429489603024575, + "grad_norm": 1.5078125, + "learning_rate": 8.976275870327928e-06, + "loss": 0.6324, + "num_tokens": 17550789.0, + "step": 2790 + }, + { + "epoch": 8.432514177693761, + "grad_norm": 1.8359375, + "learning_rate": 8.97347756703442e-06, + "loss": 0.6229, + "num_tokens": 17561022.0, + "step": 2791 + }, + { + "epoch": 8.435538752362948, + "grad_norm": 1.2578125, + "learning_rate": 8.970675936730618e-06, + "loss": 0.8569, + "num_tokens": 17601368.0, + "step": 2792 + }, + { + "epoch": 8.438563327032137, + "grad_norm": 1.8515625, + "learning_rate": 8.967870982100017e-06, + "loss": 0.7379, + "num_tokens": 17616365.0, + "step": 2793 + }, + { + "epoch": 8.441587901701324, + "grad_norm": 1.453125, + "learning_rate": 8.965062705829295e-06, + "loss": 0.805, + "num_tokens": 17639674.0, + "step": 2794 + }, + { + "epoch": 8.44461247637051, + "grad_norm": 1.8203125, + "learning_rate": 8.96225111060831e-06, + "loss": 0.6447, + "num_tokens": 17651506.0, + "step": 2795 + }, + { + "epoch": 8.447637051039697, + "grad_norm": 1.3671875, + "learning_rate": 8.959436199130102e-06, + "loss": 0.8462, + "num_tokens": 17681824.0, + "step": 2796 + }, + { + "epoch": 8.450661625708884, + "grad_norm": 1.9375, + "learning_rate": 8.956617974090879e-06, + "loss": 0.618, + "num_tokens": 17689597.0, + "step": 2797 + }, + { + "epoch": 8.453686200378073, + "grad_norm": 1.5390625, + "learning_rate": 8.953796438190038e-06, + "loss": 0.7892, + "num_tokens": 17713000.0, + "step": 2798 + }, + { + "epoch": 8.45671077504726, + "grad_norm": 1.7578125, + "learning_rate": 8.950971594130131e-06, + "loss": 0.654, + "num_tokens": 17727286.0, + "step": 2799 + }, + { + "epoch": 8.459735349716446, + "grad_norm": 1.0234375, + "learning_rate": 8.948143444616889e-06, + "loss": 0.847, + "num_tokens": 17782438.0, + "step": 2800 + }, + { + "epoch": 8.462759924385633, + "grad_norm": 1.921875, + "learning_rate": 8.945311992359206e-06, + "loss": 0.6432, + "num_tokens": 17789500.0, + "step": 2801 + }, + { + "epoch": 8.46578449905482, + "grad_norm": 1.8125, + "learning_rate": 8.942477240069137e-06, + "loss": 0.6019, + "num_tokens": 17797804.0, + "step": 2802 + }, + { + "epoch": 8.468809073724007, + "grad_norm": 1.3046875, + "learning_rate": 8.939639190461904e-06, + "loss": 0.6992, + "num_tokens": 17822185.0, + "step": 2803 + }, + { + "epoch": 8.471833648393195, + "grad_norm": 1.671875, + "learning_rate": 8.936797846255882e-06, + "loss": 0.7689, + "num_tokens": 17846288.0, + "step": 2804 + }, + { + "epoch": 8.474858223062382, + "grad_norm": 1.8671875, + "learning_rate": 8.933953210172604e-06, + "loss": 0.5639, + "num_tokens": 17853900.0, + "step": 2805 + }, + { + "epoch": 8.477882797731569, + "grad_norm": 1.8828125, + "learning_rate": 8.931105284936752e-06, + "loss": 0.5619, + "num_tokens": 17861666.0, + "step": 2806 + }, + { + "epoch": 8.480907372400756, + "grad_norm": 1.9609375, + "learning_rate": 8.928254073276166e-06, + "loss": 0.5939, + "num_tokens": 17869391.0, + "step": 2807 + }, + { + "epoch": 8.483931947069943, + "grad_norm": 1.5390625, + "learning_rate": 8.925399577921827e-06, + "loss": 0.7926, + "num_tokens": 17892366.0, + "step": 2808 + }, + { + "epoch": 8.486956521739131, + "grad_norm": 1.5078125, + "learning_rate": 8.922541801607866e-06, + "loss": 0.8114, + "num_tokens": 17916130.0, + "step": 2809 + }, + { + "epoch": 8.489981096408318, + "grad_norm": 1.4609375, + "learning_rate": 8.919680747071554e-06, + "loss": 0.8234, + "num_tokens": 17944050.0, + "step": 2810 + }, + { + "epoch": 8.493005671077505, + "grad_norm": 1.9375, + "learning_rate": 8.916816417053302e-06, + "loss": 0.6347, + "num_tokens": 17951553.0, + "step": 2811 + }, + { + "epoch": 8.496030245746692, + "grad_norm": 2.109375, + "learning_rate": 8.913948814296658e-06, + "loss": 0.6332, + "num_tokens": 17958607.0, + "step": 2812 + }, + { + "epoch": 8.499054820415878, + "grad_norm": 1.3125, + "learning_rate": 8.911077941548306e-06, + "loss": 0.807, + "num_tokens": 17990259.0, + "step": 2813 + }, + { + "epoch": 8.502079395085065, + "grad_norm": 1.0546875, + "learning_rate": 8.908203801558065e-06, + "loss": 0.8244, + "num_tokens": 18040839.0, + "step": 2814 + }, + { + "epoch": 8.505103969754254, + "grad_norm": 2.046875, + "learning_rate": 8.905326397078873e-06, + "loss": 0.611, + "num_tokens": 18047497.0, + "step": 2815 + }, + { + "epoch": 8.50812854442344, + "grad_norm": 1.4765625, + "learning_rate": 8.90244573086681e-06, + "loss": 0.8072, + "num_tokens": 18070399.0, + "step": 2816 + }, + { + "epoch": 8.511153119092628, + "grad_norm": 1.5078125, + "learning_rate": 8.899561805681067e-06, + "loss": 0.7743, + "num_tokens": 18093216.0, + "step": 2817 + }, + { + "epoch": 8.514177693761814, + "grad_norm": 1.4453125, + "learning_rate": 8.896674624283965e-06, + "loss": 0.7702, + "num_tokens": 18116679.0, + "step": 2818 + }, + { + "epoch": 8.517202268431001, + "grad_norm": 1.265625, + "learning_rate": 8.893784189440937e-06, + "loss": 0.8679, + "num_tokens": 18154899.0, + "step": 2819 + }, + { + "epoch": 8.52022684310019, + "grad_norm": 1.84375, + "learning_rate": 8.890890503920538e-06, + "loss": 0.7251, + "num_tokens": 18168419.0, + "step": 2820 + }, + { + "epoch": 8.523251417769377, + "grad_norm": 1.3828125, + "learning_rate": 8.887993570494436e-06, + "loss": 0.7134, + "num_tokens": 18200443.0, + "step": 2821 + }, + { + "epoch": 8.526275992438563, + "grad_norm": 1.0625, + "learning_rate": 8.885093391937406e-06, + "loss": 0.8779, + "num_tokens": 18251824.0, + "step": 2822 + }, + { + "epoch": 8.52930056710775, + "grad_norm": 0.9921875, + "learning_rate": 8.882189971027335e-06, + "loss": 0.8416, + "num_tokens": 18322632.0, + "step": 2823 + }, + { + "epoch": 8.532325141776937, + "grad_norm": 1.390625, + "learning_rate": 8.879283310545214e-06, + "loss": 0.6864, + "num_tokens": 18352034.0, + "step": 2824 + }, + { + "epoch": 8.535349716446124, + "grad_norm": 1.65625, + "learning_rate": 8.876373413275139e-06, + "loss": 0.8662, + "num_tokens": 18374393.0, + "step": 2825 + }, + { + "epoch": 8.538374291115312, + "grad_norm": 1.359375, + "learning_rate": 8.873460282004302e-06, + "loss": 0.6945, + "num_tokens": 18399824.0, + "step": 2826 + }, + { + "epoch": 8.5413988657845, + "grad_norm": 1.8671875, + "learning_rate": 8.870543919522996e-06, + "loss": 0.5885, + "num_tokens": 18408333.0, + "step": 2827 + }, + { + "epoch": 8.544423440453686, + "grad_norm": 2.015625, + "learning_rate": 8.86762432862461e-06, + "loss": 0.612, + "num_tokens": 18416116.0, + "step": 2828 + }, + { + "epoch": 8.547448015122873, + "grad_norm": 1.5234375, + "learning_rate": 8.86470151210562e-06, + "loss": 0.6953, + "num_tokens": 18433821.0, + "step": 2829 + }, + { + "epoch": 8.55047258979206, + "grad_norm": 1.8359375, + "learning_rate": 8.861775472765599e-06, + "loss": 0.6342, + "num_tokens": 18443546.0, + "step": 2830 + }, + { + "epoch": 8.553497164461248, + "grad_norm": 2.0625, + "learning_rate": 8.858846213407201e-06, + "loss": 0.6551, + "num_tokens": 18450735.0, + "step": 2831 + }, + { + "epoch": 8.556521739130435, + "grad_norm": 1.4609375, + "learning_rate": 8.855913736836168e-06, + "loss": 0.7669, + "num_tokens": 18473115.0, + "step": 2832 + }, + { + "epoch": 8.559546313799622, + "grad_norm": 1.21875, + "learning_rate": 8.85297804586132e-06, + "loss": 0.894, + "num_tokens": 18513488.0, + "step": 2833 + }, + { + "epoch": 8.562570888468809, + "grad_norm": 1.4921875, + "learning_rate": 8.85003914329456e-06, + "loss": 0.8005, + "num_tokens": 18533829.0, + "step": 2834 + }, + { + "epoch": 8.565595463137996, + "grad_norm": 1.3125, + "learning_rate": 8.84709703195086e-06, + "loss": 0.7287, + "num_tokens": 18564046.0, + "step": 2835 + }, + { + "epoch": 8.568620037807184, + "grad_norm": 1.1875, + "learning_rate": 8.844151714648274e-06, + "loss": 0.7669, + "num_tokens": 18604899.0, + "step": 2836 + }, + { + "epoch": 8.571644612476371, + "grad_norm": 1.3359375, + "learning_rate": 8.841203194207925e-06, + "loss": 0.8669, + "num_tokens": 18637406.0, + "step": 2837 + }, + { + "epoch": 8.574669187145558, + "grad_norm": 1.5625, + "learning_rate": 8.838251473453998e-06, + "loss": 0.7585, + "num_tokens": 18656892.0, + "step": 2838 + }, + { + "epoch": 8.577693761814745, + "grad_norm": 1.9453125, + "learning_rate": 8.835296555213753e-06, + "loss": 0.552, + "num_tokens": 18663954.0, + "step": 2839 + }, + { + "epoch": 8.580718336483931, + "grad_norm": 1.4140625, + "learning_rate": 8.832338442317499e-06, + "loss": 0.7484, + "num_tokens": 18688998.0, + "step": 2840 + }, + { + "epoch": 8.583742911153118, + "grad_norm": 1.2578125, + "learning_rate": 8.829377137598621e-06, + "loss": 0.8343, + "num_tokens": 18727143.0, + "step": 2841 + }, + { + "epoch": 8.586767485822307, + "grad_norm": 1.8046875, + "learning_rate": 8.82641264389355e-06, + "loss": 0.7174, + "num_tokens": 18739889.0, + "step": 2842 + }, + { + "epoch": 8.589792060491494, + "grad_norm": 1.5390625, + "learning_rate": 8.823444964041777e-06, + "loss": 0.8046, + "num_tokens": 18761334.0, + "step": 2843 + }, + { + "epoch": 8.59281663516068, + "grad_norm": 1.2421875, + "learning_rate": 8.820474100885839e-06, + "loss": 0.8722, + "num_tokens": 18801084.0, + "step": 2844 + }, + { + "epoch": 8.595841209829867, + "grad_norm": 1.6328125, + "learning_rate": 8.817500057271328e-06, + "loss": 0.7835, + "num_tokens": 18818408.0, + "step": 2845 + }, + { + "epoch": 8.598865784499054, + "grad_norm": 1.65625, + "learning_rate": 8.81452283604688e-06, + "loss": 0.7826, + "num_tokens": 18839076.0, + "step": 2846 + }, + { + "epoch": 8.601890359168243, + "grad_norm": 0.96484375, + "learning_rate": 8.811542440064174e-06, + "loss": 0.8013, + "num_tokens": 18903897.0, + "step": 2847 + }, + { + "epoch": 8.60491493383743, + "grad_norm": 1.921875, + "learning_rate": 8.80855887217793e-06, + "loss": 0.5836, + "num_tokens": 18911790.0, + "step": 2848 + }, + { + "epoch": 8.607939508506616, + "grad_norm": 1.625, + "learning_rate": 8.805572135245911e-06, + "loss": 0.6527, + "num_tokens": 18924856.0, + "step": 2849 + }, + { + "epoch": 8.610964083175803, + "grad_norm": 2.296875, + "learning_rate": 8.802582232128908e-06, + "loss": 0.5914, + "num_tokens": 18931186.0, + "step": 2850 + }, + { + "epoch": 8.61398865784499, + "grad_norm": 1.5390625, + "learning_rate": 8.799589165690749e-06, + "loss": 0.6908, + "num_tokens": 18953828.0, + "step": 2851 + }, + { + "epoch": 8.617013232514179, + "grad_norm": 1.8828125, + "learning_rate": 8.796592938798287e-06, + "loss": 0.5621, + "num_tokens": 18961280.0, + "step": 2852 + }, + { + "epoch": 8.620037807183365, + "grad_norm": 1.703125, + "learning_rate": 8.793593554321413e-06, + "loss": 0.5386, + "num_tokens": 18969822.0, + "step": 2853 + }, + { + "epoch": 8.623062381852552, + "grad_norm": 1.46875, + "learning_rate": 8.790591015133029e-06, + "loss": 0.8312, + "num_tokens": 18993199.0, + "step": 2854 + }, + { + "epoch": 8.626086956521739, + "grad_norm": 1.8125, + "learning_rate": 8.787585324109067e-06, + "loss": 0.7168, + "num_tokens": 19003702.0, + "step": 2855 + }, + { + "epoch": 8.629111531190926, + "grad_norm": 2.15625, + "learning_rate": 8.784576484128477e-06, + "loss": 0.6501, + "num_tokens": 19010250.0, + "step": 2856 + }, + { + "epoch": 8.632136105860113, + "grad_norm": 1.890625, + "learning_rate": 8.781564498073226e-06, + "loss": 0.598, + "num_tokens": 19018752.0, + "step": 2857 + }, + { + "epoch": 8.635160680529301, + "grad_norm": 1.1640625, + "learning_rate": 8.77854936882829e-06, + "loss": 0.7991, + "num_tokens": 19063071.0, + "step": 2858 + }, + { + "epoch": 8.638185255198488, + "grad_norm": 1.6015625, + "learning_rate": 8.775531099281657e-06, + "loss": 0.7697, + "num_tokens": 19085417.0, + "step": 2859 + }, + { + "epoch": 8.641209829867675, + "grad_norm": 1.984375, + "learning_rate": 8.772509692324325e-06, + "loss": 0.6312, + "num_tokens": 19093072.0, + "step": 2860 + }, + { + "epoch": 8.644234404536862, + "grad_norm": 1.9453125, + "learning_rate": 8.7694851508503e-06, + "loss": 0.6167, + "num_tokens": 19102121.0, + "step": 2861 + }, + { + "epoch": 8.647258979206049, + "grad_norm": 2.03125, + "learning_rate": 8.766457477756583e-06, + "loss": 0.6509, + "num_tokens": 19109651.0, + "step": 2862 + }, + { + "epoch": 8.650283553875237, + "grad_norm": 1.7421875, + "learning_rate": 8.76342667594318e-06, + "loss": 0.7016, + "num_tokens": 19122709.0, + "step": 2863 + }, + { + "epoch": 8.653308128544424, + "grad_norm": 1.1171875, + "learning_rate": 8.760392748313092e-06, + "loss": 0.8464, + "num_tokens": 19169485.0, + "step": 2864 + }, + { + "epoch": 8.65633270321361, + "grad_norm": 1.984375, + "learning_rate": 8.75735569777231e-06, + "loss": 0.531, + "num_tokens": 19176129.0, + "step": 2865 + }, + { + "epoch": 8.659357277882798, + "grad_norm": 1.1015625, + "learning_rate": 8.754315527229825e-06, + "loss": 0.8505, + "num_tokens": 19237677.0, + "step": 2866 + }, + { + "epoch": 8.662381852551984, + "grad_norm": 1.25, + "learning_rate": 8.751272239597612e-06, + "loss": 0.8657, + "num_tokens": 19273545.0, + "step": 2867 + }, + { + "epoch": 8.665406427221171, + "grad_norm": 1.9453125, + "learning_rate": 8.748225837790626e-06, + "loss": 0.6758, + "num_tokens": 19282256.0, + "step": 2868 + }, + { + "epoch": 8.66843100189036, + "grad_norm": 1.390625, + "learning_rate": 8.745176324726816e-06, + "loss": 0.7724, + "num_tokens": 19309937.0, + "step": 2869 + }, + { + "epoch": 8.671455576559547, + "grad_norm": 1.75, + "learning_rate": 8.7421237033271e-06, + "loss": 0.7501, + "num_tokens": 19325414.0, + "step": 2870 + }, + { + "epoch": 8.674480151228734, + "grad_norm": 1.9609375, + "learning_rate": 8.739067976515385e-06, + "loss": 0.5656, + "num_tokens": 19333080.0, + "step": 2871 + }, + { + "epoch": 8.67750472589792, + "grad_norm": 1.5546875, + "learning_rate": 8.73600914721854e-06, + "loss": 0.7966, + "num_tokens": 19357152.0, + "step": 2872 + }, + { + "epoch": 8.680529300567107, + "grad_norm": 1.9921875, + "learning_rate": 8.732947218366414e-06, + "loss": 0.924, + "num_tokens": 19380607.0, + "step": 2873 + }, + { + "epoch": 8.683553875236296, + "grad_norm": 1.4140625, + "learning_rate": 8.729882192891823e-06, + "loss": 0.8137, + "num_tokens": 19406951.0, + "step": 2874 + }, + { + "epoch": 8.686578449905483, + "grad_norm": 1.5859375, + "learning_rate": 8.726814073730548e-06, + "loss": 0.8204, + "num_tokens": 19429284.0, + "step": 2875 + }, + { + "epoch": 8.68960302457467, + "grad_norm": 1.203125, + "learning_rate": 8.723742863821332e-06, + "loss": 0.851, + "num_tokens": 19473575.0, + "step": 2876 + }, + { + "epoch": 8.692627599243856, + "grad_norm": 1.3125, + "learning_rate": 8.720668566105883e-06, + "loss": 0.6756, + "num_tokens": 19497888.0, + "step": 2877 + }, + { + "epoch": 8.695652173913043, + "grad_norm": 1.921875, + "learning_rate": 8.717591183528865e-06, + "loss": 0.5825, + "num_tokens": 19505512.0, + "step": 2878 + }, + { + "epoch": 8.69867674858223, + "grad_norm": 1.9921875, + "learning_rate": 8.71451071903789e-06, + "loss": 0.6673, + "num_tokens": 19515365.0, + "step": 2879 + }, + { + "epoch": 8.701701323251418, + "grad_norm": 1.4765625, + "learning_rate": 8.711427175583531e-06, + "loss": 0.7483, + "num_tokens": 19537648.0, + "step": 2880 + }, + { + "epoch": 8.704725897920605, + "grad_norm": 1.71875, + "learning_rate": 8.708340556119307e-06, + "loss": 0.7696, + "num_tokens": 19552827.0, + "step": 2881 + }, + { + "epoch": 8.707750472589792, + "grad_norm": 1.8359375, + "learning_rate": 8.705250863601684e-06, + "loss": 0.5693, + "num_tokens": 19560321.0, + "step": 2882 + }, + { + "epoch": 8.710775047258979, + "grad_norm": 1.15625, + "learning_rate": 8.702158100990066e-06, + "loss": 0.9062, + "num_tokens": 19603860.0, + "step": 2883 + }, + { + "epoch": 8.713799621928166, + "grad_norm": 1.34375, + "learning_rate": 8.699062271246806e-06, + "loss": 0.643, + "num_tokens": 19626351.0, + "step": 2884 + }, + { + "epoch": 8.716824196597354, + "grad_norm": 1.578125, + "learning_rate": 8.695963377337191e-06, + "loss": 0.7043, + "num_tokens": 19642869.0, + "step": 2885 + }, + { + "epoch": 8.719848771266541, + "grad_norm": 2.171875, + "learning_rate": 8.69286142222944e-06, + "loss": 0.6272, + "num_tokens": 19650620.0, + "step": 2886 + }, + { + "epoch": 8.722873345935728, + "grad_norm": 1.8515625, + "learning_rate": 8.689756408894709e-06, + "loss": 0.6389, + "num_tokens": 19659725.0, + "step": 2887 + }, + { + "epoch": 8.725897920604915, + "grad_norm": 1.5703125, + "learning_rate": 8.686648340307081e-06, + "loss": 0.68, + "num_tokens": 19679304.0, + "step": 2888 + }, + { + "epoch": 8.728922495274102, + "grad_norm": 1.7109375, + "learning_rate": 8.683537219443565e-06, + "loss": 0.8071, + "num_tokens": 19695435.0, + "step": 2889 + }, + { + "epoch": 8.731947069943288, + "grad_norm": 1.796875, + "learning_rate": 8.680423049284096e-06, + "loss": 0.564, + "num_tokens": 19703615.0, + "step": 2890 + }, + { + "epoch": 8.734971644612477, + "grad_norm": 1.2734375, + "learning_rate": 8.677305832811524e-06, + "loss": 0.7764, + "num_tokens": 19734161.0, + "step": 2891 + }, + { + "epoch": 8.737996219281664, + "grad_norm": 1.4296875, + "learning_rate": 8.674185573011625e-06, + "loss": 0.8671, + "num_tokens": 19762975.0, + "step": 2892 + }, + { + "epoch": 8.74102079395085, + "grad_norm": 1.8515625, + "learning_rate": 8.671062272873084e-06, + "loss": 0.667, + "num_tokens": 19774296.0, + "step": 2893 + }, + { + "epoch": 8.744045368620037, + "grad_norm": 1.8125, + "learning_rate": 8.667935935387497e-06, + "loss": 0.6508, + "num_tokens": 19783876.0, + "step": 2894 + }, + { + "epoch": 8.747069943289224, + "grad_norm": 1.6484375, + "learning_rate": 8.664806563549379e-06, + "loss": 0.6533, + "num_tokens": 19797802.0, + "step": 2895 + }, + { + "epoch": 8.750094517958413, + "grad_norm": 1.609375, + "learning_rate": 8.66167416035614e-06, + "loss": 0.7024, + "num_tokens": 19813588.0, + "step": 2896 + }, + { + "epoch": 8.7531190926276, + "grad_norm": 1.046875, + "learning_rate": 8.658538728808097e-06, + "loss": 0.9034, + "num_tokens": 19877441.0, + "step": 2897 + }, + { + "epoch": 8.756143667296787, + "grad_norm": 1.25, + "learning_rate": 8.655400271908472e-06, + "loss": 0.92, + "num_tokens": 19916577.0, + "step": 2898 + }, + { + "epoch": 8.759168241965973, + "grad_norm": 1.8359375, + "learning_rate": 8.65225879266338e-06, + "loss": 0.5875, + "num_tokens": 19924275.0, + "step": 2899 + }, + { + "epoch": 8.76219281663516, + "grad_norm": 1.8359375, + "learning_rate": 8.649114294081833e-06, + "loss": 0.5627, + "num_tokens": 19932318.0, + "step": 2900 + }, + { + "epoch": 8.765217391304347, + "grad_norm": 1.96875, + "learning_rate": 8.645966779175734e-06, + "loss": 0.606, + "num_tokens": 19940506.0, + "step": 2901 + }, + { + "epoch": 8.768241965973536, + "grad_norm": 1.953125, + "learning_rate": 8.642816250959874e-06, + "loss": 0.6289, + "num_tokens": 19948422.0, + "step": 2902 + }, + { + "epoch": 8.771266540642722, + "grad_norm": 1.1953125, + "learning_rate": 8.639662712451935e-06, + "loss": 0.741, + "num_tokens": 19981745.0, + "step": 2903 + }, + { + "epoch": 8.77429111531191, + "grad_norm": 1.7578125, + "learning_rate": 8.636506166672478e-06, + "loss": 0.7447, + "num_tokens": 19998008.0, + "step": 2904 + }, + { + "epoch": 8.777315689981096, + "grad_norm": 1.71875, + "learning_rate": 8.633346616644944e-06, + "loss": 0.7209, + "num_tokens": 20015096.0, + "step": 2905 + }, + { + "epoch": 8.780340264650283, + "grad_norm": 1.859375, + "learning_rate": 8.630184065395655e-06, + "loss": 0.697, + "num_tokens": 20027634.0, + "step": 2906 + }, + { + "epoch": 8.783364839319471, + "grad_norm": 1.2109375, + "learning_rate": 8.627018515953807e-06, + "loss": 0.8251, + "num_tokens": 20068275.0, + "step": 2907 + }, + { + "epoch": 8.786389413988658, + "grad_norm": 1.3984375, + "learning_rate": 8.623849971351464e-06, + "loss": 0.7711, + "num_tokens": 20092985.0, + "step": 2908 + }, + { + "epoch": 8.789413988657845, + "grad_norm": 1.890625, + "learning_rate": 8.620678434623563e-06, + "loss": 0.5751, + "num_tokens": 20101024.0, + "step": 2909 + }, + { + "epoch": 8.792438563327032, + "grad_norm": 1.078125, + "learning_rate": 8.617503908807905e-06, + "loss": 0.8212, + "num_tokens": 20147134.0, + "step": 2910 + }, + { + "epoch": 8.795463137996219, + "grad_norm": 1.875, + "learning_rate": 8.614326396945154e-06, + "loss": 0.5399, + "num_tokens": 20154738.0, + "step": 2911 + }, + { + "epoch": 8.798487712665406, + "grad_norm": 2.046875, + "learning_rate": 8.611145902078838e-06, + "loss": 0.5578, + "num_tokens": 20161866.0, + "step": 2912 + }, + { + "epoch": 8.801512287334594, + "grad_norm": 2.015625, + "learning_rate": 8.607962427255334e-06, + "loss": 0.5574, + "num_tokens": 20169349.0, + "step": 2913 + }, + { + "epoch": 8.804536862003781, + "grad_norm": 1.390625, + "learning_rate": 8.60477597552388e-06, + "loss": 0.8133, + "num_tokens": 20198407.0, + "step": 2914 + }, + { + "epoch": 8.807561436672968, + "grad_norm": 1.65625, + "learning_rate": 8.601586549936567e-06, + "loss": 0.7886, + "num_tokens": 20223414.0, + "step": 2915 + }, + { + "epoch": 8.810586011342155, + "grad_norm": 1.5546875, + "learning_rate": 8.598394153548328e-06, + "loss": 0.8361, + "num_tokens": 20246649.0, + "step": 2916 + }, + { + "epoch": 8.813610586011341, + "grad_norm": 1.3828125, + "learning_rate": 8.595198789416943e-06, + "loss": 0.7568, + "num_tokens": 20270893.0, + "step": 2917 + }, + { + "epoch": 8.81663516068053, + "grad_norm": 1.7265625, + "learning_rate": 8.592000460603039e-06, + "loss": 0.6748, + "num_tokens": 20285638.0, + "step": 2918 + }, + { + "epoch": 8.819659735349717, + "grad_norm": 1.2421875, + "learning_rate": 8.588799170170077e-06, + "loss": 0.7674, + "num_tokens": 20324222.0, + "step": 2919 + }, + { + "epoch": 8.822684310018904, + "grad_norm": 1.8671875, + "learning_rate": 8.58559492118436e-06, + "loss": 0.5313, + "num_tokens": 20331490.0, + "step": 2920 + }, + { + "epoch": 8.82570888468809, + "grad_norm": 1.3984375, + "learning_rate": 8.582387716715021e-06, + "loss": 0.8077, + "num_tokens": 20361052.0, + "step": 2921 + }, + { + "epoch": 8.828733459357277, + "grad_norm": 1.59375, + "learning_rate": 8.579177559834027e-06, + "loss": 0.9107, + "num_tokens": 20386200.0, + "step": 2922 + }, + { + "epoch": 8.831758034026464, + "grad_norm": 1.1328125, + "learning_rate": 8.57596445361617e-06, + "loss": 0.8834, + "num_tokens": 20440908.0, + "step": 2923 + }, + { + "epoch": 8.834782608695653, + "grad_norm": 1.9453125, + "learning_rate": 8.572748401139066e-06, + "loss": 0.576, + "num_tokens": 20448699.0, + "step": 2924 + }, + { + "epoch": 8.83780718336484, + "grad_norm": 1.7109375, + "learning_rate": 8.56952940548316e-06, + "loss": 0.7644, + "num_tokens": 20465118.0, + "step": 2925 + }, + { + "epoch": 8.840831758034026, + "grad_norm": 1.28125, + "learning_rate": 8.566307469731706e-06, + "loss": 0.8556, + "num_tokens": 20502647.0, + "step": 2926 + }, + { + "epoch": 8.843856332703213, + "grad_norm": 1.4453125, + "learning_rate": 8.563082596970785e-06, + "loss": 0.7135, + "num_tokens": 20526237.0, + "step": 2927 + }, + { + "epoch": 8.8468809073724, + "grad_norm": 1.4296875, + "learning_rate": 8.559854790289278e-06, + "loss": 0.8469, + "num_tokens": 20556314.0, + "step": 2928 + }, + { + "epoch": 8.849905482041589, + "grad_norm": 1.25, + "learning_rate": 8.55662405277889e-06, + "loss": 0.8173, + "num_tokens": 20594373.0, + "step": 2929 + }, + { + "epoch": 8.852930056710775, + "grad_norm": 1.3046875, + "learning_rate": 8.553390387534126e-06, + "loss": 0.8735, + "num_tokens": 20632884.0, + "step": 2930 + }, + { + "epoch": 8.855954631379962, + "grad_norm": 1.7578125, + "learning_rate": 8.550153797652292e-06, + "loss": 0.5684, + "num_tokens": 20643949.0, + "step": 2931 + }, + { + "epoch": 8.858979206049149, + "grad_norm": 1.6484375, + "learning_rate": 8.546914286233505e-06, + "loss": 0.7827, + "num_tokens": 20663797.0, + "step": 2932 + }, + { + "epoch": 8.862003780718336, + "grad_norm": 1.625, + "learning_rate": 8.543671856380672e-06, + "loss": 0.7474, + "num_tokens": 20681138.0, + "step": 2933 + }, + { + "epoch": 8.865028355387524, + "grad_norm": 1.671875, + "learning_rate": 8.540426511199498e-06, + "loss": 0.719, + "num_tokens": 20696315.0, + "step": 2934 + }, + { + "epoch": 8.868052930056711, + "grad_norm": 1.3359375, + "learning_rate": 8.537178253798483e-06, + "loss": 0.7547, + "num_tokens": 20721101.0, + "step": 2935 + }, + { + "epoch": 8.871077504725898, + "grad_norm": 1.8359375, + "learning_rate": 8.53392708728891e-06, + "loss": 0.5489, + "num_tokens": 20729194.0, + "step": 2936 + }, + { + "epoch": 8.874102079395085, + "grad_norm": 1.21875, + "learning_rate": 8.530673014784857e-06, + "loss": 0.7586, + "num_tokens": 20760385.0, + "step": 2937 + }, + { + "epoch": 8.877126654064272, + "grad_norm": 1.765625, + "learning_rate": 8.527416039403178e-06, + "loss": 0.6371, + "num_tokens": 20768956.0, + "step": 2938 + }, + { + "epoch": 8.88015122873346, + "grad_norm": 1.3046875, + "learning_rate": 8.524156164263509e-06, + "loss": 0.7654, + "num_tokens": 20803880.0, + "step": 2939 + }, + { + "epoch": 8.883175803402647, + "grad_norm": 1.90625, + "learning_rate": 8.52089339248827e-06, + "loss": 0.5813, + "num_tokens": 20811493.0, + "step": 2940 + }, + { + "epoch": 8.886200378071834, + "grad_norm": 1.859375, + "learning_rate": 8.517627727202647e-06, + "loss": 0.6324, + "num_tokens": 20823505.0, + "step": 2941 + }, + { + "epoch": 8.88922495274102, + "grad_norm": 1.5234375, + "learning_rate": 8.514359171534603e-06, + "loss": 0.7164, + "num_tokens": 20845916.0, + "step": 2942 + }, + { + "epoch": 8.892249527410208, + "grad_norm": 1.359375, + "learning_rate": 8.511087728614863e-06, + "loss": 0.6756, + "num_tokens": 20869781.0, + "step": 2943 + }, + { + "epoch": 8.895274102079394, + "grad_norm": 1.921875, + "learning_rate": 8.507813401576925e-06, + "loss": 0.596, + "num_tokens": 20877453.0, + "step": 2944 + }, + { + "epoch": 8.898298676748583, + "grad_norm": 1.484375, + "learning_rate": 8.504536193557049e-06, + "loss": 0.7732, + "num_tokens": 20900830.0, + "step": 2945 + }, + { + "epoch": 8.90132325141777, + "grad_norm": 1.3828125, + "learning_rate": 8.501256107694246e-06, + "loss": 0.8599, + "num_tokens": 20932745.0, + "step": 2946 + }, + { + "epoch": 8.904347826086957, + "grad_norm": 1.5078125, + "learning_rate": 8.497973147130296e-06, + "loss": 0.8514, + "num_tokens": 20956134.0, + "step": 2947 + }, + { + "epoch": 8.907372400756143, + "grad_norm": 1.859375, + "learning_rate": 8.494687315009718e-06, + "loss": 0.5898, + "num_tokens": 20964788.0, + "step": 2948 + }, + { + "epoch": 8.91039697542533, + "grad_norm": 1.7421875, + "learning_rate": 8.491398614479795e-06, + "loss": 0.7574, + "num_tokens": 20980058.0, + "step": 2949 + }, + { + "epoch": 8.913421550094519, + "grad_norm": 1.3828125, + "learning_rate": 8.488107048690554e-06, + "loss": 0.8426, + "num_tokens": 21008370.0, + "step": 2950 + }, + { + "epoch": 8.916446124763706, + "grad_norm": 1.1875, + "learning_rate": 8.484812620794757e-06, + "loss": 0.8017, + "num_tokens": 21048935.0, + "step": 2951 + }, + { + "epoch": 8.919470699432893, + "grad_norm": 1.5078125, + "learning_rate": 8.481515333947921e-06, + "loss": 0.7857, + "num_tokens": 21072927.0, + "step": 2952 + }, + { + "epoch": 8.92249527410208, + "grad_norm": 1.1875, + "learning_rate": 8.478215191308289e-06, + "loss": 0.7829, + "num_tokens": 21106827.0, + "step": 2953 + }, + { + "epoch": 8.925519848771266, + "grad_norm": 1.34375, + "learning_rate": 8.474912196036852e-06, + "loss": 0.7973, + "num_tokens": 21136044.0, + "step": 2954 + }, + { + "epoch": 8.928544423440453, + "grad_norm": 1.8046875, + "learning_rate": 8.471606351297323e-06, + "loss": 0.633, + "num_tokens": 21147218.0, + "step": 2955 + }, + { + "epoch": 8.931568998109642, + "grad_norm": 1.8515625, + "learning_rate": 8.468297660256148e-06, + "loss": 0.5173, + "num_tokens": 21154787.0, + "step": 2956 + }, + { + "epoch": 8.934593572778828, + "grad_norm": 2.03125, + "learning_rate": 8.4649861260825e-06, + "loss": 0.617, + "num_tokens": 21161603.0, + "step": 2957 + }, + { + "epoch": 8.937618147448015, + "grad_norm": 1.7421875, + "learning_rate": 8.461671751948277e-06, + "loss": 0.681, + "num_tokens": 21174413.0, + "step": 2958 + }, + { + "epoch": 8.940642722117202, + "grad_norm": 1.765625, + "learning_rate": 8.458354541028091e-06, + "loss": 0.6746, + "num_tokens": 21185378.0, + "step": 2959 + }, + { + "epoch": 8.943667296786389, + "grad_norm": 1.1953125, + "learning_rate": 8.455034496499276e-06, + "loss": 0.8237, + "num_tokens": 21224733.0, + "step": 2960 + }, + { + "epoch": 8.946691871455577, + "grad_norm": 1.421875, + "learning_rate": 8.451711621541881e-06, + "loss": 0.6716, + "num_tokens": 21246378.0, + "step": 2961 + }, + { + "epoch": 8.949716446124764, + "grad_norm": 1.2421875, + "learning_rate": 8.448385919338664e-06, + "loss": 0.7847, + "num_tokens": 21284578.0, + "step": 2962 + }, + { + "epoch": 8.952741020793951, + "grad_norm": 1.875, + "learning_rate": 8.445057393075088e-06, + "loss": 0.5673, + "num_tokens": 21293060.0, + "step": 2963 + }, + { + "epoch": 8.955765595463138, + "grad_norm": 1.53125, + "learning_rate": 8.441726045939327e-06, + "loss": 0.7435, + "num_tokens": 21311793.0, + "step": 2964 + }, + { + "epoch": 8.958790170132325, + "grad_norm": 1.96875, + "learning_rate": 8.438391881122255e-06, + "loss": 0.6864, + "num_tokens": 21323616.0, + "step": 2965 + }, + { + "epoch": 8.961814744801512, + "grad_norm": 2.09375, + "learning_rate": 8.435054901817438e-06, + "loss": 0.7895, + "num_tokens": 21336294.0, + "step": 2966 + }, + { + "epoch": 8.9648393194707, + "grad_norm": 2.09375, + "learning_rate": 8.431715111221152e-06, + "loss": 0.5467, + "num_tokens": 21343029.0, + "step": 2967 + }, + { + "epoch": 8.967863894139887, + "grad_norm": 1.828125, + "learning_rate": 8.428372512532354e-06, + "loss": 0.6573, + "num_tokens": 21354003.0, + "step": 2968 + }, + { + "epoch": 8.970888468809074, + "grad_norm": 1.203125, + "learning_rate": 8.425027108952693e-06, + "loss": 0.8796, + "num_tokens": 21400142.0, + "step": 2969 + }, + { + "epoch": 8.97391304347826, + "grad_norm": 1.921875, + "learning_rate": 8.421678903686508e-06, + "loss": 0.5479, + "num_tokens": 21407861.0, + "step": 2970 + }, + { + "epoch": 8.976937618147447, + "grad_norm": 1.5625, + "learning_rate": 8.418327899940816e-06, + "loss": 0.6866, + "num_tokens": 21427389.0, + "step": 2971 + }, + { + "epoch": 8.979962192816636, + "grad_norm": 1.3515625, + "learning_rate": 8.414974100925321e-06, + "loss": 0.6988, + "num_tokens": 21451414.0, + "step": 2972 + }, + { + "epoch": 8.982986767485823, + "grad_norm": 1.4140625, + "learning_rate": 8.411617509852401e-06, + "loss": 0.8137, + "num_tokens": 21475198.0, + "step": 2973 + }, + { + "epoch": 8.98601134215501, + "grad_norm": 1.390625, + "learning_rate": 8.408258129937108e-06, + "loss": 0.8257, + "num_tokens": 21504193.0, + "step": 2974 + }, + { + "epoch": 8.989035916824196, + "grad_norm": 1.8984375, + "learning_rate": 8.404895964397166e-06, + "loss": 0.6951, + "num_tokens": 21516963.0, + "step": 2975 + }, + { + "epoch": 8.992060491493383, + "grad_norm": 1.3515625, + "learning_rate": 8.401531016452967e-06, + "loss": 0.7464, + "num_tokens": 21545093.0, + "step": 2976 + }, + { + "epoch": 8.99508506616257, + "grad_norm": 1.8359375, + "learning_rate": 8.398163289327568e-06, + "loss": 0.6491, + "num_tokens": 21556989.0, + "step": 2977 + }, + { + "epoch": 8.998109640831759, + "grad_norm": 2.25, + "learning_rate": 8.394792786246686e-06, + "loss": 0.5801, + "num_tokens": 21564193.0, + "step": 2978 + }, + { + "epoch": 9.0, + "grad_norm": 1.4765625, + "learning_rate": 8.391419510438703e-06, + "loss": 0.8342, + "num_tokens": 21585939.0, + "step": 2979 + }, + { + "epoch": 9.003024574669187, + "grad_norm": 2.109375, + "learning_rate": 8.388043465134649e-06, + "loss": 0.6033, + "num_tokens": 21592185.0, + "step": 2980 + }, + { + "epoch": 9.006049149338374, + "grad_norm": 1.9453125, + "learning_rate": 8.384664653568213e-06, + "loss": 0.5653, + "num_tokens": 21600156.0, + "step": 2981 + }, + { + "epoch": 9.009073724007562, + "grad_norm": 1.828125, + "learning_rate": 8.38128307897573e-06, + "loss": 0.5897, + "num_tokens": 21608752.0, + "step": 2982 + }, + { + "epoch": 9.012098298676749, + "grad_norm": 1.515625, + "learning_rate": 8.377898744596185e-06, + "loss": 0.7994, + "num_tokens": 21634650.0, + "step": 2983 + }, + { + "epoch": 9.015122873345936, + "grad_norm": 1.484375, + "learning_rate": 8.3745116536712e-06, + "loss": 0.786, + "num_tokens": 21658249.0, + "step": 2984 + }, + { + "epoch": 9.018147448015123, + "grad_norm": 1.4921875, + "learning_rate": 8.371121809445043e-06, + "loss": 0.6614, + "num_tokens": 21677924.0, + "step": 2985 + }, + { + "epoch": 9.02117202268431, + "grad_norm": 1.484375, + "learning_rate": 8.367729215164618e-06, + "loss": 0.9605, + "num_tokens": 21716450.0, + "step": 2986 + }, + { + "epoch": 9.024196597353496, + "grad_norm": 1.5234375, + "learning_rate": 8.364333874079462e-06, + "loss": 0.8047, + "num_tokens": 21739803.0, + "step": 2987 + }, + { + "epoch": 9.027221172022685, + "grad_norm": 1.4921875, + "learning_rate": 8.360935789441746e-06, + "loss": 0.8079, + "num_tokens": 21761741.0, + "step": 2988 + }, + { + "epoch": 9.030245746691872, + "grad_norm": 1.2890625, + "learning_rate": 8.357534964506262e-06, + "loss": 0.7829, + "num_tokens": 21791338.0, + "step": 2989 + }, + { + "epoch": 9.033270321361059, + "grad_norm": 1.65625, + "learning_rate": 8.354131402530436e-06, + "loss": 0.661, + "num_tokens": 21803664.0, + "step": 2990 + }, + { + "epoch": 9.036294896030245, + "grad_norm": 1.7109375, + "learning_rate": 8.350725106774308e-06, + "loss": 0.5485, + "num_tokens": 21813177.0, + "step": 2991 + }, + { + "epoch": 9.039319470699432, + "grad_norm": 1.5859375, + "learning_rate": 8.34731608050054e-06, + "loss": 0.8637, + "num_tokens": 21835739.0, + "step": 2992 + }, + { + "epoch": 9.04234404536862, + "grad_norm": 1.5078125, + "learning_rate": 8.343904326974409e-06, + "loss": 0.7539, + "num_tokens": 21857276.0, + "step": 2993 + }, + { + "epoch": 9.045368620037808, + "grad_norm": 1.53125, + "learning_rate": 8.340489849463804e-06, + "loss": 0.8047, + "num_tokens": 21880226.0, + "step": 2994 + }, + { + "epoch": 9.048393194706994, + "grad_norm": 1.5078125, + "learning_rate": 8.33707265123922e-06, + "loss": 0.7289, + "num_tokens": 21900823.0, + "step": 2995 + }, + { + "epoch": 9.051417769376181, + "grad_norm": 1.1796875, + "learning_rate": 8.333652735573767e-06, + "loss": 0.7827, + "num_tokens": 21938221.0, + "step": 2996 + }, + { + "epoch": 9.054442344045368, + "grad_norm": 1.421875, + "learning_rate": 8.330230105743147e-06, + "loss": 0.7681, + "num_tokens": 21962033.0, + "step": 2997 + }, + { + "epoch": 9.057466918714555, + "grad_norm": 1.328125, + "learning_rate": 8.32680476502567e-06, + "loss": 0.726, + "num_tokens": 21990172.0, + "step": 2998 + }, + { + "epoch": 9.060491493383743, + "grad_norm": 1.78125, + "learning_rate": 8.323376716702236e-06, + "loss": 0.5209, + "num_tokens": 21998618.0, + "step": 2999 + }, + { + "epoch": 9.06351606805293, + "grad_norm": 1.3359375, + "learning_rate": 8.319945964056346e-06, + "loss": 0.7002, + "num_tokens": 22022426.0, + "step": 3000 + }, + { + "epoch": 9.066540642722117, + "grad_norm": 1.5390625, + "learning_rate": 8.316512510374083e-06, + "loss": 0.816, + "num_tokens": 22046271.0, + "step": 3001 + }, + { + "epoch": 9.069565217391304, + "grad_norm": 1.9765625, + "learning_rate": 8.313076358944122e-06, + "loss": 0.6104, + "num_tokens": 22053945.0, + "step": 3002 + }, + { + "epoch": 9.07258979206049, + "grad_norm": 2.203125, + "learning_rate": 8.309637513057723e-06, + "loss": 0.652, + "num_tokens": 22061025.0, + "step": 3003 + }, + { + "epoch": 9.07561436672968, + "grad_norm": 1.5078125, + "learning_rate": 8.306195976008723e-06, + "loss": 0.7706, + "num_tokens": 22084317.0, + "step": 3004 + }, + { + "epoch": 9.078638941398866, + "grad_norm": 2.015625, + "learning_rate": 8.302751751093539e-06, + "loss": 0.5992, + "num_tokens": 22092099.0, + "step": 3005 + }, + { + "epoch": 9.081663516068053, + "grad_norm": 2.046875, + "learning_rate": 8.299304841611162e-06, + "loss": 0.6167, + "num_tokens": 22099685.0, + "step": 3006 + }, + { + "epoch": 9.08468809073724, + "grad_norm": 1.59375, + "learning_rate": 8.295855250863156e-06, + "loss": 0.7755, + "num_tokens": 22120386.0, + "step": 3007 + }, + { + "epoch": 9.087712665406427, + "grad_norm": 1.265625, + "learning_rate": 8.29240298215365e-06, + "loss": 0.8534, + "num_tokens": 22160100.0, + "step": 3008 + }, + { + "epoch": 9.090737240075615, + "grad_norm": 1.6953125, + "learning_rate": 8.288948038789341e-06, + "loss": 0.689, + "num_tokens": 22173835.0, + "step": 3009 + }, + { + "epoch": 9.093761814744802, + "grad_norm": 1.3359375, + "learning_rate": 8.285490424079486e-06, + "loss": 0.7392, + "num_tokens": 22198276.0, + "step": 3010 + }, + { + "epoch": 9.096786389413989, + "grad_norm": 1.7890625, + "learning_rate": 8.282030141335899e-06, + "loss": 0.5838, + "num_tokens": 22207678.0, + "step": 3011 + }, + { + "epoch": 9.099810964083176, + "grad_norm": 1.5703125, + "learning_rate": 8.278567193872955e-06, + "loss": 0.7746, + "num_tokens": 22232052.0, + "step": 3012 + }, + { + "epoch": 9.102835538752363, + "grad_norm": 1.2734375, + "learning_rate": 8.275101585007578e-06, + "loss": 0.8395, + "num_tokens": 22266108.0, + "step": 3013 + }, + { + "epoch": 9.10586011342155, + "grad_norm": 1.15625, + "learning_rate": 8.271633318059241e-06, + "loss": 0.8629, + "num_tokens": 22311049.0, + "step": 3014 + }, + { + "epoch": 9.108884688090738, + "grad_norm": 1.8203125, + "learning_rate": 8.268162396349963e-06, + "loss": 0.5689, + "num_tokens": 22318648.0, + "step": 3015 + }, + { + "epoch": 9.111909262759925, + "grad_norm": 2.09375, + "learning_rate": 8.264688823204307e-06, + "loss": 0.6225, + "num_tokens": 22326179.0, + "step": 3016 + }, + { + "epoch": 9.114933837429112, + "grad_norm": 1.390625, + "learning_rate": 8.261212601949374e-06, + "loss": 0.7662, + "num_tokens": 22350606.0, + "step": 3017 + }, + { + "epoch": 9.117958412098298, + "grad_norm": 1.34375, + "learning_rate": 8.2577337359148e-06, + "loss": 0.7442, + "num_tokens": 22373431.0, + "step": 3018 + }, + { + "epoch": 9.120982986767485, + "grad_norm": 1.421875, + "learning_rate": 8.254252228432762e-06, + "loss": 0.7196, + "num_tokens": 22396835.0, + "step": 3019 + }, + { + "epoch": 9.124007561436674, + "grad_norm": 1.7265625, + "learning_rate": 8.250768082837957e-06, + "loss": 0.6476, + "num_tokens": 22409064.0, + "step": 3020 + }, + { + "epoch": 9.12703213610586, + "grad_norm": 1.765625, + "learning_rate": 8.247281302467616e-06, + "loss": 0.5522, + "num_tokens": 22417541.0, + "step": 3021 + }, + { + "epoch": 9.130056710775047, + "grad_norm": 1.5078125, + "learning_rate": 8.24379189066149e-06, + "loss": 0.7195, + "num_tokens": 22439806.0, + "step": 3022 + }, + { + "epoch": 9.133081285444234, + "grad_norm": 1.4296875, + "learning_rate": 8.240299850761851e-06, + "loss": 0.8764, + "num_tokens": 22470501.0, + "step": 3023 + }, + { + "epoch": 9.136105860113421, + "grad_norm": 1.703125, + "learning_rate": 8.236805186113492e-06, + "loss": 0.7232, + "num_tokens": 22485303.0, + "step": 3024 + }, + { + "epoch": 9.139130434782608, + "grad_norm": 2.0625, + "learning_rate": 8.233307900063714e-06, + "loss": 0.5792, + "num_tokens": 22492462.0, + "step": 3025 + }, + { + "epoch": 9.142155009451796, + "grad_norm": 1.5546875, + "learning_rate": 8.229807995962332e-06, + "loss": 0.8052, + "num_tokens": 22515891.0, + "step": 3026 + }, + { + "epoch": 9.145179584120983, + "grad_norm": 1.484375, + "learning_rate": 8.226305477161669e-06, + "loss": 0.654, + "num_tokens": 22539301.0, + "step": 3027 + }, + { + "epoch": 9.14820415879017, + "grad_norm": 1.171875, + "learning_rate": 8.222800347016553e-06, + "loss": 0.7879, + "num_tokens": 22578820.0, + "step": 3028 + }, + { + "epoch": 9.151228733459357, + "grad_norm": 1.34375, + "learning_rate": 8.219292608884309e-06, + "loss": 0.7646, + "num_tokens": 22609277.0, + "step": 3029 + }, + { + "epoch": 9.154253308128544, + "grad_norm": 1.2890625, + "learning_rate": 8.215782266124766e-06, + "loss": 0.9064, + "num_tokens": 22644556.0, + "step": 3030 + }, + { + "epoch": 9.157277882797732, + "grad_norm": 1.5234375, + "learning_rate": 8.212269322100244e-06, + "loss": 0.8331, + "num_tokens": 22670470.0, + "step": 3031 + }, + { + "epoch": 9.16030245746692, + "grad_norm": 1.5625, + "learning_rate": 8.208753780175557e-06, + "loss": 0.8964, + "num_tokens": 22694369.0, + "step": 3032 + }, + { + "epoch": 9.163327032136106, + "grad_norm": 1.265625, + "learning_rate": 8.205235643718004e-06, + "loss": 0.8562, + "num_tokens": 22731773.0, + "step": 3033 + }, + { + "epoch": 9.166351606805293, + "grad_norm": 1.46875, + "learning_rate": 8.201714916097372e-06, + "loss": 0.7519, + "num_tokens": 22755414.0, + "step": 3034 + }, + { + "epoch": 9.16937618147448, + "grad_norm": 1.1015625, + "learning_rate": 8.198191600685931e-06, + "loss": 0.784, + "num_tokens": 22795112.0, + "step": 3035 + }, + { + "epoch": 9.172400756143666, + "grad_norm": 2.09375, + "learning_rate": 8.194665700858422e-06, + "loss": 0.5512, + "num_tokens": 22802018.0, + "step": 3036 + }, + { + "epoch": 9.175425330812855, + "grad_norm": 1.09375, + "learning_rate": 8.191137219992073e-06, + "loss": 0.8245, + "num_tokens": 22842161.0, + "step": 3037 + }, + { + "epoch": 9.178449905482042, + "grad_norm": 1.5546875, + "learning_rate": 8.187606161466575e-06, + "loss": 0.814, + "num_tokens": 22865630.0, + "step": 3038 + }, + { + "epoch": 9.181474480151229, + "grad_norm": 1.9296875, + "learning_rate": 8.184072528664091e-06, + "loss": 0.597, + "num_tokens": 22873528.0, + "step": 3039 + }, + { + "epoch": 9.184499054820416, + "grad_norm": 1.9921875, + "learning_rate": 8.180536324969251e-06, + "loss": 0.6344, + "num_tokens": 22881564.0, + "step": 3040 + }, + { + "epoch": 9.187523629489602, + "grad_norm": 1.46875, + "learning_rate": 8.176997553769146e-06, + "loss": 0.8043, + "num_tokens": 22905503.0, + "step": 3041 + }, + { + "epoch": 9.190548204158791, + "grad_norm": 1.4609375, + "learning_rate": 8.17345621845333e-06, + "loss": 0.7538, + "num_tokens": 22929122.0, + "step": 3042 + }, + { + "epoch": 9.193572778827978, + "grad_norm": 1.0703125, + "learning_rate": 8.169912322413803e-06, + "loss": 0.8213, + "num_tokens": 22983150.0, + "step": 3043 + }, + { + "epoch": 9.196597353497165, + "grad_norm": 1.4296875, + "learning_rate": 8.166365869045029e-06, + "loss": 0.7838, + "num_tokens": 23006309.0, + "step": 3044 + }, + { + "epoch": 9.199621928166351, + "grad_norm": 1.1328125, + "learning_rate": 8.162816861743917e-06, + "loss": 0.8161, + "num_tokens": 23054885.0, + "step": 3045 + }, + { + "epoch": 9.202646502835538, + "grad_norm": 1.578125, + "learning_rate": 8.159265303909818e-06, + "loss": 0.5955, + "num_tokens": 23068516.0, + "step": 3046 + }, + { + "epoch": 9.205671077504725, + "grad_norm": 1.3671875, + "learning_rate": 8.155711198944536e-06, + "loss": 0.8324, + "num_tokens": 23100858.0, + "step": 3047 + }, + { + "epoch": 9.208695652173914, + "grad_norm": 1.2109375, + "learning_rate": 8.152154550252305e-06, + "loss": 0.8574, + "num_tokens": 23139859.0, + "step": 3048 + }, + { + "epoch": 9.2117202268431, + "grad_norm": 1.375, + "learning_rate": 8.1485953612398e-06, + "loss": 0.7949, + "num_tokens": 23167792.0, + "step": 3049 + }, + { + "epoch": 9.214744801512287, + "grad_norm": 1.9140625, + "learning_rate": 8.14503363531613e-06, + "loss": 0.769, + "num_tokens": 23179197.0, + "step": 3050 + }, + { + "epoch": 9.217769376181474, + "grad_norm": 1.03125, + "learning_rate": 8.14146937589283e-06, + "loss": 0.8204, + "num_tokens": 23232272.0, + "step": 3051 + }, + { + "epoch": 9.220793950850661, + "grad_norm": 1.65625, + "learning_rate": 8.137902586383866e-06, + "loss": 0.6889, + "num_tokens": 23248560.0, + "step": 3052 + }, + { + "epoch": 9.22381852551985, + "grad_norm": 1.7421875, + "learning_rate": 8.134333270205624e-06, + "loss": 0.6627, + "num_tokens": 23262540.0, + "step": 3053 + }, + { + "epoch": 9.226843100189036, + "grad_norm": 1.875, + "learning_rate": 8.130761430776915e-06, + "loss": 0.5437, + "num_tokens": 23270859.0, + "step": 3054 + }, + { + "epoch": 9.229867674858223, + "grad_norm": 1.1875, + "learning_rate": 8.12718707151896e-06, + "loss": 0.8089, + "num_tokens": 23310033.0, + "step": 3055 + }, + { + "epoch": 9.23289224952741, + "grad_norm": 1.65625, + "learning_rate": 8.123610195855395e-06, + "loss": 0.7595, + "num_tokens": 23325546.0, + "step": 3056 + }, + { + "epoch": 9.235916824196597, + "grad_norm": 1.984375, + "learning_rate": 8.120030807212276e-06, + "loss": 0.5814, + "num_tokens": 23333248.0, + "step": 3057 + }, + { + "epoch": 9.238941398865785, + "grad_norm": 1.796875, + "learning_rate": 8.116448909018051e-06, + "loss": 0.5988, + "num_tokens": 23345601.0, + "step": 3058 + }, + { + "epoch": 9.241965973534972, + "grad_norm": 1.8125, + "learning_rate": 8.112864504703582e-06, + "loss": 0.5351, + "num_tokens": 23353780.0, + "step": 3059 + }, + { + "epoch": 9.244990548204159, + "grad_norm": 1.546875, + "learning_rate": 8.10927759770213e-06, + "loss": 0.7707, + "num_tokens": 23376281.0, + "step": 3060 + }, + { + "epoch": 9.248015122873346, + "grad_norm": 1.78125, + "learning_rate": 8.10568819144935e-06, + "loss": 0.6858, + "num_tokens": 23389620.0, + "step": 3061 + }, + { + "epoch": 9.251039697542533, + "grad_norm": 1.5078125, + "learning_rate": 8.102096289383293e-06, + "loss": 0.7436, + "num_tokens": 23412652.0, + "step": 3062 + }, + { + "epoch": 9.25406427221172, + "grad_norm": 1.8671875, + "learning_rate": 8.0985018949444e-06, + "loss": 0.6167, + "num_tokens": 23421116.0, + "step": 3063 + }, + { + "epoch": 9.257088846880908, + "grad_norm": 1.4609375, + "learning_rate": 8.094905011575496e-06, + "loss": 0.7699, + "num_tokens": 23444597.0, + "step": 3064 + }, + { + "epoch": 9.260113421550095, + "grad_norm": 1.953125, + "learning_rate": 8.0913056427218e-06, + "loss": 0.6091, + "num_tokens": 23451986.0, + "step": 3065 + }, + { + "epoch": 9.263137996219282, + "grad_norm": 1.9453125, + "learning_rate": 8.087703791830902e-06, + "loss": 0.5383, + "num_tokens": 23459668.0, + "step": 3066 + }, + { + "epoch": 9.266162570888469, + "grad_norm": 1.390625, + "learning_rate": 8.084099462352772e-06, + "loss": 0.7525, + "num_tokens": 23481381.0, + "step": 3067 + }, + { + "epoch": 9.269187145557655, + "grad_norm": 1.15625, + "learning_rate": 8.080492657739754e-06, + "loss": 0.8454, + "num_tokens": 23524179.0, + "step": 3068 + }, + { + "epoch": 9.272211720226844, + "grad_norm": 1.9765625, + "learning_rate": 8.076883381446564e-06, + "loss": 0.5929, + "num_tokens": 23531645.0, + "step": 3069 + }, + { + "epoch": 9.27523629489603, + "grad_norm": 1.796875, + "learning_rate": 8.073271636930287e-06, + "loss": 0.7319, + "num_tokens": 23543922.0, + "step": 3070 + }, + { + "epoch": 9.278260869565218, + "grad_norm": 1.484375, + "learning_rate": 8.069657427650364e-06, + "loss": 0.7728, + "num_tokens": 23566940.0, + "step": 3071 + }, + { + "epoch": 9.281285444234404, + "grad_norm": 1.6171875, + "learning_rate": 8.06604075706861e-06, + "loss": 0.6479, + "num_tokens": 23585403.0, + "step": 3072 + }, + { + "epoch": 9.284310018903591, + "grad_norm": 1.8359375, + "learning_rate": 8.062421628649187e-06, + "loss": 0.6481, + "num_tokens": 23593715.0, + "step": 3073 + }, + { + "epoch": 9.287334593572778, + "grad_norm": 1.4921875, + "learning_rate": 8.058800045858613e-06, + "loss": 0.8263, + "num_tokens": 23624745.0, + "step": 3074 + }, + { + "epoch": 9.290359168241967, + "grad_norm": 1.75, + "learning_rate": 8.055176012165758e-06, + "loss": 0.6591, + "num_tokens": 23637811.0, + "step": 3075 + }, + { + "epoch": 9.293383742911153, + "grad_norm": 1.8671875, + "learning_rate": 8.051549531041843e-06, + "loss": 0.6032, + "num_tokens": 23645818.0, + "step": 3076 + }, + { + "epoch": 9.29640831758034, + "grad_norm": 1.9375, + "learning_rate": 8.047920605960428e-06, + "loss": 0.5322, + "num_tokens": 23652922.0, + "step": 3077 + }, + { + "epoch": 9.299432892249527, + "grad_norm": 1.40625, + "learning_rate": 8.044289240397415e-06, + "loss": 0.6968, + "num_tokens": 23677123.0, + "step": 3078 + }, + { + "epoch": 9.302457466918714, + "grad_norm": 0.87109375, + "learning_rate": 8.040655437831047e-06, + "loss": 0.8761, + "num_tokens": 23756853.0, + "step": 3079 + }, + { + "epoch": 9.305482041587902, + "grad_norm": 1.390625, + "learning_rate": 8.037019201741901e-06, + "loss": 0.7789, + "num_tokens": 23783289.0, + "step": 3080 + }, + { + "epoch": 9.30850661625709, + "grad_norm": 1.3046875, + "learning_rate": 8.033380535612875e-06, + "loss": 0.9072, + "num_tokens": 23822464.0, + "step": 3081 + }, + { + "epoch": 9.311531190926276, + "grad_norm": 1.5625, + "learning_rate": 8.02973944292921e-06, + "loss": 0.805, + "num_tokens": 23844734.0, + "step": 3082 + }, + { + "epoch": 9.314555765595463, + "grad_norm": 1.390625, + "learning_rate": 8.026095927178458e-06, + "loss": 0.7082, + "num_tokens": 23867435.0, + "step": 3083 + }, + { + "epoch": 9.31758034026465, + "grad_norm": 1.0625, + "learning_rate": 8.022449991850501e-06, + "loss": 0.8484, + "num_tokens": 23922175.0, + "step": 3084 + }, + { + "epoch": 9.320604914933837, + "grad_norm": 1.1875, + "learning_rate": 8.018801640437535e-06, + "loss": 0.8156, + "num_tokens": 23954889.0, + "step": 3085 + }, + { + "epoch": 9.323629489603025, + "grad_norm": 1.3203125, + "learning_rate": 8.015150876434066e-06, + "loss": 0.8445, + "num_tokens": 23994301.0, + "step": 3086 + }, + { + "epoch": 9.326654064272212, + "grad_norm": 1.84375, + "learning_rate": 8.011497703336921e-06, + "loss": 0.5663, + "num_tokens": 24002097.0, + "step": 3087 + }, + { + "epoch": 9.329678638941399, + "grad_norm": 1.25, + "learning_rate": 8.007842124645225e-06, + "loss": 0.7693, + "num_tokens": 24038996.0, + "step": 3088 + }, + { + "epoch": 9.332703213610586, + "grad_norm": 1.453125, + "learning_rate": 8.004184143860408e-06, + "loss": 0.8549, + "num_tokens": 24063420.0, + "step": 3089 + }, + { + "epoch": 9.335727788279772, + "grad_norm": 1.4609375, + "learning_rate": 8.00052376448621e-06, + "loss": 0.7741, + "num_tokens": 24091120.0, + "step": 3090 + }, + { + "epoch": 9.338752362948961, + "grad_norm": 1.2265625, + "learning_rate": 7.996860990028658e-06, + "loss": 0.751, + "num_tokens": 24131890.0, + "step": 3091 + }, + { + "epoch": 9.341776937618148, + "grad_norm": 1.1953125, + "learning_rate": 7.993195823996077e-06, + "loss": 0.7836, + "num_tokens": 24171229.0, + "step": 3092 + }, + { + "epoch": 9.344801512287335, + "grad_norm": 1.875, + "learning_rate": 7.989528269899085e-06, + "loss": 0.5293, + "num_tokens": 24178518.0, + "step": 3093 + }, + { + "epoch": 9.347826086956522, + "grad_norm": 1.0546875, + "learning_rate": 7.98585833125058e-06, + "loss": 0.8882, + "num_tokens": 24234049.0, + "step": 3094 + }, + { + "epoch": 9.350850661625708, + "grad_norm": 1.2421875, + "learning_rate": 7.982186011565755e-06, + "loss": 0.8168, + "num_tokens": 24272042.0, + "step": 3095 + }, + { + "epoch": 9.353875236294897, + "grad_norm": 1.3828125, + "learning_rate": 7.978511314362077e-06, + "loss": 0.7277, + "num_tokens": 24295483.0, + "step": 3096 + }, + { + "epoch": 9.356899810964084, + "grad_norm": 2.046875, + "learning_rate": 7.974834243159288e-06, + "loss": 0.6501, + "num_tokens": 24303171.0, + "step": 3097 + }, + { + "epoch": 9.35992438563327, + "grad_norm": 1.8515625, + "learning_rate": 7.971154801479409e-06, + "loss": 0.6753, + "num_tokens": 24313102.0, + "step": 3098 + }, + { + "epoch": 9.362948960302457, + "grad_norm": 1.2578125, + "learning_rate": 7.96747299284673e-06, + "loss": 0.8528, + "num_tokens": 24352124.0, + "step": 3099 + }, + { + "epoch": 9.365973534971644, + "grad_norm": 1.5234375, + "learning_rate": 7.963788820787805e-06, + "loss": 0.7011, + "num_tokens": 24375032.0, + "step": 3100 + }, + { + "epoch": 9.368998109640831, + "grad_norm": 1.734375, + "learning_rate": 7.960102288831454e-06, + "loss": 0.6492, + "num_tokens": 24384686.0, + "step": 3101 + }, + { + "epoch": 9.37202268431002, + "grad_norm": 1.640625, + "learning_rate": 7.956413400508763e-06, + "loss": 0.8622, + "num_tokens": 24407985.0, + "step": 3102 + }, + { + "epoch": 9.375047258979206, + "grad_norm": 2.015625, + "learning_rate": 7.952722159353065e-06, + "loss": 0.6184, + "num_tokens": 24417963.0, + "step": 3103 + }, + { + "epoch": 9.378071833648393, + "grad_norm": 1.3984375, + "learning_rate": 7.949028568899951e-06, + "loss": 0.6385, + "num_tokens": 24438079.0, + "step": 3104 + }, + { + "epoch": 9.38109640831758, + "grad_norm": 1.390625, + "learning_rate": 7.945332632687268e-06, + "loss": 0.6919, + "num_tokens": 24458929.0, + "step": 3105 + }, + { + "epoch": 9.384120982986767, + "grad_norm": 1.1640625, + "learning_rate": 7.941634354255098e-06, + "loss": 0.7877, + "num_tokens": 24495113.0, + "step": 3106 + }, + { + "epoch": 9.387145557655955, + "grad_norm": 1.40625, + "learning_rate": 7.937933737145777e-06, + "loss": 0.8228, + "num_tokens": 24522239.0, + "step": 3107 + }, + { + "epoch": 9.390170132325142, + "grad_norm": 1.3515625, + "learning_rate": 7.934230784903875e-06, + "loss": 0.6899, + "num_tokens": 24543994.0, + "step": 3108 + }, + { + "epoch": 9.39319470699433, + "grad_norm": 1.171875, + "learning_rate": 7.9305255010762e-06, + "loss": 0.764, + "num_tokens": 24581159.0, + "step": 3109 + }, + { + "epoch": 9.396219281663516, + "grad_norm": 1.8828125, + "learning_rate": 7.926817889211798e-06, + "loss": 0.5934, + "num_tokens": 24591590.0, + "step": 3110 + }, + { + "epoch": 9.399243856332703, + "grad_norm": 1.5859375, + "learning_rate": 7.923107952861938e-06, + "loss": 0.8304, + "num_tokens": 24614423.0, + "step": 3111 + }, + { + "epoch": 9.40226843100189, + "grad_norm": 1.4375, + "learning_rate": 7.919395695580114e-06, + "loss": 0.7657, + "num_tokens": 24637787.0, + "step": 3112 + }, + { + "epoch": 9.405293005671078, + "grad_norm": 1.8671875, + "learning_rate": 7.915681120922055e-06, + "loss": 0.5502, + "num_tokens": 24645543.0, + "step": 3113 + }, + { + "epoch": 9.408317580340265, + "grad_norm": 1.9765625, + "learning_rate": 7.911964232445698e-06, + "loss": 0.5917, + "num_tokens": 24652935.0, + "step": 3114 + }, + { + "epoch": 9.411342155009452, + "grad_norm": 1.28125, + "learning_rate": 7.908245033711201e-06, + "loss": 0.7318, + "num_tokens": 24676922.0, + "step": 3115 + }, + { + "epoch": 9.414366729678639, + "grad_norm": 1.7109375, + "learning_rate": 7.904523528280932e-06, + "loss": 0.5307, + "num_tokens": 24685009.0, + "step": 3116 + }, + { + "epoch": 9.417391304347825, + "grad_norm": 1.796875, + "learning_rate": 7.900799719719473e-06, + "loss": 0.4968, + "num_tokens": 24692869.0, + "step": 3117 + }, + { + "epoch": 9.420415879017014, + "grad_norm": 1.6875, + "learning_rate": 7.89707361159361e-06, + "loss": 0.7193, + "num_tokens": 24707007.0, + "step": 3118 + }, + { + "epoch": 9.4234404536862, + "grad_norm": 2.09375, + "learning_rate": 7.893345207472329e-06, + "loss": 0.5928, + "num_tokens": 24714537.0, + "step": 3119 + }, + { + "epoch": 9.426465028355388, + "grad_norm": 1.546875, + "learning_rate": 7.88961451092682e-06, + "loss": 0.8425, + "num_tokens": 24737171.0, + "step": 3120 + }, + { + "epoch": 9.429489603024575, + "grad_norm": 1.328125, + "learning_rate": 7.88588152553046e-06, + "loss": 0.7559, + "num_tokens": 24761645.0, + "step": 3121 + }, + { + "epoch": 9.432514177693761, + "grad_norm": 1.359375, + "learning_rate": 7.882146254858832e-06, + "loss": 0.7667, + "num_tokens": 24789472.0, + "step": 3122 + }, + { + "epoch": 9.435538752362948, + "grad_norm": 2.0625, + "learning_rate": 7.878408702489695e-06, + "loss": 0.5902, + "num_tokens": 24796266.0, + "step": 3123 + }, + { + "epoch": 9.438563327032137, + "grad_norm": 0.9453125, + "learning_rate": 7.874668872003e-06, + "loss": 0.8411, + "num_tokens": 24856498.0, + "step": 3124 + }, + { + "epoch": 9.441587901701324, + "grad_norm": 1.140625, + "learning_rate": 7.870926766980879e-06, + "loss": 0.8294, + "num_tokens": 24895769.0, + "step": 3125 + }, + { + "epoch": 9.44461247637051, + "grad_norm": 1.546875, + "learning_rate": 7.867182391007644e-06, + "loss": 0.8171, + "num_tokens": 24922021.0, + "step": 3126 + }, + { + "epoch": 9.447637051039697, + "grad_norm": 1.2890625, + "learning_rate": 7.863435747669781e-06, + "loss": 0.8891, + "num_tokens": 24962323.0, + "step": 3127 + }, + { + "epoch": 9.450661625708884, + "grad_norm": 1.421875, + "learning_rate": 7.859686840555944e-06, + "loss": 0.7543, + "num_tokens": 24987156.0, + "step": 3128 + }, + { + "epoch": 9.453686200378073, + "grad_norm": 1.5625, + "learning_rate": 7.85593567325696e-06, + "loss": 0.7673, + "num_tokens": 25011088.0, + "step": 3129 + }, + { + "epoch": 9.45671077504726, + "grad_norm": 1.6015625, + "learning_rate": 7.85218224936582e-06, + "loss": 0.8004, + "num_tokens": 25033704.0, + "step": 3130 + }, + { + "epoch": 9.459735349716446, + "grad_norm": 1.3203125, + "learning_rate": 7.848426572477677e-06, + "loss": 0.8117, + "num_tokens": 25072481.0, + "step": 3131 + }, + { + "epoch": 9.462759924385633, + "grad_norm": 1.5, + "learning_rate": 7.844668646189842e-06, + "loss": 0.6725, + "num_tokens": 25090857.0, + "step": 3132 + }, + { + "epoch": 9.46578449905482, + "grad_norm": 1.140625, + "learning_rate": 7.840908474101774e-06, + "loss": 0.7644, + "num_tokens": 25131339.0, + "step": 3133 + }, + { + "epoch": 9.468809073724007, + "grad_norm": 1.40625, + "learning_rate": 7.837146059815096e-06, + "loss": 0.6736, + "num_tokens": 25154174.0, + "step": 3134 + }, + { + "epoch": 9.471833648393195, + "grad_norm": 2.0625, + "learning_rate": 7.833381406933564e-06, + "loss": 0.6081, + "num_tokens": 25161332.0, + "step": 3135 + }, + { + "epoch": 9.474858223062382, + "grad_norm": 0.93359375, + "learning_rate": 7.829614519063093e-06, + "loss": 0.8907, + "num_tokens": 25230320.0, + "step": 3136 + }, + { + "epoch": 9.477882797731569, + "grad_norm": 1.40625, + "learning_rate": 7.825845399811723e-06, + "loss": 0.7339, + "num_tokens": 25250897.0, + "step": 3137 + }, + { + "epoch": 9.480907372400756, + "grad_norm": 2.171875, + "learning_rate": 7.822074052789648e-06, + "loss": 0.6618, + "num_tokens": 25257711.0, + "step": 3138 + }, + { + "epoch": 9.483931947069943, + "grad_norm": 0.953125, + "learning_rate": 7.818300481609184e-06, + "loss": 0.8091, + "num_tokens": 25326655.0, + "step": 3139 + }, + { + "epoch": 9.486956521739131, + "grad_norm": 1.578125, + "learning_rate": 7.814524689884776e-06, + "loss": 0.8007, + "num_tokens": 25344368.0, + "step": 3140 + }, + { + "epoch": 9.489981096408318, + "grad_norm": 1.390625, + "learning_rate": 7.810746681233008e-06, + "loss": 0.7758, + "num_tokens": 25367021.0, + "step": 3141 + }, + { + "epoch": 9.493005671077505, + "grad_norm": 1.875, + "learning_rate": 7.806966459272577e-06, + "loss": 0.5427, + "num_tokens": 25374467.0, + "step": 3142 + }, + { + "epoch": 9.496030245746692, + "grad_norm": 1.53125, + "learning_rate": 7.8031840276243e-06, + "loss": 0.7739, + "num_tokens": 25399696.0, + "step": 3143 + }, + { + "epoch": 9.499054820415878, + "grad_norm": 1.921875, + "learning_rate": 7.799399389911122e-06, + "loss": 0.5693, + "num_tokens": 25407424.0, + "step": 3144 + }, + { + "epoch": 9.502079395085065, + "grad_norm": 1.5, + "learning_rate": 7.795612549758083e-06, + "loss": 0.7009, + "num_tokens": 25428412.0, + "step": 3145 + }, + { + "epoch": 9.505103969754254, + "grad_norm": 2.21875, + "learning_rate": 7.791823510792348e-06, + "loss": 0.6237, + "num_tokens": 25435214.0, + "step": 3146 + }, + { + "epoch": 9.50812854442344, + "grad_norm": 1.53125, + "learning_rate": 7.788032276643181e-06, + "loss": 0.8012, + "num_tokens": 25458559.0, + "step": 3147 + }, + { + "epoch": 9.511153119092628, + "grad_norm": 1.21875, + "learning_rate": 7.784238850941948e-06, + "loss": 0.7905, + "num_tokens": 25498436.0, + "step": 3148 + }, + { + "epoch": 9.514177693761814, + "grad_norm": 1.6640625, + "learning_rate": 7.78044323732212e-06, + "loss": 0.7269, + "num_tokens": 25515180.0, + "step": 3149 + }, + { + "epoch": 9.517202268431001, + "grad_norm": 1.5, + "learning_rate": 7.776645439419259e-06, + "loss": 0.7412, + "num_tokens": 25536616.0, + "step": 3150 + }, + { + "epoch": 9.52022684310019, + "grad_norm": 1.6015625, + "learning_rate": 7.772845460871016e-06, + "loss": 0.7965, + "num_tokens": 25559544.0, + "step": 3151 + }, + { + "epoch": 9.523251417769377, + "grad_norm": 1.15625, + "learning_rate": 7.769043305317138e-06, + "loss": 0.854, + "num_tokens": 25605253.0, + "step": 3152 + }, + { + "epoch": 9.526275992438563, + "grad_norm": 2.046875, + "learning_rate": 7.765238976399451e-06, + "loss": 0.6069, + "num_tokens": 25612640.0, + "step": 3153 + }, + { + "epoch": 9.52930056710775, + "grad_norm": 1.5, + "learning_rate": 7.761432477761869e-06, + "loss": 0.818, + "num_tokens": 25635551.0, + "step": 3154 + }, + { + "epoch": 9.532325141776937, + "grad_norm": 1.609375, + "learning_rate": 7.75762381305038e-06, + "loss": 0.7126, + "num_tokens": 25652769.0, + "step": 3155 + }, + { + "epoch": 9.535349716446124, + "grad_norm": 1.5703125, + "learning_rate": 7.753812985913043e-06, + "loss": 0.6908, + "num_tokens": 25669782.0, + "step": 3156 + }, + { + "epoch": 9.538374291115312, + "grad_norm": 1.5078125, + "learning_rate": 7.75e-06, + "loss": 0.805, + "num_tokens": 25694291.0, + "step": 3157 + }, + { + "epoch": 9.5413988657845, + "grad_norm": 1.0703125, + "learning_rate": 7.74618485896345e-06, + "loss": 0.8601, + "num_tokens": 25748658.0, + "step": 3158 + }, + { + "epoch": 9.544423440453686, + "grad_norm": 1.84375, + "learning_rate": 7.742367566457656e-06, + "loss": 0.6289, + "num_tokens": 25758870.0, + "step": 3159 + }, + { + "epoch": 9.547448015122873, + "grad_norm": 2.015625, + "learning_rate": 7.738548126138954e-06, + "loss": 0.7155, + "num_tokens": 25768686.0, + "step": 3160 + }, + { + "epoch": 9.55047258979206, + "grad_norm": 1.3203125, + "learning_rate": 7.734726541665722e-06, + "loss": 0.7037, + "num_tokens": 25792974.0, + "step": 3161 + }, + { + "epoch": 9.553497164461248, + "grad_norm": 1.5625, + "learning_rate": 7.7309028166984e-06, + "loss": 0.7469, + "num_tokens": 25814686.0, + "step": 3162 + }, + { + "epoch": 9.556521739130435, + "grad_norm": 1.921875, + "learning_rate": 7.727076954899475e-06, + "loss": 0.5959, + "num_tokens": 25823302.0, + "step": 3163 + }, + { + "epoch": 9.559546313799622, + "grad_norm": 1.96875, + "learning_rate": 7.723248959933486e-06, + "loss": 0.583, + "num_tokens": 25831723.0, + "step": 3164 + }, + { + "epoch": 9.562570888468809, + "grad_norm": 2.171875, + "learning_rate": 7.719418835467009e-06, + "loss": 0.6553, + "num_tokens": 25838608.0, + "step": 3165 + }, + { + "epoch": 9.565595463137996, + "grad_norm": 1.484375, + "learning_rate": 7.715586585168666e-06, + "loss": 0.7573, + "num_tokens": 25859361.0, + "step": 3166 + }, + { + "epoch": 9.568620037807184, + "grad_norm": 1.515625, + "learning_rate": 7.711752212709106e-06, + "loss": 0.7991, + "num_tokens": 25885881.0, + "step": 3167 + }, + { + "epoch": 9.571644612476371, + "grad_norm": 1.1875, + "learning_rate": 7.70791572176102e-06, + "loss": 0.7988, + "num_tokens": 25930873.0, + "step": 3168 + }, + { + "epoch": 9.574669187145558, + "grad_norm": 1.6875, + "learning_rate": 7.704077115999124e-06, + "loss": 0.6932, + "num_tokens": 25945838.0, + "step": 3169 + }, + { + "epoch": 9.577693761814745, + "grad_norm": 1.234375, + "learning_rate": 7.700236399100159e-06, + "loss": 0.7831, + "num_tokens": 25982687.0, + "step": 3170 + }, + { + "epoch": 9.580718336483931, + "grad_norm": 1.578125, + "learning_rate": 7.696393574742889e-06, + "loss": 0.6648, + "num_tokens": 26001888.0, + "step": 3171 + }, + { + "epoch": 9.583742911153118, + "grad_norm": 1.328125, + "learning_rate": 7.692548646608099e-06, + "loss": 0.7078, + "num_tokens": 26027335.0, + "step": 3172 + }, + { + "epoch": 9.586767485822307, + "grad_norm": 2.046875, + "learning_rate": 7.688701618378583e-06, + "loss": 0.6235, + "num_tokens": 26034466.0, + "step": 3173 + }, + { + "epoch": 9.589792060491494, + "grad_norm": 1.8984375, + "learning_rate": 7.684852493739153e-06, + "loss": 0.6003, + "num_tokens": 26042395.0, + "step": 3174 + }, + { + "epoch": 9.59281663516068, + "grad_norm": 1.015625, + "learning_rate": 7.681001276376625e-06, + "loss": 0.8528, + "num_tokens": 26101353.0, + "step": 3175 + }, + { + "epoch": 9.595841209829867, + "grad_norm": 1.5625, + "learning_rate": 7.677147969979824e-06, + "loss": 0.7641, + "num_tokens": 26119705.0, + "step": 3176 + }, + { + "epoch": 9.598865784499054, + "grad_norm": 1.1796875, + "learning_rate": 7.673292578239569e-06, + "loss": 0.8344, + "num_tokens": 26156679.0, + "step": 3177 + }, + { + "epoch": 9.601890359168243, + "grad_norm": 1.75, + "learning_rate": 7.66943510484868e-06, + "loss": 0.7756, + "num_tokens": 26172839.0, + "step": 3178 + }, + { + "epoch": 9.60491493383743, + "grad_norm": 1.21875, + "learning_rate": 7.665575553501973e-06, + "loss": 0.8114, + "num_tokens": 26211621.0, + "step": 3179 + }, + { + "epoch": 9.607939508506616, + "grad_norm": 1.7109375, + "learning_rate": 7.661713927896254e-06, + "loss": 0.7466, + "num_tokens": 26226005.0, + "step": 3180 + }, + { + "epoch": 9.610964083175803, + "grad_norm": 2.09375, + "learning_rate": 7.65785023173031e-06, + "loss": 0.5573, + "num_tokens": 26232391.0, + "step": 3181 + }, + { + "epoch": 9.61398865784499, + "grad_norm": 1.53125, + "learning_rate": 7.653984468704918e-06, + "loss": 0.8783, + "num_tokens": 26255529.0, + "step": 3182 + }, + { + "epoch": 9.617013232514179, + "grad_norm": 1.7734375, + "learning_rate": 7.650116642522831e-06, + "loss": 0.5253, + "num_tokens": 26264348.0, + "step": 3183 + }, + { + "epoch": 9.620037807183365, + "grad_norm": 1.0390625, + "learning_rate": 7.646246756888781e-06, + "loss": 0.817, + "num_tokens": 26322802.0, + "step": 3184 + }, + { + "epoch": 9.623062381852552, + "grad_norm": 1.8203125, + "learning_rate": 7.64237481550947e-06, + "loss": 0.6451, + "num_tokens": 26333600.0, + "step": 3185 + }, + { + "epoch": 9.626086956521739, + "grad_norm": 1.6484375, + "learning_rate": 7.638500822093566e-06, + "loss": 0.6941, + "num_tokens": 26347961.0, + "step": 3186 + }, + { + "epoch": 9.629111531190926, + "grad_norm": 1.515625, + "learning_rate": 7.634624780351715e-06, + "loss": 0.8382, + "num_tokens": 26371717.0, + "step": 3187 + }, + { + "epoch": 9.632136105860113, + "grad_norm": 1.890625, + "learning_rate": 7.63074669399651e-06, + "loss": 0.5084, + "num_tokens": 26378823.0, + "step": 3188 + }, + { + "epoch": 9.635160680529301, + "grad_norm": 1.71875, + "learning_rate": 7.626866566742508e-06, + "loss": 0.679, + "num_tokens": 26391159.0, + "step": 3189 + }, + { + "epoch": 9.638185255198488, + "grad_norm": 1.4140625, + "learning_rate": 7.622984402306226e-06, + "loss": 0.7174, + "num_tokens": 26411235.0, + "step": 3190 + }, + { + "epoch": 9.641209829867675, + "grad_norm": 1.8359375, + "learning_rate": 7.619100204406127e-06, + "loss": 0.6045, + "num_tokens": 26419256.0, + "step": 3191 + }, + { + "epoch": 9.644234404536862, + "grad_norm": 1.0234375, + "learning_rate": 7.615213976762619e-06, + "loss": 0.8797, + "num_tokens": 26477709.0, + "step": 3192 + }, + { + "epoch": 9.647258979206049, + "grad_norm": 1.40625, + "learning_rate": 7.6113257230980615e-06, + "loss": 0.7457, + "num_tokens": 26505016.0, + "step": 3193 + }, + { + "epoch": 9.650283553875237, + "grad_norm": 1.6875, + "learning_rate": 7.60743544713675e-06, + "loss": 0.7423, + "num_tokens": 26520172.0, + "step": 3194 + }, + { + "epoch": 9.653308128544424, + "grad_norm": 1.5078125, + "learning_rate": 7.603543152604918e-06, + "loss": 0.8187, + "num_tokens": 26544005.0, + "step": 3195 + }, + { + "epoch": 9.65633270321361, + "grad_norm": 1.75, + "learning_rate": 7.59964884323073e-06, + "loss": 0.5817, + "num_tokens": 26554808.0, + "step": 3196 + }, + { + "epoch": 9.659357277882798, + "grad_norm": 1.6171875, + "learning_rate": 7.595752522744287e-06, + "loss": 0.8701, + "num_tokens": 26578093.0, + "step": 3197 + }, + { + "epoch": 9.662381852551984, + "grad_norm": 1.1328125, + "learning_rate": 7.591854194877609e-06, + "loss": 0.8817, + "num_tokens": 26618096.0, + "step": 3198 + }, + { + "epoch": 9.665406427221171, + "grad_norm": 1.4453125, + "learning_rate": 7.587953863364642e-06, + "loss": 0.7861, + "num_tokens": 26643618.0, + "step": 3199 + }, + { + "epoch": 9.66843100189036, + "grad_norm": 2.015625, + "learning_rate": 7.584051531941252e-06, + "loss": 0.6471, + "num_tokens": 26651661.0, + "step": 3200 + }, + { + "epoch": 9.671455576559547, + "grad_norm": 1.4453125, + "learning_rate": 7.5801472043452195e-06, + "loss": 0.8134, + "num_tokens": 26681850.0, + "step": 3201 + }, + { + "epoch": 9.674480151228734, + "grad_norm": 2.171875, + "learning_rate": 7.5762408843162364e-06, + "loss": 0.6341, + "num_tokens": 26688375.0, + "step": 3202 + }, + { + "epoch": 9.67750472589792, + "grad_norm": 1.5625, + "learning_rate": 7.572332575595904e-06, + "loss": 0.8787, + "num_tokens": 26713395.0, + "step": 3203 + }, + { + "epoch": 9.680529300567107, + "grad_norm": 1.296875, + "learning_rate": 7.5684222819277265e-06, + "loss": 0.8254, + "num_tokens": 26745458.0, + "step": 3204 + }, + { + "epoch": 9.683553875236296, + "grad_norm": 1.5078125, + "learning_rate": 7.564510007057114e-06, + "loss": 0.7518, + "num_tokens": 26768730.0, + "step": 3205 + }, + { + "epoch": 9.686578449905483, + "grad_norm": 1.9140625, + "learning_rate": 7.5605957547313695e-06, + "loss": 0.5475, + "num_tokens": 26776638.0, + "step": 3206 + }, + { + "epoch": 9.68960302457467, + "grad_norm": 1.65625, + "learning_rate": 7.556679528699692e-06, + "loss": 0.7794, + "num_tokens": 26801204.0, + "step": 3207 + }, + { + "epoch": 9.692627599243856, + "grad_norm": 1.5234375, + "learning_rate": 7.5527613327131685e-06, + "loss": 0.803, + "num_tokens": 26824297.0, + "step": 3208 + }, + { + "epoch": 9.695652173913043, + "grad_norm": 1.359375, + "learning_rate": 7.548841170524779e-06, + "loss": 0.7818, + "num_tokens": 26854465.0, + "step": 3209 + }, + { + "epoch": 9.69867674858223, + "grad_norm": 1.5390625, + "learning_rate": 7.544919045889382e-06, + "loss": 0.7228, + "num_tokens": 26874904.0, + "step": 3210 + }, + { + "epoch": 9.701701323251418, + "grad_norm": 1.625, + "learning_rate": 7.540994962563714e-06, + "loss": 0.8178, + "num_tokens": 26898108.0, + "step": 3211 + }, + { + "epoch": 9.704725897920605, + "grad_norm": 1.625, + "learning_rate": 7.537068924306392e-06, + "loss": 0.7865, + "num_tokens": 26918555.0, + "step": 3212 + }, + { + "epoch": 9.707750472589792, + "grad_norm": 1.421875, + "learning_rate": 7.5331409348779035e-06, + "loss": 0.7385, + "num_tokens": 26940689.0, + "step": 3213 + }, + { + "epoch": 9.710775047258979, + "grad_norm": 1.9375, + "learning_rate": 7.529210998040604e-06, + "loss": 0.5197, + "num_tokens": 26948160.0, + "step": 3214 + }, + { + "epoch": 9.713799621928166, + "grad_norm": 1.8125, + "learning_rate": 7.525279117558719e-06, + "loss": 0.708, + "num_tokens": 26959020.0, + "step": 3215 + }, + { + "epoch": 9.716824196597354, + "grad_norm": 0.890625, + "learning_rate": 7.521345297198327e-06, + "loss": 0.8512, + "num_tokens": 27025675.0, + "step": 3216 + }, + { + "epoch": 9.719848771266541, + "grad_norm": 1.390625, + "learning_rate": 7.5174095407273765e-06, + "loss": 0.7183, + "num_tokens": 27048730.0, + "step": 3217 + }, + { + "epoch": 9.722873345935728, + "grad_norm": 1.4140625, + "learning_rate": 7.513471851915657e-06, + "loss": 0.7632, + "num_tokens": 27072010.0, + "step": 3218 + }, + { + "epoch": 9.725897920604915, + "grad_norm": 2.109375, + "learning_rate": 7.509532234534818e-06, + "loss": 0.5733, + "num_tokens": 27079036.0, + "step": 3219 + }, + { + "epoch": 9.728922495274102, + "grad_norm": 2.0625, + "learning_rate": 7.505590692358353e-06, + "loss": 0.5814, + "num_tokens": 27085529.0, + "step": 3220 + }, + { + "epoch": 9.731947069943288, + "grad_norm": 1.1015625, + "learning_rate": 7.501647229161599e-06, + "loss": 0.7825, + "num_tokens": 27124923.0, + "step": 3221 + }, + { + "epoch": 9.734971644612477, + "grad_norm": 1.3671875, + "learning_rate": 7.497701848721738e-06, + "loss": 0.7887, + "num_tokens": 27149130.0, + "step": 3222 + }, + { + "epoch": 9.737996219281664, + "grad_norm": 1.65625, + "learning_rate": 7.493754554817782e-06, + "loss": 0.7666, + "num_tokens": 27172281.0, + "step": 3223 + }, + { + "epoch": 9.74102079395085, + "grad_norm": 1.6328125, + "learning_rate": 7.489805351230575e-06, + "loss": 0.5904, + "num_tokens": 27191215.0, + "step": 3224 + }, + { + "epoch": 9.744045368620037, + "grad_norm": 1.484375, + "learning_rate": 7.485854241742799e-06, + "loss": 0.7662, + "num_tokens": 27214686.0, + "step": 3225 + }, + { + "epoch": 9.747069943289224, + "grad_norm": 1.3671875, + "learning_rate": 7.4819012301389514e-06, + "loss": 0.7768, + "num_tokens": 27240177.0, + "step": 3226 + }, + { + "epoch": 9.750094517958413, + "grad_norm": 1.671875, + "learning_rate": 7.477946320205358e-06, + "loss": 0.5684, + "num_tokens": 27252510.0, + "step": 3227 + }, + { + "epoch": 9.7531190926276, + "grad_norm": 1.640625, + "learning_rate": 7.473989515730159e-06, + "loss": 0.9315, + "num_tokens": 27276025.0, + "step": 3228 + }, + { + "epoch": 9.756143667296787, + "grad_norm": 2.328125, + "learning_rate": 7.470030820503312e-06, + "loss": 0.6107, + "num_tokens": 27281915.0, + "step": 3229 + }, + { + "epoch": 9.759168241965973, + "grad_norm": 1.6328125, + "learning_rate": 7.466070238316584e-06, + "loss": 0.7201, + "num_tokens": 27297631.0, + "step": 3230 + }, + { + "epoch": 9.76219281663516, + "grad_norm": 2.078125, + "learning_rate": 7.4621077729635516e-06, + "loss": 0.5664, + "num_tokens": 27304737.0, + "step": 3231 + }, + { + "epoch": 9.765217391304347, + "grad_norm": 1.9375, + "learning_rate": 7.458143428239593e-06, + "loss": 0.549, + "num_tokens": 27312478.0, + "step": 3232 + }, + { + "epoch": 9.768241965973536, + "grad_norm": 1.4375, + "learning_rate": 7.454177207941884e-06, + "loss": 0.7393, + "num_tokens": 27335364.0, + "step": 3233 + }, + { + "epoch": 9.771266540642722, + "grad_norm": 1.3515625, + "learning_rate": 7.450209115869407e-06, + "loss": 0.8149, + "num_tokens": 27369367.0, + "step": 3234 + }, + { + "epoch": 9.77429111531191, + "grad_norm": 1.53125, + "learning_rate": 7.446239155822921e-06, + "loss": 0.8362, + "num_tokens": 27392829.0, + "step": 3235 + }, + { + "epoch": 9.777315689981096, + "grad_norm": 1.78125, + "learning_rate": 7.44226733160499e-06, + "loss": 0.5913, + "num_tokens": 27402895.0, + "step": 3236 + }, + { + "epoch": 9.780340264650283, + "grad_norm": 2.0, + "learning_rate": 7.438293647019958e-06, + "loss": 0.5794, + "num_tokens": 27410019.0, + "step": 3237 + }, + { + "epoch": 9.783364839319471, + "grad_norm": 1.6640625, + "learning_rate": 7.434318105873943e-06, + "loss": 0.6752, + "num_tokens": 27424359.0, + "step": 3238 + }, + { + "epoch": 9.786389413988658, + "grad_norm": 1.53125, + "learning_rate": 7.430340711974855e-06, + "loss": 0.8166, + "num_tokens": 27447814.0, + "step": 3239 + }, + { + "epoch": 9.789413988657845, + "grad_norm": 1.9921875, + "learning_rate": 7.426361469132367e-06, + "loss": 0.6062, + "num_tokens": 27454860.0, + "step": 3240 + }, + { + "epoch": 9.792438563327032, + "grad_norm": 1.4296875, + "learning_rate": 7.422380381157931e-06, + "loss": 0.6749, + "num_tokens": 27474361.0, + "step": 3241 + }, + { + "epoch": 9.795463137996219, + "grad_norm": 1.5546875, + "learning_rate": 7.418397451864763e-06, + "loss": 0.8224, + "num_tokens": 27497936.0, + "step": 3242 + }, + { + "epoch": 9.798487712665406, + "grad_norm": 1.546875, + "learning_rate": 7.4144126850678396e-06, + "loss": 0.8329, + "num_tokens": 27522024.0, + "step": 3243 + }, + { + "epoch": 9.801512287334594, + "grad_norm": 1.75, + "learning_rate": 7.410426084583906e-06, + "loss": 0.7487, + "num_tokens": 27535977.0, + "step": 3244 + }, + { + "epoch": 9.804536862003781, + "grad_norm": 2.078125, + "learning_rate": 7.406437654231453e-06, + "loss": 0.5859, + "num_tokens": 27543643.0, + "step": 3245 + }, + { + "epoch": 9.807561436672968, + "grad_norm": 1.8828125, + "learning_rate": 7.402447397830732e-06, + "loss": 0.5982, + "num_tokens": 27551671.0, + "step": 3246 + }, + { + "epoch": 9.810586011342155, + "grad_norm": 1.7109375, + "learning_rate": 7.398455319203741e-06, + "loss": 0.7147, + "num_tokens": 27564841.0, + "step": 3247 + }, + { + "epoch": 9.813610586011341, + "grad_norm": 1.21875, + "learning_rate": 7.394461422174224e-06, + "loss": 0.7327, + "num_tokens": 27596677.0, + "step": 3248 + }, + { + "epoch": 9.81663516068053, + "grad_norm": 1.484375, + "learning_rate": 7.390465710567664e-06, + "loss": 0.6388, + "num_tokens": 27614551.0, + "step": 3249 + }, + { + "epoch": 9.819659735349717, + "grad_norm": 1.6953125, + "learning_rate": 7.386468188211285e-06, + "loss": 0.5429, + "num_tokens": 27623261.0, + "step": 3250 + }, + { + "epoch": 9.822684310018904, + "grad_norm": 1.6640625, + "learning_rate": 7.382468858934046e-06, + "loss": 0.6052, + "num_tokens": 27636147.0, + "step": 3251 + }, + { + "epoch": 9.82570888468809, + "grad_norm": 1.59375, + "learning_rate": 7.378467726566632e-06, + "loss": 0.7438, + "num_tokens": 27654787.0, + "step": 3252 + }, + { + "epoch": 9.828733459357277, + "grad_norm": 1.9453125, + "learning_rate": 7.374464794941461e-06, + "loss": 0.5718, + "num_tokens": 27662224.0, + "step": 3253 + }, + { + "epoch": 9.831758034026464, + "grad_norm": 2.15625, + "learning_rate": 7.370460067892672e-06, + "loss": 0.5502, + "num_tokens": 27668503.0, + "step": 3254 + }, + { + "epoch": 9.834782608695653, + "grad_norm": 1.8125, + "learning_rate": 7.36645354925612e-06, + "loss": 0.6265, + "num_tokens": 27679005.0, + "step": 3255 + }, + { + "epoch": 9.83780718336484, + "grad_norm": 1.96875, + "learning_rate": 7.362445242869384e-06, + "loss": 0.5146, + "num_tokens": 27685671.0, + "step": 3256 + }, + { + "epoch": 9.840831758034026, + "grad_norm": 1.8125, + "learning_rate": 7.358435152571749e-06, + "loss": 0.6094, + "num_tokens": 27698400.0, + "step": 3257 + }, + { + "epoch": 9.843856332703213, + "grad_norm": 1.3671875, + "learning_rate": 7.3544232822042084e-06, + "loss": 0.8058, + "num_tokens": 27721882.0, + "step": 3258 + }, + { + "epoch": 9.8468809073724, + "grad_norm": 1.6171875, + "learning_rate": 7.350409635609466e-06, + "loss": 0.704, + "num_tokens": 27736909.0, + "step": 3259 + }, + { + "epoch": 9.849905482041589, + "grad_norm": 1.515625, + "learning_rate": 7.34639421663192e-06, + "loss": 0.797, + "num_tokens": 27759342.0, + "step": 3260 + }, + { + "epoch": 9.852930056710775, + "grad_norm": 1.7734375, + "learning_rate": 7.342377029117673e-06, + "loss": 0.5661, + "num_tokens": 27767648.0, + "step": 3261 + }, + { + "epoch": 9.855954631379962, + "grad_norm": 1.96875, + "learning_rate": 7.338358076914517e-06, + "loss": 0.5887, + "num_tokens": 27775338.0, + "step": 3262 + }, + { + "epoch": 9.858979206049149, + "grad_norm": 1.3046875, + "learning_rate": 7.334337363871936e-06, + "loss": 0.7574, + "num_tokens": 27805189.0, + "step": 3263 + }, + { + "epoch": 9.862003780718336, + "grad_norm": 1.7265625, + "learning_rate": 7.330314893841102e-06, + "loss": 0.8554, + "num_tokens": 27828226.0, + "step": 3264 + }, + { + "epoch": 9.865028355387524, + "grad_norm": 1.3359375, + "learning_rate": 7.326290670674864e-06, + "loss": 0.7141, + "num_tokens": 27850110.0, + "step": 3265 + }, + { + "epoch": 9.868052930056711, + "grad_norm": 1.484375, + "learning_rate": 7.322264698227759e-06, + "loss": 0.7474, + "num_tokens": 27870603.0, + "step": 3266 + }, + { + "epoch": 9.871077504725898, + "grad_norm": 1.9296875, + "learning_rate": 7.318236980355993e-06, + "loss": 0.5511, + "num_tokens": 27878405.0, + "step": 3267 + }, + { + "epoch": 9.874102079395085, + "grad_norm": 2.046875, + "learning_rate": 7.314207520917445e-06, + "loss": 0.5894, + "num_tokens": 27885645.0, + "step": 3268 + }, + { + "epoch": 9.877126654064272, + "grad_norm": 1.8046875, + "learning_rate": 7.310176323771663e-06, + "loss": 0.7769, + "num_tokens": 27900972.0, + "step": 3269 + }, + { + "epoch": 9.88015122873346, + "grad_norm": 1.9921875, + "learning_rate": 7.306143392779862e-06, + "loss": 0.5836, + "num_tokens": 27908184.0, + "step": 3270 + }, + { + "epoch": 9.883175803402647, + "grad_norm": 1.609375, + "learning_rate": 7.30210873180491e-06, + "loss": 0.6699, + "num_tokens": 27923688.0, + "step": 3271 + }, + { + "epoch": 9.886200378071834, + "grad_norm": 1.0859375, + "learning_rate": 7.298072344711342e-06, + "loss": 0.7878, + "num_tokens": 27968389.0, + "step": 3272 + }, + { + "epoch": 9.88922495274102, + "grad_norm": 1.796875, + "learning_rate": 7.2940342353653394e-06, + "loss": 0.5332, + "num_tokens": 27976599.0, + "step": 3273 + }, + { + "epoch": 9.892249527410208, + "grad_norm": 1.5078125, + "learning_rate": 7.289994407634735e-06, + "loss": 0.8755, + "num_tokens": 28000209.0, + "step": 3274 + }, + { + "epoch": 9.895274102079394, + "grad_norm": 1.3203125, + "learning_rate": 7.285952865389007e-06, + "loss": 0.8885, + "num_tokens": 28039118.0, + "step": 3275 + }, + { + "epoch": 9.898298676748583, + "grad_norm": 1.8671875, + "learning_rate": 7.28190961249928e-06, + "loss": 0.5734, + "num_tokens": 28046583.0, + "step": 3276 + }, + { + "epoch": 9.90132325141777, + "grad_norm": 1.640625, + "learning_rate": 7.2778646528383105e-06, + "loss": 0.8202, + "num_tokens": 28067957.0, + "step": 3277 + }, + { + "epoch": 9.904347826086957, + "grad_norm": 1.984375, + "learning_rate": 7.273817990280494e-06, + "loss": 0.5926, + "num_tokens": 28075420.0, + "step": 3278 + }, + { + "epoch": 9.907372400756143, + "grad_norm": 1.4140625, + "learning_rate": 7.269769628701855e-06, + "loss": 0.7003, + "num_tokens": 28096369.0, + "step": 3279 + }, + { + "epoch": 9.91039697542533, + "grad_norm": 1.1484375, + "learning_rate": 7.265719571980048e-06, + "loss": 0.793, + "num_tokens": 28135131.0, + "step": 3280 + }, + { + "epoch": 9.913421550094519, + "grad_norm": 2.203125, + "learning_rate": 7.261667823994351e-06, + "loss": 0.6715, + "num_tokens": 28141572.0, + "step": 3281 + }, + { + "epoch": 9.916446124763706, + "grad_norm": 1.5234375, + "learning_rate": 7.257614388625656e-06, + "loss": 0.8213, + "num_tokens": 28167573.0, + "step": 3282 + }, + { + "epoch": 9.919470699432893, + "grad_norm": 1.625, + "learning_rate": 7.253559269756481e-06, + "loss": 0.6448, + "num_tokens": 28181614.0, + "step": 3283 + }, + { + "epoch": 9.92249527410208, + "grad_norm": 2.109375, + "learning_rate": 7.249502471270947e-06, + "loss": 0.6656, + "num_tokens": 28189710.0, + "step": 3284 + }, + { + "epoch": 9.925519848771266, + "grad_norm": 1.421875, + "learning_rate": 7.245443997054791e-06, + "loss": 0.785, + "num_tokens": 28213759.0, + "step": 3285 + }, + { + "epoch": 9.928544423440453, + "grad_norm": 1.390625, + "learning_rate": 7.2413838509953504e-06, + "loss": 0.8928, + "num_tokens": 28247287.0, + "step": 3286 + }, + { + "epoch": 9.931568998109642, + "grad_norm": 1.6015625, + "learning_rate": 7.237322036981568e-06, + "loss": 0.7575, + "num_tokens": 28267586.0, + "step": 3287 + }, + { + "epoch": 9.934593572778828, + "grad_norm": 1.5859375, + "learning_rate": 7.23325855890398e-06, + "loss": 0.6924, + "num_tokens": 28284903.0, + "step": 3288 + }, + { + "epoch": 9.937618147448015, + "grad_norm": 1.8203125, + "learning_rate": 7.22919342065472e-06, + "loss": 0.6832, + "num_tokens": 28297541.0, + "step": 3289 + }, + { + "epoch": 9.940642722117202, + "grad_norm": 1.1171875, + "learning_rate": 7.225126626127507e-06, + "loss": 0.8179, + "num_tokens": 28350482.0, + "step": 3290 + }, + { + "epoch": 9.943667296786389, + "grad_norm": 1.6640625, + "learning_rate": 7.221058179217653e-06, + "loss": 0.6493, + "num_tokens": 28364927.0, + "step": 3291 + }, + { + "epoch": 9.946691871455577, + "grad_norm": 1.96875, + "learning_rate": 7.216988083822048e-06, + "loss": 0.5811, + "num_tokens": 28372904.0, + "step": 3292 + }, + { + "epoch": 9.949716446124764, + "grad_norm": 1.1796875, + "learning_rate": 7.212916343839163e-06, + "loss": 0.8419, + "num_tokens": 28413786.0, + "step": 3293 + }, + { + "epoch": 9.952741020793951, + "grad_norm": 1.6796875, + "learning_rate": 7.208842963169044e-06, + "loss": 0.76, + "num_tokens": 28428218.0, + "step": 3294 + }, + { + "epoch": 9.955765595463138, + "grad_norm": 1.8828125, + "learning_rate": 7.204767945713304e-06, + "loss": 0.6505, + "num_tokens": 28437434.0, + "step": 3295 + }, + { + "epoch": 9.958790170132325, + "grad_norm": 1.921875, + "learning_rate": 7.200691295375134e-06, + "loss": 0.5979, + "num_tokens": 28445567.0, + "step": 3296 + }, + { + "epoch": 9.961814744801512, + "grad_norm": 1.5390625, + "learning_rate": 7.196613016059281e-06, + "loss": 0.7829, + "num_tokens": 28466388.0, + "step": 3297 + }, + { + "epoch": 9.9648393194707, + "grad_norm": 1.4765625, + "learning_rate": 7.19253311167205e-06, + "loss": 0.8216, + "num_tokens": 28495526.0, + "step": 3298 + }, + { + "epoch": 9.967863894139887, + "grad_norm": 1.3046875, + "learning_rate": 7.188451586121312e-06, + "loss": 0.7747, + "num_tokens": 28527529.0, + "step": 3299 + }, + { + "epoch": 9.970888468809074, + "grad_norm": 1.84375, + "learning_rate": 7.184368443316481e-06, + "loss": 0.6148, + "num_tokens": 28535797.0, + "step": 3300 + }, + { + "epoch": 9.97391304347826, + "grad_norm": 1.703125, + "learning_rate": 7.18028368716853e-06, + "loss": 0.7684, + "num_tokens": 28551290.0, + "step": 3301 + }, + { + "epoch": 9.976937618147447, + "grad_norm": 0.9140625, + "learning_rate": 7.176197321589967e-06, + "loss": 0.782, + "num_tokens": 28611536.0, + "step": 3302 + }, + { + "epoch": 9.979962192816636, + "grad_norm": 1.890625, + "learning_rate": 7.172109350494849e-06, + "loss": 0.6127, + "num_tokens": 28619077.0, + "step": 3303 + }, + { + "epoch": 9.982986767485823, + "grad_norm": 1.578125, + "learning_rate": 7.168019777798767e-06, + "loss": 0.7445, + "num_tokens": 28641199.0, + "step": 3304 + }, + { + "epoch": 9.98601134215501, + "grad_norm": 1.8828125, + "learning_rate": 7.163928607418849e-06, + "loss": 0.5837, + "num_tokens": 28649562.0, + "step": 3305 + }, + { + "epoch": 9.989035916824196, + "grad_norm": 1.4296875, + "learning_rate": 7.159835843273748e-06, + "loss": 0.7619, + "num_tokens": 28670524.0, + "step": 3306 + }, + { + "epoch": 9.992060491493383, + "grad_norm": 1.0703125, + "learning_rate": 7.155741489283653e-06, + "loss": 0.815, + "num_tokens": 28725020.0, + "step": 3307 + }, + { + "epoch": 9.99508506616257, + "grad_norm": 1.7109375, + "learning_rate": 7.1516455493702654e-06, + "loss": 0.8046, + "num_tokens": 28747997.0, + "step": 3308 + }, + { + "epoch": 9.998109640831759, + "grad_norm": 1.28125, + "learning_rate": 7.147548027456811e-06, + "loss": 0.7555, + "num_tokens": 28776330.0, + "step": 3309 + }, + { + "epoch": 10.0, + "grad_norm": 2.453125, + "learning_rate": 7.143448927468031e-06, + "loss": 0.6673, + "num_tokens": 28781252.0, + "step": 3310 + }, + { + "epoch": 10.0, + "step": 3310, + "total_flos": 5.6569864867021824e+17, + "train_loss": 0.2920910031773893, + "train_runtime": 7360.5417, + "train_samples_per_second": 7.187, + "train_steps_per_second": 0.45 + } + ], + "logging_steps": 1, + "max_steps": 3310, + "num_input_tokens_seen": 0, + "num_train_epochs": 10, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 5.6569864867021824e+17, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}