| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 11.0, | |
| "eval_steps": 500, | |
| "global_step": 2794, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.019714144898965006, | |
| "grad_norm": 2.673164129257202, | |
| "learning_rate": 1.3114754098360657e-06, | |
| "loss": 3.8588, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.03942828979793001, | |
| "grad_norm": 2.886237621307373, | |
| "learning_rate": 2.9508196721311478e-06, | |
| "loss": 3.8163, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.05914243469689502, | |
| "grad_norm": 2.4650022983551025, | |
| "learning_rate": 4.59016393442623e-06, | |
| "loss": 3.7443, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.07885657959586002, | |
| "grad_norm": 2.4647622108459473, | |
| "learning_rate": 6.229508196721312e-06, | |
| "loss": 3.7573, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.09857072449482504, | |
| "grad_norm": 2.5846645832061768, | |
| "learning_rate": 7.868852459016394e-06, | |
| "loss": 3.6941, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.11828486939379004, | |
| "grad_norm": 2.3263237476348877, | |
| "learning_rate": 9.508196721311476e-06, | |
| "loss": 3.5506, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.13799901429275505, | |
| "grad_norm": 2.181896448135376, | |
| "learning_rate": 1.1147540983606558e-05, | |
| "loss": 3.4485, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.15771315919172005, | |
| "grad_norm": 2.374994993209839, | |
| "learning_rate": 1.2786885245901639e-05, | |
| "loss": 3.269, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.17742730409068508, | |
| "grad_norm": 2.852327585220337, | |
| "learning_rate": 1.4426229508196722e-05, | |
| "loss": 3.1914, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.19714144898965008, | |
| "grad_norm": 2.855822801589966, | |
| "learning_rate": 1.6065573770491805e-05, | |
| "loss": 2.9482, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.21685559388861508, | |
| "grad_norm": 3.2098145484924316, | |
| "learning_rate": 1.7704918032786887e-05, | |
| "loss": 2.7581, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.23656973878758009, | |
| "grad_norm": 2.033719062805176, | |
| "learning_rate": 1.934426229508197e-05, | |
| "loss": 2.5988, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.2562838836865451, | |
| "grad_norm": 2.0503222942352295, | |
| "learning_rate": 2.098360655737705e-05, | |
| "loss": 2.6184, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.2759980285855101, | |
| "grad_norm": 1.858931303024292, | |
| "learning_rate": 2.262295081967213e-05, | |
| "loss": 2.5111, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.2957121734844751, | |
| "grad_norm": 1.6557573080062866, | |
| "learning_rate": 2.4262295081967215e-05, | |
| "loss": 2.4524, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.3154263183834401, | |
| "grad_norm": 1.706704020500183, | |
| "learning_rate": 2.5901639344262297e-05, | |
| "loss": 2.4317, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.3351404632824051, | |
| "grad_norm": 1.4746320247650146, | |
| "learning_rate": 2.754098360655738e-05, | |
| "loss": 2.3915, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.35485460818137016, | |
| "grad_norm": 1.7562377452850342, | |
| "learning_rate": 2.9180327868852458e-05, | |
| "loss": 2.3791, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.37456875308033516, | |
| "grad_norm": 1.7075871229171753, | |
| "learning_rate": 3.0819672131147544e-05, | |
| "loss": 2.3717, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.39428289797930016, | |
| "grad_norm": 1.5588061809539795, | |
| "learning_rate": 3.245901639344263e-05, | |
| "loss": 2.3924, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.41399704287826516, | |
| "grad_norm": 1.5635536909103394, | |
| "learning_rate": 3.409836065573771e-05, | |
| "loss": 2.3031, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.43371118777723017, | |
| "grad_norm": 1.7249213457107544, | |
| "learning_rate": 3.5737704918032786e-05, | |
| "loss": 2.3186, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.45342533267619517, | |
| "grad_norm": 1.868545651435852, | |
| "learning_rate": 3.737704918032787e-05, | |
| "loss": 2.1715, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.47313947757516017, | |
| "grad_norm": 1.9135463237762451, | |
| "learning_rate": 3.901639344262295e-05, | |
| "loss": 2.2488, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.4928536224741252, | |
| "grad_norm": 1.895492434501648, | |
| "learning_rate": 4.0655737704918036e-05, | |
| "loss": 2.2387, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.5125677673730902, | |
| "grad_norm": 1.767385721206665, | |
| "learning_rate": 4.229508196721312e-05, | |
| "loss": 2.2541, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.5322819122720552, | |
| "grad_norm": 2.0554378032684326, | |
| "learning_rate": 4.3934426229508194e-05, | |
| "loss": 2.1672, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.5519960571710202, | |
| "grad_norm": 2.3369956016540527, | |
| "learning_rate": 4.557377049180328e-05, | |
| "loss": 2.223, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.5717102020699852, | |
| "grad_norm": 2.0532915592193604, | |
| "learning_rate": 4.7213114754098365e-05, | |
| "loss": 2.1007, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.5914243469689502, | |
| "grad_norm": 2.2023046016693115, | |
| "learning_rate": 4.885245901639344e-05, | |
| "loss": 2.1027, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.6111384918679152, | |
| "grad_norm": 2.194356918334961, | |
| "learning_rate": 5.049180327868853e-05, | |
| "loss": 2.1188, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.6308526367668802, | |
| "grad_norm": 2.6036267280578613, | |
| "learning_rate": 5.213114754098361e-05, | |
| "loss": 1.9685, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.6505667816658453, | |
| "grad_norm": 2.6643617153167725, | |
| "learning_rate": 5.3770491803278686e-05, | |
| "loss": 2.0843, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.6702809265648102, | |
| "grad_norm": 2.3738605976104736, | |
| "learning_rate": 5.540983606557377e-05, | |
| "loss": 2.0489, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.6899950714637753, | |
| "grad_norm": 2.43137526512146, | |
| "learning_rate": 5.704918032786886e-05, | |
| "loss": 1.976, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.7097092163627403, | |
| "grad_norm": 3.0450685024261475, | |
| "learning_rate": 5.868852459016394e-05, | |
| "loss": 1.8117, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.7294233612617053, | |
| "grad_norm": 2.9214789867401123, | |
| "learning_rate": 6.032786885245902e-05, | |
| "loss": 1.8229, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.7491375061606703, | |
| "grad_norm": 2.7570457458496094, | |
| "learning_rate": 6.19672131147541e-05, | |
| "loss": 1.8662, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.7688516510596353, | |
| "grad_norm": 4.113077640533447, | |
| "learning_rate": 6.360655737704918e-05, | |
| "loss": 1.7928, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.7885657959586003, | |
| "grad_norm": 3.127991199493408, | |
| "learning_rate": 6.524590163934427e-05, | |
| "loss": 1.8187, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.8082799408575653, | |
| "grad_norm": 3.4384043216705322, | |
| "learning_rate": 6.688524590163935e-05, | |
| "loss": 1.7548, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.8279940857565303, | |
| "grad_norm": 3.2650253772735596, | |
| "learning_rate": 6.852459016393443e-05, | |
| "loss": 1.8635, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.8477082306554953, | |
| "grad_norm": 3.676208019256592, | |
| "learning_rate": 7.016393442622952e-05, | |
| "loss": 1.6462, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.8674223755544603, | |
| "grad_norm": 3.6363656520843506, | |
| "learning_rate": 7.180327868852459e-05, | |
| "loss": 1.6826, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.8871365204534253, | |
| "grad_norm": 3.487661123275757, | |
| "learning_rate": 7.344262295081968e-05, | |
| "loss": 1.7096, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.9068506653523903, | |
| "grad_norm": 4.129843235015869, | |
| "learning_rate": 7.508196721311476e-05, | |
| "loss": 1.6489, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.9265648102513554, | |
| "grad_norm": 3.7981042861938477, | |
| "learning_rate": 7.672131147540984e-05, | |
| "loss": 1.5388, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.9462789551503203, | |
| "grad_norm": 4.129542827606201, | |
| "learning_rate": 7.836065573770493e-05, | |
| "loss": 1.5883, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.9659931000492854, | |
| "grad_norm": 3.7202744483947754, | |
| "learning_rate": 8e-05, | |
| "loss": 1.5327, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.9857072449482503, | |
| "grad_norm": 5.224013805389404, | |
| "learning_rate": 8.163934426229509e-05, | |
| "loss": 1.5075, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.003942828979793, | |
| "grad_norm": 4.3526787757873535, | |
| "learning_rate": 8.327868852459016e-05, | |
| "loss": 1.4976, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 1.023656973878758, | |
| "grad_norm": 5.239654541015625, | |
| "learning_rate": 8.491803278688524e-05, | |
| "loss": 1.3648, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.043371118777723, | |
| "grad_norm": 4.757369518280029, | |
| "learning_rate": 8.655737704918033e-05, | |
| "loss": 1.3312, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 1.063085263676688, | |
| "grad_norm": 4.107004642486572, | |
| "learning_rate": 8.819672131147541e-05, | |
| "loss": 1.3081, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.082799408575653, | |
| "grad_norm": 3.8866994380950928, | |
| "learning_rate": 8.98360655737705e-05, | |
| "loss": 1.3473, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 1.102513553474618, | |
| "grad_norm": 4.225423812866211, | |
| "learning_rate": 9.147540983606557e-05, | |
| "loss": 1.3429, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.122227698373583, | |
| "grad_norm": 4.068089485168457, | |
| "learning_rate": 9.311475409836066e-05, | |
| "loss": 1.3851, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 1.141941843272548, | |
| "grad_norm": 4.4080986976623535, | |
| "learning_rate": 9.475409836065574e-05, | |
| "loss": 1.1531, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.1616559881715132, | |
| "grad_norm": 3.994210720062256, | |
| "learning_rate": 9.639344262295082e-05, | |
| "loss": 1.1968, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 1.181370133070478, | |
| "grad_norm": 4.020788669586182, | |
| "learning_rate": 9.803278688524591e-05, | |
| "loss": 1.1876, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.201084277969443, | |
| "grad_norm": 5.0161848068237305, | |
| "learning_rate": 9.967213114754099e-05, | |
| "loss": 1.135, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 1.2207984228684081, | |
| "grad_norm": 4.258621692657471, | |
| "learning_rate": 9.999947530447293e-05, | |
| "loss": 1.0355, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.2405125677673732, | |
| "grad_norm": 4.326329708099365, | |
| "learning_rate": 9.99973437477677e-05, | |
| "loss": 1.0635, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 1.2602267126663382, | |
| "grad_norm": 3.8772778511047363, | |
| "learning_rate": 9.999357260626116e-05, | |
| "loss": 1.1091, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.279940857565303, | |
| "grad_norm": 4.675949573516846, | |
| "learning_rate": 9.998816200362199e-05, | |
| "loss": 1.0763, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 1.2996550024642681, | |
| "grad_norm": 4.871826171875, | |
| "learning_rate": 9.998111211728248e-05, | |
| "loss": 1.016, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.3193691473632332, | |
| "grad_norm": 4.208927154541016, | |
| "learning_rate": 9.99724231784326e-05, | |
| "loss": 0.9334, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 1.339083292262198, | |
| "grad_norm": 3.859124183654785, | |
| "learning_rate": 9.996209547201258e-05, | |
| "loss": 0.9612, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.358797437161163, | |
| "grad_norm": 4.167286396026611, | |
| "learning_rate": 9.99501293367034e-05, | |
| "loss": 1.071, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 1.3785115820601281, | |
| "grad_norm": 4.229668617248535, | |
| "learning_rate": 9.993652516491579e-05, | |
| "loss": 0.9526, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.3982257269590932, | |
| "grad_norm": 6.326845645904541, | |
| "learning_rate": 9.992128340277729e-05, | |
| "loss": 0.8992, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.4179398718580583, | |
| "grad_norm": 4.119451522827148, | |
| "learning_rate": 9.990440455011769e-05, | |
| "loss": 0.9638, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.437654016757023, | |
| "grad_norm": 4.288192272186279, | |
| "learning_rate": 9.988588916045263e-05, | |
| "loss": 0.849, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.4573681616559881, | |
| "grad_norm": 3.985097885131836, | |
| "learning_rate": 9.986573784096536e-05, | |
| "loss": 0.8638, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.4770823065549532, | |
| "grad_norm": 3.398552894592285, | |
| "learning_rate": 9.984395125248695e-05, | |
| "loss": 0.8869, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.4967964514539183, | |
| "grad_norm": 5.5323486328125, | |
| "learning_rate": 9.982053010947455e-05, | |
| "loss": 0.8194, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.5165105963528833, | |
| "grad_norm": 4.199021816253662, | |
| "learning_rate": 9.979547517998796e-05, | |
| "loss": 0.7118, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 1.5362247412518482, | |
| "grad_norm": 4.152290344238281, | |
| "learning_rate": 9.976878728566443e-05, | |
| "loss": 0.8509, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.5559388861508132, | |
| "grad_norm": 3.820864200592041, | |
| "learning_rate": 9.974046730169183e-05, | |
| "loss": 0.768, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.5756530310497783, | |
| "grad_norm": 3.8033151626586914, | |
| "learning_rate": 9.971051615677974e-05, | |
| "loss": 0.7861, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.595367175948743, | |
| "grad_norm": 4.548523902893066, | |
| "learning_rate": 9.967893483312922e-05, | |
| "loss": 0.8347, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.6150813208477084, | |
| "grad_norm": 3.6572163105010986, | |
| "learning_rate": 9.964572436640045e-05, | |
| "loss": 0.7201, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.6347954657466732, | |
| "grad_norm": 3.980424404144287, | |
| "learning_rate": 9.961088584567882e-05, | |
| "loss": 0.7655, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.6545096106456383, | |
| "grad_norm": 3.2318334579467773, | |
| "learning_rate": 9.95744204134392e-05, | |
| "loss": 0.8285, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.6742237555446033, | |
| "grad_norm": 4.426093578338623, | |
| "learning_rate": 9.953632926550847e-05, | |
| "loss": 0.6971, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.6939379004435682, | |
| "grad_norm": 3.0529074668884277, | |
| "learning_rate": 9.949661365102637e-05, | |
| "loss": 0.7098, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.7136520453425332, | |
| "grad_norm": 3.948803424835205, | |
| "learning_rate": 9.945527487240442e-05, | |
| "loss": 0.7815, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.7333661902414983, | |
| "grad_norm": 3.2549209594726562, | |
| "learning_rate": 9.941231428528332e-05, | |
| "loss": 0.7385, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.7530803351404631, | |
| "grad_norm": 3.8000028133392334, | |
| "learning_rate": 9.93677332984884e-05, | |
| "loss": 0.7357, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.7727944800394284, | |
| "grad_norm": 3.5841259956359863, | |
| "learning_rate": 9.932153337398354e-05, | |
| "loss": 0.6641, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.7925086249383932, | |
| "grad_norm": 4.315596580505371, | |
| "learning_rate": 9.927371602682305e-05, | |
| "loss": 0.6193, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.8122227698373583, | |
| "grad_norm": 3.2834675312042236, | |
| "learning_rate": 9.92242828251022e-05, | |
| "loss": 0.5768, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.8319369147363234, | |
| "grad_norm": 3.127601385116577, | |
| "learning_rate": 9.917323538990561e-05, | |
| "loss": 0.5912, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.8516510596352882, | |
| "grad_norm": 3.2609682083129883, | |
| "learning_rate": 9.912057539525419e-05, | |
| "loss": 0.6093, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.8713652045342535, | |
| "grad_norm": 3.4914708137512207, | |
| "learning_rate": 9.906630456805024e-05, | |
| "loss": 0.5924, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.8910793494332183, | |
| "grad_norm": 3.5819380283355713, | |
| "learning_rate": 9.901042468802074e-05, | |
| "loss": 0.6426, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.9107934943321834, | |
| "grad_norm": 3.237508773803711, | |
| "learning_rate": 9.89529375876591e-05, | |
| "loss": 0.5933, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.9305076392311484, | |
| "grad_norm": 3.832315444946289, | |
| "learning_rate": 9.889384515216501e-05, | |
| "loss": 0.5767, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.9502217841301133, | |
| "grad_norm": 3.097031593322754, | |
| "learning_rate": 9.883314931938258e-05, | |
| "loss": 0.5225, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.9699359290290783, | |
| "grad_norm": 3.5058581829071045, | |
| "learning_rate": 9.877085207973684e-05, | |
| "loss": 0.5591, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.9896500739280434, | |
| "grad_norm": 3.0875792503356934, | |
| "learning_rate": 9.870695547616851e-05, | |
| "loss": 0.537, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 2.007885657959586, | |
| "grad_norm": 3.042051315307617, | |
| "learning_rate": 9.864146160406684e-05, | |
| "loss": 0.5084, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.027599802858551, | |
| "grad_norm": 3.661208152770996, | |
| "learning_rate": 9.857437261120115e-05, | |
| "loss": 0.4259, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 2.047313947757516, | |
| "grad_norm": 2.7599899768829346, | |
| "learning_rate": 9.850569069765012e-05, | |
| "loss": 0.4659, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.0670280926564812, | |
| "grad_norm": 3.0007238388061523, | |
| "learning_rate": 9.843541811572988e-05, | |
| "loss": 0.4666, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 2.086742237555446, | |
| "grad_norm": 3.00451397895813, | |
| "learning_rate": 9.836355716991999e-05, | |
| "loss": 0.4842, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.106456382454411, | |
| "grad_norm": 3.2380623817443848, | |
| "learning_rate": 9.82901102167879e-05, | |
| "loss": 0.4566, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 2.126170527353376, | |
| "grad_norm": 3.130038022994995, | |
| "learning_rate": 9.821507966491177e-05, | |
| "loss": 0.4407, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.145884672252341, | |
| "grad_norm": 3.2306203842163086, | |
| "learning_rate": 9.813846797480134e-05, | |
| "loss": 0.4002, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 2.165598817151306, | |
| "grad_norm": 3.315145969390869, | |
| "learning_rate": 9.806027765881734e-05, | |
| "loss": 0.4392, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.185312962050271, | |
| "grad_norm": 2.3235769271850586, | |
| "learning_rate": 9.798051128108907e-05, | |
| "loss": 0.4443, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 2.205027106949236, | |
| "grad_norm": 2.693516731262207, | |
| "learning_rate": 9.78991714574303e-05, | |
| "loss": 0.4259, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.2247412518482013, | |
| "grad_norm": 2.472679853439331, | |
| "learning_rate": 9.781626085525352e-05, | |
| "loss": 0.4294, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 2.244455396747166, | |
| "grad_norm": 2.6586391925811768, | |
| "learning_rate": 9.773178219348242e-05, | |
| "loss": 0.4111, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 2.264169541646131, | |
| "grad_norm": 2.869844913482666, | |
| "learning_rate": 9.764573824246277e-05, | |
| "loss": 0.4128, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 2.283883686545096, | |
| "grad_norm": 2.2448887825012207, | |
| "learning_rate": 9.755813182387158e-05, | |
| "loss": 0.4833, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.303597831444061, | |
| "grad_norm": 2.5775742530822754, | |
| "learning_rate": 9.74689658106245e-05, | |
| "loss": 0.4466, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 2.3233119763430263, | |
| "grad_norm": 2.392392158508301, | |
| "learning_rate": 9.73782431267817e-05, | |
| "loss": 0.4098, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 2.343026121241991, | |
| "grad_norm": 2.7074978351593018, | |
| "learning_rate": 9.728596674745187e-05, | |
| "loss": 0.4585, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 2.362740266140956, | |
| "grad_norm": 3.289433240890503, | |
| "learning_rate": 9.719213969869478e-05, | |
| "loss": 0.3965, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 2.3824544110399213, | |
| "grad_norm": 2.337214708328247, | |
| "learning_rate": 9.709676505742193e-05, | |
| "loss": 0.3742, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 2.402168555938886, | |
| "grad_norm": 2.2772057056427, | |
| "learning_rate": 9.699984595129575e-05, | |
| "loss": 0.3871, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 2.421882700837851, | |
| "grad_norm": 2.8387715816497803, | |
| "learning_rate": 9.690138555862691e-05, | |
| "loss": 0.3897, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 2.4415968457368162, | |
| "grad_norm": 2.725186586380005, | |
| "learning_rate": 9.680138710827025e-05, | |
| "loss": 0.4116, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 2.461310990635781, | |
| "grad_norm": 2.4635608196258545, | |
| "learning_rate": 9.669985387951877e-05, | |
| "loss": 0.3498, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 2.4810251355347464, | |
| "grad_norm": 3.7746026515960693, | |
| "learning_rate": 9.659678920199612e-05, | |
| "loss": 0.351, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 2.500739280433711, | |
| "grad_norm": 2.5471270084381104, | |
| "learning_rate": 9.649219645554741e-05, | |
| "loss": 0.3738, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 2.5204534253326765, | |
| "grad_norm": 2.378277063369751, | |
| "learning_rate": 9.638607907012844e-05, | |
| "loss": 0.403, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.5401675702316413, | |
| "grad_norm": 2.655937433242798, | |
| "learning_rate": 9.627844052569312e-05, | |
| "loss": 0.4509, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 2.559881715130606, | |
| "grad_norm": 2.7430179119110107, | |
| "learning_rate": 9.616928435207938e-05, | |
| "loss": 0.4314, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 2.5795958600295714, | |
| "grad_norm": 3.2100863456726074, | |
| "learning_rate": 9.605861412889347e-05, | |
| "loss": 0.3806, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 2.5993100049285363, | |
| "grad_norm": 2.856989860534668, | |
| "learning_rate": 9.594643348539251e-05, | |
| "loss": 0.4194, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 2.619024149827501, | |
| "grad_norm": 2.215460777282715, | |
| "learning_rate": 9.583274610036549e-05, | |
| "loss": 0.3617, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 2.6387382947264664, | |
| "grad_norm": 2.89630389213562, | |
| "learning_rate": 9.571755570201266e-05, | |
| "loss": 0.4555, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 2.658452439625431, | |
| "grad_norm": 2.8467066287994385, | |
| "learning_rate": 9.560086606782322e-05, | |
| "loss": 0.3644, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 2.678166584524396, | |
| "grad_norm": 2.772758722305298, | |
| "learning_rate": 9.548268102445145e-05, | |
| "loss": 0.3363, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 2.6978807294233613, | |
| "grad_norm": 2.4616968631744385, | |
| "learning_rate": 9.53630044475913e-05, | |
| "loss": 0.3377, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 2.717594874322326, | |
| "grad_norm": 2.449794292449951, | |
| "learning_rate": 9.524184026184917e-05, | |
| "loss": 0.4089, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 2.7373090192212914, | |
| "grad_norm": 2.1945290565490723, | |
| "learning_rate": 9.511919244061532e-05, | |
| "loss": 0.3989, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 2.7570231641202563, | |
| "grad_norm": 2.0283315181732178, | |
| "learning_rate": 9.499506500593345e-05, | |
| "loss": 0.4989, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 2.7767373090192216, | |
| "grad_norm": 2.6042144298553467, | |
| "learning_rate": 9.486946202836898e-05, | |
| "loss": 0.3996, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 2.7964514539181864, | |
| "grad_norm": 2.1515965461730957, | |
| "learning_rate": 9.474238762687532e-05, | |
| "loss": 0.3421, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 2.8161655988171512, | |
| "grad_norm": 2.147844076156616, | |
| "learning_rate": 9.461384596865904e-05, | |
| "loss": 0.4217, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 2.8358797437161165, | |
| "grad_norm": 2.502347707748413, | |
| "learning_rate": 9.448384126904306e-05, | |
| "loss": 0.4129, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 2.8555938886150813, | |
| "grad_norm": 2.373370409011841, | |
| "learning_rate": 9.435237779132845e-05, | |
| "loss": 0.3447, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 2.875308033514046, | |
| "grad_norm": 2.6700592041015625, | |
| "learning_rate": 9.421945984665465e-05, | |
| "loss": 0.355, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 2.8950221784130115, | |
| "grad_norm": 2.2917869091033936, | |
| "learning_rate": 9.408509179385806e-05, | |
| "loss": 0.3461, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 2.9147363233119763, | |
| "grad_norm": 1.9684882164001465, | |
| "learning_rate": 9.394927803932909e-05, | |
| "loss": 0.3912, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 2.934450468210941, | |
| "grad_norm": 2.7318103313446045, | |
| "learning_rate": 9.381202303686773e-05, | |
| "loss": 0.385, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 2.9541646131099064, | |
| "grad_norm": 2.490328311920166, | |
| "learning_rate": 9.367333128753741e-05, | |
| "loss": 0.3136, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 2.9738787580088712, | |
| "grad_norm": 1.991101861000061, | |
| "learning_rate": 9.353320733951745e-05, | |
| "loss": 0.3771, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 2.9935929029078365, | |
| "grad_norm": 2.4289438724517822, | |
| "learning_rate": 9.339165578795383e-05, | |
| "loss": 0.3484, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 3.011828486939379, | |
| "grad_norm": 2.3784472942352295, | |
| "learning_rate": 9.324868127480867e-05, | |
| "loss": 0.2815, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 3.031542631838344, | |
| "grad_norm": 2.365229606628418, | |
| "learning_rate": 9.310428848870778e-05, | |
| "loss": 0.307, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 3.051256776737309, | |
| "grad_norm": 2.1624045372009277, | |
| "learning_rate": 9.295848216478709e-05, | |
| "loss": 0.2851, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 3.070970921636274, | |
| "grad_norm": 1.6744097471237183, | |
| "learning_rate": 9.281126708453724e-05, | |
| "loss": 0.2774, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 3.0906850665352392, | |
| "grad_norm": 2.131070613861084, | |
| "learning_rate": 9.26626480756469e-05, | |
| "loss": 0.3181, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 3.110399211434204, | |
| "grad_norm": 1.9431699514389038, | |
| "learning_rate": 9.251263001184435e-05, | |
| "loss": 0.2512, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 3.130113356333169, | |
| "grad_norm": 1.7866734266281128, | |
| "learning_rate": 9.236121781273766e-05, | |
| "loss": 0.2931, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 3.149827501232134, | |
| "grad_norm": 2.349666118621826, | |
| "learning_rate": 9.220841644365343e-05, | |
| "loss": 0.3041, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 3.169541646131099, | |
| "grad_norm": 2.101146936416626, | |
| "learning_rate": 9.205423091547393e-05, | |
| "loss": 0.2767, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 3.189255791030064, | |
| "grad_norm": 1.6372120380401611, | |
| "learning_rate": 9.189866628447272e-05, | |
| "loss": 0.3017, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 3.208969935929029, | |
| "grad_norm": 1.7807250022888184, | |
| "learning_rate": 9.174172765214893e-05, | |
| "loss": 0.2727, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 3.228684080827994, | |
| "grad_norm": 2.0267744064331055, | |
| "learning_rate": 9.15834201650599e-05, | |
| "loss": 0.3501, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 3.2483982257269592, | |
| "grad_norm": 2.022195339202881, | |
| "learning_rate": 9.142374901465238e-05, | |
| "loss": 0.3237, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 3.268112370625924, | |
| "grad_norm": 2.5703470706939697, | |
| "learning_rate": 9.126271943709244e-05, | |
| "loss": 0.241, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 3.287826515524889, | |
| "grad_norm": 2.4614508152008057, | |
| "learning_rate": 9.110033671309352e-05, | |
| "loss": 0.2835, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 3.307540660423854, | |
| "grad_norm": 1.5219550132751465, | |
| "learning_rate": 9.093660616774344e-05, | |
| "loss": 0.2858, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 3.327254805322819, | |
| "grad_norm": 1.7633792161941528, | |
| "learning_rate": 9.077153317032976e-05, | |
| "loss": 0.2278, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 3.3469689502217843, | |
| "grad_norm": 1.829990267753601, | |
| "learning_rate": 9.060512313416358e-05, | |
| "loss": 0.3449, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 3.366683095120749, | |
| "grad_norm": 1.9619853496551514, | |
| "learning_rate": 9.043738151640216e-05, | |
| "loss": 0.3066, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 3.386397240019714, | |
| "grad_norm": 2.067197561264038, | |
| "learning_rate": 9.026831381786984e-05, | |
| "loss": 0.2594, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 3.4061113849186793, | |
| "grad_norm": 1.9810595512390137, | |
| "learning_rate": 9.009792558287777e-05, | |
| "loss": 0.294, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 3.425825529817644, | |
| "grad_norm": 2.0421338081359863, | |
| "learning_rate": 8.992622239904198e-05, | |
| "loss": 0.3122, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 3.445539674716609, | |
| "grad_norm": 1.50675368309021, | |
| "learning_rate": 8.975320989710022e-05, | |
| "loss": 0.3048, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 3.465253819615574, | |
| "grad_norm": 1.730079174041748, | |
| "learning_rate": 8.957889375072728e-05, | |
| "loss": 0.2534, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 3.484967964514539, | |
| "grad_norm": 2.360069751739502, | |
| "learning_rate": 8.940327967634893e-05, | |
| "loss": 0.2415, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 3.5046821094135043, | |
| "grad_norm": 2.379964828491211, | |
| "learning_rate": 8.922637343295446e-05, | |
| "loss": 0.3176, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 3.524396254312469, | |
| "grad_norm": 1.2987922430038452, | |
| "learning_rate": 8.904818082190783e-05, | |
| "loss": 0.3124, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 3.5441103992114344, | |
| "grad_norm": 2.5630524158477783, | |
| "learning_rate": 8.88687076867574e-05, | |
| "loss": 0.3022, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 3.5638245441103993, | |
| "grad_norm": 2.043360948562622, | |
| "learning_rate": 8.868795991304436e-05, | |
| "loss": 0.2802, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 3.583538689009364, | |
| "grad_norm": 2.51525616645813, | |
| "learning_rate": 8.850594342810961e-05, | |
| "loss": 0.2965, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 3.6032528339083294, | |
| "grad_norm": 2.4079837799072266, | |
| "learning_rate": 8.832266420089953e-05, | |
| "loss": 0.3132, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 3.6229669788072942, | |
| "grad_norm": 1.6233525276184082, | |
| "learning_rate": 8.813812824177012e-05, | |
| "loss": 0.2818, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 3.642681123706259, | |
| "grad_norm": 1.8031409978866577, | |
| "learning_rate": 8.795234160228995e-05, | |
| "loss": 0.2706, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 3.6623952686052244, | |
| "grad_norm": 2.136138439178467, | |
| "learning_rate": 8.776531037504169e-05, | |
| "loss": 0.287, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 3.682109413504189, | |
| "grad_norm": 1.9893665313720703, | |
| "learning_rate": 8.757704069342234e-05, | |
| "loss": 0.2901, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 3.701823558403154, | |
| "grad_norm": 1.5316953659057617, | |
| "learning_rate": 8.738753873144205e-05, | |
| "loss": 0.2958, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 3.7215377033021193, | |
| "grad_norm": 1.774769902229309, | |
| "learning_rate": 8.71968107035217e-05, | |
| "loss": 0.2604, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 3.741251848201084, | |
| "grad_norm": 1.890213966369629, | |
| "learning_rate": 8.700486286428909e-05, | |
| "loss": 0.258, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 3.7609659931000494, | |
| "grad_norm": 2.19962215423584, | |
| "learning_rate": 8.681170150837382e-05, | |
| "loss": 0.3343, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 3.7806801379990143, | |
| "grad_norm": 1.7772096395492554, | |
| "learning_rate": 8.661733297020085e-05, | |
| "loss": 0.2605, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 3.8003942828979795, | |
| "grad_norm": 1.5164539813995361, | |
| "learning_rate": 8.642176362378285e-05, | |
| "loss": 0.3217, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 3.8201084277969444, | |
| "grad_norm": 2.329491138458252, | |
| "learning_rate": 8.622499988251106e-05, | |
| "loss": 0.3079, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 3.839822572695909, | |
| "grad_norm": 2.1188507080078125, | |
| "learning_rate": 8.602704819894511e-05, | |
| "loss": 0.3327, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 3.8595367175948745, | |
| "grad_norm": 1.6314496994018555, | |
| "learning_rate": 8.582791506460126e-05, | |
| "loss": 0.2557, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 3.8792508624938393, | |
| "grad_norm": 1.6020344495773315, | |
| "learning_rate": 8.562760700973966e-05, | |
| "loss": 0.286, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 3.898965007392804, | |
| "grad_norm": 1.744199275970459, | |
| "learning_rate": 8.54261306031501e-05, | |
| "loss": 0.264, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 3.9186791522917694, | |
| "grad_norm": 2.455364942550659, | |
| "learning_rate": 8.52234924519367e-05, | |
| "loss": 0.2795, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 3.9383932971907343, | |
| "grad_norm": 2.127448558807373, | |
| "learning_rate": 8.501969920130111e-05, | |
| "loss": 0.3015, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 3.958107442089699, | |
| "grad_norm": 1.9393630027770996, | |
| "learning_rate": 8.481475753432473e-05, | |
| "loss": 0.3879, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 3.9778215869886644, | |
| "grad_norm": 1.3514901399612427, | |
| "learning_rate": 8.46086741717494e-05, | |
| "loss": 0.2484, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 3.9975357318876292, | |
| "grad_norm": 1.2901164293289185, | |
| "learning_rate": 8.440145587175718e-05, | |
| "loss": 0.2712, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 4.015771315919172, | |
| "grad_norm": 1.8801252841949463, | |
| "learning_rate": 8.419310942974855e-05, | |
| "loss": 0.2422, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 4.035485460818137, | |
| "grad_norm": 1.428691029548645, | |
| "learning_rate": 8.398364167811967e-05, | |
| "loss": 0.2097, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 4.055199605717102, | |
| "grad_norm": 1.7484722137451172, | |
| "learning_rate": 8.377305948603834e-05, | |
| "loss": 0.2155, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 4.074913750616067, | |
| "grad_norm": 1.617615818977356, | |
| "learning_rate": 8.356136975921863e-05, | |
| "loss": 0.2147, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 4.094627895515032, | |
| "grad_norm": 1.8283522129058838, | |
| "learning_rate": 8.334857943969451e-05, | |
| "loss": 0.214, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 4.114342040413997, | |
| "grad_norm": 1.7181671857833862, | |
| "learning_rate": 8.313469550559219e-05, | |
| "loss": 0.2102, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 4.1340561853129625, | |
| "grad_norm": 1.7264341115951538, | |
| "learning_rate": 8.291972497090119e-05, | |
| "loss": 0.2665, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 4.153770330211927, | |
| "grad_norm": 2.047731876373291, | |
| "learning_rate": 8.27036748852445e-05, | |
| "loss": 0.2314, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 4.173484475110892, | |
| "grad_norm": 1.5944523811340332, | |
| "learning_rate": 8.248655233364724e-05, | |
| "loss": 0.2129, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 4.193198620009857, | |
| "grad_norm": 1.3731194734573364, | |
| "learning_rate": 8.226836443630437e-05, | |
| "loss": 0.2286, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 4.212912764908822, | |
| "grad_norm": 1.8190358877182007, | |
| "learning_rate": 8.204911834834722e-05, | |
| "loss": 0.2564, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 4.232626909807787, | |
| "grad_norm": 3.020439624786377, | |
| "learning_rate": 8.182882125960882e-05, | |
| "loss": 0.2337, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 4.252341054706752, | |
| "grad_norm": 2.168088436126709, | |
| "learning_rate": 8.160748039438811e-05, | |
| "loss": 0.2177, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 4.272055199605717, | |
| "grad_norm": 1.3565387725830078, | |
| "learning_rate": 8.138510301121312e-05, | |
| "loss": 0.2267, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 4.291769344504682, | |
| "grad_norm": 1.3000303506851196, | |
| "learning_rate": 8.116169640260276e-05, | |
| "loss": 0.2212, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 4.311483489403647, | |
| "grad_norm": 2.5623159408569336, | |
| "learning_rate": 8.093726789482788e-05, | |
| "loss": 0.2447, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 4.331197634302612, | |
| "grad_norm": 1.2196520566940308, | |
| "learning_rate": 8.071182484767085e-05, | |
| "loss": 0.2146, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 4.350911779201577, | |
| "grad_norm": 2.003669261932373, | |
| "learning_rate": 8.04853746541843e-05, | |
| "loss": 0.2048, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 4.370625924100542, | |
| "grad_norm": 1.9316153526306152, | |
| "learning_rate": 8.025792474044869e-05, | |
| "loss": 0.2451, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 4.390340068999507, | |
| "grad_norm": 1.6847680807113647, | |
| "learning_rate": 8.002948256532865e-05, | |
| "loss": 0.2213, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 4.410054213898472, | |
| "grad_norm": 2.238589286804199, | |
| "learning_rate": 7.980005562022859e-05, | |
| "loss": 0.2175, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 4.429768358797437, | |
| "grad_norm": 1.7301565408706665, | |
| "learning_rate": 7.956965142884677e-05, | |
| "loss": 0.2245, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 4.4494825036964025, | |
| "grad_norm": 1.7427107095718384, | |
| "learning_rate": 7.933827754692886e-05, | |
| "loss": 0.245, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 4.469196648595367, | |
| "grad_norm": 1.6464322805404663, | |
| "learning_rate": 7.910594156201993e-05, | |
| "loss": 0.21, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 4.488910793494332, | |
| "grad_norm": 2.1638522148132324, | |
| "learning_rate": 7.887265109321576e-05, | |
| "loss": 0.2183, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 4.5086249383932975, | |
| "grad_norm": 1.441428542137146, | |
| "learning_rate": 7.863841379091289e-05, | |
| "loss": 0.237, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 4.528339083292262, | |
| "grad_norm": 1.6750849485397339, | |
| "learning_rate": 7.840323733655778e-05, | |
| "loss": 0.2347, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 4.548053228191227, | |
| "grad_norm": 1.4054741859436035, | |
| "learning_rate": 7.816712944239503e-05, | |
| "loss": 0.1962, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 4.567767373090192, | |
| "grad_norm": 1.8165724277496338, | |
| "learning_rate": 7.793009785121421e-05, | |
| "loss": 0.229, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 4.587481517989158, | |
| "grad_norm": 2.425626754760742, | |
| "learning_rate": 7.769215033609618e-05, | |
| "loss": 0.2182, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 4.607195662888122, | |
| "grad_norm": 2.2778477668762207, | |
| "learning_rate": 7.745329470015805e-05, | |
| "loss": 0.1964, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 4.626909807787087, | |
| "grad_norm": 1.7259466648101807, | |
| "learning_rate": 7.721353877629736e-05, | |
| "loss": 0.235, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 4.646623952686053, | |
| "grad_norm": 2.020500659942627, | |
| "learning_rate": 7.697289042693518e-05, | |
| "loss": 0.2297, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 4.666338097585017, | |
| "grad_norm": 1.7260327339172363, | |
| "learning_rate": 7.673135754375828e-05, | |
| "loss": 0.2256, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 4.686052242483982, | |
| "grad_norm": 1.3809380531311035, | |
| "learning_rate": 7.648894804746031e-05, | |
| "loss": 0.2165, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 4.705766387382948, | |
| "grad_norm": 1.692401647567749, | |
| "learning_rate": 7.624566988748211e-05, | |
| "loss": 0.2163, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 4.725480532281912, | |
| "grad_norm": 1.4471834897994995, | |
| "learning_rate": 7.6001531041751e-05, | |
| "loss": 0.2106, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 4.745194677180877, | |
| "grad_norm": 1.6524702310562134, | |
| "learning_rate": 7.575653951641911e-05, | |
| "loss": 0.228, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 4.764908822079843, | |
| "grad_norm": 1.7597817182540894, | |
| "learning_rate": 7.551070334560088e-05, | |
| "loss": 0.1982, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 4.784622966978807, | |
| "grad_norm": 1.2614564895629883, | |
| "learning_rate": 7.526403059110956e-05, | |
| "loss": 0.182, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 4.804337111877772, | |
| "grad_norm": 1.6442838907241821, | |
| "learning_rate": 7.501652934219292e-05, | |
| "loss": 0.2434, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 4.8240512567767375, | |
| "grad_norm": 1.6709420680999756, | |
| "learning_rate": 7.476820771526784e-05, | |
| "loss": 0.2059, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 4.843765401675702, | |
| "grad_norm": 1.393500804901123, | |
| "learning_rate": 7.451907385365423e-05, | |
| "loss": 0.2612, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 4.863479546574667, | |
| "grad_norm": 1.7826026678085327, | |
| "learning_rate": 7.426913592730799e-05, | |
| "loss": 0.2208, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 4.8831936914736325, | |
| "grad_norm": 2.0293965339660645, | |
| "learning_rate": 7.401840213255305e-05, | |
| "loss": 0.2141, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 4.902907836372597, | |
| "grad_norm": 1.216652512550354, | |
| "learning_rate": 7.376688069181263e-05, | |
| "loss": 0.2188, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 4.922621981271562, | |
| "grad_norm": 1.2291804552078247, | |
| "learning_rate": 7.351457985333954e-05, | |
| "loss": 0.2567, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 4.942336126170527, | |
| "grad_norm": 1.022298812866211, | |
| "learning_rate": 7.32615078909457e-05, | |
| "loss": 0.1975, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 4.962050271069493, | |
| "grad_norm": 1.1458094120025635, | |
| "learning_rate": 7.30076731037309e-05, | |
| "loss": 0.2115, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 4.981764415968457, | |
| "grad_norm": 1.5030829906463623, | |
| "learning_rate": 7.275308381581055e-05, | |
| "loss": 0.2531, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 1.6188287734985352, | |
| "learning_rate": 7.249774837604266e-05, | |
| "loss": 0.216, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 5.019714144898965, | |
| "grad_norm": 1.6201726198196411, | |
| "learning_rate": 7.224167515775425e-05, | |
| "loss": 0.1635, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 5.03942828979793, | |
| "grad_norm": 2.038351535797119, | |
| "learning_rate": 7.198487255846655e-05, | |
| "loss": 0.1592, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 5.059142434696895, | |
| "grad_norm": 1.7551215887069702, | |
| "learning_rate": 7.172734899961973e-05, | |
| "loss": 0.1796, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 5.07885657959586, | |
| "grad_norm": 1.3391064405441284, | |
| "learning_rate": 7.14691129262967e-05, | |
| "loss": 0.1772, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 5.098570724494825, | |
| "grad_norm": 1.2368159294128418, | |
| "learning_rate": 7.121017280694615e-05, | |
| "loss": 0.1759, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 5.11828486939379, | |
| "grad_norm": 1.5747733116149902, | |
| "learning_rate": 7.09505371331049e-05, | |
| "loss": 0.167, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 5.137999014292755, | |
| "grad_norm": 1.57236909866333, | |
| "learning_rate": 7.069021441911934e-05, | |
| "loss": 0.1727, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 5.1577131591917205, | |
| "grad_norm": 1.917982816696167, | |
| "learning_rate": 7.042921320186634e-05, | |
| "loss": 0.1756, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 5.177427304090685, | |
| "grad_norm": 1.263602375984192, | |
| "learning_rate": 7.016754204047314e-05, | |
| "loss": 0.1758, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 5.19714144898965, | |
| "grad_norm": 1.1524710655212402, | |
| "learning_rate": 6.990520951603682e-05, | |
| "loss": 0.1602, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 5.216855593888615, | |
| "grad_norm": 0.9892676472663879, | |
| "learning_rate": 6.964222423134278e-05, | |
| "loss": 0.1587, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 5.23656973878758, | |
| "grad_norm": 1.5549856424331665, | |
| "learning_rate": 6.937859481058273e-05, | |
| "loss": 0.1734, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 5.256283883686545, | |
| "grad_norm": 1.2518091201782227, | |
| "learning_rate": 6.911432989907174e-05, | |
| "loss": 0.1744, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 5.27599802858551, | |
| "grad_norm": 2.184462070465088, | |
| "learning_rate": 6.884943816296487e-05, | |
| "loss": 0.1536, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 5.295712173484475, | |
| "grad_norm": 1.6105055809020996, | |
| "learning_rate": 6.85839282889729e-05, | |
| "loss": 0.172, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 5.31542631838344, | |
| "grad_norm": 1.1322895288467407, | |
| "learning_rate": 6.831780898407746e-05, | |
| "loss": 0.1622, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 5.335140463282405, | |
| "grad_norm": 1.8505653142929077, | |
| "learning_rate": 6.805108897524552e-05, | |
| "loss": 0.1709, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 5.35485460818137, | |
| "grad_norm": 1.405929446220398, | |
| "learning_rate": 6.77837770091432e-05, | |
| "loss": 0.1807, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 5.374568753080335, | |
| "grad_norm": 1.7219219207763672, | |
| "learning_rate": 6.751588185184896e-05, | |
| "loss": 0.2026, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 5.3942828979793, | |
| "grad_norm": 1.5548940896987915, | |
| "learning_rate": 6.724741228856604e-05, | |
| "loss": 0.1671, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 5.4139970428782656, | |
| "grad_norm": 1.397202730178833, | |
| "learning_rate": 6.69783771233345e-05, | |
| "loss": 0.1675, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 5.43371118777723, | |
| "grad_norm": 1.0036927461624146, | |
| "learning_rate": 6.670878517874238e-05, | |
| "loss": 0.1553, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 5.453425332676195, | |
| "grad_norm": 1.6738075017929077, | |
| "learning_rate": 6.643864529563644e-05, | |
| "loss": 0.195, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 5.4731394775751605, | |
| "grad_norm": 1.7666957378387451, | |
| "learning_rate": 6.616796633283227e-05, | |
| "loss": 0.16, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 5.492853622474125, | |
| "grad_norm": 1.3601677417755127, | |
| "learning_rate": 6.589675716682366e-05, | |
| "loss": 0.1695, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 5.51256776737309, | |
| "grad_norm": 2.429863691329956, | |
| "learning_rate": 6.562502669149166e-05, | |
| "loss": 0.1816, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 5.5322819122720555, | |
| "grad_norm": 1.3872816562652588, | |
| "learning_rate": 6.535278381781281e-05, | |
| "loss": 0.1683, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 5.55199605717102, | |
| "grad_norm": 1.4309865236282349, | |
| "learning_rate": 6.508003747356695e-05, | |
| "loss": 0.1693, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 5.571710202069985, | |
| "grad_norm": 1.8246867656707764, | |
| "learning_rate": 6.480679660304448e-05, | |
| "loss": 0.1562, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 5.59142434696895, | |
| "grad_norm": 1.3177618980407715, | |
| "learning_rate": 6.453307016675296e-05, | |
| "loss": 0.1812, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 5.611138491867916, | |
| "grad_norm": 1.7910345792770386, | |
| "learning_rate": 6.425886714112337e-05, | |
| "loss": 0.1902, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 5.63085263676688, | |
| "grad_norm": 1.1931573152542114, | |
| "learning_rate": 6.398419651821568e-05, | |
| "loss": 0.1838, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 5.650566781665845, | |
| "grad_norm": 1.5768245458602905, | |
| "learning_rate": 6.3709067305424e-05, | |
| "loss": 0.1715, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 5.670280926564811, | |
| "grad_norm": 2.553699493408203, | |
| "learning_rate": 6.343348852518114e-05, | |
| "loss": 0.1749, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 5.689995071463775, | |
| "grad_norm": 1.0562894344329834, | |
| "learning_rate": 6.315746921466279e-05, | |
| "loss": 0.1668, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 5.70970921636274, | |
| "grad_norm": 1.594449758529663, | |
| "learning_rate": 6.288101842549116e-05, | |
| "loss": 0.1741, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 5.729423361261706, | |
| "grad_norm": 1.732932448387146, | |
| "learning_rate": 6.260414522343813e-05, | |
| "loss": 0.1593, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 5.74913750616067, | |
| "grad_norm": 1.8460274934768677, | |
| "learning_rate": 6.232685868812789e-05, | |
| "loss": 0.1868, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 5.768851651059635, | |
| "grad_norm": 1.6852288246154785, | |
| "learning_rate": 6.204916791273932e-05, | |
| "loss": 0.173, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 5.7885657959586005, | |
| "grad_norm": 1.645918369293213, | |
| "learning_rate": 6.177108200370771e-05, | |
| "loss": 0.162, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 5.808279940857565, | |
| "grad_norm": 1.5185496807098389, | |
| "learning_rate": 6.14926100804261e-05, | |
| "loss": 0.1844, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 5.82799408575653, | |
| "grad_norm": 1.133602499961853, | |
| "learning_rate": 6.12137612749463e-05, | |
| "loss": 0.1861, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 5.8477082306554955, | |
| "grad_norm": 1.1715219020843506, | |
| "learning_rate": 6.093454473167938e-05, | |
| "loss": 0.1699, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 5.86742237555446, | |
| "grad_norm": 1.3760854005813599, | |
| "learning_rate": 6.065496960709578e-05, | |
| "loss": 0.1621, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 5.887136520453425, | |
| "grad_norm": 1.3439675569534302, | |
| "learning_rate": 6.0375045069425084e-05, | |
| "loss": 0.1486, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 5.9068506653523905, | |
| "grad_norm": 1.2491891384124756, | |
| "learning_rate": 6.009478029835528e-05, | |
| "loss": 0.1602, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 5.926564810251356, | |
| "grad_norm": 1.6365981101989746, | |
| "learning_rate": 5.981418448473184e-05, | |
| "loss": 0.1691, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 5.94627895515032, | |
| "grad_norm": 1.0324171781539917, | |
| "learning_rate": 5.953326683025621e-05, | |
| "loss": 0.1651, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 5.965993100049285, | |
| "grad_norm": 1.453691840171814, | |
| "learning_rate": 5.925203654718415e-05, | |
| "loss": 0.1892, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 5.985707244948251, | |
| "grad_norm": 1.7985492944717407, | |
| "learning_rate": 5.8970502858023545e-05, | |
| "loss": 0.1992, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 6.003942828979793, | |
| "grad_norm": 1.3267945051193237, | |
| "learning_rate": 5.868867499523203e-05, | |
| "loss": 0.1616, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 6.023656973878758, | |
| "grad_norm": 1.3626664876937866, | |
| "learning_rate": 5.8406562200914196e-05, | |
| "loss": 0.1427, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 6.043371118777723, | |
| "grad_norm": 1.3885642290115356, | |
| "learning_rate": 5.812417372651854e-05, | |
| "loss": 0.138, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 6.063085263676688, | |
| "grad_norm": 1.6768662929534912, | |
| "learning_rate": 5.7841518832534014e-05, | |
| "loss": 0.1392, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 6.082799408575653, | |
| "grad_norm": 1.214672565460205, | |
| "learning_rate": 5.755860678818642e-05, | |
| "loss": 0.142, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 6.102513553474618, | |
| "grad_norm": 1.4035040140151978, | |
| "learning_rate": 5.7275446871134405e-05, | |
| "loss": 0.1353, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 6.1222276983735835, | |
| "grad_norm": 1.6018613576889038, | |
| "learning_rate": 5.69920483671652e-05, | |
| "loss": 0.1378, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 6.141941843272548, | |
| "grad_norm": 1.4302403926849365, | |
| "learning_rate": 5.670842056989012e-05, | |
| "loss": 0.1381, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 6.161655988171513, | |
| "grad_norm": 0.9095770716667175, | |
| "learning_rate": 5.64245727804398e-05, | |
| "loss": 0.1409, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 6.1813701330704784, | |
| "grad_norm": 1.0363128185272217, | |
| "learning_rate": 5.614051430715915e-05, | |
| "loss": 0.1297, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 6.201084277969443, | |
| "grad_norm": 1.1587727069854736, | |
| "learning_rate": 5.5856254465302225e-05, | |
| "loss": 0.1352, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 6.220798422868408, | |
| "grad_norm": 1.387983798980713, | |
| "learning_rate": 5.5571802576726507e-05, | |
| "loss": 0.1356, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 6.240512567767373, | |
| "grad_norm": 1.4277117252349854, | |
| "learning_rate": 5.528716796958745e-05, | |
| "loss": 0.1356, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 6.260226712666338, | |
| "grad_norm": 1.3528679609298706, | |
| "learning_rate": 5.5002359978032435e-05, | |
| "loss": 0.1408, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 6.279940857565303, | |
| "grad_norm": 1.6311657428741455, | |
| "learning_rate": 5.471738794189474e-05, | |
| "loss": 0.1426, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 6.299655002464268, | |
| "grad_norm": 0.8196179866790771, | |
| "learning_rate": 5.443226120638719e-05, | |
| "loss": 0.1346, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 6.319369147363233, | |
| "grad_norm": 1.4031368494033813, | |
| "learning_rate": 5.414698912179579e-05, | |
| "loss": 0.1327, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 6.339083292262198, | |
| "grad_norm": 1.8704664707183838, | |
| "learning_rate": 5.386158104317298e-05, | |
| "loss": 0.1397, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 6.358797437161163, | |
| "grad_norm": 1.4920119047164917, | |
| "learning_rate": 5.3576046330030974e-05, | |
| "loss": 0.1325, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 6.378511582060128, | |
| "grad_norm": 1.0346848964691162, | |
| "learning_rate": 5.329039434603471e-05, | |
| "loss": 0.1482, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 6.398225726959093, | |
| "grad_norm": 1.9379348754882812, | |
| "learning_rate": 5.300463445869488e-05, | |
| "loss": 0.1524, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 6.417939871858058, | |
| "grad_norm": 0.7949770092964172, | |
| "learning_rate": 5.271877603906067e-05, | |
| "loss": 0.1282, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 6.4376540167570235, | |
| "grad_norm": 0.9802160859107971, | |
| "learning_rate": 5.24328284614125e-05, | |
| "loss": 0.1346, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 6.457368161655988, | |
| "grad_norm": 1.578125, | |
| "learning_rate": 5.214680110295459e-05, | |
| "loss": 0.1398, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 6.477082306554953, | |
| "grad_norm": 1.2905495166778564, | |
| "learning_rate": 5.186070334350741e-05, | |
| "loss": 0.1278, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 6.4967964514539185, | |
| "grad_norm": 1.4263631105422974, | |
| "learning_rate": 5.1574544565200175e-05, | |
| "loss": 0.1263, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 6.516510596352883, | |
| "grad_norm": 0.8475696444511414, | |
| "learning_rate": 5.128833415216309e-05, | |
| "loss": 0.1322, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 6.536224741251848, | |
| "grad_norm": 0.9275863170623779, | |
| "learning_rate": 5.100208149021962e-05, | |
| "loss": 0.1311, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 6.555938886150813, | |
| "grad_norm": 1.2516257762908936, | |
| "learning_rate": 5.071579596657876e-05, | |
| "loss": 0.1381, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 6.575653031049778, | |
| "grad_norm": 1.048720359802246, | |
| "learning_rate": 5.042948696952711e-05, | |
| "loss": 0.1374, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 6.595367175948743, | |
| "grad_norm": 1.3890330791473389, | |
| "learning_rate": 5.014316388812108e-05, | |
| "loss": 0.139, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 6.615081320847708, | |
| "grad_norm": 1.5058406591415405, | |
| "learning_rate": 4.985683611187895e-05, | |
| "loss": 0.152, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 6.634795465746674, | |
| "grad_norm": 1.5658972263336182, | |
| "learning_rate": 4.9570513030472896e-05, | |
| "loss": 0.1406, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 6.654509610645638, | |
| "grad_norm": 1.483197808265686, | |
| "learning_rate": 4.9284204033421266e-05, | |
| "loss": 0.1321, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 6.674223755544603, | |
| "grad_norm": 1.8198909759521484, | |
| "learning_rate": 4.8997918509780386e-05, | |
| "loss": 0.1367, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 6.693937900443569, | |
| "grad_norm": 1.1847096681594849, | |
| "learning_rate": 4.871166584783693e-05, | |
| "loss": 0.1372, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 6.713652045342533, | |
| "grad_norm": 0.8408295512199402, | |
| "learning_rate": 4.842545543479983e-05, | |
| "loss": 0.1242, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 6.733366190241498, | |
| "grad_norm": 1.6216832399368286, | |
| "learning_rate": 4.8139296656492605e-05, | |
| "loss": 0.1388, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 6.753080335140464, | |
| "grad_norm": 1.05503511428833, | |
| "learning_rate": 4.7853198897045424e-05, | |
| "loss": 0.137, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 6.772794480039428, | |
| "grad_norm": 1.3355563879013062, | |
| "learning_rate": 4.7567171538587514e-05, | |
| "loss": 0.1248, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 6.792508624938393, | |
| "grad_norm": 1.3196101188659668, | |
| "learning_rate": 4.728122396093935e-05, | |
| "loss": 0.1383, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 6.8122227698373585, | |
| "grad_norm": 0.7643747329711914, | |
| "learning_rate": 4.6995365541305145e-05, | |
| "loss": 0.1394, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 6.831936914736323, | |
| "grad_norm": 1.3718485832214355, | |
| "learning_rate": 4.6709605653965297e-05, | |
| "loss": 0.1421, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 6.851651059635288, | |
| "grad_norm": 0.7023568153381348, | |
| "learning_rate": 4.642395366996903e-05, | |
| "loss": 0.1338, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 6.8713652045342535, | |
| "grad_norm": 1.994978666305542, | |
| "learning_rate": 4.6138418956827026e-05, | |
| "loss": 0.1487, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 6.891079349433218, | |
| "grad_norm": 1.417363166809082, | |
| "learning_rate": 4.585301087820421e-05, | |
| "loss": 0.1271, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 6.910793494332183, | |
| "grad_norm": 1.2611218690872192, | |
| "learning_rate": 4.5567738793612813e-05, | |
| "loss": 0.1353, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 6.930507639231148, | |
| "grad_norm": 1.1835087537765503, | |
| "learning_rate": 4.528261205810527e-05, | |
| "loss": 0.1398, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 6.950221784130114, | |
| "grad_norm": 1.0265967845916748, | |
| "learning_rate": 4.499764002196758e-05, | |
| "loss": 0.1482, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 6.969935929029078, | |
| "grad_norm": 0.9783711433410645, | |
| "learning_rate": 4.471283203041255e-05, | |
| "loss": 0.1346, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 6.989650073928043, | |
| "grad_norm": 1.1431827545166016, | |
| "learning_rate": 4.4428197423273505e-05, | |
| "loss": 0.1409, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 7.007885657959586, | |
| "grad_norm": 0.9335059523582458, | |
| "learning_rate": 4.4143745534697787e-05, | |
| "loss": 0.1411, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 7.027599802858551, | |
| "grad_norm": 1.3468759059906006, | |
| "learning_rate": 4.385948569284085e-05, | |
| "loss": 0.1184, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 7.047313947757516, | |
| "grad_norm": 0.603523313999176, | |
| "learning_rate": 4.357542721956021e-05, | |
| "loss": 0.1128, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 7.067028092656481, | |
| "grad_norm": 1.1114201545715332, | |
| "learning_rate": 4.32915794301099e-05, | |
| "loss": 0.1159, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 7.086742237555446, | |
| "grad_norm": 0.6488471031188965, | |
| "learning_rate": 4.300795163283482e-05, | |
| "loss": 0.1166, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 7.106456382454411, | |
| "grad_norm": 0.9025667309761047, | |
| "learning_rate": 4.272455312886562e-05, | |
| "loss": 0.1183, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 7.126170527353376, | |
| "grad_norm": 0.7467586994171143, | |
| "learning_rate": 4.244139321181358e-05, | |
| "loss": 0.1068, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 7.1458846722523415, | |
| "grad_norm": 0.8195639848709106, | |
| "learning_rate": 4.2158481167466004e-05, | |
| "loss": 0.1164, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 7.165598817151306, | |
| "grad_norm": 1.6819171905517578, | |
| "learning_rate": 4.1875826273481474e-05, | |
| "loss": 0.1146, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 7.185312962050271, | |
| "grad_norm": 1.1047701835632324, | |
| "learning_rate": 4.1593437799085816e-05, | |
| "loss": 0.1204, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 7.205027106949236, | |
| "grad_norm": 0.9841985106468201, | |
| "learning_rate": 4.131132500476799e-05, | |
| "loss": 0.1222, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 7.224741251848201, | |
| "grad_norm": 0.8494741916656494, | |
| "learning_rate": 4.102949714197648e-05, | |
| "loss": 0.1196, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 7.244455396747166, | |
| "grad_norm": 1.0732312202453613, | |
| "learning_rate": 4.0747963452815854e-05, | |
| "loss": 0.1171, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 7.264169541646131, | |
| "grad_norm": 0.6279380917549133, | |
| "learning_rate": 4.046673316974379e-05, | |
| "loss": 0.1178, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 7.283883686545096, | |
| "grad_norm": 2.309483766555786, | |
| "learning_rate": 4.018581551526818e-05, | |
| "loss": 0.1137, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 7.303597831444061, | |
| "grad_norm": 0.8398231267929077, | |
| "learning_rate": 3.990521970164472e-05, | |
| "loss": 0.1174, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 7.323311976343026, | |
| "grad_norm": 1.108704924583435, | |
| "learning_rate": 3.9624954930574934e-05, | |
| "loss": 0.1133, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 7.343026121241991, | |
| "grad_norm": 0.9234030842781067, | |
| "learning_rate": 3.934503039290422e-05, | |
| "loss": 0.1097, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 7.362740266140956, | |
| "grad_norm": 0.9003049731254578, | |
| "learning_rate": 3.9065455268320635e-05, | |
| "loss": 0.1152, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 7.382454411039921, | |
| "grad_norm": 1.0665256977081299, | |
| "learning_rate": 3.87862387250537e-05, | |
| "loss": 0.1178, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 7.402168555938887, | |
| "grad_norm": 0.7589354515075684, | |
| "learning_rate": 3.850738991957391e-05, | |
| "loss": 0.113, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 7.421882700837851, | |
| "grad_norm": 0.8894014954566956, | |
| "learning_rate": 3.82289179962923e-05, | |
| "loss": 0.1161, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 7.441596845736816, | |
| "grad_norm": 1.0884486436843872, | |
| "learning_rate": 3.795083208726069e-05, | |
| "loss": 0.1167, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 7.4613109906357815, | |
| "grad_norm": 1.075456976890564, | |
| "learning_rate": 3.767314131187211e-05, | |
| "loss": 0.1106, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 7.481025135534746, | |
| "grad_norm": 0.5905814170837402, | |
| "learning_rate": 3.7395854776561886e-05, | |
| "loss": 0.1126, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 7.500739280433711, | |
| "grad_norm": 0.6967124342918396, | |
| "learning_rate": 3.711898157450884e-05, | |
| "loss": 0.1182, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 7.5204534253326765, | |
| "grad_norm": 0.8434427976608276, | |
| "learning_rate": 3.684253078533722e-05, | |
| "loss": 0.1213, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 7.540167570231641, | |
| "grad_norm": 0.7524256110191345, | |
| "learning_rate": 3.6566511474818874e-05, | |
| "loss": 0.1105, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 7.559881715130606, | |
| "grad_norm": 0.5648545622825623, | |
| "learning_rate": 3.629093269457602e-05, | |
| "loss": 0.1167, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 7.579595860029571, | |
| "grad_norm": 1.1991491317749023, | |
| "learning_rate": 3.6015803481784334e-05, | |
| "loss": 0.1195, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 7.599310004928537, | |
| "grad_norm": 0.7800848484039307, | |
| "learning_rate": 3.574113285887666e-05, | |
| "loss": 0.1183, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 7.619024149827501, | |
| "grad_norm": 0.8941491842269897, | |
| "learning_rate": 3.546692983324706e-05, | |
| "loss": 0.1161, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 7.638738294726466, | |
| "grad_norm": 0.6570645570755005, | |
| "learning_rate": 3.519320339695555e-05, | |
| "loss": 0.1135, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 7.658452439625432, | |
| "grad_norm": 0.989401638507843, | |
| "learning_rate": 3.491996252643305e-05, | |
| "loss": 0.111, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 7.678166584524396, | |
| "grad_norm": 1.1077444553375244, | |
| "learning_rate": 3.4647216182187205e-05, | |
| "loss": 0.1281, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 7.697880729423361, | |
| "grad_norm": 1.2130508422851562, | |
| "learning_rate": 3.437497330850835e-05, | |
| "loss": 0.1149, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 7.717594874322327, | |
| "grad_norm": 0.8594215512275696, | |
| "learning_rate": 3.410324283317633e-05, | |
| "loss": 0.1116, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 7.737309019221291, | |
| "grad_norm": 0.6998811364173889, | |
| "learning_rate": 3.3832033667167745e-05, | |
| "loss": 0.1198, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 7.757023164120256, | |
| "grad_norm": 1.1719328165054321, | |
| "learning_rate": 3.3561354704363564e-05, | |
| "loss": 0.1161, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 7.7767373090192216, | |
| "grad_norm": 0.5164292454719543, | |
| "learning_rate": 3.3291214821257635e-05, | |
| "loss": 0.1187, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 7.796451453918186, | |
| "grad_norm": 0.9778832197189331, | |
| "learning_rate": 3.30216228766655e-05, | |
| "loss": 0.1123, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 7.816165598817151, | |
| "grad_norm": 0.8570555448532104, | |
| "learning_rate": 3.2752587711433963e-05, | |
| "loss": 0.1187, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 7.8358797437161165, | |
| "grad_norm": 0.6279511451721191, | |
| "learning_rate": 3.248411814815104e-05, | |
| "loss": 0.1135, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 7.855593888615081, | |
| "grad_norm": 1.0085203647613525, | |
| "learning_rate": 3.22162229908568e-05, | |
| "loss": 0.1129, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 7.875308033514046, | |
| "grad_norm": 0.7156813144683838, | |
| "learning_rate": 3.194891102475448e-05, | |
| "loss": 0.1149, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 7.8950221784130115, | |
| "grad_norm": 0.5609560012817383, | |
| "learning_rate": 3.1682191015922554e-05, | |
| "loss": 0.1149, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 7.914736323311976, | |
| "grad_norm": 1.008686900138855, | |
| "learning_rate": 3.141607171102711e-05, | |
| "loss": 0.1174, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 7.934450468210941, | |
| "grad_norm": 0.7896909713745117, | |
| "learning_rate": 3.1150561837035146e-05, | |
| "loss": 0.1144, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 7.954164613109906, | |
| "grad_norm": 0.8853268027305603, | |
| "learning_rate": 3.088567010092827e-05, | |
| "loss": 0.1228, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 7.973878758008872, | |
| "grad_norm": 1.2333228588104248, | |
| "learning_rate": 3.0621405189417294e-05, | |
| "loss": 0.1165, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 7.993592902907836, | |
| "grad_norm": 0.6127500534057617, | |
| "learning_rate": 3.0357775768657227e-05, | |
| "loss": 0.119, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 8.011828486939379, | |
| "grad_norm": 0.6676454544067383, | |
| "learning_rate": 3.009479048396321e-05, | |
| "loss": 0.1097, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 8.031542631838343, | |
| "grad_norm": 0.5246162414550781, | |
| "learning_rate": 2.9832457959526866e-05, | |
| "loss": 0.1057, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 8.05125677673731, | |
| "grad_norm": 0.6145051121711731, | |
| "learning_rate": 2.9570786798133678e-05, | |
| "loss": 0.1045, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 8.070970921636274, | |
| "grad_norm": 0.6032987236976624, | |
| "learning_rate": 2.930978558088066e-05, | |
| "loss": 0.1099, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 8.090685066535238, | |
| "grad_norm": 0.7035672068595886, | |
| "learning_rate": 2.9049462866895104e-05, | |
| "loss": 0.1099, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 8.110399211434205, | |
| "grad_norm": 0.9585490822792053, | |
| "learning_rate": 2.8789827193053853e-05, | |
| "loss": 0.1069, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 8.130113356333169, | |
| "grad_norm": 0.7242105603218079, | |
| "learning_rate": 2.8530887073703304e-05, | |
| "loss": 0.1027, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 8.149827501232133, | |
| "grad_norm": 0.7860429883003235, | |
| "learning_rate": 2.8272651000380274e-05, | |
| "loss": 0.1033, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 8.1695416461311, | |
| "grad_norm": 0.5597553253173828, | |
| "learning_rate": 2.801512744153345e-05, | |
| "loss": 0.1046, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 8.189255791030064, | |
| "grad_norm": 0.5892042517662048, | |
| "learning_rate": 2.7758324842245755e-05, | |
| "loss": 0.1085, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 8.208969935929028, | |
| "grad_norm": 0.5419883728027344, | |
| "learning_rate": 2.7502251623957343e-05, | |
| "loss": 0.1003, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 8.228684080827994, | |
| "grad_norm": 0.5444959402084351, | |
| "learning_rate": 2.7246916184189485e-05, | |
| "loss": 0.1066, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 8.248398225726959, | |
| "grad_norm": 0.6137767434120178, | |
| "learning_rate": 2.699232689626909e-05, | |
| "loss": 0.1061, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 8.268112370625925, | |
| "grad_norm": 0.4853847920894623, | |
| "learning_rate": 2.6738492109054302e-05, | |
| "loss": 0.1024, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 8.28782651552489, | |
| "grad_norm": 1.2990504503250122, | |
| "learning_rate": 2.648542014666048e-05, | |
| "loss": 0.0968, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 8.307540660423854, | |
| "grad_norm": 0.7490946054458618, | |
| "learning_rate": 2.623311930818737e-05, | |
| "loss": 0.1099, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 8.32725480532282, | |
| "grad_norm": 0.5346263647079468, | |
| "learning_rate": 2.598159786744695e-05, | |
| "loss": 0.1018, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 8.346968950221784, | |
| "grad_norm": 0.5948059558868408, | |
| "learning_rate": 2.5730864072692018e-05, | |
| "loss": 0.1032, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 8.366683095120749, | |
| "grad_norm": 0.623589277267456, | |
| "learning_rate": 2.5480926146345785e-05, | |
| "loss": 0.1038, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 8.386397240019715, | |
| "grad_norm": 0.47425466775894165, | |
| "learning_rate": 2.523179228473219e-05, | |
| "loss": 0.1011, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 8.40611138491868, | |
| "grad_norm": 0.6056951880455017, | |
| "learning_rate": 2.4983470657807085e-05, | |
| "loss": 0.0997, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 8.425825529817644, | |
| "grad_norm": 0.5924261808395386, | |
| "learning_rate": 2.473596940889045e-05, | |
| "loss": 0.1098, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 8.44553967471661, | |
| "grad_norm": 0.6256517171859741, | |
| "learning_rate": 2.4489296654399148e-05, | |
| "loss": 0.111, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 8.465253819615574, | |
| "grad_norm": 1.0766382217407227, | |
| "learning_rate": 2.424346048358091e-05, | |
| "loss": 0.1093, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 8.484967964514539, | |
| "grad_norm": 0.6356780529022217, | |
| "learning_rate": 2.3998468958249017e-05, | |
| "loss": 0.1073, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 8.504682109413505, | |
| "grad_norm": 0.6873161196708679, | |
| "learning_rate": 2.3754330112517898e-05, | |
| "loss": 0.1044, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 8.52439625431247, | |
| "grad_norm": 0.670346736907959, | |
| "learning_rate": 2.3511051952539702e-05, | |
| "loss": 0.106, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 8.544110399211434, | |
| "grad_norm": 0.626423180103302, | |
| "learning_rate": 2.3268642456241752e-05, | |
| "loss": 0.0985, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 8.5638245441104, | |
| "grad_norm": 0.6810294985771179, | |
| "learning_rate": 2.302710957306482e-05, | |
| "loss": 0.1065, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 8.583538689009364, | |
| "grad_norm": 0.7520167827606201, | |
| "learning_rate": 2.278646122370263e-05, | |
| "loss": 0.105, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 8.603252833908329, | |
| "grad_norm": 0.8657585382461548, | |
| "learning_rate": 2.2546705299841958e-05, | |
| "loss": 0.1014, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 8.622966978807295, | |
| "grad_norm": 0.7666080594062805, | |
| "learning_rate": 2.230784966390381e-05, | |
| "loss": 0.1058, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 8.642681123706259, | |
| "grad_norm": 0.7878192067146301, | |
| "learning_rate": 2.20699021487858e-05, | |
| "loss": 0.1022, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 8.662395268605223, | |
| "grad_norm": 0.601855993270874, | |
| "learning_rate": 2.183287055760496e-05, | |
| "loss": 0.0985, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 8.68210941350419, | |
| "grad_norm": 0.6196221709251404, | |
| "learning_rate": 2.1596762663442218e-05, | |
| "loss": 0.1007, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 8.701823558403154, | |
| "grad_norm": 0.6262930631637573, | |
| "learning_rate": 2.1361586209087136e-05, | |
| "loss": 0.1034, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 8.721537703302118, | |
| "grad_norm": 0.8374412059783936, | |
| "learning_rate": 2.1127348906784255e-05, | |
| "loss": 0.1043, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 8.741251848201085, | |
| "grad_norm": 0.5195895433425903, | |
| "learning_rate": 2.0894058437980068e-05, | |
| "loss": 0.1029, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 8.760965993100049, | |
| "grad_norm": 0.7010400891304016, | |
| "learning_rate": 2.0661722453071143e-05, | |
| "loss": 0.1047, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 8.780680137999013, | |
| "grad_norm": 0.5973253846168518, | |
| "learning_rate": 2.043034857115323e-05, | |
| "loss": 0.1112, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 8.80039428289798, | |
| "grad_norm": 0.7826218008995056, | |
| "learning_rate": 2.0199944379771447e-05, | |
| "loss": 0.0996, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 8.820108427796944, | |
| "grad_norm": 1.0917187929153442, | |
| "learning_rate": 1.9970517434671336e-05, | |
| "loss": 0.1088, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 8.83982257269591, | |
| "grad_norm": 0.6169748306274414, | |
| "learning_rate": 1.974207525955133e-05, | |
| "loss": 0.111, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 8.859536717594874, | |
| "grad_norm": 1.227509617805481, | |
| "learning_rate": 1.95146253458157e-05, | |
| "loss": 0.1077, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 8.879250862493839, | |
| "grad_norm": 0.6013702750205994, | |
| "learning_rate": 1.9288175152329163e-05, | |
| "loss": 0.1038, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 8.898965007392805, | |
| "grad_norm": 0.622795820236206, | |
| "learning_rate": 1.9062732105172138e-05, | |
| "loss": 0.105, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 8.91867915229177, | |
| "grad_norm": 0.5975040793418884, | |
| "learning_rate": 1.883830359739725e-05, | |
| "loss": 0.108, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 8.938393297190734, | |
| "grad_norm": 0.646460771560669, | |
| "learning_rate": 1.8614896988786902e-05, | |
| "loss": 0.1038, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 8.9581074420897, | |
| "grad_norm": 0.5937023162841797, | |
| "learning_rate": 1.8392519605611907e-05, | |
| "loss": 0.1072, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 8.977821586988664, | |
| "grad_norm": 0.6777051687240601, | |
| "learning_rate": 1.8171178740391194e-05, | |
| "loss": 0.1056, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 8.997535731887629, | |
| "grad_norm": 1.1725229024887085, | |
| "learning_rate": 1.7950881651652783e-05, | |
| "loss": 0.1014, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 9.015771315919173, | |
| "grad_norm": 0.47024473547935486, | |
| "learning_rate": 1.7731635563695643e-05, | |
| "loss": 0.0966, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 9.035485460818137, | |
| "grad_norm": 0.564164936542511, | |
| "learning_rate": 1.7513447666352752e-05, | |
| "loss": 0.1032, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 9.055199605717101, | |
| "grad_norm": 0.6236717700958252, | |
| "learning_rate": 1.7296325114755503e-05, | |
| "loss": 0.0964, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 9.074913750616068, | |
| "grad_norm": 0.5206897854804993, | |
| "learning_rate": 1.7080275029098812e-05, | |
| "loss": 0.1033, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 9.094627895515032, | |
| "grad_norm": 0.633885383605957, | |
| "learning_rate": 1.686530449440783e-05, | |
| "loss": 0.0939, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 9.114342040413996, | |
| "grad_norm": 0.6955597400665283, | |
| "learning_rate": 1.6651420560305497e-05, | |
| "loss": 0.0992, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 9.134056185312962, | |
| "grad_norm": 0.5905971527099609, | |
| "learning_rate": 1.6438630240781378e-05, | |
| "loss": 0.0989, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 9.153770330211927, | |
| "grad_norm": 0.501326322555542, | |
| "learning_rate": 1.6226940513961664e-05, | |
| "loss": 0.091, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 9.173484475110891, | |
| "grad_norm": 0.5881623029708862, | |
| "learning_rate": 1.6016358321880326e-05, | |
| "loss": 0.0976, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 9.193198620009857, | |
| "grad_norm": 0.6199246048927307, | |
| "learning_rate": 1.580689057025146e-05, | |
| "loss": 0.1007, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 9.212912764908822, | |
| "grad_norm": 0.6353612542152405, | |
| "learning_rate": 1.559854412824284e-05, | |
| "loss": 0.0959, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 9.232626909807786, | |
| "grad_norm": 0.5803267359733582, | |
| "learning_rate": 1.5391325828250596e-05, | |
| "loss": 0.0962, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 9.252341054706752, | |
| "grad_norm": 0.5344117283821106, | |
| "learning_rate": 1.5185242465675297e-05, | |
| "loss": 0.0961, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 9.272055199605717, | |
| "grad_norm": 0.5603470802307129, | |
| "learning_rate": 1.4980300798698904e-05, | |
| "loss": 0.1016, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 9.291769344504683, | |
| "grad_norm": 0.6176356077194214, | |
| "learning_rate": 1.4776507548063318e-05, | |
| "loss": 0.1071, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 9.311483489403647, | |
| "grad_norm": 0.6276616454124451, | |
| "learning_rate": 1.4573869396849904e-05, | |
| "loss": 0.0968, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 9.331197634302612, | |
| "grad_norm": 0.614722490310669, | |
| "learning_rate": 1.4372392990260353e-05, | |
| "loss": 0.0987, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 9.350911779201578, | |
| "grad_norm": 0.6439083218574524, | |
| "learning_rate": 1.4172084935398744e-05, | |
| "loss": 0.0985, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 9.370625924100542, | |
| "grad_norm": 0.6016230583190918, | |
| "learning_rate": 1.3972951801054912e-05, | |
| "loss": 0.1004, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 9.390340068999507, | |
| "grad_norm": 0.5583289861679077, | |
| "learning_rate": 1.3775000117488934e-05, | |
| "loss": 0.0924, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 9.410054213898473, | |
| "grad_norm": 0.5593780875205994, | |
| "learning_rate": 1.3578236376217173e-05, | |
| "loss": 0.0939, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 9.429768358797437, | |
| "grad_norm": 0.6913323998451233, | |
| "learning_rate": 1.3382667029799167e-05, | |
| "loss": 0.0977, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 9.449482503696402, | |
| "grad_norm": 0.7171539664268494, | |
| "learning_rate": 1.3188298491626183e-05, | |
| "loss": 0.0974, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 9.469196648595368, | |
| "grad_norm": 0.6020176410675049, | |
| "learning_rate": 1.2995137135710916e-05, | |
| "loss": 0.0978, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 9.488910793494332, | |
| "grad_norm": 0.633091151714325, | |
| "learning_rate": 1.2803189296478302e-05, | |
| "loss": 0.1047, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 9.508624938393297, | |
| "grad_norm": 0.5339157581329346, | |
| "learning_rate": 1.2612461268557957e-05, | |
| "loss": 0.095, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 9.528339083292263, | |
| "grad_norm": 0.6085830926895142, | |
| "learning_rate": 1.2422959306577664e-05, | |
| "loss": 0.0987, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 9.548053228191227, | |
| "grad_norm": 0.5931851267814636, | |
| "learning_rate": 1.2234689624958306e-05, | |
| "loss": 0.0985, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 9.567767373090192, | |
| "grad_norm": 0.611231803894043, | |
| "learning_rate": 1.204765839771005e-05, | |
| "loss": 0.1024, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 9.587481517989158, | |
| "grad_norm": 0.5894888043403625, | |
| "learning_rate": 1.1861871758229893e-05, | |
| "loss": 0.0946, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 9.607195662888122, | |
| "grad_norm": 0.5542821288108826, | |
| "learning_rate": 1.1677335799100469e-05, | |
| "loss": 0.094, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 9.626909807787086, | |
| "grad_norm": 0.5735796093940735, | |
| "learning_rate": 1.1494056571890405e-05, | |
| "loss": 0.103, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 9.646623952686053, | |
| "grad_norm": 0.6378740072250366, | |
| "learning_rate": 1.1312040086955656e-05, | |
| "loss": 0.0993, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 9.666338097585017, | |
| "grad_norm": 0.5262072682380676, | |
| "learning_rate": 1.1131292313242614e-05, | |
| "loss": 0.0919, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 9.686052242483981, | |
| "grad_norm": 0.6029082536697388, | |
| "learning_rate": 1.0951819178092188e-05, | |
| "loss": 0.094, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 9.705766387382948, | |
| "grad_norm": 0.6229907870292664, | |
| "learning_rate": 1.0773626567045547e-05, | |
| "loss": 0.0973, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 9.725480532281912, | |
| "grad_norm": 0.5560586452484131, | |
| "learning_rate": 1.0596720323651077e-05, | |
| "loss": 0.0943, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 9.745194677180876, | |
| "grad_norm": 0.651473343372345, | |
| "learning_rate": 1.0421106249272722e-05, | |
| "loss": 0.0961, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 9.764908822079843, | |
| "grad_norm": 0.7311715483665466, | |
| "learning_rate": 1.0246790102899784e-05, | |
| "loss": 0.1024, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 9.784622966978807, | |
| "grad_norm": 0.5268740653991699, | |
| "learning_rate": 1.007377760095804e-05, | |
| "loss": 0.0907, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 9.804337111877773, | |
| "grad_norm": 0.6558658480644226, | |
| "learning_rate": 9.90207441712223e-06, | |
| "loss": 0.105, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 9.824051256776738, | |
| "grad_norm": 0.7688642740249634, | |
| "learning_rate": 9.731686182130168e-06, | |
| "loss": 0.1001, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 9.843765401675702, | |
| "grad_norm": 0.5601643919944763, | |
| "learning_rate": 9.562618483597852e-06, | |
| "loss": 0.0959, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 9.863479546574668, | |
| "grad_norm": 0.6482700705528259, | |
| "learning_rate": 9.394876865836411e-06, | |
| "loss": 0.1044, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 9.883193691473632, | |
| "grad_norm": 0.6158556938171387, | |
| "learning_rate": 9.228466829670246e-06, | |
| "loss": 0.0998, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 9.902907836372597, | |
| "grad_norm": 0.6211926341056824, | |
| "learning_rate": 9.063393832256562e-06, | |
| "loss": 0.1007, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 9.922621981271563, | |
| "grad_norm": 0.577765703201294, | |
| "learning_rate": 8.899663286906496e-06, | |
| "loss": 0.0908, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 9.942336126170527, | |
| "grad_norm": 0.6280847787857056, | |
| "learning_rate": 8.737280562907574e-06, | |
| "loss": 0.1005, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 9.962050271069492, | |
| "grad_norm": 0.6178187727928162, | |
| "learning_rate": 8.576250985347616e-06, | |
| "loss": 0.1055, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 9.981764415968458, | |
| "grad_norm": 0.7428198456764221, | |
| "learning_rate": 8.416579834940113e-06, | |
| "loss": 0.1005, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 0.7962881326675415, | |
| "learning_rate": 8.258272347851087e-06, | |
| "loss": 0.0921, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 10.019714144898964, | |
| "grad_norm": 0.6026802659034729, | |
| "learning_rate": 8.101333715527283e-06, | |
| "loss": 0.0943, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 10.03942828979793, | |
| "grad_norm": 0.6051628589630127, | |
| "learning_rate": 7.945769084526088e-06, | |
| "loss": 0.0881, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 10.059142434696895, | |
| "grad_norm": 0.6173446774482727, | |
| "learning_rate": 7.791583556346576e-06, | |
| "loss": 0.0965, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 10.07885657959586, | |
| "grad_norm": 0.5841408967971802, | |
| "learning_rate": 7.638782187262355e-06, | |
| "loss": 0.0964, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 10.098570724494826, | |
| "grad_norm": 0.5438782572746277, | |
| "learning_rate": 7.487369988155662e-06, | |
| "loss": 0.094, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 10.11828486939379, | |
| "grad_norm": 0.5726168751716614, | |
| "learning_rate": 7.3373519243530975e-06, | |
| "loss": 0.0977, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 10.137999014292754, | |
| "grad_norm": 0.5545468926429749, | |
| "learning_rate": 7.188732915462754e-06, | |
| "loss": 0.0885, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 10.15771315919172, | |
| "grad_norm": 0.600556492805481, | |
| "learning_rate": 7.0415178352129185e-06, | |
| "loss": 0.0913, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 10.177427304090685, | |
| "grad_norm": 0.6833725571632385, | |
| "learning_rate": 6.8957115112922245e-06, | |
| "loss": 0.0997, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 10.19714144898965, | |
| "grad_norm": 0.6429631114006042, | |
| "learning_rate": 6.7513187251913466e-06, | |
| "loss": 0.0939, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 10.216855593888615, | |
| "grad_norm": 0.5843133330345154, | |
| "learning_rate": 6.608344212046164e-06, | |
| "loss": 0.0969, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 10.23656973878758, | |
| "grad_norm": 0.5661614537239075, | |
| "learning_rate": 6.466792660482574e-06, | |
| "loss": 0.0945, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 10.256283883686546, | |
| "grad_norm": 0.6150486469268799, | |
| "learning_rate": 6.326668712462597e-06, | |
| "loss": 0.0907, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 10.27599802858551, | |
| "grad_norm": 0.5596334934234619, | |
| "learning_rate": 6.187976963132275e-06, | |
| "loss": 0.09, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 10.295712173484475, | |
| "grad_norm": 0.7830381393432617, | |
| "learning_rate": 6.050721960670918e-06, | |
| "loss": 0.1001, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 10.315426318383441, | |
| "grad_norm": 0.5723216533660889, | |
| "learning_rate": 5.914908206141956e-06, | |
| "loss": 0.088, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 10.335140463282405, | |
| "grad_norm": 0.6770078539848328, | |
| "learning_rate": 5.780540153345354e-06, | |
| "loss": 0.0976, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 10.35485460818137, | |
| "grad_norm": 0.5593023896217346, | |
| "learning_rate": 5.647622208671555e-06, | |
| "loss": 0.0909, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 10.374568753080336, | |
| "grad_norm": 0.5808446407318115, | |
| "learning_rate": 5.516158730956944e-06, | |
| "loss": 0.0956, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 10.3942828979793, | |
| "grad_norm": 0.59349524974823, | |
| "learning_rate": 5.386154031340962e-06, | |
| "loss": 0.0912, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 10.413997042878265, | |
| "grad_norm": 0.5928412675857544, | |
| "learning_rate": 5.257612373124699e-06, | |
| "loss": 0.0913, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 10.43371118777723, | |
| "grad_norm": 0.618300199508667, | |
| "learning_rate": 5.130537971631039e-06, | |
| "loss": 0.0902, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 10.453425332676195, | |
| "grad_norm": 0.6695954203605652, | |
| "learning_rate": 5.004934994066557e-06, | |
| "loss": 0.0961, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 10.47313947757516, | |
| "grad_norm": 0.7120895385742188, | |
| "learning_rate": 4.880807559384704e-06, | |
| "loss": 0.0961, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 10.492853622474126, | |
| "grad_norm": 0.5676897764205933, | |
| "learning_rate": 4.758159738150841e-06, | |
| "loss": 0.0956, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 10.51256776737309, | |
| "grad_norm": 0.5924949645996094, | |
| "learning_rate": 4.636995552408713e-06, | |
| "loss": 0.0914, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 10.532281912272055, | |
| "grad_norm": 0.5598572492599487, | |
| "learning_rate": 4.517318975548557e-06, | |
| "loss": 0.091, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 10.55199605717102, | |
| "grad_norm": 0.5901981592178345, | |
| "learning_rate": 4.399133932176797e-06, | |
| "loss": 0.0885, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 10.571710202069985, | |
| "grad_norm": 0.6229949593544006, | |
| "learning_rate": 4.282444297987359e-06, | |
| "loss": 0.0924, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 10.59142434696895, | |
| "grad_norm": 0.6218413710594177, | |
| "learning_rate": 4.16725389963451e-06, | |
| "loss": 0.0923, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 10.611138491867916, | |
| "grad_norm": 0.7330309748649597, | |
| "learning_rate": 4.053566514607504e-06, | |
| "loss": 0.0962, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 10.63085263676688, | |
| "grad_norm": 0.6477744579315186, | |
| "learning_rate": 3.941385871106534e-06, | |
| "loss": 0.0925, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 10.650566781665844, | |
| "grad_norm": 0.582679033279419, | |
| "learning_rate": 3.830715647920629e-06, | |
| "loss": 0.0958, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 10.67028092656481, | |
| "grad_norm": 0.5494212508201599, | |
| "learning_rate": 3.721559474306896e-06, | |
| "loss": 0.0984, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 10.689995071463775, | |
| "grad_norm": 0.5505300164222717, | |
| "learning_rate": 3.6139209298715616e-06, | |
| "loss": 0.094, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 10.70970921636274, | |
| "grad_norm": 0.752632200717926, | |
| "learning_rate": 3.507803544452587e-06, | |
| "loss": 0.0945, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 10.729423361261706, | |
| "grad_norm": 0.6900667548179626, | |
| "learning_rate": 3.403210798003892e-06, | |
| "loss": 0.0923, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 10.74913750616067, | |
| "grad_norm": 0.66816246509552, | |
| "learning_rate": 3.300146120481229e-06, | |
| "loss": 0.0939, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 10.768851651059634, | |
| "grad_norm": 0.6422622203826904, | |
| "learning_rate": 3.1986128917297464e-06, | |
| "loss": 0.0926, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 10.7885657959586, | |
| "grad_norm": 0.6510041356086731, | |
| "learning_rate": 3.0986144413730876e-06, | |
| "loss": 0.0929, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 10.808279940857565, | |
| "grad_norm": 0.5580385327339172, | |
| "learning_rate": 3.000154048704268e-06, | |
| "loss": 0.096, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 10.827994085756531, | |
| "grad_norm": 0.6091373562812805, | |
| "learning_rate": 2.9032349425780813e-06, | |
| "loss": 0.0931, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 10.847708230655495, | |
| "grad_norm": 0.6932477951049805, | |
| "learning_rate": 2.8078603013052263e-06, | |
| "loss": 0.0966, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 10.86742237555446, | |
| "grad_norm": 0.592022716999054, | |
| "learning_rate": 2.714033252548137e-06, | |
| "loss": 0.0964, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 10.887136520453426, | |
| "grad_norm": 0.6746100187301636, | |
| "learning_rate": 2.621756873218312e-06, | |
| "loss": 0.0924, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 10.90685066535239, | |
| "grad_norm": 0.6170192360877991, | |
| "learning_rate": 2.531034189375503e-06, | |
| "loss": 0.0915, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 10.926564810251355, | |
| "grad_norm": 0.6232552528381348, | |
| "learning_rate": 2.4418681761284255e-06, | |
| "loss": 0.0916, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 10.946278955150321, | |
| "grad_norm": 0.5538495779037476, | |
| "learning_rate": 2.354261757537235e-06, | |
| "loss": 0.0926, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 10.965993100049285, | |
| "grad_norm": 0.5978652834892273, | |
| "learning_rate": 2.268217806517592e-06, | |
| "loss": 0.0878, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 10.98570724494825, | |
| "grad_norm": 0.5522828102111816, | |
| "learning_rate": 2.183739144746494e-06, | |
| "loss": 0.093, | |
| "step": 2790 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 3048, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 12, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 4.384719421932749e+16, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |