sanjaylb's picture
Upload folder using huggingface_hub
65f1386 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 11.0,
"eval_steps": 500,
"global_step": 2794,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.019714144898965006,
"grad_norm": 2.673164129257202,
"learning_rate": 1.3114754098360657e-06,
"loss": 3.8588,
"step": 5
},
{
"epoch": 0.03942828979793001,
"grad_norm": 2.886237621307373,
"learning_rate": 2.9508196721311478e-06,
"loss": 3.8163,
"step": 10
},
{
"epoch": 0.05914243469689502,
"grad_norm": 2.4650022983551025,
"learning_rate": 4.59016393442623e-06,
"loss": 3.7443,
"step": 15
},
{
"epoch": 0.07885657959586002,
"grad_norm": 2.4647622108459473,
"learning_rate": 6.229508196721312e-06,
"loss": 3.7573,
"step": 20
},
{
"epoch": 0.09857072449482504,
"grad_norm": 2.5846645832061768,
"learning_rate": 7.868852459016394e-06,
"loss": 3.6941,
"step": 25
},
{
"epoch": 0.11828486939379004,
"grad_norm": 2.3263237476348877,
"learning_rate": 9.508196721311476e-06,
"loss": 3.5506,
"step": 30
},
{
"epoch": 0.13799901429275505,
"grad_norm": 2.181896448135376,
"learning_rate": 1.1147540983606558e-05,
"loss": 3.4485,
"step": 35
},
{
"epoch": 0.15771315919172005,
"grad_norm": 2.374994993209839,
"learning_rate": 1.2786885245901639e-05,
"loss": 3.269,
"step": 40
},
{
"epoch": 0.17742730409068508,
"grad_norm": 2.852327585220337,
"learning_rate": 1.4426229508196722e-05,
"loss": 3.1914,
"step": 45
},
{
"epoch": 0.19714144898965008,
"grad_norm": 2.855822801589966,
"learning_rate": 1.6065573770491805e-05,
"loss": 2.9482,
"step": 50
},
{
"epoch": 0.21685559388861508,
"grad_norm": 3.2098145484924316,
"learning_rate": 1.7704918032786887e-05,
"loss": 2.7581,
"step": 55
},
{
"epoch": 0.23656973878758009,
"grad_norm": 2.033719062805176,
"learning_rate": 1.934426229508197e-05,
"loss": 2.5988,
"step": 60
},
{
"epoch": 0.2562838836865451,
"grad_norm": 2.0503222942352295,
"learning_rate": 2.098360655737705e-05,
"loss": 2.6184,
"step": 65
},
{
"epoch": 0.2759980285855101,
"grad_norm": 1.858931303024292,
"learning_rate": 2.262295081967213e-05,
"loss": 2.5111,
"step": 70
},
{
"epoch": 0.2957121734844751,
"grad_norm": 1.6557573080062866,
"learning_rate": 2.4262295081967215e-05,
"loss": 2.4524,
"step": 75
},
{
"epoch": 0.3154263183834401,
"grad_norm": 1.706704020500183,
"learning_rate": 2.5901639344262297e-05,
"loss": 2.4317,
"step": 80
},
{
"epoch": 0.3351404632824051,
"grad_norm": 1.4746320247650146,
"learning_rate": 2.754098360655738e-05,
"loss": 2.3915,
"step": 85
},
{
"epoch": 0.35485460818137016,
"grad_norm": 1.7562377452850342,
"learning_rate": 2.9180327868852458e-05,
"loss": 2.3791,
"step": 90
},
{
"epoch": 0.37456875308033516,
"grad_norm": 1.7075871229171753,
"learning_rate": 3.0819672131147544e-05,
"loss": 2.3717,
"step": 95
},
{
"epoch": 0.39428289797930016,
"grad_norm": 1.5588061809539795,
"learning_rate": 3.245901639344263e-05,
"loss": 2.3924,
"step": 100
},
{
"epoch": 0.41399704287826516,
"grad_norm": 1.5635536909103394,
"learning_rate": 3.409836065573771e-05,
"loss": 2.3031,
"step": 105
},
{
"epoch": 0.43371118777723017,
"grad_norm": 1.7249213457107544,
"learning_rate": 3.5737704918032786e-05,
"loss": 2.3186,
"step": 110
},
{
"epoch": 0.45342533267619517,
"grad_norm": 1.868545651435852,
"learning_rate": 3.737704918032787e-05,
"loss": 2.1715,
"step": 115
},
{
"epoch": 0.47313947757516017,
"grad_norm": 1.9135463237762451,
"learning_rate": 3.901639344262295e-05,
"loss": 2.2488,
"step": 120
},
{
"epoch": 0.4928536224741252,
"grad_norm": 1.895492434501648,
"learning_rate": 4.0655737704918036e-05,
"loss": 2.2387,
"step": 125
},
{
"epoch": 0.5125677673730902,
"grad_norm": 1.767385721206665,
"learning_rate": 4.229508196721312e-05,
"loss": 2.2541,
"step": 130
},
{
"epoch": 0.5322819122720552,
"grad_norm": 2.0554378032684326,
"learning_rate": 4.3934426229508194e-05,
"loss": 2.1672,
"step": 135
},
{
"epoch": 0.5519960571710202,
"grad_norm": 2.3369956016540527,
"learning_rate": 4.557377049180328e-05,
"loss": 2.223,
"step": 140
},
{
"epoch": 0.5717102020699852,
"grad_norm": 2.0532915592193604,
"learning_rate": 4.7213114754098365e-05,
"loss": 2.1007,
"step": 145
},
{
"epoch": 0.5914243469689502,
"grad_norm": 2.2023046016693115,
"learning_rate": 4.885245901639344e-05,
"loss": 2.1027,
"step": 150
},
{
"epoch": 0.6111384918679152,
"grad_norm": 2.194356918334961,
"learning_rate": 5.049180327868853e-05,
"loss": 2.1188,
"step": 155
},
{
"epoch": 0.6308526367668802,
"grad_norm": 2.6036267280578613,
"learning_rate": 5.213114754098361e-05,
"loss": 1.9685,
"step": 160
},
{
"epoch": 0.6505667816658453,
"grad_norm": 2.6643617153167725,
"learning_rate": 5.3770491803278686e-05,
"loss": 2.0843,
"step": 165
},
{
"epoch": 0.6702809265648102,
"grad_norm": 2.3738605976104736,
"learning_rate": 5.540983606557377e-05,
"loss": 2.0489,
"step": 170
},
{
"epoch": 0.6899950714637753,
"grad_norm": 2.43137526512146,
"learning_rate": 5.704918032786886e-05,
"loss": 1.976,
"step": 175
},
{
"epoch": 0.7097092163627403,
"grad_norm": 3.0450685024261475,
"learning_rate": 5.868852459016394e-05,
"loss": 1.8117,
"step": 180
},
{
"epoch": 0.7294233612617053,
"grad_norm": 2.9214789867401123,
"learning_rate": 6.032786885245902e-05,
"loss": 1.8229,
"step": 185
},
{
"epoch": 0.7491375061606703,
"grad_norm": 2.7570457458496094,
"learning_rate": 6.19672131147541e-05,
"loss": 1.8662,
"step": 190
},
{
"epoch": 0.7688516510596353,
"grad_norm": 4.113077640533447,
"learning_rate": 6.360655737704918e-05,
"loss": 1.7928,
"step": 195
},
{
"epoch": 0.7885657959586003,
"grad_norm": 3.127991199493408,
"learning_rate": 6.524590163934427e-05,
"loss": 1.8187,
"step": 200
},
{
"epoch": 0.8082799408575653,
"grad_norm": 3.4384043216705322,
"learning_rate": 6.688524590163935e-05,
"loss": 1.7548,
"step": 205
},
{
"epoch": 0.8279940857565303,
"grad_norm": 3.2650253772735596,
"learning_rate": 6.852459016393443e-05,
"loss": 1.8635,
"step": 210
},
{
"epoch": 0.8477082306554953,
"grad_norm": 3.676208019256592,
"learning_rate": 7.016393442622952e-05,
"loss": 1.6462,
"step": 215
},
{
"epoch": 0.8674223755544603,
"grad_norm": 3.6363656520843506,
"learning_rate": 7.180327868852459e-05,
"loss": 1.6826,
"step": 220
},
{
"epoch": 0.8871365204534253,
"grad_norm": 3.487661123275757,
"learning_rate": 7.344262295081968e-05,
"loss": 1.7096,
"step": 225
},
{
"epoch": 0.9068506653523903,
"grad_norm": 4.129843235015869,
"learning_rate": 7.508196721311476e-05,
"loss": 1.6489,
"step": 230
},
{
"epoch": 0.9265648102513554,
"grad_norm": 3.7981042861938477,
"learning_rate": 7.672131147540984e-05,
"loss": 1.5388,
"step": 235
},
{
"epoch": 0.9462789551503203,
"grad_norm": 4.129542827606201,
"learning_rate": 7.836065573770493e-05,
"loss": 1.5883,
"step": 240
},
{
"epoch": 0.9659931000492854,
"grad_norm": 3.7202744483947754,
"learning_rate": 8e-05,
"loss": 1.5327,
"step": 245
},
{
"epoch": 0.9857072449482503,
"grad_norm": 5.224013805389404,
"learning_rate": 8.163934426229509e-05,
"loss": 1.5075,
"step": 250
},
{
"epoch": 1.003942828979793,
"grad_norm": 4.3526787757873535,
"learning_rate": 8.327868852459016e-05,
"loss": 1.4976,
"step": 255
},
{
"epoch": 1.023656973878758,
"grad_norm": 5.239654541015625,
"learning_rate": 8.491803278688524e-05,
"loss": 1.3648,
"step": 260
},
{
"epoch": 1.043371118777723,
"grad_norm": 4.757369518280029,
"learning_rate": 8.655737704918033e-05,
"loss": 1.3312,
"step": 265
},
{
"epoch": 1.063085263676688,
"grad_norm": 4.107004642486572,
"learning_rate": 8.819672131147541e-05,
"loss": 1.3081,
"step": 270
},
{
"epoch": 1.082799408575653,
"grad_norm": 3.8866994380950928,
"learning_rate": 8.98360655737705e-05,
"loss": 1.3473,
"step": 275
},
{
"epoch": 1.102513553474618,
"grad_norm": 4.225423812866211,
"learning_rate": 9.147540983606557e-05,
"loss": 1.3429,
"step": 280
},
{
"epoch": 1.122227698373583,
"grad_norm": 4.068089485168457,
"learning_rate": 9.311475409836066e-05,
"loss": 1.3851,
"step": 285
},
{
"epoch": 1.141941843272548,
"grad_norm": 4.4080986976623535,
"learning_rate": 9.475409836065574e-05,
"loss": 1.1531,
"step": 290
},
{
"epoch": 1.1616559881715132,
"grad_norm": 3.994210720062256,
"learning_rate": 9.639344262295082e-05,
"loss": 1.1968,
"step": 295
},
{
"epoch": 1.181370133070478,
"grad_norm": 4.020788669586182,
"learning_rate": 9.803278688524591e-05,
"loss": 1.1876,
"step": 300
},
{
"epoch": 1.201084277969443,
"grad_norm": 5.0161848068237305,
"learning_rate": 9.967213114754099e-05,
"loss": 1.135,
"step": 305
},
{
"epoch": 1.2207984228684081,
"grad_norm": 4.258621692657471,
"learning_rate": 9.999947530447293e-05,
"loss": 1.0355,
"step": 310
},
{
"epoch": 1.2405125677673732,
"grad_norm": 4.326329708099365,
"learning_rate": 9.99973437477677e-05,
"loss": 1.0635,
"step": 315
},
{
"epoch": 1.2602267126663382,
"grad_norm": 3.8772778511047363,
"learning_rate": 9.999357260626116e-05,
"loss": 1.1091,
"step": 320
},
{
"epoch": 1.279940857565303,
"grad_norm": 4.675949573516846,
"learning_rate": 9.998816200362199e-05,
"loss": 1.0763,
"step": 325
},
{
"epoch": 1.2996550024642681,
"grad_norm": 4.871826171875,
"learning_rate": 9.998111211728248e-05,
"loss": 1.016,
"step": 330
},
{
"epoch": 1.3193691473632332,
"grad_norm": 4.208927154541016,
"learning_rate": 9.99724231784326e-05,
"loss": 0.9334,
"step": 335
},
{
"epoch": 1.339083292262198,
"grad_norm": 3.859124183654785,
"learning_rate": 9.996209547201258e-05,
"loss": 0.9612,
"step": 340
},
{
"epoch": 1.358797437161163,
"grad_norm": 4.167286396026611,
"learning_rate": 9.99501293367034e-05,
"loss": 1.071,
"step": 345
},
{
"epoch": 1.3785115820601281,
"grad_norm": 4.229668617248535,
"learning_rate": 9.993652516491579e-05,
"loss": 0.9526,
"step": 350
},
{
"epoch": 1.3982257269590932,
"grad_norm": 6.326845645904541,
"learning_rate": 9.992128340277729e-05,
"loss": 0.8992,
"step": 355
},
{
"epoch": 1.4179398718580583,
"grad_norm": 4.119451522827148,
"learning_rate": 9.990440455011769e-05,
"loss": 0.9638,
"step": 360
},
{
"epoch": 1.437654016757023,
"grad_norm": 4.288192272186279,
"learning_rate": 9.988588916045263e-05,
"loss": 0.849,
"step": 365
},
{
"epoch": 1.4573681616559881,
"grad_norm": 3.985097885131836,
"learning_rate": 9.986573784096536e-05,
"loss": 0.8638,
"step": 370
},
{
"epoch": 1.4770823065549532,
"grad_norm": 3.398552894592285,
"learning_rate": 9.984395125248695e-05,
"loss": 0.8869,
"step": 375
},
{
"epoch": 1.4967964514539183,
"grad_norm": 5.5323486328125,
"learning_rate": 9.982053010947455e-05,
"loss": 0.8194,
"step": 380
},
{
"epoch": 1.5165105963528833,
"grad_norm": 4.199021816253662,
"learning_rate": 9.979547517998796e-05,
"loss": 0.7118,
"step": 385
},
{
"epoch": 1.5362247412518482,
"grad_norm": 4.152290344238281,
"learning_rate": 9.976878728566443e-05,
"loss": 0.8509,
"step": 390
},
{
"epoch": 1.5559388861508132,
"grad_norm": 3.820864200592041,
"learning_rate": 9.974046730169183e-05,
"loss": 0.768,
"step": 395
},
{
"epoch": 1.5756530310497783,
"grad_norm": 3.8033151626586914,
"learning_rate": 9.971051615677974e-05,
"loss": 0.7861,
"step": 400
},
{
"epoch": 1.595367175948743,
"grad_norm": 4.548523902893066,
"learning_rate": 9.967893483312922e-05,
"loss": 0.8347,
"step": 405
},
{
"epoch": 1.6150813208477084,
"grad_norm": 3.6572163105010986,
"learning_rate": 9.964572436640045e-05,
"loss": 0.7201,
"step": 410
},
{
"epoch": 1.6347954657466732,
"grad_norm": 3.980424404144287,
"learning_rate": 9.961088584567882e-05,
"loss": 0.7655,
"step": 415
},
{
"epoch": 1.6545096106456383,
"grad_norm": 3.2318334579467773,
"learning_rate": 9.95744204134392e-05,
"loss": 0.8285,
"step": 420
},
{
"epoch": 1.6742237555446033,
"grad_norm": 4.426093578338623,
"learning_rate": 9.953632926550847e-05,
"loss": 0.6971,
"step": 425
},
{
"epoch": 1.6939379004435682,
"grad_norm": 3.0529074668884277,
"learning_rate": 9.949661365102637e-05,
"loss": 0.7098,
"step": 430
},
{
"epoch": 1.7136520453425332,
"grad_norm": 3.948803424835205,
"learning_rate": 9.945527487240442e-05,
"loss": 0.7815,
"step": 435
},
{
"epoch": 1.7333661902414983,
"grad_norm": 3.2549209594726562,
"learning_rate": 9.941231428528332e-05,
"loss": 0.7385,
"step": 440
},
{
"epoch": 1.7530803351404631,
"grad_norm": 3.8000028133392334,
"learning_rate": 9.93677332984884e-05,
"loss": 0.7357,
"step": 445
},
{
"epoch": 1.7727944800394284,
"grad_norm": 3.5841259956359863,
"learning_rate": 9.932153337398354e-05,
"loss": 0.6641,
"step": 450
},
{
"epoch": 1.7925086249383932,
"grad_norm": 4.315596580505371,
"learning_rate": 9.927371602682305e-05,
"loss": 0.6193,
"step": 455
},
{
"epoch": 1.8122227698373583,
"grad_norm": 3.2834675312042236,
"learning_rate": 9.92242828251022e-05,
"loss": 0.5768,
"step": 460
},
{
"epoch": 1.8319369147363234,
"grad_norm": 3.127601385116577,
"learning_rate": 9.917323538990561e-05,
"loss": 0.5912,
"step": 465
},
{
"epoch": 1.8516510596352882,
"grad_norm": 3.2609682083129883,
"learning_rate": 9.912057539525419e-05,
"loss": 0.6093,
"step": 470
},
{
"epoch": 1.8713652045342535,
"grad_norm": 3.4914708137512207,
"learning_rate": 9.906630456805024e-05,
"loss": 0.5924,
"step": 475
},
{
"epoch": 1.8910793494332183,
"grad_norm": 3.5819380283355713,
"learning_rate": 9.901042468802074e-05,
"loss": 0.6426,
"step": 480
},
{
"epoch": 1.9107934943321834,
"grad_norm": 3.237508773803711,
"learning_rate": 9.89529375876591e-05,
"loss": 0.5933,
"step": 485
},
{
"epoch": 1.9305076392311484,
"grad_norm": 3.832315444946289,
"learning_rate": 9.889384515216501e-05,
"loss": 0.5767,
"step": 490
},
{
"epoch": 1.9502217841301133,
"grad_norm": 3.097031593322754,
"learning_rate": 9.883314931938258e-05,
"loss": 0.5225,
"step": 495
},
{
"epoch": 1.9699359290290783,
"grad_norm": 3.5058581829071045,
"learning_rate": 9.877085207973684e-05,
"loss": 0.5591,
"step": 500
},
{
"epoch": 1.9896500739280434,
"grad_norm": 3.0875792503356934,
"learning_rate": 9.870695547616851e-05,
"loss": 0.537,
"step": 505
},
{
"epoch": 2.007885657959586,
"grad_norm": 3.042051315307617,
"learning_rate": 9.864146160406684e-05,
"loss": 0.5084,
"step": 510
},
{
"epoch": 2.027599802858551,
"grad_norm": 3.661208152770996,
"learning_rate": 9.857437261120115e-05,
"loss": 0.4259,
"step": 515
},
{
"epoch": 2.047313947757516,
"grad_norm": 2.7599899768829346,
"learning_rate": 9.850569069765012e-05,
"loss": 0.4659,
"step": 520
},
{
"epoch": 2.0670280926564812,
"grad_norm": 3.0007238388061523,
"learning_rate": 9.843541811572988e-05,
"loss": 0.4666,
"step": 525
},
{
"epoch": 2.086742237555446,
"grad_norm": 3.00451397895813,
"learning_rate": 9.836355716991999e-05,
"loss": 0.4842,
"step": 530
},
{
"epoch": 2.106456382454411,
"grad_norm": 3.2380623817443848,
"learning_rate": 9.82901102167879e-05,
"loss": 0.4566,
"step": 535
},
{
"epoch": 2.126170527353376,
"grad_norm": 3.130038022994995,
"learning_rate": 9.821507966491177e-05,
"loss": 0.4407,
"step": 540
},
{
"epoch": 2.145884672252341,
"grad_norm": 3.2306203842163086,
"learning_rate": 9.813846797480134e-05,
"loss": 0.4002,
"step": 545
},
{
"epoch": 2.165598817151306,
"grad_norm": 3.315145969390869,
"learning_rate": 9.806027765881734e-05,
"loss": 0.4392,
"step": 550
},
{
"epoch": 2.185312962050271,
"grad_norm": 2.3235769271850586,
"learning_rate": 9.798051128108907e-05,
"loss": 0.4443,
"step": 555
},
{
"epoch": 2.205027106949236,
"grad_norm": 2.693516731262207,
"learning_rate": 9.78991714574303e-05,
"loss": 0.4259,
"step": 560
},
{
"epoch": 2.2247412518482013,
"grad_norm": 2.472679853439331,
"learning_rate": 9.781626085525352e-05,
"loss": 0.4294,
"step": 565
},
{
"epoch": 2.244455396747166,
"grad_norm": 2.6586391925811768,
"learning_rate": 9.773178219348242e-05,
"loss": 0.4111,
"step": 570
},
{
"epoch": 2.264169541646131,
"grad_norm": 2.869844913482666,
"learning_rate": 9.764573824246277e-05,
"loss": 0.4128,
"step": 575
},
{
"epoch": 2.283883686545096,
"grad_norm": 2.2448887825012207,
"learning_rate": 9.755813182387158e-05,
"loss": 0.4833,
"step": 580
},
{
"epoch": 2.303597831444061,
"grad_norm": 2.5775742530822754,
"learning_rate": 9.74689658106245e-05,
"loss": 0.4466,
"step": 585
},
{
"epoch": 2.3233119763430263,
"grad_norm": 2.392392158508301,
"learning_rate": 9.73782431267817e-05,
"loss": 0.4098,
"step": 590
},
{
"epoch": 2.343026121241991,
"grad_norm": 2.7074978351593018,
"learning_rate": 9.728596674745187e-05,
"loss": 0.4585,
"step": 595
},
{
"epoch": 2.362740266140956,
"grad_norm": 3.289433240890503,
"learning_rate": 9.719213969869478e-05,
"loss": 0.3965,
"step": 600
},
{
"epoch": 2.3824544110399213,
"grad_norm": 2.337214708328247,
"learning_rate": 9.709676505742193e-05,
"loss": 0.3742,
"step": 605
},
{
"epoch": 2.402168555938886,
"grad_norm": 2.2772057056427,
"learning_rate": 9.699984595129575e-05,
"loss": 0.3871,
"step": 610
},
{
"epoch": 2.421882700837851,
"grad_norm": 2.8387715816497803,
"learning_rate": 9.690138555862691e-05,
"loss": 0.3897,
"step": 615
},
{
"epoch": 2.4415968457368162,
"grad_norm": 2.725186586380005,
"learning_rate": 9.680138710827025e-05,
"loss": 0.4116,
"step": 620
},
{
"epoch": 2.461310990635781,
"grad_norm": 2.4635608196258545,
"learning_rate": 9.669985387951877e-05,
"loss": 0.3498,
"step": 625
},
{
"epoch": 2.4810251355347464,
"grad_norm": 3.7746026515960693,
"learning_rate": 9.659678920199612e-05,
"loss": 0.351,
"step": 630
},
{
"epoch": 2.500739280433711,
"grad_norm": 2.5471270084381104,
"learning_rate": 9.649219645554741e-05,
"loss": 0.3738,
"step": 635
},
{
"epoch": 2.5204534253326765,
"grad_norm": 2.378277063369751,
"learning_rate": 9.638607907012844e-05,
"loss": 0.403,
"step": 640
},
{
"epoch": 2.5401675702316413,
"grad_norm": 2.655937433242798,
"learning_rate": 9.627844052569312e-05,
"loss": 0.4509,
"step": 645
},
{
"epoch": 2.559881715130606,
"grad_norm": 2.7430179119110107,
"learning_rate": 9.616928435207938e-05,
"loss": 0.4314,
"step": 650
},
{
"epoch": 2.5795958600295714,
"grad_norm": 3.2100863456726074,
"learning_rate": 9.605861412889347e-05,
"loss": 0.3806,
"step": 655
},
{
"epoch": 2.5993100049285363,
"grad_norm": 2.856989860534668,
"learning_rate": 9.594643348539251e-05,
"loss": 0.4194,
"step": 660
},
{
"epoch": 2.619024149827501,
"grad_norm": 2.215460777282715,
"learning_rate": 9.583274610036549e-05,
"loss": 0.3617,
"step": 665
},
{
"epoch": 2.6387382947264664,
"grad_norm": 2.89630389213562,
"learning_rate": 9.571755570201266e-05,
"loss": 0.4555,
"step": 670
},
{
"epoch": 2.658452439625431,
"grad_norm": 2.8467066287994385,
"learning_rate": 9.560086606782322e-05,
"loss": 0.3644,
"step": 675
},
{
"epoch": 2.678166584524396,
"grad_norm": 2.772758722305298,
"learning_rate": 9.548268102445145e-05,
"loss": 0.3363,
"step": 680
},
{
"epoch": 2.6978807294233613,
"grad_norm": 2.4616968631744385,
"learning_rate": 9.53630044475913e-05,
"loss": 0.3377,
"step": 685
},
{
"epoch": 2.717594874322326,
"grad_norm": 2.449794292449951,
"learning_rate": 9.524184026184917e-05,
"loss": 0.4089,
"step": 690
},
{
"epoch": 2.7373090192212914,
"grad_norm": 2.1945290565490723,
"learning_rate": 9.511919244061532e-05,
"loss": 0.3989,
"step": 695
},
{
"epoch": 2.7570231641202563,
"grad_norm": 2.0283315181732178,
"learning_rate": 9.499506500593345e-05,
"loss": 0.4989,
"step": 700
},
{
"epoch": 2.7767373090192216,
"grad_norm": 2.6042144298553467,
"learning_rate": 9.486946202836898e-05,
"loss": 0.3996,
"step": 705
},
{
"epoch": 2.7964514539181864,
"grad_norm": 2.1515965461730957,
"learning_rate": 9.474238762687532e-05,
"loss": 0.3421,
"step": 710
},
{
"epoch": 2.8161655988171512,
"grad_norm": 2.147844076156616,
"learning_rate": 9.461384596865904e-05,
"loss": 0.4217,
"step": 715
},
{
"epoch": 2.8358797437161165,
"grad_norm": 2.502347707748413,
"learning_rate": 9.448384126904306e-05,
"loss": 0.4129,
"step": 720
},
{
"epoch": 2.8555938886150813,
"grad_norm": 2.373370409011841,
"learning_rate": 9.435237779132845e-05,
"loss": 0.3447,
"step": 725
},
{
"epoch": 2.875308033514046,
"grad_norm": 2.6700592041015625,
"learning_rate": 9.421945984665465e-05,
"loss": 0.355,
"step": 730
},
{
"epoch": 2.8950221784130115,
"grad_norm": 2.2917869091033936,
"learning_rate": 9.408509179385806e-05,
"loss": 0.3461,
"step": 735
},
{
"epoch": 2.9147363233119763,
"grad_norm": 1.9684882164001465,
"learning_rate": 9.394927803932909e-05,
"loss": 0.3912,
"step": 740
},
{
"epoch": 2.934450468210941,
"grad_norm": 2.7318103313446045,
"learning_rate": 9.381202303686773e-05,
"loss": 0.385,
"step": 745
},
{
"epoch": 2.9541646131099064,
"grad_norm": 2.490328311920166,
"learning_rate": 9.367333128753741e-05,
"loss": 0.3136,
"step": 750
},
{
"epoch": 2.9738787580088712,
"grad_norm": 1.991101861000061,
"learning_rate": 9.353320733951745e-05,
"loss": 0.3771,
"step": 755
},
{
"epoch": 2.9935929029078365,
"grad_norm": 2.4289438724517822,
"learning_rate": 9.339165578795383e-05,
"loss": 0.3484,
"step": 760
},
{
"epoch": 3.011828486939379,
"grad_norm": 2.3784472942352295,
"learning_rate": 9.324868127480867e-05,
"loss": 0.2815,
"step": 765
},
{
"epoch": 3.031542631838344,
"grad_norm": 2.365229606628418,
"learning_rate": 9.310428848870778e-05,
"loss": 0.307,
"step": 770
},
{
"epoch": 3.051256776737309,
"grad_norm": 2.1624045372009277,
"learning_rate": 9.295848216478709e-05,
"loss": 0.2851,
"step": 775
},
{
"epoch": 3.070970921636274,
"grad_norm": 1.6744097471237183,
"learning_rate": 9.281126708453724e-05,
"loss": 0.2774,
"step": 780
},
{
"epoch": 3.0906850665352392,
"grad_norm": 2.131070613861084,
"learning_rate": 9.26626480756469e-05,
"loss": 0.3181,
"step": 785
},
{
"epoch": 3.110399211434204,
"grad_norm": 1.9431699514389038,
"learning_rate": 9.251263001184435e-05,
"loss": 0.2512,
"step": 790
},
{
"epoch": 3.130113356333169,
"grad_norm": 1.7866734266281128,
"learning_rate": 9.236121781273766e-05,
"loss": 0.2931,
"step": 795
},
{
"epoch": 3.149827501232134,
"grad_norm": 2.349666118621826,
"learning_rate": 9.220841644365343e-05,
"loss": 0.3041,
"step": 800
},
{
"epoch": 3.169541646131099,
"grad_norm": 2.101146936416626,
"learning_rate": 9.205423091547393e-05,
"loss": 0.2767,
"step": 805
},
{
"epoch": 3.189255791030064,
"grad_norm": 1.6372120380401611,
"learning_rate": 9.189866628447272e-05,
"loss": 0.3017,
"step": 810
},
{
"epoch": 3.208969935929029,
"grad_norm": 1.7807250022888184,
"learning_rate": 9.174172765214893e-05,
"loss": 0.2727,
"step": 815
},
{
"epoch": 3.228684080827994,
"grad_norm": 2.0267744064331055,
"learning_rate": 9.15834201650599e-05,
"loss": 0.3501,
"step": 820
},
{
"epoch": 3.2483982257269592,
"grad_norm": 2.022195339202881,
"learning_rate": 9.142374901465238e-05,
"loss": 0.3237,
"step": 825
},
{
"epoch": 3.268112370625924,
"grad_norm": 2.5703470706939697,
"learning_rate": 9.126271943709244e-05,
"loss": 0.241,
"step": 830
},
{
"epoch": 3.287826515524889,
"grad_norm": 2.4614508152008057,
"learning_rate": 9.110033671309352e-05,
"loss": 0.2835,
"step": 835
},
{
"epoch": 3.307540660423854,
"grad_norm": 1.5219550132751465,
"learning_rate": 9.093660616774344e-05,
"loss": 0.2858,
"step": 840
},
{
"epoch": 3.327254805322819,
"grad_norm": 1.7633792161941528,
"learning_rate": 9.077153317032976e-05,
"loss": 0.2278,
"step": 845
},
{
"epoch": 3.3469689502217843,
"grad_norm": 1.829990267753601,
"learning_rate": 9.060512313416358e-05,
"loss": 0.3449,
"step": 850
},
{
"epoch": 3.366683095120749,
"grad_norm": 1.9619853496551514,
"learning_rate": 9.043738151640216e-05,
"loss": 0.3066,
"step": 855
},
{
"epoch": 3.386397240019714,
"grad_norm": 2.067197561264038,
"learning_rate": 9.026831381786984e-05,
"loss": 0.2594,
"step": 860
},
{
"epoch": 3.4061113849186793,
"grad_norm": 1.9810595512390137,
"learning_rate": 9.009792558287777e-05,
"loss": 0.294,
"step": 865
},
{
"epoch": 3.425825529817644,
"grad_norm": 2.0421338081359863,
"learning_rate": 8.992622239904198e-05,
"loss": 0.3122,
"step": 870
},
{
"epoch": 3.445539674716609,
"grad_norm": 1.50675368309021,
"learning_rate": 8.975320989710022e-05,
"loss": 0.3048,
"step": 875
},
{
"epoch": 3.465253819615574,
"grad_norm": 1.730079174041748,
"learning_rate": 8.957889375072728e-05,
"loss": 0.2534,
"step": 880
},
{
"epoch": 3.484967964514539,
"grad_norm": 2.360069751739502,
"learning_rate": 8.940327967634893e-05,
"loss": 0.2415,
"step": 885
},
{
"epoch": 3.5046821094135043,
"grad_norm": 2.379964828491211,
"learning_rate": 8.922637343295446e-05,
"loss": 0.3176,
"step": 890
},
{
"epoch": 3.524396254312469,
"grad_norm": 1.2987922430038452,
"learning_rate": 8.904818082190783e-05,
"loss": 0.3124,
"step": 895
},
{
"epoch": 3.5441103992114344,
"grad_norm": 2.5630524158477783,
"learning_rate": 8.88687076867574e-05,
"loss": 0.3022,
"step": 900
},
{
"epoch": 3.5638245441103993,
"grad_norm": 2.043360948562622,
"learning_rate": 8.868795991304436e-05,
"loss": 0.2802,
"step": 905
},
{
"epoch": 3.583538689009364,
"grad_norm": 2.51525616645813,
"learning_rate": 8.850594342810961e-05,
"loss": 0.2965,
"step": 910
},
{
"epoch": 3.6032528339083294,
"grad_norm": 2.4079837799072266,
"learning_rate": 8.832266420089953e-05,
"loss": 0.3132,
"step": 915
},
{
"epoch": 3.6229669788072942,
"grad_norm": 1.6233525276184082,
"learning_rate": 8.813812824177012e-05,
"loss": 0.2818,
"step": 920
},
{
"epoch": 3.642681123706259,
"grad_norm": 1.8031409978866577,
"learning_rate": 8.795234160228995e-05,
"loss": 0.2706,
"step": 925
},
{
"epoch": 3.6623952686052244,
"grad_norm": 2.136138439178467,
"learning_rate": 8.776531037504169e-05,
"loss": 0.287,
"step": 930
},
{
"epoch": 3.682109413504189,
"grad_norm": 1.9893665313720703,
"learning_rate": 8.757704069342234e-05,
"loss": 0.2901,
"step": 935
},
{
"epoch": 3.701823558403154,
"grad_norm": 1.5316953659057617,
"learning_rate": 8.738753873144205e-05,
"loss": 0.2958,
"step": 940
},
{
"epoch": 3.7215377033021193,
"grad_norm": 1.774769902229309,
"learning_rate": 8.71968107035217e-05,
"loss": 0.2604,
"step": 945
},
{
"epoch": 3.741251848201084,
"grad_norm": 1.890213966369629,
"learning_rate": 8.700486286428909e-05,
"loss": 0.258,
"step": 950
},
{
"epoch": 3.7609659931000494,
"grad_norm": 2.19962215423584,
"learning_rate": 8.681170150837382e-05,
"loss": 0.3343,
"step": 955
},
{
"epoch": 3.7806801379990143,
"grad_norm": 1.7772096395492554,
"learning_rate": 8.661733297020085e-05,
"loss": 0.2605,
"step": 960
},
{
"epoch": 3.8003942828979795,
"grad_norm": 1.5164539813995361,
"learning_rate": 8.642176362378285e-05,
"loss": 0.3217,
"step": 965
},
{
"epoch": 3.8201084277969444,
"grad_norm": 2.329491138458252,
"learning_rate": 8.622499988251106e-05,
"loss": 0.3079,
"step": 970
},
{
"epoch": 3.839822572695909,
"grad_norm": 2.1188507080078125,
"learning_rate": 8.602704819894511e-05,
"loss": 0.3327,
"step": 975
},
{
"epoch": 3.8595367175948745,
"grad_norm": 1.6314496994018555,
"learning_rate": 8.582791506460126e-05,
"loss": 0.2557,
"step": 980
},
{
"epoch": 3.8792508624938393,
"grad_norm": 1.6020344495773315,
"learning_rate": 8.562760700973966e-05,
"loss": 0.286,
"step": 985
},
{
"epoch": 3.898965007392804,
"grad_norm": 1.744199275970459,
"learning_rate": 8.54261306031501e-05,
"loss": 0.264,
"step": 990
},
{
"epoch": 3.9186791522917694,
"grad_norm": 2.455364942550659,
"learning_rate": 8.52234924519367e-05,
"loss": 0.2795,
"step": 995
},
{
"epoch": 3.9383932971907343,
"grad_norm": 2.127448558807373,
"learning_rate": 8.501969920130111e-05,
"loss": 0.3015,
"step": 1000
},
{
"epoch": 3.958107442089699,
"grad_norm": 1.9393630027770996,
"learning_rate": 8.481475753432473e-05,
"loss": 0.3879,
"step": 1005
},
{
"epoch": 3.9778215869886644,
"grad_norm": 1.3514901399612427,
"learning_rate": 8.46086741717494e-05,
"loss": 0.2484,
"step": 1010
},
{
"epoch": 3.9975357318876292,
"grad_norm": 1.2901164293289185,
"learning_rate": 8.440145587175718e-05,
"loss": 0.2712,
"step": 1015
},
{
"epoch": 4.015771315919172,
"grad_norm": 1.8801252841949463,
"learning_rate": 8.419310942974855e-05,
"loss": 0.2422,
"step": 1020
},
{
"epoch": 4.035485460818137,
"grad_norm": 1.428691029548645,
"learning_rate": 8.398364167811967e-05,
"loss": 0.2097,
"step": 1025
},
{
"epoch": 4.055199605717102,
"grad_norm": 1.7484722137451172,
"learning_rate": 8.377305948603834e-05,
"loss": 0.2155,
"step": 1030
},
{
"epoch": 4.074913750616067,
"grad_norm": 1.617615818977356,
"learning_rate": 8.356136975921863e-05,
"loss": 0.2147,
"step": 1035
},
{
"epoch": 4.094627895515032,
"grad_norm": 1.8283522129058838,
"learning_rate": 8.334857943969451e-05,
"loss": 0.214,
"step": 1040
},
{
"epoch": 4.114342040413997,
"grad_norm": 1.7181671857833862,
"learning_rate": 8.313469550559219e-05,
"loss": 0.2102,
"step": 1045
},
{
"epoch": 4.1340561853129625,
"grad_norm": 1.7264341115951538,
"learning_rate": 8.291972497090119e-05,
"loss": 0.2665,
"step": 1050
},
{
"epoch": 4.153770330211927,
"grad_norm": 2.047731876373291,
"learning_rate": 8.27036748852445e-05,
"loss": 0.2314,
"step": 1055
},
{
"epoch": 4.173484475110892,
"grad_norm": 1.5944523811340332,
"learning_rate": 8.248655233364724e-05,
"loss": 0.2129,
"step": 1060
},
{
"epoch": 4.193198620009857,
"grad_norm": 1.3731194734573364,
"learning_rate": 8.226836443630437e-05,
"loss": 0.2286,
"step": 1065
},
{
"epoch": 4.212912764908822,
"grad_norm": 1.8190358877182007,
"learning_rate": 8.204911834834722e-05,
"loss": 0.2564,
"step": 1070
},
{
"epoch": 4.232626909807787,
"grad_norm": 3.020439624786377,
"learning_rate": 8.182882125960882e-05,
"loss": 0.2337,
"step": 1075
},
{
"epoch": 4.252341054706752,
"grad_norm": 2.168088436126709,
"learning_rate": 8.160748039438811e-05,
"loss": 0.2177,
"step": 1080
},
{
"epoch": 4.272055199605717,
"grad_norm": 1.3565387725830078,
"learning_rate": 8.138510301121312e-05,
"loss": 0.2267,
"step": 1085
},
{
"epoch": 4.291769344504682,
"grad_norm": 1.3000303506851196,
"learning_rate": 8.116169640260276e-05,
"loss": 0.2212,
"step": 1090
},
{
"epoch": 4.311483489403647,
"grad_norm": 2.5623159408569336,
"learning_rate": 8.093726789482788e-05,
"loss": 0.2447,
"step": 1095
},
{
"epoch": 4.331197634302612,
"grad_norm": 1.2196520566940308,
"learning_rate": 8.071182484767085e-05,
"loss": 0.2146,
"step": 1100
},
{
"epoch": 4.350911779201577,
"grad_norm": 2.003669261932373,
"learning_rate": 8.04853746541843e-05,
"loss": 0.2048,
"step": 1105
},
{
"epoch": 4.370625924100542,
"grad_norm": 1.9316153526306152,
"learning_rate": 8.025792474044869e-05,
"loss": 0.2451,
"step": 1110
},
{
"epoch": 4.390340068999507,
"grad_norm": 1.6847680807113647,
"learning_rate": 8.002948256532865e-05,
"loss": 0.2213,
"step": 1115
},
{
"epoch": 4.410054213898472,
"grad_norm": 2.238589286804199,
"learning_rate": 7.980005562022859e-05,
"loss": 0.2175,
"step": 1120
},
{
"epoch": 4.429768358797437,
"grad_norm": 1.7301565408706665,
"learning_rate": 7.956965142884677e-05,
"loss": 0.2245,
"step": 1125
},
{
"epoch": 4.4494825036964025,
"grad_norm": 1.7427107095718384,
"learning_rate": 7.933827754692886e-05,
"loss": 0.245,
"step": 1130
},
{
"epoch": 4.469196648595367,
"grad_norm": 1.6464322805404663,
"learning_rate": 7.910594156201993e-05,
"loss": 0.21,
"step": 1135
},
{
"epoch": 4.488910793494332,
"grad_norm": 2.1638522148132324,
"learning_rate": 7.887265109321576e-05,
"loss": 0.2183,
"step": 1140
},
{
"epoch": 4.5086249383932975,
"grad_norm": 1.441428542137146,
"learning_rate": 7.863841379091289e-05,
"loss": 0.237,
"step": 1145
},
{
"epoch": 4.528339083292262,
"grad_norm": 1.6750849485397339,
"learning_rate": 7.840323733655778e-05,
"loss": 0.2347,
"step": 1150
},
{
"epoch": 4.548053228191227,
"grad_norm": 1.4054741859436035,
"learning_rate": 7.816712944239503e-05,
"loss": 0.1962,
"step": 1155
},
{
"epoch": 4.567767373090192,
"grad_norm": 1.8165724277496338,
"learning_rate": 7.793009785121421e-05,
"loss": 0.229,
"step": 1160
},
{
"epoch": 4.587481517989158,
"grad_norm": 2.425626754760742,
"learning_rate": 7.769215033609618e-05,
"loss": 0.2182,
"step": 1165
},
{
"epoch": 4.607195662888122,
"grad_norm": 2.2778477668762207,
"learning_rate": 7.745329470015805e-05,
"loss": 0.1964,
"step": 1170
},
{
"epoch": 4.626909807787087,
"grad_norm": 1.7259466648101807,
"learning_rate": 7.721353877629736e-05,
"loss": 0.235,
"step": 1175
},
{
"epoch": 4.646623952686053,
"grad_norm": 2.020500659942627,
"learning_rate": 7.697289042693518e-05,
"loss": 0.2297,
"step": 1180
},
{
"epoch": 4.666338097585017,
"grad_norm": 1.7260327339172363,
"learning_rate": 7.673135754375828e-05,
"loss": 0.2256,
"step": 1185
},
{
"epoch": 4.686052242483982,
"grad_norm": 1.3809380531311035,
"learning_rate": 7.648894804746031e-05,
"loss": 0.2165,
"step": 1190
},
{
"epoch": 4.705766387382948,
"grad_norm": 1.692401647567749,
"learning_rate": 7.624566988748211e-05,
"loss": 0.2163,
"step": 1195
},
{
"epoch": 4.725480532281912,
"grad_norm": 1.4471834897994995,
"learning_rate": 7.6001531041751e-05,
"loss": 0.2106,
"step": 1200
},
{
"epoch": 4.745194677180877,
"grad_norm": 1.6524702310562134,
"learning_rate": 7.575653951641911e-05,
"loss": 0.228,
"step": 1205
},
{
"epoch": 4.764908822079843,
"grad_norm": 1.7597817182540894,
"learning_rate": 7.551070334560088e-05,
"loss": 0.1982,
"step": 1210
},
{
"epoch": 4.784622966978807,
"grad_norm": 1.2614564895629883,
"learning_rate": 7.526403059110956e-05,
"loss": 0.182,
"step": 1215
},
{
"epoch": 4.804337111877772,
"grad_norm": 1.6442838907241821,
"learning_rate": 7.501652934219292e-05,
"loss": 0.2434,
"step": 1220
},
{
"epoch": 4.8240512567767375,
"grad_norm": 1.6709420680999756,
"learning_rate": 7.476820771526784e-05,
"loss": 0.2059,
"step": 1225
},
{
"epoch": 4.843765401675702,
"grad_norm": 1.393500804901123,
"learning_rate": 7.451907385365423e-05,
"loss": 0.2612,
"step": 1230
},
{
"epoch": 4.863479546574667,
"grad_norm": 1.7826026678085327,
"learning_rate": 7.426913592730799e-05,
"loss": 0.2208,
"step": 1235
},
{
"epoch": 4.8831936914736325,
"grad_norm": 2.0293965339660645,
"learning_rate": 7.401840213255305e-05,
"loss": 0.2141,
"step": 1240
},
{
"epoch": 4.902907836372597,
"grad_norm": 1.216652512550354,
"learning_rate": 7.376688069181263e-05,
"loss": 0.2188,
"step": 1245
},
{
"epoch": 4.922621981271562,
"grad_norm": 1.2291804552078247,
"learning_rate": 7.351457985333954e-05,
"loss": 0.2567,
"step": 1250
},
{
"epoch": 4.942336126170527,
"grad_norm": 1.022298812866211,
"learning_rate": 7.32615078909457e-05,
"loss": 0.1975,
"step": 1255
},
{
"epoch": 4.962050271069493,
"grad_norm": 1.1458094120025635,
"learning_rate": 7.30076731037309e-05,
"loss": 0.2115,
"step": 1260
},
{
"epoch": 4.981764415968457,
"grad_norm": 1.5030829906463623,
"learning_rate": 7.275308381581055e-05,
"loss": 0.2531,
"step": 1265
},
{
"epoch": 5.0,
"grad_norm": 1.6188287734985352,
"learning_rate": 7.249774837604266e-05,
"loss": 0.216,
"step": 1270
},
{
"epoch": 5.019714144898965,
"grad_norm": 1.6201726198196411,
"learning_rate": 7.224167515775425e-05,
"loss": 0.1635,
"step": 1275
},
{
"epoch": 5.03942828979793,
"grad_norm": 2.038351535797119,
"learning_rate": 7.198487255846655e-05,
"loss": 0.1592,
"step": 1280
},
{
"epoch": 5.059142434696895,
"grad_norm": 1.7551215887069702,
"learning_rate": 7.172734899961973e-05,
"loss": 0.1796,
"step": 1285
},
{
"epoch": 5.07885657959586,
"grad_norm": 1.3391064405441284,
"learning_rate": 7.14691129262967e-05,
"loss": 0.1772,
"step": 1290
},
{
"epoch": 5.098570724494825,
"grad_norm": 1.2368159294128418,
"learning_rate": 7.121017280694615e-05,
"loss": 0.1759,
"step": 1295
},
{
"epoch": 5.11828486939379,
"grad_norm": 1.5747733116149902,
"learning_rate": 7.09505371331049e-05,
"loss": 0.167,
"step": 1300
},
{
"epoch": 5.137999014292755,
"grad_norm": 1.57236909866333,
"learning_rate": 7.069021441911934e-05,
"loss": 0.1727,
"step": 1305
},
{
"epoch": 5.1577131591917205,
"grad_norm": 1.917982816696167,
"learning_rate": 7.042921320186634e-05,
"loss": 0.1756,
"step": 1310
},
{
"epoch": 5.177427304090685,
"grad_norm": 1.263602375984192,
"learning_rate": 7.016754204047314e-05,
"loss": 0.1758,
"step": 1315
},
{
"epoch": 5.19714144898965,
"grad_norm": 1.1524710655212402,
"learning_rate": 6.990520951603682e-05,
"loss": 0.1602,
"step": 1320
},
{
"epoch": 5.216855593888615,
"grad_norm": 0.9892676472663879,
"learning_rate": 6.964222423134278e-05,
"loss": 0.1587,
"step": 1325
},
{
"epoch": 5.23656973878758,
"grad_norm": 1.5549856424331665,
"learning_rate": 6.937859481058273e-05,
"loss": 0.1734,
"step": 1330
},
{
"epoch": 5.256283883686545,
"grad_norm": 1.2518091201782227,
"learning_rate": 6.911432989907174e-05,
"loss": 0.1744,
"step": 1335
},
{
"epoch": 5.27599802858551,
"grad_norm": 2.184462070465088,
"learning_rate": 6.884943816296487e-05,
"loss": 0.1536,
"step": 1340
},
{
"epoch": 5.295712173484475,
"grad_norm": 1.6105055809020996,
"learning_rate": 6.85839282889729e-05,
"loss": 0.172,
"step": 1345
},
{
"epoch": 5.31542631838344,
"grad_norm": 1.1322895288467407,
"learning_rate": 6.831780898407746e-05,
"loss": 0.1622,
"step": 1350
},
{
"epoch": 5.335140463282405,
"grad_norm": 1.8505653142929077,
"learning_rate": 6.805108897524552e-05,
"loss": 0.1709,
"step": 1355
},
{
"epoch": 5.35485460818137,
"grad_norm": 1.405929446220398,
"learning_rate": 6.77837770091432e-05,
"loss": 0.1807,
"step": 1360
},
{
"epoch": 5.374568753080335,
"grad_norm": 1.7219219207763672,
"learning_rate": 6.751588185184896e-05,
"loss": 0.2026,
"step": 1365
},
{
"epoch": 5.3942828979793,
"grad_norm": 1.5548940896987915,
"learning_rate": 6.724741228856604e-05,
"loss": 0.1671,
"step": 1370
},
{
"epoch": 5.4139970428782656,
"grad_norm": 1.397202730178833,
"learning_rate": 6.69783771233345e-05,
"loss": 0.1675,
"step": 1375
},
{
"epoch": 5.43371118777723,
"grad_norm": 1.0036927461624146,
"learning_rate": 6.670878517874238e-05,
"loss": 0.1553,
"step": 1380
},
{
"epoch": 5.453425332676195,
"grad_norm": 1.6738075017929077,
"learning_rate": 6.643864529563644e-05,
"loss": 0.195,
"step": 1385
},
{
"epoch": 5.4731394775751605,
"grad_norm": 1.7666957378387451,
"learning_rate": 6.616796633283227e-05,
"loss": 0.16,
"step": 1390
},
{
"epoch": 5.492853622474125,
"grad_norm": 1.3601677417755127,
"learning_rate": 6.589675716682366e-05,
"loss": 0.1695,
"step": 1395
},
{
"epoch": 5.51256776737309,
"grad_norm": 2.429863691329956,
"learning_rate": 6.562502669149166e-05,
"loss": 0.1816,
"step": 1400
},
{
"epoch": 5.5322819122720555,
"grad_norm": 1.3872816562652588,
"learning_rate": 6.535278381781281e-05,
"loss": 0.1683,
"step": 1405
},
{
"epoch": 5.55199605717102,
"grad_norm": 1.4309865236282349,
"learning_rate": 6.508003747356695e-05,
"loss": 0.1693,
"step": 1410
},
{
"epoch": 5.571710202069985,
"grad_norm": 1.8246867656707764,
"learning_rate": 6.480679660304448e-05,
"loss": 0.1562,
"step": 1415
},
{
"epoch": 5.59142434696895,
"grad_norm": 1.3177618980407715,
"learning_rate": 6.453307016675296e-05,
"loss": 0.1812,
"step": 1420
},
{
"epoch": 5.611138491867916,
"grad_norm": 1.7910345792770386,
"learning_rate": 6.425886714112337e-05,
"loss": 0.1902,
"step": 1425
},
{
"epoch": 5.63085263676688,
"grad_norm": 1.1931573152542114,
"learning_rate": 6.398419651821568e-05,
"loss": 0.1838,
"step": 1430
},
{
"epoch": 5.650566781665845,
"grad_norm": 1.5768245458602905,
"learning_rate": 6.3709067305424e-05,
"loss": 0.1715,
"step": 1435
},
{
"epoch": 5.670280926564811,
"grad_norm": 2.553699493408203,
"learning_rate": 6.343348852518114e-05,
"loss": 0.1749,
"step": 1440
},
{
"epoch": 5.689995071463775,
"grad_norm": 1.0562894344329834,
"learning_rate": 6.315746921466279e-05,
"loss": 0.1668,
"step": 1445
},
{
"epoch": 5.70970921636274,
"grad_norm": 1.594449758529663,
"learning_rate": 6.288101842549116e-05,
"loss": 0.1741,
"step": 1450
},
{
"epoch": 5.729423361261706,
"grad_norm": 1.732932448387146,
"learning_rate": 6.260414522343813e-05,
"loss": 0.1593,
"step": 1455
},
{
"epoch": 5.74913750616067,
"grad_norm": 1.8460274934768677,
"learning_rate": 6.232685868812789e-05,
"loss": 0.1868,
"step": 1460
},
{
"epoch": 5.768851651059635,
"grad_norm": 1.6852288246154785,
"learning_rate": 6.204916791273932e-05,
"loss": 0.173,
"step": 1465
},
{
"epoch": 5.7885657959586005,
"grad_norm": 1.645918369293213,
"learning_rate": 6.177108200370771e-05,
"loss": 0.162,
"step": 1470
},
{
"epoch": 5.808279940857565,
"grad_norm": 1.5185496807098389,
"learning_rate": 6.14926100804261e-05,
"loss": 0.1844,
"step": 1475
},
{
"epoch": 5.82799408575653,
"grad_norm": 1.133602499961853,
"learning_rate": 6.12137612749463e-05,
"loss": 0.1861,
"step": 1480
},
{
"epoch": 5.8477082306554955,
"grad_norm": 1.1715219020843506,
"learning_rate": 6.093454473167938e-05,
"loss": 0.1699,
"step": 1485
},
{
"epoch": 5.86742237555446,
"grad_norm": 1.3760854005813599,
"learning_rate": 6.065496960709578e-05,
"loss": 0.1621,
"step": 1490
},
{
"epoch": 5.887136520453425,
"grad_norm": 1.3439675569534302,
"learning_rate": 6.0375045069425084e-05,
"loss": 0.1486,
"step": 1495
},
{
"epoch": 5.9068506653523905,
"grad_norm": 1.2491891384124756,
"learning_rate": 6.009478029835528e-05,
"loss": 0.1602,
"step": 1500
},
{
"epoch": 5.926564810251356,
"grad_norm": 1.6365981101989746,
"learning_rate": 5.981418448473184e-05,
"loss": 0.1691,
"step": 1505
},
{
"epoch": 5.94627895515032,
"grad_norm": 1.0324171781539917,
"learning_rate": 5.953326683025621e-05,
"loss": 0.1651,
"step": 1510
},
{
"epoch": 5.965993100049285,
"grad_norm": 1.453691840171814,
"learning_rate": 5.925203654718415e-05,
"loss": 0.1892,
"step": 1515
},
{
"epoch": 5.985707244948251,
"grad_norm": 1.7985492944717407,
"learning_rate": 5.8970502858023545e-05,
"loss": 0.1992,
"step": 1520
},
{
"epoch": 6.003942828979793,
"grad_norm": 1.3267945051193237,
"learning_rate": 5.868867499523203e-05,
"loss": 0.1616,
"step": 1525
},
{
"epoch": 6.023656973878758,
"grad_norm": 1.3626664876937866,
"learning_rate": 5.8406562200914196e-05,
"loss": 0.1427,
"step": 1530
},
{
"epoch": 6.043371118777723,
"grad_norm": 1.3885642290115356,
"learning_rate": 5.812417372651854e-05,
"loss": 0.138,
"step": 1535
},
{
"epoch": 6.063085263676688,
"grad_norm": 1.6768662929534912,
"learning_rate": 5.7841518832534014e-05,
"loss": 0.1392,
"step": 1540
},
{
"epoch": 6.082799408575653,
"grad_norm": 1.214672565460205,
"learning_rate": 5.755860678818642e-05,
"loss": 0.142,
"step": 1545
},
{
"epoch": 6.102513553474618,
"grad_norm": 1.4035040140151978,
"learning_rate": 5.7275446871134405e-05,
"loss": 0.1353,
"step": 1550
},
{
"epoch": 6.1222276983735835,
"grad_norm": 1.6018613576889038,
"learning_rate": 5.69920483671652e-05,
"loss": 0.1378,
"step": 1555
},
{
"epoch": 6.141941843272548,
"grad_norm": 1.4302403926849365,
"learning_rate": 5.670842056989012e-05,
"loss": 0.1381,
"step": 1560
},
{
"epoch": 6.161655988171513,
"grad_norm": 0.9095770716667175,
"learning_rate": 5.64245727804398e-05,
"loss": 0.1409,
"step": 1565
},
{
"epoch": 6.1813701330704784,
"grad_norm": 1.0363128185272217,
"learning_rate": 5.614051430715915e-05,
"loss": 0.1297,
"step": 1570
},
{
"epoch": 6.201084277969443,
"grad_norm": 1.1587727069854736,
"learning_rate": 5.5856254465302225e-05,
"loss": 0.1352,
"step": 1575
},
{
"epoch": 6.220798422868408,
"grad_norm": 1.387983798980713,
"learning_rate": 5.5571802576726507e-05,
"loss": 0.1356,
"step": 1580
},
{
"epoch": 6.240512567767373,
"grad_norm": 1.4277117252349854,
"learning_rate": 5.528716796958745e-05,
"loss": 0.1356,
"step": 1585
},
{
"epoch": 6.260226712666338,
"grad_norm": 1.3528679609298706,
"learning_rate": 5.5002359978032435e-05,
"loss": 0.1408,
"step": 1590
},
{
"epoch": 6.279940857565303,
"grad_norm": 1.6311657428741455,
"learning_rate": 5.471738794189474e-05,
"loss": 0.1426,
"step": 1595
},
{
"epoch": 6.299655002464268,
"grad_norm": 0.8196179866790771,
"learning_rate": 5.443226120638719e-05,
"loss": 0.1346,
"step": 1600
},
{
"epoch": 6.319369147363233,
"grad_norm": 1.4031368494033813,
"learning_rate": 5.414698912179579e-05,
"loss": 0.1327,
"step": 1605
},
{
"epoch": 6.339083292262198,
"grad_norm": 1.8704664707183838,
"learning_rate": 5.386158104317298e-05,
"loss": 0.1397,
"step": 1610
},
{
"epoch": 6.358797437161163,
"grad_norm": 1.4920119047164917,
"learning_rate": 5.3576046330030974e-05,
"loss": 0.1325,
"step": 1615
},
{
"epoch": 6.378511582060128,
"grad_norm": 1.0346848964691162,
"learning_rate": 5.329039434603471e-05,
"loss": 0.1482,
"step": 1620
},
{
"epoch": 6.398225726959093,
"grad_norm": 1.9379348754882812,
"learning_rate": 5.300463445869488e-05,
"loss": 0.1524,
"step": 1625
},
{
"epoch": 6.417939871858058,
"grad_norm": 0.7949770092964172,
"learning_rate": 5.271877603906067e-05,
"loss": 0.1282,
"step": 1630
},
{
"epoch": 6.4376540167570235,
"grad_norm": 0.9802160859107971,
"learning_rate": 5.24328284614125e-05,
"loss": 0.1346,
"step": 1635
},
{
"epoch": 6.457368161655988,
"grad_norm": 1.578125,
"learning_rate": 5.214680110295459e-05,
"loss": 0.1398,
"step": 1640
},
{
"epoch": 6.477082306554953,
"grad_norm": 1.2905495166778564,
"learning_rate": 5.186070334350741e-05,
"loss": 0.1278,
"step": 1645
},
{
"epoch": 6.4967964514539185,
"grad_norm": 1.4263631105422974,
"learning_rate": 5.1574544565200175e-05,
"loss": 0.1263,
"step": 1650
},
{
"epoch": 6.516510596352883,
"grad_norm": 0.8475696444511414,
"learning_rate": 5.128833415216309e-05,
"loss": 0.1322,
"step": 1655
},
{
"epoch": 6.536224741251848,
"grad_norm": 0.9275863170623779,
"learning_rate": 5.100208149021962e-05,
"loss": 0.1311,
"step": 1660
},
{
"epoch": 6.555938886150813,
"grad_norm": 1.2516257762908936,
"learning_rate": 5.071579596657876e-05,
"loss": 0.1381,
"step": 1665
},
{
"epoch": 6.575653031049778,
"grad_norm": 1.048720359802246,
"learning_rate": 5.042948696952711e-05,
"loss": 0.1374,
"step": 1670
},
{
"epoch": 6.595367175948743,
"grad_norm": 1.3890330791473389,
"learning_rate": 5.014316388812108e-05,
"loss": 0.139,
"step": 1675
},
{
"epoch": 6.615081320847708,
"grad_norm": 1.5058406591415405,
"learning_rate": 4.985683611187895e-05,
"loss": 0.152,
"step": 1680
},
{
"epoch": 6.634795465746674,
"grad_norm": 1.5658972263336182,
"learning_rate": 4.9570513030472896e-05,
"loss": 0.1406,
"step": 1685
},
{
"epoch": 6.654509610645638,
"grad_norm": 1.483197808265686,
"learning_rate": 4.9284204033421266e-05,
"loss": 0.1321,
"step": 1690
},
{
"epoch": 6.674223755544603,
"grad_norm": 1.8198909759521484,
"learning_rate": 4.8997918509780386e-05,
"loss": 0.1367,
"step": 1695
},
{
"epoch": 6.693937900443569,
"grad_norm": 1.1847096681594849,
"learning_rate": 4.871166584783693e-05,
"loss": 0.1372,
"step": 1700
},
{
"epoch": 6.713652045342533,
"grad_norm": 0.8408295512199402,
"learning_rate": 4.842545543479983e-05,
"loss": 0.1242,
"step": 1705
},
{
"epoch": 6.733366190241498,
"grad_norm": 1.6216832399368286,
"learning_rate": 4.8139296656492605e-05,
"loss": 0.1388,
"step": 1710
},
{
"epoch": 6.753080335140464,
"grad_norm": 1.05503511428833,
"learning_rate": 4.7853198897045424e-05,
"loss": 0.137,
"step": 1715
},
{
"epoch": 6.772794480039428,
"grad_norm": 1.3355563879013062,
"learning_rate": 4.7567171538587514e-05,
"loss": 0.1248,
"step": 1720
},
{
"epoch": 6.792508624938393,
"grad_norm": 1.3196101188659668,
"learning_rate": 4.728122396093935e-05,
"loss": 0.1383,
"step": 1725
},
{
"epoch": 6.8122227698373585,
"grad_norm": 0.7643747329711914,
"learning_rate": 4.6995365541305145e-05,
"loss": 0.1394,
"step": 1730
},
{
"epoch": 6.831936914736323,
"grad_norm": 1.3718485832214355,
"learning_rate": 4.6709605653965297e-05,
"loss": 0.1421,
"step": 1735
},
{
"epoch": 6.851651059635288,
"grad_norm": 0.7023568153381348,
"learning_rate": 4.642395366996903e-05,
"loss": 0.1338,
"step": 1740
},
{
"epoch": 6.8713652045342535,
"grad_norm": 1.994978666305542,
"learning_rate": 4.6138418956827026e-05,
"loss": 0.1487,
"step": 1745
},
{
"epoch": 6.891079349433218,
"grad_norm": 1.417363166809082,
"learning_rate": 4.585301087820421e-05,
"loss": 0.1271,
"step": 1750
},
{
"epoch": 6.910793494332183,
"grad_norm": 1.2611218690872192,
"learning_rate": 4.5567738793612813e-05,
"loss": 0.1353,
"step": 1755
},
{
"epoch": 6.930507639231148,
"grad_norm": 1.1835087537765503,
"learning_rate": 4.528261205810527e-05,
"loss": 0.1398,
"step": 1760
},
{
"epoch": 6.950221784130114,
"grad_norm": 1.0265967845916748,
"learning_rate": 4.499764002196758e-05,
"loss": 0.1482,
"step": 1765
},
{
"epoch": 6.969935929029078,
"grad_norm": 0.9783711433410645,
"learning_rate": 4.471283203041255e-05,
"loss": 0.1346,
"step": 1770
},
{
"epoch": 6.989650073928043,
"grad_norm": 1.1431827545166016,
"learning_rate": 4.4428197423273505e-05,
"loss": 0.1409,
"step": 1775
},
{
"epoch": 7.007885657959586,
"grad_norm": 0.9335059523582458,
"learning_rate": 4.4143745534697787e-05,
"loss": 0.1411,
"step": 1780
},
{
"epoch": 7.027599802858551,
"grad_norm": 1.3468759059906006,
"learning_rate": 4.385948569284085e-05,
"loss": 0.1184,
"step": 1785
},
{
"epoch": 7.047313947757516,
"grad_norm": 0.603523313999176,
"learning_rate": 4.357542721956021e-05,
"loss": 0.1128,
"step": 1790
},
{
"epoch": 7.067028092656481,
"grad_norm": 1.1114201545715332,
"learning_rate": 4.32915794301099e-05,
"loss": 0.1159,
"step": 1795
},
{
"epoch": 7.086742237555446,
"grad_norm": 0.6488471031188965,
"learning_rate": 4.300795163283482e-05,
"loss": 0.1166,
"step": 1800
},
{
"epoch": 7.106456382454411,
"grad_norm": 0.9025667309761047,
"learning_rate": 4.272455312886562e-05,
"loss": 0.1183,
"step": 1805
},
{
"epoch": 7.126170527353376,
"grad_norm": 0.7467586994171143,
"learning_rate": 4.244139321181358e-05,
"loss": 0.1068,
"step": 1810
},
{
"epoch": 7.1458846722523415,
"grad_norm": 0.8195639848709106,
"learning_rate": 4.2158481167466004e-05,
"loss": 0.1164,
"step": 1815
},
{
"epoch": 7.165598817151306,
"grad_norm": 1.6819171905517578,
"learning_rate": 4.1875826273481474e-05,
"loss": 0.1146,
"step": 1820
},
{
"epoch": 7.185312962050271,
"grad_norm": 1.1047701835632324,
"learning_rate": 4.1593437799085816e-05,
"loss": 0.1204,
"step": 1825
},
{
"epoch": 7.205027106949236,
"grad_norm": 0.9841985106468201,
"learning_rate": 4.131132500476799e-05,
"loss": 0.1222,
"step": 1830
},
{
"epoch": 7.224741251848201,
"grad_norm": 0.8494741916656494,
"learning_rate": 4.102949714197648e-05,
"loss": 0.1196,
"step": 1835
},
{
"epoch": 7.244455396747166,
"grad_norm": 1.0732312202453613,
"learning_rate": 4.0747963452815854e-05,
"loss": 0.1171,
"step": 1840
},
{
"epoch": 7.264169541646131,
"grad_norm": 0.6279380917549133,
"learning_rate": 4.046673316974379e-05,
"loss": 0.1178,
"step": 1845
},
{
"epoch": 7.283883686545096,
"grad_norm": 2.309483766555786,
"learning_rate": 4.018581551526818e-05,
"loss": 0.1137,
"step": 1850
},
{
"epoch": 7.303597831444061,
"grad_norm": 0.8398231267929077,
"learning_rate": 3.990521970164472e-05,
"loss": 0.1174,
"step": 1855
},
{
"epoch": 7.323311976343026,
"grad_norm": 1.108704924583435,
"learning_rate": 3.9624954930574934e-05,
"loss": 0.1133,
"step": 1860
},
{
"epoch": 7.343026121241991,
"grad_norm": 0.9234030842781067,
"learning_rate": 3.934503039290422e-05,
"loss": 0.1097,
"step": 1865
},
{
"epoch": 7.362740266140956,
"grad_norm": 0.9003049731254578,
"learning_rate": 3.9065455268320635e-05,
"loss": 0.1152,
"step": 1870
},
{
"epoch": 7.382454411039921,
"grad_norm": 1.0665256977081299,
"learning_rate": 3.87862387250537e-05,
"loss": 0.1178,
"step": 1875
},
{
"epoch": 7.402168555938887,
"grad_norm": 0.7589354515075684,
"learning_rate": 3.850738991957391e-05,
"loss": 0.113,
"step": 1880
},
{
"epoch": 7.421882700837851,
"grad_norm": 0.8894014954566956,
"learning_rate": 3.82289179962923e-05,
"loss": 0.1161,
"step": 1885
},
{
"epoch": 7.441596845736816,
"grad_norm": 1.0884486436843872,
"learning_rate": 3.795083208726069e-05,
"loss": 0.1167,
"step": 1890
},
{
"epoch": 7.4613109906357815,
"grad_norm": 1.075456976890564,
"learning_rate": 3.767314131187211e-05,
"loss": 0.1106,
"step": 1895
},
{
"epoch": 7.481025135534746,
"grad_norm": 0.5905814170837402,
"learning_rate": 3.7395854776561886e-05,
"loss": 0.1126,
"step": 1900
},
{
"epoch": 7.500739280433711,
"grad_norm": 0.6967124342918396,
"learning_rate": 3.711898157450884e-05,
"loss": 0.1182,
"step": 1905
},
{
"epoch": 7.5204534253326765,
"grad_norm": 0.8434427976608276,
"learning_rate": 3.684253078533722e-05,
"loss": 0.1213,
"step": 1910
},
{
"epoch": 7.540167570231641,
"grad_norm": 0.7524256110191345,
"learning_rate": 3.6566511474818874e-05,
"loss": 0.1105,
"step": 1915
},
{
"epoch": 7.559881715130606,
"grad_norm": 0.5648545622825623,
"learning_rate": 3.629093269457602e-05,
"loss": 0.1167,
"step": 1920
},
{
"epoch": 7.579595860029571,
"grad_norm": 1.1991491317749023,
"learning_rate": 3.6015803481784334e-05,
"loss": 0.1195,
"step": 1925
},
{
"epoch": 7.599310004928537,
"grad_norm": 0.7800848484039307,
"learning_rate": 3.574113285887666e-05,
"loss": 0.1183,
"step": 1930
},
{
"epoch": 7.619024149827501,
"grad_norm": 0.8941491842269897,
"learning_rate": 3.546692983324706e-05,
"loss": 0.1161,
"step": 1935
},
{
"epoch": 7.638738294726466,
"grad_norm": 0.6570645570755005,
"learning_rate": 3.519320339695555e-05,
"loss": 0.1135,
"step": 1940
},
{
"epoch": 7.658452439625432,
"grad_norm": 0.989401638507843,
"learning_rate": 3.491996252643305e-05,
"loss": 0.111,
"step": 1945
},
{
"epoch": 7.678166584524396,
"grad_norm": 1.1077444553375244,
"learning_rate": 3.4647216182187205e-05,
"loss": 0.1281,
"step": 1950
},
{
"epoch": 7.697880729423361,
"grad_norm": 1.2130508422851562,
"learning_rate": 3.437497330850835e-05,
"loss": 0.1149,
"step": 1955
},
{
"epoch": 7.717594874322327,
"grad_norm": 0.8594215512275696,
"learning_rate": 3.410324283317633e-05,
"loss": 0.1116,
"step": 1960
},
{
"epoch": 7.737309019221291,
"grad_norm": 0.6998811364173889,
"learning_rate": 3.3832033667167745e-05,
"loss": 0.1198,
"step": 1965
},
{
"epoch": 7.757023164120256,
"grad_norm": 1.1719328165054321,
"learning_rate": 3.3561354704363564e-05,
"loss": 0.1161,
"step": 1970
},
{
"epoch": 7.7767373090192216,
"grad_norm": 0.5164292454719543,
"learning_rate": 3.3291214821257635e-05,
"loss": 0.1187,
"step": 1975
},
{
"epoch": 7.796451453918186,
"grad_norm": 0.9778832197189331,
"learning_rate": 3.30216228766655e-05,
"loss": 0.1123,
"step": 1980
},
{
"epoch": 7.816165598817151,
"grad_norm": 0.8570555448532104,
"learning_rate": 3.2752587711433963e-05,
"loss": 0.1187,
"step": 1985
},
{
"epoch": 7.8358797437161165,
"grad_norm": 0.6279511451721191,
"learning_rate": 3.248411814815104e-05,
"loss": 0.1135,
"step": 1990
},
{
"epoch": 7.855593888615081,
"grad_norm": 1.0085203647613525,
"learning_rate": 3.22162229908568e-05,
"loss": 0.1129,
"step": 1995
},
{
"epoch": 7.875308033514046,
"grad_norm": 0.7156813144683838,
"learning_rate": 3.194891102475448e-05,
"loss": 0.1149,
"step": 2000
},
{
"epoch": 7.8950221784130115,
"grad_norm": 0.5609560012817383,
"learning_rate": 3.1682191015922554e-05,
"loss": 0.1149,
"step": 2005
},
{
"epoch": 7.914736323311976,
"grad_norm": 1.008686900138855,
"learning_rate": 3.141607171102711e-05,
"loss": 0.1174,
"step": 2010
},
{
"epoch": 7.934450468210941,
"grad_norm": 0.7896909713745117,
"learning_rate": 3.1150561837035146e-05,
"loss": 0.1144,
"step": 2015
},
{
"epoch": 7.954164613109906,
"grad_norm": 0.8853268027305603,
"learning_rate": 3.088567010092827e-05,
"loss": 0.1228,
"step": 2020
},
{
"epoch": 7.973878758008872,
"grad_norm": 1.2333228588104248,
"learning_rate": 3.0621405189417294e-05,
"loss": 0.1165,
"step": 2025
},
{
"epoch": 7.993592902907836,
"grad_norm": 0.6127500534057617,
"learning_rate": 3.0357775768657227e-05,
"loss": 0.119,
"step": 2030
},
{
"epoch": 8.011828486939379,
"grad_norm": 0.6676454544067383,
"learning_rate": 3.009479048396321e-05,
"loss": 0.1097,
"step": 2035
},
{
"epoch": 8.031542631838343,
"grad_norm": 0.5246162414550781,
"learning_rate": 2.9832457959526866e-05,
"loss": 0.1057,
"step": 2040
},
{
"epoch": 8.05125677673731,
"grad_norm": 0.6145051121711731,
"learning_rate": 2.9570786798133678e-05,
"loss": 0.1045,
"step": 2045
},
{
"epoch": 8.070970921636274,
"grad_norm": 0.6032987236976624,
"learning_rate": 2.930978558088066e-05,
"loss": 0.1099,
"step": 2050
},
{
"epoch": 8.090685066535238,
"grad_norm": 0.7035672068595886,
"learning_rate": 2.9049462866895104e-05,
"loss": 0.1099,
"step": 2055
},
{
"epoch": 8.110399211434205,
"grad_norm": 0.9585490822792053,
"learning_rate": 2.8789827193053853e-05,
"loss": 0.1069,
"step": 2060
},
{
"epoch": 8.130113356333169,
"grad_norm": 0.7242105603218079,
"learning_rate": 2.8530887073703304e-05,
"loss": 0.1027,
"step": 2065
},
{
"epoch": 8.149827501232133,
"grad_norm": 0.7860429883003235,
"learning_rate": 2.8272651000380274e-05,
"loss": 0.1033,
"step": 2070
},
{
"epoch": 8.1695416461311,
"grad_norm": 0.5597553253173828,
"learning_rate": 2.801512744153345e-05,
"loss": 0.1046,
"step": 2075
},
{
"epoch": 8.189255791030064,
"grad_norm": 0.5892042517662048,
"learning_rate": 2.7758324842245755e-05,
"loss": 0.1085,
"step": 2080
},
{
"epoch": 8.208969935929028,
"grad_norm": 0.5419883728027344,
"learning_rate": 2.7502251623957343e-05,
"loss": 0.1003,
"step": 2085
},
{
"epoch": 8.228684080827994,
"grad_norm": 0.5444959402084351,
"learning_rate": 2.7246916184189485e-05,
"loss": 0.1066,
"step": 2090
},
{
"epoch": 8.248398225726959,
"grad_norm": 0.6137767434120178,
"learning_rate": 2.699232689626909e-05,
"loss": 0.1061,
"step": 2095
},
{
"epoch": 8.268112370625925,
"grad_norm": 0.4853847920894623,
"learning_rate": 2.6738492109054302e-05,
"loss": 0.1024,
"step": 2100
},
{
"epoch": 8.28782651552489,
"grad_norm": 1.2990504503250122,
"learning_rate": 2.648542014666048e-05,
"loss": 0.0968,
"step": 2105
},
{
"epoch": 8.307540660423854,
"grad_norm": 0.7490946054458618,
"learning_rate": 2.623311930818737e-05,
"loss": 0.1099,
"step": 2110
},
{
"epoch": 8.32725480532282,
"grad_norm": 0.5346263647079468,
"learning_rate": 2.598159786744695e-05,
"loss": 0.1018,
"step": 2115
},
{
"epoch": 8.346968950221784,
"grad_norm": 0.5948059558868408,
"learning_rate": 2.5730864072692018e-05,
"loss": 0.1032,
"step": 2120
},
{
"epoch": 8.366683095120749,
"grad_norm": 0.623589277267456,
"learning_rate": 2.5480926146345785e-05,
"loss": 0.1038,
"step": 2125
},
{
"epoch": 8.386397240019715,
"grad_norm": 0.47425466775894165,
"learning_rate": 2.523179228473219e-05,
"loss": 0.1011,
"step": 2130
},
{
"epoch": 8.40611138491868,
"grad_norm": 0.6056951880455017,
"learning_rate": 2.4983470657807085e-05,
"loss": 0.0997,
"step": 2135
},
{
"epoch": 8.425825529817644,
"grad_norm": 0.5924261808395386,
"learning_rate": 2.473596940889045e-05,
"loss": 0.1098,
"step": 2140
},
{
"epoch": 8.44553967471661,
"grad_norm": 0.6256517171859741,
"learning_rate": 2.4489296654399148e-05,
"loss": 0.111,
"step": 2145
},
{
"epoch": 8.465253819615574,
"grad_norm": 1.0766382217407227,
"learning_rate": 2.424346048358091e-05,
"loss": 0.1093,
"step": 2150
},
{
"epoch": 8.484967964514539,
"grad_norm": 0.6356780529022217,
"learning_rate": 2.3998468958249017e-05,
"loss": 0.1073,
"step": 2155
},
{
"epoch": 8.504682109413505,
"grad_norm": 0.6873161196708679,
"learning_rate": 2.3754330112517898e-05,
"loss": 0.1044,
"step": 2160
},
{
"epoch": 8.52439625431247,
"grad_norm": 0.670346736907959,
"learning_rate": 2.3511051952539702e-05,
"loss": 0.106,
"step": 2165
},
{
"epoch": 8.544110399211434,
"grad_norm": 0.626423180103302,
"learning_rate": 2.3268642456241752e-05,
"loss": 0.0985,
"step": 2170
},
{
"epoch": 8.5638245441104,
"grad_norm": 0.6810294985771179,
"learning_rate": 2.302710957306482e-05,
"loss": 0.1065,
"step": 2175
},
{
"epoch": 8.583538689009364,
"grad_norm": 0.7520167827606201,
"learning_rate": 2.278646122370263e-05,
"loss": 0.105,
"step": 2180
},
{
"epoch": 8.603252833908329,
"grad_norm": 0.8657585382461548,
"learning_rate": 2.2546705299841958e-05,
"loss": 0.1014,
"step": 2185
},
{
"epoch": 8.622966978807295,
"grad_norm": 0.7666080594062805,
"learning_rate": 2.230784966390381e-05,
"loss": 0.1058,
"step": 2190
},
{
"epoch": 8.642681123706259,
"grad_norm": 0.7878192067146301,
"learning_rate": 2.20699021487858e-05,
"loss": 0.1022,
"step": 2195
},
{
"epoch": 8.662395268605223,
"grad_norm": 0.601855993270874,
"learning_rate": 2.183287055760496e-05,
"loss": 0.0985,
"step": 2200
},
{
"epoch": 8.68210941350419,
"grad_norm": 0.6196221709251404,
"learning_rate": 2.1596762663442218e-05,
"loss": 0.1007,
"step": 2205
},
{
"epoch": 8.701823558403154,
"grad_norm": 0.6262930631637573,
"learning_rate": 2.1361586209087136e-05,
"loss": 0.1034,
"step": 2210
},
{
"epoch": 8.721537703302118,
"grad_norm": 0.8374412059783936,
"learning_rate": 2.1127348906784255e-05,
"loss": 0.1043,
"step": 2215
},
{
"epoch": 8.741251848201085,
"grad_norm": 0.5195895433425903,
"learning_rate": 2.0894058437980068e-05,
"loss": 0.1029,
"step": 2220
},
{
"epoch": 8.760965993100049,
"grad_norm": 0.7010400891304016,
"learning_rate": 2.0661722453071143e-05,
"loss": 0.1047,
"step": 2225
},
{
"epoch": 8.780680137999013,
"grad_norm": 0.5973253846168518,
"learning_rate": 2.043034857115323e-05,
"loss": 0.1112,
"step": 2230
},
{
"epoch": 8.80039428289798,
"grad_norm": 0.7826218008995056,
"learning_rate": 2.0199944379771447e-05,
"loss": 0.0996,
"step": 2235
},
{
"epoch": 8.820108427796944,
"grad_norm": 1.0917187929153442,
"learning_rate": 1.9970517434671336e-05,
"loss": 0.1088,
"step": 2240
},
{
"epoch": 8.83982257269591,
"grad_norm": 0.6169748306274414,
"learning_rate": 1.974207525955133e-05,
"loss": 0.111,
"step": 2245
},
{
"epoch": 8.859536717594874,
"grad_norm": 1.227509617805481,
"learning_rate": 1.95146253458157e-05,
"loss": 0.1077,
"step": 2250
},
{
"epoch": 8.879250862493839,
"grad_norm": 0.6013702750205994,
"learning_rate": 1.9288175152329163e-05,
"loss": 0.1038,
"step": 2255
},
{
"epoch": 8.898965007392805,
"grad_norm": 0.622795820236206,
"learning_rate": 1.9062732105172138e-05,
"loss": 0.105,
"step": 2260
},
{
"epoch": 8.91867915229177,
"grad_norm": 0.5975040793418884,
"learning_rate": 1.883830359739725e-05,
"loss": 0.108,
"step": 2265
},
{
"epoch": 8.938393297190734,
"grad_norm": 0.646460771560669,
"learning_rate": 1.8614896988786902e-05,
"loss": 0.1038,
"step": 2270
},
{
"epoch": 8.9581074420897,
"grad_norm": 0.5937023162841797,
"learning_rate": 1.8392519605611907e-05,
"loss": 0.1072,
"step": 2275
},
{
"epoch": 8.977821586988664,
"grad_norm": 0.6777051687240601,
"learning_rate": 1.8171178740391194e-05,
"loss": 0.1056,
"step": 2280
},
{
"epoch": 8.997535731887629,
"grad_norm": 1.1725229024887085,
"learning_rate": 1.7950881651652783e-05,
"loss": 0.1014,
"step": 2285
},
{
"epoch": 9.015771315919173,
"grad_norm": 0.47024473547935486,
"learning_rate": 1.7731635563695643e-05,
"loss": 0.0966,
"step": 2290
},
{
"epoch": 9.035485460818137,
"grad_norm": 0.564164936542511,
"learning_rate": 1.7513447666352752e-05,
"loss": 0.1032,
"step": 2295
},
{
"epoch": 9.055199605717101,
"grad_norm": 0.6236717700958252,
"learning_rate": 1.7296325114755503e-05,
"loss": 0.0964,
"step": 2300
},
{
"epoch": 9.074913750616068,
"grad_norm": 0.5206897854804993,
"learning_rate": 1.7080275029098812e-05,
"loss": 0.1033,
"step": 2305
},
{
"epoch": 9.094627895515032,
"grad_norm": 0.633885383605957,
"learning_rate": 1.686530449440783e-05,
"loss": 0.0939,
"step": 2310
},
{
"epoch": 9.114342040413996,
"grad_norm": 0.6955597400665283,
"learning_rate": 1.6651420560305497e-05,
"loss": 0.0992,
"step": 2315
},
{
"epoch": 9.134056185312962,
"grad_norm": 0.5905971527099609,
"learning_rate": 1.6438630240781378e-05,
"loss": 0.0989,
"step": 2320
},
{
"epoch": 9.153770330211927,
"grad_norm": 0.501326322555542,
"learning_rate": 1.6226940513961664e-05,
"loss": 0.091,
"step": 2325
},
{
"epoch": 9.173484475110891,
"grad_norm": 0.5881623029708862,
"learning_rate": 1.6016358321880326e-05,
"loss": 0.0976,
"step": 2330
},
{
"epoch": 9.193198620009857,
"grad_norm": 0.6199246048927307,
"learning_rate": 1.580689057025146e-05,
"loss": 0.1007,
"step": 2335
},
{
"epoch": 9.212912764908822,
"grad_norm": 0.6353612542152405,
"learning_rate": 1.559854412824284e-05,
"loss": 0.0959,
"step": 2340
},
{
"epoch": 9.232626909807786,
"grad_norm": 0.5803267359733582,
"learning_rate": 1.5391325828250596e-05,
"loss": 0.0962,
"step": 2345
},
{
"epoch": 9.252341054706752,
"grad_norm": 0.5344117283821106,
"learning_rate": 1.5185242465675297e-05,
"loss": 0.0961,
"step": 2350
},
{
"epoch": 9.272055199605717,
"grad_norm": 0.5603470802307129,
"learning_rate": 1.4980300798698904e-05,
"loss": 0.1016,
"step": 2355
},
{
"epoch": 9.291769344504683,
"grad_norm": 0.6176356077194214,
"learning_rate": 1.4776507548063318e-05,
"loss": 0.1071,
"step": 2360
},
{
"epoch": 9.311483489403647,
"grad_norm": 0.6276616454124451,
"learning_rate": 1.4573869396849904e-05,
"loss": 0.0968,
"step": 2365
},
{
"epoch": 9.331197634302612,
"grad_norm": 0.614722490310669,
"learning_rate": 1.4372392990260353e-05,
"loss": 0.0987,
"step": 2370
},
{
"epoch": 9.350911779201578,
"grad_norm": 0.6439083218574524,
"learning_rate": 1.4172084935398744e-05,
"loss": 0.0985,
"step": 2375
},
{
"epoch": 9.370625924100542,
"grad_norm": 0.6016230583190918,
"learning_rate": 1.3972951801054912e-05,
"loss": 0.1004,
"step": 2380
},
{
"epoch": 9.390340068999507,
"grad_norm": 0.5583289861679077,
"learning_rate": 1.3775000117488934e-05,
"loss": 0.0924,
"step": 2385
},
{
"epoch": 9.410054213898473,
"grad_norm": 0.5593780875205994,
"learning_rate": 1.3578236376217173e-05,
"loss": 0.0939,
"step": 2390
},
{
"epoch": 9.429768358797437,
"grad_norm": 0.6913323998451233,
"learning_rate": 1.3382667029799167e-05,
"loss": 0.0977,
"step": 2395
},
{
"epoch": 9.449482503696402,
"grad_norm": 0.7171539664268494,
"learning_rate": 1.3188298491626183e-05,
"loss": 0.0974,
"step": 2400
},
{
"epoch": 9.469196648595368,
"grad_norm": 0.6020176410675049,
"learning_rate": 1.2995137135710916e-05,
"loss": 0.0978,
"step": 2405
},
{
"epoch": 9.488910793494332,
"grad_norm": 0.633091151714325,
"learning_rate": 1.2803189296478302e-05,
"loss": 0.1047,
"step": 2410
},
{
"epoch": 9.508624938393297,
"grad_norm": 0.5339157581329346,
"learning_rate": 1.2612461268557957e-05,
"loss": 0.095,
"step": 2415
},
{
"epoch": 9.528339083292263,
"grad_norm": 0.6085830926895142,
"learning_rate": 1.2422959306577664e-05,
"loss": 0.0987,
"step": 2420
},
{
"epoch": 9.548053228191227,
"grad_norm": 0.5931851267814636,
"learning_rate": 1.2234689624958306e-05,
"loss": 0.0985,
"step": 2425
},
{
"epoch": 9.567767373090192,
"grad_norm": 0.611231803894043,
"learning_rate": 1.204765839771005e-05,
"loss": 0.1024,
"step": 2430
},
{
"epoch": 9.587481517989158,
"grad_norm": 0.5894888043403625,
"learning_rate": 1.1861871758229893e-05,
"loss": 0.0946,
"step": 2435
},
{
"epoch": 9.607195662888122,
"grad_norm": 0.5542821288108826,
"learning_rate": 1.1677335799100469e-05,
"loss": 0.094,
"step": 2440
},
{
"epoch": 9.626909807787086,
"grad_norm": 0.5735796093940735,
"learning_rate": 1.1494056571890405e-05,
"loss": 0.103,
"step": 2445
},
{
"epoch": 9.646623952686053,
"grad_norm": 0.6378740072250366,
"learning_rate": 1.1312040086955656e-05,
"loss": 0.0993,
"step": 2450
},
{
"epoch": 9.666338097585017,
"grad_norm": 0.5262072682380676,
"learning_rate": 1.1131292313242614e-05,
"loss": 0.0919,
"step": 2455
},
{
"epoch": 9.686052242483981,
"grad_norm": 0.6029082536697388,
"learning_rate": 1.0951819178092188e-05,
"loss": 0.094,
"step": 2460
},
{
"epoch": 9.705766387382948,
"grad_norm": 0.6229907870292664,
"learning_rate": 1.0773626567045547e-05,
"loss": 0.0973,
"step": 2465
},
{
"epoch": 9.725480532281912,
"grad_norm": 0.5560586452484131,
"learning_rate": 1.0596720323651077e-05,
"loss": 0.0943,
"step": 2470
},
{
"epoch": 9.745194677180876,
"grad_norm": 0.651473343372345,
"learning_rate": 1.0421106249272722e-05,
"loss": 0.0961,
"step": 2475
},
{
"epoch": 9.764908822079843,
"grad_norm": 0.7311715483665466,
"learning_rate": 1.0246790102899784e-05,
"loss": 0.1024,
"step": 2480
},
{
"epoch": 9.784622966978807,
"grad_norm": 0.5268740653991699,
"learning_rate": 1.007377760095804e-05,
"loss": 0.0907,
"step": 2485
},
{
"epoch": 9.804337111877773,
"grad_norm": 0.6558658480644226,
"learning_rate": 9.90207441712223e-06,
"loss": 0.105,
"step": 2490
},
{
"epoch": 9.824051256776738,
"grad_norm": 0.7688642740249634,
"learning_rate": 9.731686182130168e-06,
"loss": 0.1001,
"step": 2495
},
{
"epoch": 9.843765401675702,
"grad_norm": 0.5601643919944763,
"learning_rate": 9.562618483597852e-06,
"loss": 0.0959,
"step": 2500
},
{
"epoch": 9.863479546574668,
"grad_norm": 0.6482700705528259,
"learning_rate": 9.394876865836411e-06,
"loss": 0.1044,
"step": 2505
},
{
"epoch": 9.883193691473632,
"grad_norm": 0.6158556938171387,
"learning_rate": 9.228466829670246e-06,
"loss": 0.0998,
"step": 2510
},
{
"epoch": 9.902907836372597,
"grad_norm": 0.6211926341056824,
"learning_rate": 9.063393832256562e-06,
"loss": 0.1007,
"step": 2515
},
{
"epoch": 9.922621981271563,
"grad_norm": 0.577765703201294,
"learning_rate": 8.899663286906496e-06,
"loss": 0.0908,
"step": 2520
},
{
"epoch": 9.942336126170527,
"grad_norm": 0.6280847787857056,
"learning_rate": 8.737280562907574e-06,
"loss": 0.1005,
"step": 2525
},
{
"epoch": 9.962050271069492,
"grad_norm": 0.6178187727928162,
"learning_rate": 8.576250985347616e-06,
"loss": 0.1055,
"step": 2530
},
{
"epoch": 9.981764415968458,
"grad_norm": 0.7428198456764221,
"learning_rate": 8.416579834940113e-06,
"loss": 0.1005,
"step": 2535
},
{
"epoch": 10.0,
"grad_norm": 0.7962881326675415,
"learning_rate": 8.258272347851087e-06,
"loss": 0.0921,
"step": 2540
},
{
"epoch": 10.019714144898964,
"grad_norm": 0.6026802659034729,
"learning_rate": 8.101333715527283e-06,
"loss": 0.0943,
"step": 2545
},
{
"epoch": 10.03942828979793,
"grad_norm": 0.6051628589630127,
"learning_rate": 7.945769084526088e-06,
"loss": 0.0881,
"step": 2550
},
{
"epoch": 10.059142434696895,
"grad_norm": 0.6173446774482727,
"learning_rate": 7.791583556346576e-06,
"loss": 0.0965,
"step": 2555
},
{
"epoch": 10.07885657959586,
"grad_norm": 0.5841408967971802,
"learning_rate": 7.638782187262355e-06,
"loss": 0.0964,
"step": 2560
},
{
"epoch": 10.098570724494826,
"grad_norm": 0.5438782572746277,
"learning_rate": 7.487369988155662e-06,
"loss": 0.094,
"step": 2565
},
{
"epoch": 10.11828486939379,
"grad_norm": 0.5726168751716614,
"learning_rate": 7.3373519243530975e-06,
"loss": 0.0977,
"step": 2570
},
{
"epoch": 10.137999014292754,
"grad_norm": 0.5545468926429749,
"learning_rate": 7.188732915462754e-06,
"loss": 0.0885,
"step": 2575
},
{
"epoch": 10.15771315919172,
"grad_norm": 0.600556492805481,
"learning_rate": 7.0415178352129185e-06,
"loss": 0.0913,
"step": 2580
},
{
"epoch": 10.177427304090685,
"grad_norm": 0.6833725571632385,
"learning_rate": 6.8957115112922245e-06,
"loss": 0.0997,
"step": 2585
},
{
"epoch": 10.19714144898965,
"grad_norm": 0.6429631114006042,
"learning_rate": 6.7513187251913466e-06,
"loss": 0.0939,
"step": 2590
},
{
"epoch": 10.216855593888615,
"grad_norm": 0.5843133330345154,
"learning_rate": 6.608344212046164e-06,
"loss": 0.0969,
"step": 2595
},
{
"epoch": 10.23656973878758,
"grad_norm": 0.5661614537239075,
"learning_rate": 6.466792660482574e-06,
"loss": 0.0945,
"step": 2600
},
{
"epoch": 10.256283883686546,
"grad_norm": 0.6150486469268799,
"learning_rate": 6.326668712462597e-06,
"loss": 0.0907,
"step": 2605
},
{
"epoch": 10.27599802858551,
"grad_norm": 0.5596334934234619,
"learning_rate": 6.187976963132275e-06,
"loss": 0.09,
"step": 2610
},
{
"epoch": 10.295712173484475,
"grad_norm": 0.7830381393432617,
"learning_rate": 6.050721960670918e-06,
"loss": 0.1001,
"step": 2615
},
{
"epoch": 10.315426318383441,
"grad_norm": 0.5723216533660889,
"learning_rate": 5.914908206141956e-06,
"loss": 0.088,
"step": 2620
},
{
"epoch": 10.335140463282405,
"grad_norm": 0.6770078539848328,
"learning_rate": 5.780540153345354e-06,
"loss": 0.0976,
"step": 2625
},
{
"epoch": 10.35485460818137,
"grad_norm": 0.5593023896217346,
"learning_rate": 5.647622208671555e-06,
"loss": 0.0909,
"step": 2630
},
{
"epoch": 10.374568753080336,
"grad_norm": 0.5808446407318115,
"learning_rate": 5.516158730956944e-06,
"loss": 0.0956,
"step": 2635
},
{
"epoch": 10.3942828979793,
"grad_norm": 0.59349524974823,
"learning_rate": 5.386154031340962e-06,
"loss": 0.0912,
"step": 2640
},
{
"epoch": 10.413997042878265,
"grad_norm": 0.5928412675857544,
"learning_rate": 5.257612373124699e-06,
"loss": 0.0913,
"step": 2645
},
{
"epoch": 10.43371118777723,
"grad_norm": 0.618300199508667,
"learning_rate": 5.130537971631039e-06,
"loss": 0.0902,
"step": 2650
},
{
"epoch": 10.453425332676195,
"grad_norm": 0.6695954203605652,
"learning_rate": 5.004934994066557e-06,
"loss": 0.0961,
"step": 2655
},
{
"epoch": 10.47313947757516,
"grad_norm": 0.7120895385742188,
"learning_rate": 4.880807559384704e-06,
"loss": 0.0961,
"step": 2660
},
{
"epoch": 10.492853622474126,
"grad_norm": 0.5676897764205933,
"learning_rate": 4.758159738150841e-06,
"loss": 0.0956,
"step": 2665
},
{
"epoch": 10.51256776737309,
"grad_norm": 0.5924949645996094,
"learning_rate": 4.636995552408713e-06,
"loss": 0.0914,
"step": 2670
},
{
"epoch": 10.532281912272055,
"grad_norm": 0.5598572492599487,
"learning_rate": 4.517318975548557e-06,
"loss": 0.091,
"step": 2675
},
{
"epoch": 10.55199605717102,
"grad_norm": 0.5901981592178345,
"learning_rate": 4.399133932176797e-06,
"loss": 0.0885,
"step": 2680
},
{
"epoch": 10.571710202069985,
"grad_norm": 0.6229949593544006,
"learning_rate": 4.282444297987359e-06,
"loss": 0.0924,
"step": 2685
},
{
"epoch": 10.59142434696895,
"grad_norm": 0.6218413710594177,
"learning_rate": 4.16725389963451e-06,
"loss": 0.0923,
"step": 2690
},
{
"epoch": 10.611138491867916,
"grad_norm": 0.7330309748649597,
"learning_rate": 4.053566514607504e-06,
"loss": 0.0962,
"step": 2695
},
{
"epoch": 10.63085263676688,
"grad_norm": 0.6477744579315186,
"learning_rate": 3.941385871106534e-06,
"loss": 0.0925,
"step": 2700
},
{
"epoch": 10.650566781665844,
"grad_norm": 0.582679033279419,
"learning_rate": 3.830715647920629e-06,
"loss": 0.0958,
"step": 2705
},
{
"epoch": 10.67028092656481,
"grad_norm": 0.5494212508201599,
"learning_rate": 3.721559474306896e-06,
"loss": 0.0984,
"step": 2710
},
{
"epoch": 10.689995071463775,
"grad_norm": 0.5505300164222717,
"learning_rate": 3.6139209298715616e-06,
"loss": 0.094,
"step": 2715
},
{
"epoch": 10.70970921636274,
"grad_norm": 0.752632200717926,
"learning_rate": 3.507803544452587e-06,
"loss": 0.0945,
"step": 2720
},
{
"epoch": 10.729423361261706,
"grad_norm": 0.6900667548179626,
"learning_rate": 3.403210798003892e-06,
"loss": 0.0923,
"step": 2725
},
{
"epoch": 10.74913750616067,
"grad_norm": 0.66816246509552,
"learning_rate": 3.300146120481229e-06,
"loss": 0.0939,
"step": 2730
},
{
"epoch": 10.768851651059634,
"grad_norm": 0.6422622203826904,
"learning_rate": 3.1986128917297464e-06,
"loss": 0.0926,
"step": 2735
},
{
"epoch": 10.7885657959586,
"grad_norm": 0.6510041356086731,
"learning_rate": 3.0986144413730876e-06,
"loss": 0.0929,
"step": 2740
},
{
"epoch": 10.808279940857565,
"grad_norm": 0.5580385327339172,
"learning_rate": 3.000154048704268e-06,
"loss": 0.096,
"step": 2745
},
{
"epoch": 10.827994085756531,
"grad_norm": 0.6091373562812805,
"learning_rate": 2.9032349425780813e-06,
"loss": 0.0931,
"step": 2750
},
{
"epoch": 10.847708230655495,
"grad_norm": 0.6932477951049805,
"learning_rate": 2.8078603013052263e-06,
"loss": 0.0966,
"step": 2755
},
{
"epoch": 10.86742237555446,
"grad_norm": 0.592022716999054,
"learning_rate": 2.714033252548137e-06,
"loss": 0.0964,
"step": 2760
},
{
"epoch": 10.887136520453426,
"grad_norm": 0.6746100187301636,
"learning_rate": 2.621756873218312e-06,
"loss": 0.0924,
"step": 2765
},
{
"epoch": 10.90685066535239,
"grad_norm": 0.6170192360877991,
"learning_rate": 2.531034189375503e-06,
"loss": 0.0915,
"step": 2770
},
{
"epoch": 10.926564810251355,
"grad_norm": 0.6232552528381348,
"learning_rate": 2.4418681761284255e-06,
"loss": 0.0916,
"step": 2775
},
{
"epoch": 10.946278955150321,
"grad_norm": 0.5538495779037476,
"learning_rate": 2.354261757537235e-06,
"loss": 0.0926,
"step": 2780
},
{
"epoch": 10.965993100049285,
"grad_norm": 0.5978652834892273,
"learning_rate": 2.268217806517592e-06,
"loss": 0.0878,
"step": 2785
},
{
"epoch": 10.98570724494825,
"grad_norm": 0.5522828102111816,
"learning_rate": 2.183739144746494e-06,
"loss": 0.093,
"step": 2790
}
],
"logging_steps": 5,
"max_steps": 3048,
"num_input_tokens_seen": 0,
"num_train_epochs": 12,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 4.384719421932749e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}