| [ | |
| { | |
| "loss": 2.096086181640625, | |
| "learning_rate": 4.4444444444444447e-05, | |
| "epoch": 0.2222222222222222, | |
| "total_flos": 1529116360704000, | |
| "step": 500 | |
| }, | |
| { | |
| "loss": 1.851991455078125, | |
| "learning_rate": 3.888888888888889e-05, | |
| "epoch": 0.4444444444444444, | |
| "total_flos": 3058232721408000, | |
| "step": 1000 | |
| }, | |
| { | |
| "loss": 1.8009189453125, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "epoch": 0.6666666666666666, | |
| "total_flos": 4587349082112000, | |
| "step": 1500 | |
| }, | |
| { | |
| "loss": 1.770126953125, | |
| "learning_rate": 2.777777777777778e-05, | |
| "epoch": 0.8888888888888888, | |
| "total_flos": 6116465442816000, | |
| "step": 2000 | |
| }, | |
| { | |
| "loss": 1.71751513671875, | |
| "learning_rate": 2.2222222222222223e-05, | |
| "epoch": 1.1111111111111112, | |
| "total_flos": 7645581803520000, | |
| "step": 2500 | |
| }, | |
| { | |
| "loss": 1.6680078125, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "epoch": 1.3333333333333333, | |
| "total_flos": 9174698164224000, | |
| "step": 3000 | |
| }, | |
| { | |
| "loss": 1.6722744140625, | |
| "learning_rate": 1.1111111111111112e-05, | |
| "epoch": 1.5555555555555556, | |
| "total_flos": 10703814524928000, | |
| "step": 3500 | |
| }, | |
| { | |
| "loss": 1.64707421875, | |
| "learning_rate": 5.555555555555556e-06, | |
| "epoch": 1.7777777777777777, | |
| "total_flos": 12232930885632000, | |
| "step": 4000 | |
| }, | |
| { | |
| "loss": 1.667453125, | |
| "learning_rate": 0.0, | |
| "epoch": 2.0, | |
| "total_flos": 13762047246336000, | |
| "step": 4500 | |
| } | |
| ] |