| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 1689, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.9229, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.0000000000000004e-07, | |
| "loss": 0.7591, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 6.000000000000001e-07, | |
| "loss": 0.7555, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 9e-07, | |
| "loss": 0.7155, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.2000000000000002e-06, | |
| "loss": 0.6815, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.5e-06, | |
| "loss": 0.6505, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.8e-06, | |
| "loss": 0.5808, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.1e-06, | |
| "loss": 0.6016, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.4000000000000003e-06, | |
| "loss": 0.5805, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.7e-06, | |
| "loss": 0.5634, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3e-06, | |
| "loss": 0.563, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.3e-06, | |
| "loss": 0.5584, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.6e-06, | |
| "loss": 0.5985, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.9e-06, | |
| "loss": 0.5701, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.2e-06, | |
| "loss": 0.581, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.5e-06, | |
| "loss": 0.5583, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.800000000000001e-06, | |
| "loss": 0.5464, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.1e-06, | |
| "loss": 0.5283, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.4e-06, | |
| "loss": 0.5393, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 5.7e-06, | |
| "loss": 0.5583, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 6e-06, | |
| "loss": 0.5323, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 5.9964050329538645e-06, | |
| "loss": 0.5669, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 5.99281006590773e-06, | |
| "loss": 0.5353, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 5.989215098861594e-06, | |
| "loss": 0.5198, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 5.985620131815458e-06, | |
| "loss": 0.5225, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 5.9820251647693235e-06, | |
| "loss": 0.5484, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 5.978430197723188e-06, | |
| "loss": 0.5469, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 5.974835230677052e-06, | |
| "loss": 0.5411, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 5.9712402636309165e-06, | |
| "loss": 0.5394, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 5.967645296584782e-06, | |
| "loss": 0.5481, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 5.964050329538646e-06, | |
| "loss": 0.5559, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 5.960455362492511e-06, | |
| "loss": 0.5317, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 5.956860395446375e-06, | |
| "loss": 0.5266, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 5.95326542840024e-06, | |
| "loss": 0.5074, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 5.949670461354104e-06, | |
| "loss": 0.542, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 5.946075494307969e-06, | |
| "loss": 0.5562, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 5.942480527261834e-06, | |
| "loss": 0.5381, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 5.938885560215698e-06, | |
| "loss": 0.5434, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 5.935290593169563e-06, | |
| "loss": 0.5447, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 5.9316956261234274e-06, | |
| "loss": 0.5155, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 5.928100659077292e-06, | |
| "loss": 0.5184, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 5.924505692031156e-06, | |
| "loss": 0.5373, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 5.920910724985021e-06, | |
| "loss": 0.549, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 5.917315757938886e-06, | |
| "loss": 0.5086, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 5.913720790892751e-06, | |
| "loss": 0.5366, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 5.910125823846615e-06, | |
| "loss": 0.5262, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 5.906530856800479e-06, | |
| "loss": 0.5318, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 5.902935889754344e-06, | |
| "loss": 0.5292, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 5.899340922708209e-06, | |
| "loss": 0.5335, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 5.895745955662073e-06, | |
| "loss": 0.5229, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 5.892150988615938e-06, | |
| "loss": 0.526, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 5.888556021569802e-06, | |
| "loss": 0.5172, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 5.884961054523667e-06, | |
| "loss": 0.5351, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 5.881366087477531e-06, | |
| "loss": 0.5248, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 5.8777711204313966e-06, | |
| "loss": 0.5031, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 5.874176153385261e-06, | |
| "loss": 0.5394, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 5.870581186339125e-06, | |
| "loss": 0.5122, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 5.8669862192929895e-06, | |
| "loss": 0.5237, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 5.863391252246855e-06, | |
| "loss": 0.5229, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 5.859796285200719e-06, | |
| "loss": 0.5297, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 5.856201318154583e-06, | |
| "loss": 0.5063, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 5.8526063511084485e-06, | |
| "loss": 0.4923, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 5.849011384062313e-06, | |
| "loss": 0.5045, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 5.845416417016178e-06, | |
| "loss": 0.5167, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 5.841821449970042e-06, | |
| "loss": 0.5266, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 5.838226482923907e-06, | |
| "loss": 0.5086, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 5.834631515877771e-06, | |
| "loss": 0.5646, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 5.831036548831636e-06, | |
| "loss": 0.5131, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 5.8274415817855005e-06, | |
| "loss": 0.5443, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 5.823846614739365e-06, | |
| "loss": 0.5233, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 5.820251647693229e-06, | |
| "loss": 0.5155, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 5.816656680647094e-06, | |
| "loss": 0.504, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 5.813061713600959e-06, | |
| "loss": 0.5143, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 5.809466746554824e-06, | |
| "loss": 0.4983, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 5.805871779508688e-06, | |
| "loss": 0.4986, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 5.8022768124625525e-06, | |
| "loss": 0.5211, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 5.798681845416417e-06, | |
| "loss": 0.5161, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 5.795086878370282e-06, | |
| "loss": 0.5029, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 5.791491911324146e-06, | |
| "loss": 0.533, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 5.787896944278011e-06, | |
| "loss": 0.5426, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 5.784301977231876e-06, | |
| "loss": 0.5069, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 5.78070701018574e-06, | |
| "loss": 0.5485, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 5.777112043139605e-06, | |
| "loss": 0.5172, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 5.773517076093469e-06, | |
| "loss": 0.5455, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 5.769922109047334e-06, | |
| "loss": 0.4776, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 5.766327142001198e-06, | |
| "loss": 0.5452, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 5.7627321749550635e-06, | |
| "loss": 0.5015, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 5.759137207908928e-06, | |
| "loss": 0.5196, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 5.755542240862792e-06, | |
| "loss": 0.5074, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 5.7519472738166564e-06, | |
| "loss": 0.4911, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 5.748352306770522e-06, | |
| "loss": 0.4974, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 5.744757339724386e-06, | |
| "loss": 0.5116, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 5.74116237267825e-06, | |
| "loss": 0.4898, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 5.7375674056321154e-06, | |
| "loss": 0.5065, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 5.73397243858598e-06, | |
| "loss": 0.54, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 5.730377471539844e-06, | |
| "loss": 0.5341, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 5.726782504493709e-06, | |
| "loss": 0.5095, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 5.723187537447574e-06, | |
| "loss": 0.4894, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 5.719592570401438e-06, | |
| "loss": 0.491, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 5.715997603355303e-06, | |
| "loss": 0.4862, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 5.712402636309167e-06, | |
| "loss": 0.4942, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 5.708807669263033e-06, | |
| "loss": 0.5128, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 5.705212702216896e-06, | |
| "loss": 0.5089, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 5.701617735170761e-06, | |
| "loss": 0.5187, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 5.6980227681246256e-06, | |
| "loss": 0.5084, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 5.694427801078491e-06, | |
| "loss": 0.5039, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 5.690832834032354e-06, | |
| "loss": 0.5023, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 5.687237866986219e-06, | |
| "loss": 0.5209, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 5.683642899940084e-06, | |
| "loss": 0.5398, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 5.680047932893949e-06, | |
| "loss": 0.4603, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 5.676452965847813e-06, | |
| "loss": 0.5129, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 5.6728579988016775e-06, | |
| "loss": 0.5323, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 5.669263031755542e-06, | |
| "loss": 0.5372, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 5.665668064709407e-06, | |
| "loss": 0.5323, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 5.662073097663271e-06, | |
| "loss": 0.504, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 5.6584781306171365e-06, | |
| "loss": 0.5196, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 5.654883163571001e-06, | |
| "loss": 0.4924, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 5.651288196524865e-06, | |
| "loss": 0.5009, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 5.64769322947873e-06, | |
| "loss": 0.5275, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 5.644098262432595e-06, | |
| "loss": 0.5115, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 5.640503295386459e-06, | |
| "loss": 0.5214, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 5.636908328340323e-06, | |
| "loss": 0.5126, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 5.6333133612941885e-06, | |
| "loss": 0.4869, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 5.629718394248053e-06, | |
| "loss": 0.5011, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 5.626123427201918e-06, | |
| "loss": 0.4977, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 5.6225284601557815e-06, | |
| "loss": 0.4864, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 5.618933493109647e-06, | |
| "loss": 0.5267, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 5.615338526063511e-06, | |
| "loss": 0.509, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 5.611743559017376e-06, | |
| "loss": 0.5008, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 5.6081485919712405e-06, | |
| "loss": 0.486, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 5.604553624925105e-06, | |
| "loss": 0.5118, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 5.600958657878969e-06, | |
| "loss": 0.4975, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 5.597363690832834e-06, | |
| "loss": 0.5084, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 5.593768723786699e-06, | |
| "loss": 0.5228, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 5.590173756740563e-06, | |
| "loss": 0.4748, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 5.586578789694428e-06, | |
| "loss": 0.5363, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 5.5829838226482925e-06, | |
| "loss": 0.5093, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 5.579388855602158e-06, | |
| "loss": 0.5086, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 5.575793888556022e-06, | |
| "loss": 0.5245, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 5.572198921509886e-06, | |
| "loss": 0.5289, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 5.568603954463751e-06, | |
| "loss": 0.5149, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 5.565008987417616e-06, | |
| "loss": 0.5001, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 5.56141402037148e-06, | |
| "loss": 0.4816, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 5.5578190533253444e-06, | |
| "loss": 0.5222, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 5.554224086279209e-06, | |
| "loss": 0.5178, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 5.550629119233074e-06, | |
| "loss": 0.53, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 5.547034152186938e-06, | |
| "loss": 0.5055, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 5.5434391851408034e-06, | |
| "loss": 0.4976, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 5.539844218094667e-06, | |
| "loss": 0.5311, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 5.536249251048532e-06, | |
| "loss": 0.5513, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 5.5326542840023964e-06, | |
| "loss": 0.5028, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 5.529059316956262e-06, | |
| "loss": 0.4915, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 5.525464349910126e-06, | |
| "loss": 0.5201, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 5.52186938286399e-06, | |
| "loss": 0.4976, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 5.518274415817855e-06, | |
| "loss": 0.4995, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 5.51467944877172e-06, | |
| "loss": 0.4812, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 5.511084481725585e-06, | |
| "loss": 0.5333, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 5.507489514679448e-06, | |
| "loss": 0.5085, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 5.5038945476333136e-06, | |
| "loss": 0.4969, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 5.500299580587178e-06, | |
| "loss": 0.5115, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 5.496704613541043e-06, | |
| "loss": 0.5235, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 5.493109646494907e-06, | |
| "loss": 0.501, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 5.489514679448772e-06, | |
| "loss": 0.4767, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 5.485919712402636e-06, | |
| "loss": 0.5356, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 5.482324745356501e-06, | |
| "loss": 0.5169, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 5.4787297783103656e-06, | |
| "loss": 0.5095, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 5.475134811264231e-06, | |
| "loss": 0.5102, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 5.471539844218094e-06, | |
| "loss": 0.5152, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 5.467944877171959e-06, | |
| "loss": 0.5144, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 5.464349910125824e-06, | |
| "loss": 0.5094, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 5.460754943079689e-06, | |
| "loss": 0.4966, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 5.457159976033553e-06, | |
| "loss": 0.5369, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 5.4535650089874175e-06, | |
| "loss": 0.5108, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 5.449970041941283e-06, | |
| "loss": 0.5016, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 5.446375074895147e-06, | |
| "loss": 0.4809, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 5.442780107849012e-06, | |
| "loss": 0.5172, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 5.439185140802876e-06, | |
| "loss": 0.5293, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 5.435590173756741e-06, | |
| "loss": 0.504, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 5.431995206710605e-06, | |
| "loss": 0.5172, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 5.42840023966447e-06, | |
| "loss": 0.5095, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 5.424805272618334e-06, | |
| "loss": 0.5142, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 5.421210305572199e-06, | |
| "loss": 0.5388, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 5.417615338526063e-06, | |
| "loss": 0.4757, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 5.4140203714799285e-06, | |
| "loss": 0.4978, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 5.410425404433793e-06, | |
| "loss": 0.4793, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 5.406830437387657e-06, | |
| "loss": 0.5079, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 5.4032354703415215e-06, | |
| "loss": 0.4696, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 5.399640503295387e-06, | |
| "loss": 0.5044, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 5.396045536249251e-06, | |
| "loss": 0.5053, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 5.392450569203116e-06, | |
| "loss": 0.5167, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 5.3888556021569805e-06, | |
| "loss": 0.5012, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 5.385260635110845e-06, | |
| "loss": 0.5129, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 5.38166566806471e-06, | |
| "loss": 0.4819, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 5.378070701018574e-06, | |
| "loss": 0.4994, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 5.374475733972439e-06, | |
| "loss": 0.5041, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 5.370880766926303e-06, | |
| "loss": 0.4841, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 5.367285799880168e-06, | |
| "loss": 0.5153, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 5.3636908328340325e-06, | |
| "loss": 0.4955, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 5.360095865787898e-06, | |
| "loss": 0.4738, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 5.356500898741761e-06, | |
| "loss": 0.5227, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 5.352905931695626e-06, | |
| "loss": 0.4975, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 5.349310964649491e-06, | |
| "loss": 0.4952, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 5.345715997603356e-06, | |
| "loss": 0.515, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 5.34212103055722e-06, | |
| "loss": 0.4993, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 5.3385260635110844e-06, | |
| "loss": 0.5063, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 5.334931096464949e-06, | |
| "loss": 0.4895, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 5.331336129418814e-06, | |
| "loss": 0.5019, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 5.327741162372678e-06, | |
| "loss": 0.4873, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 5.324146195326543e-06, | |
| "loss": 0.5416, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 5.320551228280408e-06, | |
| "loss": 0.5023, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 5.316956261234272e-06, | |
| "loss": 0.504, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 5.313361294188137e-06, | |
| "loss": 0.5139, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 5.309766327142002e-06, | |
| "loss": 0.5052, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 5.306171360095866e-06, | |
| "loss": 0.5271, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 5.30257639304973e-06, | |
| "loss": 0.4788, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 5.298981426003595e-06, | |
| "loss": 0.4837, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 5.29538645895746e-06, | |
| "loss": 0.5215, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 5.291791491911325e-06, | |
| "loss": 0.5041, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 5.288196524865188e-06, | |
| "loss": 0.4784, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 5.2846015578190536e-06, | |
| "loss": 0.5228, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 5.281006590772918e-06, | |
| "loss": 0.4943, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 5.277411623726783e-06, | |
| "loss": 0.5341, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 5.2738166566806465e-06, | |
| "loss": 0.4695, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 5.270221689634512e-06, | |
| "loss": 0.5101, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 5.266626722588376e-06, | |
| "loss": 0.4611, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 5.263031755542241e-06, | |
| "loss": 0.537, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 5.2594367884961055e-06, | |
| "loss": 0.5155, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 5.25584182144997e-06, | |
| "loss": 0.4902, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 5.252246854403835e-06, | |
| "loss": 0.4946, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 5.248651887357699e-06, | |
| "loss": 0.5033, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 5.2450569203115645e-06, | |
| "loss": 0.4956, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 5.241461953265428e-06, | |
| "loss": 0.5244, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 5.237866986219293e-06, | |
| "loss": 0.4782, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 5.2342720191731575e-06, | |
| "loss": 0.4817, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 5.230677052127023e-06, | |
| "loss": 0.5028, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 5.227082085080887e-06, | |
| "loss": 0.5069, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 5.223487118034751e-06, | |
| "loss": 0.5317, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 5.219892150988616e-06, | |
| "loss": 0.5032, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 5.216297183942481e-06, | |
| "loss": 0.493, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 5.212702216896345e-06, | |
| "loss": 0.5169, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 5.20910724985021e-06, | |
| "loss": 0.5045, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 5.205512282804074e-06, | |
| "loss": 0.5146, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 5.201917315757939e-06, | |
| "loss": 0.4969, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 5.198322348711803e-06, | |
| "loss": 0.4775, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 5.1947273816656685e-06, | |
| "loss": 0.4821, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 5.191132414619533e-06, | |
| "loss": 0.5013, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 5.187537447573397e-06, | |
| "loss": 0.4941, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 5.183942480527262e-06, | |
| "loss": 0.508, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 5.180347513481127e-06, | |
| "loss": 0.5256, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 5.176752546434992e-06, | |
| "loss": 0.4734, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 5.173157579388855e-06, | |
| "loss": 0.5134, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 5.1695626123427205e-06, | |
| "loss": 0.4926, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 5.165967645296585e-06, | |
| "loss": 0.5037, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 5.16237267825045e-06, | |
| "loss": 0.5102, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 5.158777711204314e-06, | |
| "loss": 0.4969, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 5.155182744158179e-06, | |
| "loss": 0.507, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 5.151587777112043e-06, | |
| "loss": 0.5127, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 5.147992810065908e-06, | |
| "loss": 0.4634, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 5.1443978430197724e-06, | |
| "loss": 0.5062, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 5.140802875973637e-06, | |
| "loss": 0.4894, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 5.137207908927501e-06, | |
| "loss": 0.4967, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 5.133612941881366e-06, | |
| "loss": 0.4955, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 5.130017974835231e-06, | |
| "loss": 0.4918, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 5.126423007789096e-06, | |
| "loss": 0.5135, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 5.12282804074296e-06, | |
| "loss": 0.5452, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 5.119233073696824e-06, | |
| "loss": 0.5095, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 5.11563810665069e-06, | |
| "loss": 0.4706, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 5.112043139604554e-06, | |
| "loss": 0.4796, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 5.108448172558419e-06, | |
| "loss": 0.5066, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 5.1048532055122826e-06, | |
| "loss": 0.512, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 5.101258238466148e-06, | |
| "loss": 0.4992, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 5.097663271420012e-06, | |
| "loss": 0.4764, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 5.094068304373877e-06, | |
| "loss": 0.4852, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 5.090473337327741e-06, | |
| "loss": 0.4709, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 5.086878370281606e-06, | |
| "loss": 0.4892, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 5.08328340323547e-06, | |
| "loss": 0.4951, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 5.079688436189335e-06, | |
| "loss": 0.4717, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 5.0760934691432e-06, | |
| "loss": 0.4874, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 5.072498502097064e-06, | |
| "loss": 0.4847, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 5.068903535050928e-06, | |
| "loss": 0.5009, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 5.0653085680047935e-06, | |
| "loss": 0.5244, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 5.061713600958658e-06, | |
| "loss": 0.4905, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 5.058118633912522e-06, | |
| "loss": 0.4975, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 5.054523666866387e-06, | |
| "loss": 0.4882, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 5.050928699820252e-06, | |
| "loss": 0.5256, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 5.047333732774117e-06, | |
| "loss": 0.4818, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 5.043738765727981e-06, | |
| "loss": 0.4779, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 5.0401437986818455e-06, | |
| "loss": 0.5093, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 5.03654883163571e-06, | |
| "loss": 0.4778, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 5.032953864589575e-06, | |
| "loss": 0.5105, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 5.029358897543439e-06, | |
| "loss": 0.4991, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 5.0257639304973045e-06, | |
| "loss": 0.5075, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 5.022168963451168e-06, | |
| "loss": 0.491, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 5.018573996405033e-06, | |
| "loss": 0.4922, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 5.0149790293588975e-06, | |
| "loss": 0.4915, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 5.011384062312763e-06, | |
| "loss": 0.4806, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 5.007789095266626e-06, | |
| "loss": 0.5022, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 5.004194128220491e-06, | |
| "loss": 0.4752, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 5.000599161174356e-06, | |
| "loss": 0.5015, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.997004194128221e-06, | |
| "loss": 0.5004, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.993409227082085e-06, | |
| "loss": 0.5451, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.9898142600359495e-06, | |
| "loss": 0.4991, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.986219292989815e-06, | |
| "loss": 0.4725, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.982624325943679e-06, | |
| "loss": 0.475, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.979029358897544e-06, | |
| "loss": 0.5085, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.9754343918514085e-06, | |
| "loss": 0.4764, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.971839424805273e-06, | |
| "loss": 0.4855, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.968244457759137e-06, | |
| "loss": 0.4773, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.964649490713002e-06, | |
| "loss": 0.5025, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.961054523666867e-06, | |
| "loss": 0.5135, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.957459556620731e-06, | |
| "loss": 0.4944, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.953864589574595e-06, | |
| "loss": 0.4785, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.9502696225284604e-06, | |
| "loss": 0.4955, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.946674655482325e-06, | |
| "loss": 0.4975, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.94307968843619e-06, | |
| "loss": 0.5225, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.9394847213900534e-06, | |
| "loss": 0.506, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.935889754343919e-06, | |
| "loss": 0.4723, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.932294787297783e-06, | |
| "loss": 0.4832, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.928699820251648e-06, | |
| "loss": 0.5078, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.9251048532055124e-06, | |
| "loss": 0.5033, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.921509886159377e-06, | |
| "loss": 0.4895, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.917914919113242e-06, | |
| "loss": 0.4989, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.914319952067106e-06, | |
| "loss": 0.5324, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.9107249850209714e-06, | |
| "loss": 0.491, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.907130017974835e-06, | |
| "loss": 0.4842, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.9035350509287e-06, | |
| "loss": 0.4999, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.899940083882564e-06, | |
| "loss": 0.4861, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.8963451168364296e-06, | |
| "loss": 0.4713, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.892750149790294e-06, | |
| "loss": 0.4813, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.889155182744158e-06, | |
| "loss": 0.4852, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.8855602156980226e-06, | |
| "loss": 0.5048, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.881965248651888e-06, | |
| "loss": 0.4889, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.878370281605752e-06, | |
| "loss": 0.5085, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.874775314559616e-06, | |
| "loss": 0.5087, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.871180347513481e-06, | |
| "loss": 0.5106, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.867585380467346e-06, | |
| "loss": 0.4768, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.86399041342121e-06, | |
| "loss": 0.4941, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.860395446375075e-06, | |
| "loss": 0.4809, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.85680047932894e-06, | |
| "loss": 0.4988, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.853205512282804e-06, | |
| "loss": 0.532, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.849610545236669e-06, | |
| "loss": 0.4794, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.8460155781905335e-06, | |
| "loss": 0.4897, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.842420611144398e-06, | |
| "loss": 0.5108, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.838825644098262e-06, | |
| "loss": 0.4896, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.835230677052127e-06, | |
| "loss": 0.4659, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.831635710005992e-06, | |
| "loss": 0.4569, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.828040742959857e-06, | |
| "loss": 0.4807, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.82444577591372e-06, | |
| "loss": 0.4759, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.8208508088675855e-06, | |
| "loss": 0.4879, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.81725584182145e-06, | |
| "loss": 0.4903, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.813660874775315e-06, | |
| "loss": 0.5096, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.810065907729179e-06, | |
| "loss": 0.4695, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.806470940683044e-06, | |
| "loss": 0.4806, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.802875973636908e-06, | |
| "loss": 0.4945, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.799281006590773e-06, | |
| "loss": 0.4978, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.7956860395446375e-06, | |
| "loss": 0.4941, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.792091072498503e-06, | |
| "loss": 0.5129, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.788496105452367e-06, | |
| "loss": 0.4641, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.784901138406231e-06, | |
| "loss": 0.4837, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.7813061713600965e-06, | |
| "loss": 0.4905, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.777711204313961e-06, | |
| "loss": 0.5069, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.774116237267825e-06, | |
| "loss": 0.4844, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.7705212702216895e-06, | |
| "loss": 0.4882, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.766926303175555e-06, | |
| "loss": 0.5054, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.763331336129419e-06, | |
| "loss": 0.501, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.759736369083284e-06, | |
| "loss": 0.4767, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.756141402037148e-06, | |
| "loss": 0.4982, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.752546434991013e-06, | |
| "loss": 0.5001, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.748951467944877e-06, | |
| "loss": 0.4883, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.745356500898742e-06, | |
| "loss": 0.5125, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.741761533852607e-06, | |
| "loss": 0.4736, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.738166566806471e-06, | |
| "loss": 0.5074, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.734571599760335e-06, | |
| "loss": 0.4762, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.7309766327142004e-06, | |
| "loss": 0.4984, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.727381665668065e-06, | |
| "loss": 0.4808, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.723786698621929e-06, | |
| "loss": 0.4801, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.720191731575794e-06, | |
| "loss": 0.4886, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.716596764529659e-06, | |
| "loss": 0.5019, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.713001797483523e-06, | |
| "loss": 0.5053, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.709406830437388e-06, | |
| "loss": 0.4908, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.705811863391252e-06, | |
| "loss": 0.5027, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.702216896345117e-06, | |
| "loss": 0.4523, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.698621929298982e-06, | |
| "loss": 0.4745, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.695026962252846e-06, | |
| "loss": 0.5041, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.6914319952067106e-06, | |
| "loss": 0.495, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.687837028160575e-06, | |
| "loss": 0.5137, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.68424206111444e-06, | |
| "loss": 0.4938, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.680647094068304e-06, | |
| "loss": 0.4915, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.6770521270221696e-06, | |
| "loss": 0.4986, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.673457159976033e-06, | |
| "loss": 0.4995, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.669862192929898e-06, | |
| "loss": 0.4721, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.6662672258837625e-06, | |
| "loss": 0.5002, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.662672258837628e-06, | |
| "loss": 0.4997, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.659077291791492e-06, | |
| "loss": 0.4821, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.655482324745356e-06, | |
| "loss": 0.5035, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.6518873576992215e-06, | |
| "loss": 0.5115, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.648292390653086e-06, | |
| "loss": 0.4669, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.64469742360695e-06, | |
| "loss": 0.4968, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.6411024565608145e-06, | |
| "loss": 0.4757, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.63750748951468e-06, | |
| "loss": 0.5114, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.633912522468544e-06, | |
| "loss": 0.4795, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.630317555422409e-06, | |
| "loss": 0.4783, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.6267225883762735e-06, | |
| "loss": 0.4855, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.623127621330138e-06, | |
| "loss": 0.4748, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.619532654284002e-06, | |
| "loss": 0.5073, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.615937687237867e-06, | |
| "loss": 0.5255, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.612342720191732e-06, | |
| "loss": 0.4903, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.608747753145597e-06, | |
| "loss": 0.4922, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.60515278609946e-06, | |
| "loss": 0.4746, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.6015578190533255e-06, | |
| "loss": 0.5177, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.59796285200719e-06, | |
| "loss": 0.5198, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.594367884961055e-06, | |
| "loss": 0.5077, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.590772917914919e-06, | |
| "loss": 0.473, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.587177950868784e-06, | |
| "loss": 0.5046, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.583582983822649e-06, | |
| "loss": 0.5046, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.579988016776513e-06, | |
| "loss": 0.4942, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.5763930497303775e-06, | |
| "loss": 0.4893, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.572798082684242e-06, | |
| "loss": 0.4883, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.569203115638107e-06, | |
| "loss": 0.5128, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.565608148591971e-06, | |
| "loss": 0.4822, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.5620131815458365e-06, | |
| "loss": 0.4791, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.558418214499701e-06, | |
| "loss": 0.4976, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.554823247453565e-06, | |
| "loss": 0.4866, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.5512282804074294e-06, | |
| "loss": 0.4893, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.547633313361295e-06, | |
| "loss": 0.4833, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.544038346315159e-06, | |
| "loss": 0.4658, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.540443379269023e-06, | |
| "loss": 0.5027, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.536848412222888e-06, | |
| "loss": 0.4968, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.533253445176753e-06, | |
| "loss": 0.487, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.529658478130617e-06, | |
| "loss": 0.5291, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.526063511084482e-06, | |
| "loss": 0.4751, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.522468544038347e-06, | |
| "loss": 0.4994, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.518873576992211e-06, | |
| "loss": 0.4601, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.515278609946075e-06, | |
| "loss": 0.4834, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.51168364289994e-06, | |
| "loss": 0.4551, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.508088675853805e-06, | |
| "loss": 0.5271, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.504493708807669e-06, | |
| "loss": 0.4846, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.500898741761534e-06, | |
| "loss": 0.4867, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.4973037747153986e-06, | |
| "loss": 0.476, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.493708807669264e-06, | |
| "loss": 0.5, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.490113840623127e-06, | |
| "loss": 0.5228, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.486518873576992e-06, | |
| "loss": 0.4965, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.482923906530857e-06, | |
| "loss": 0.4782, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.479328939484722e-06, | |
| "loss": 0.5109, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.475733972438586e-06, | |
| "loss": 0.4992, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.4721390053924505e-06, | |
| "loss": 0.5033, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.468544038346315e-06, | |
| "loss": 0.4931, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.46494907130018e-06, | |
| "loss": 0.4994, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.461354104254044e-06, | |
| "loss": 0.4809, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.457759137207909e-06, | |
| "loss": 0.4824, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.454164170161774e-06, | |
| "loss": 0.5227, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.450569203115638e-06, | |
| "loss": 0.4839, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.4469742360695025e-06, | |
| "loss": 0.5037, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.443379269023368e-06, | |
| "loss": 0.4911, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.439784301977232e-06, | |
| "loss": 0.4992, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.436189334931096e-06, | |
| "loss": 0.506, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.4325943678849615e-06, | |
| "loss": 0.5018, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.428999400838826e-06, | |
| "loss": 0.4862, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.425404433792691e-06, | |
| "loss": 0.4935, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.4218094667465545e-06, | |
| "loss": 0.5005, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.41821449970042e-06, | |
| "loss": 0.479, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.414619532654284e-06, | |
| "loss": 0.4998, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.411024565608149e-06, | |
| "loss": 0.4709, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.407429598562013e-06, | |
| "loss": 0.4692, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.403834631515878e-06, | |
| "loss": 0.4906, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.400239664469742e-06, | |
| "loss": 0.4811, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.396644697423607e-06, | |
| "loss": 0.4752, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.393049730377472e-06, | |
| "loss": 0.4853, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.389454763331336e-06, | |
| "loss": 0.4925, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.385859796285201e-06, | |
| "loss": 0.4842, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.3822648292390655e-06, | |
| "loss": 0.5173, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.37866986219293e-06, | |
| "loss": 0.4944, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.375074895146795e-06, | |
| "loss": 0.4875, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.371479928100659e-06, | |
| "loss": 0.488, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.367884961054524e-06, | |
| "loss": 0.4868, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.364289994008389e-06, | |
| "loss": 0.4846, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.360695026962253e-06, | |
| "loss": 0.4968, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.3571000599161175e-06, | |
| "loss": 0.4941, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.353505092869982e-06, | |
| "loss": 0.4791, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.349910125823847e-06, | |
| "loss": 0.5, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.346315158777711e-06, | |
| "loss": 0.5054, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.3427201917315765e-06, | |
| "loss": 0.4615, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.33912522468544e-06, | |
| "loss": 0.5011, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.335530257639305e-06, | |
| "loss": 0.4855, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.3319352905931694e-06, | |
| "loss": 0.473, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.328340323547035e-06, | |
| "loss": 0.4815, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.324745356500899e-06, | |
| "loss": 0.5057, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.321150389454763e-06, | |
| "loss": 0.4494, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.317555422408628e-06, | |
| "loss": 0.4738, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.313960455362493e-06, | |
| "loss": 0.4973, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.310365488316357e-06, | |
| "loss": 0.504, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.306770521270221e-06, | |
| "loss": 0.4849, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.303175554224087e-06, | |
| "loss": 0.4936, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.299580587177951e-06, | |
| "loss": 0.4821, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.295985620131816e-06, | |
| "loss": 0.467, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.29239065308568e-06, | |
| "loss": 0.4891, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.288795686039545e-06, | |
| "loss": 0.4996, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.285200718993409e-06, | |
| "loss": 0.458, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.281605751947274e-06, | |
| "loss": 0.4897, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.2780107849011386e-06, | |
| "loss": 0.4938, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.274415817855003e-06, | |
| "loss": 0.4874, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.270820850808867e-06, | |
| "loss": 0.5215, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.267225883762732e-06, | |
| "loss": 0.5014, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.263630916716597e-06, | |
| "loss": 0.4803, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.260035949670462e-06, | |
| "loss": 0.5022, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.256440982624326e-06, | |
| "loss": 0.5074, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.2528460155781905e-06, | |
| "loss": 0.4959, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.249251048532055e-06, | |
| "loss": 0.5096, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.24565608148592e-06, | |
| "loss": 0.4762, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.242061114439784e-06, | |
| "loss": 0.5014, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.238466147393649e-06, | |
| "loss": 0.4846, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.234871180347514e-06, | |
| "loss": 0.4862, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.231276213301378e-06, | |
| "loss": 0.4589, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.227681246255243e-06, | |
| "loss": 0.4955, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.224086279209107e-06, | |
| "loss": 0.4754, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.220491312162972e-06, | |
| "loss": 0.4998, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.216896345116836e-06, | |
| "loss": 0.475, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.2133013780707015e-06, | |
| "loss": 0.5176, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.209706411024566e-06, | |
| "loss": 0.5194, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.20611144397843e-06, | |
| "loss": 0.5138, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.2025164769322945e-06, | |
| "loss": 0.4778, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.19892150988616e-06, | |
| "loss": 0.4754, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.195326542840024e-06, | |
| "loss": 0.4905, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.191731575793889e-06, | |
| "loss": 0.4896, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.188136608747753e-06, | |
| "loss": 0.4694, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.184541641701618e-06, | |
| "loss": 0.5051, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.180946674655482e-06, | |
| "loss": 0.4834, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.177351707609347e-06, | |
| "loss": 0.493, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.173756740563212e-06, | |
| "loss": 0.5016, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.170161773517076e-06, | |
| "loss": 0.504, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.166566806470941e-06, | |
| "loss": 0.4803, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.1629718394248055e-06, | |
| "loss": 0.5005, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.159376872378671e-06, | |
| "loss": 0.4765, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.155781905332534e-06, | |
| "loss": 0.4913, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.152186938286399e-06, | |
| "loss": 0.4876, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.148591971240264e-06, | |
| "loss": 0.488, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.144997004194129e-06, | |
| "loss": 0.4774, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.141402037147992e-06, | |
| "loss": 0.4989, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.1378070701018574e-06, | |
| "loss": 0.4901, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.134212103055722e-06, | |
| "loss": 0.4861, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.130617136009587e-06, | |
| "loss": 0.5002, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.127022168963451e-06, | |
| "loss": 0.5004, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.123427201917316e-06, | |
| "loss": 0.4614, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.11983223487118e-06, | |
| "loss": 0.4672, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.116237267825045e-06, | |
| "loss": 0.4754, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.112642300778909e-06, | |
| "loss": 0.4927, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.109047333732775e-06, | |
| "loss": 0.4749, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.105452366686639e-06, | |
| "loss": 0.4858, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.101857399640503e-06, | |
| "loss": 0.4632, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.098262432594368e-06, | |
| "loss": 0.4892, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.094667465548233e-06, | |
| "loss": 0.5137, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.091072498502097e-06, | |
| "loss": 0.4757, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.087477531455961e-06, | |
| "loss": 0.4789, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.0838825644098266e-06, | |
| "loss": 0.468, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.080287597363691e-06, | |
| "loss": 0.4887, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.076692630317556e-06, | |
| "loss": 0.4563, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.0730976632714195e-06, | |
| "loss": 0.5165, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.069502696225285e-06, | |
| "loss": 0.52, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.065907729179149e-06, | |
| "loss": 0.5167, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.062312762133014e-06, | |
| "loss": 0.4957, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.0587177950868785e-06, | |
| "loss": 0.4721, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.055122828040743e-06, | |
| "loss": 0.5233, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.051527860994607e-06, | |
| "loss": 0.4805, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.047932893948472e-06, | |
| "loss": 0.463, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.044337926902337e-06, | |
| "loss": 0.4476, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.040742959856201e-06, | |
| "loss": 0.4295, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.037147992810066e-06, | |
| "loss": 0.4322, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.0335530257639305e-06, | |
| "loss": 0.4603, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.029958058717796e-06, | |
| "loss": 0.4301, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.02636309167166e-06, | |
| "loss": 0.4357, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.022768124625524e-06, | |
| "loss": 0.4405, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.019173157579389e-06, | |
| "loss": 0.4505, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.015578190533254e-06, | |
| "loss": 0.4326, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.011983223487118e-06, | |
| "loss": 0.4582, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.008388256440983e-06, | |
| "loss": 0.4467, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.004793289394847e-06, | |
| "loss": 0.4213, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.001198322348712e-06, | |
| "loss": 0.4507, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.997603355302576e-06, | |
| "loss": 0.4302, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.9940083882564415e-06, | |
| "loss": 0.4347, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.990413421210305e-06, | |
| "loss": 0.4342, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.98681845416417e-06, | |
| "loss": 0.413, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.9832234871180345e-06, | |
| "loss": 0.4638, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.9796285200719e-06, | |
| "loss": 0.4489, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.976033553025764e-06, | |
| "loss": 0.4497, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.972438585979628e-06, | |
| "loss": 0.4412, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.9688436189334935e-06, | |
| "loss": 0.4346, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.965248651887358e-06, | |
| "loss": 0.4329, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.961653684841223e-06, | |
| "loss": 0.421, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.9580587177950864e-06, | |
| "loss": 0.4415, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.954463750748952e-06, | |
| "loss": 0.4395, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.950868783702816e-06, | |
| "loss": 0.449, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.947273816656681e-06, | |
| "loss": 0.4494, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.9436788496105454e-06, | |
| "loss": 0.4531, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.94008388256441e-06, | |
| "loss": 0.4607, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.936488915518274e-06, | |
| "loss": 0.4379, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.932893948472139e-06, | |
| "loss": 0.4507, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.929298981426004e-06, | |
| "loss": 0.4113, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.925704014379869e-06, | |
| "loss": 0.4306, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.922109047333732e-06, | |
| "loss": 0.4348, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.918514080287597e-06, | |
| "loss": 0.4529, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.914919113241462e-06, | |
| "loss": 0.4225, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.911324146195327e-06, | |
| "loss": 0.4494, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.907729179149191e-06, | |
| "loss": 0.4389, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.9041342121030556e-06, | |
| "loss": 0.4935, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.900539245056921e-06, | |
| "loss": 0.4428, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.896944278010785e-06, | |
| "loss": 0.4121, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.89334931096465e-06, | |
| "loss": 0.4587, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.889754343918514e-06, | |
| "loss": 0.4421, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.886159376872379e-06, | |
| "loss": 0.4326, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.882564409826243e-06, | |
| "loss": 0.4266, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.878969442780108e-06, | |
| "loss": 0.4171, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.875374475733973e-06, | |
| "loss": 0.4161, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.871779508687837e-06, | |
| "loss": 0.4262, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.868184541641701e-06, | |
| "loss": 0.4314, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.8645895745955665e-06, | |
| "loss": 0.4272, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.860994607549431e-06, | |
| "loss": 0.437, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.857399640503295e-06, | |
| "loss": 0.4334, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.8538046734571595e-06, | |
| "loss": 0.4302, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.850209706411025e-06, | |
| "loss": 0.4493, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.846614739364889e-06, | |
| "loss": 0.4416, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.843019772318754e-06, | |
| "loss": 0.4456, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.8394248052726185e-06, | |
| "loss": 0.4357, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.835829838226483e-06, | |
| "loss": 0.4647, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.832234871180348e-06, | |
| "loss": 0.4495, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.828639904134212e-06, | |
| "loss": 0.4355, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.8250449370880775e-06, | |
| "loss": 0.4148, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.821449970041941e-06, | |
| "loss": 0.4451, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.817855002995806e-06, | |
| "loss": 0.4397, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.8142600359496705e-06, | |
| "loss": 0.4597, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.8106650689035353e-06, | |
| "loss": 0.423, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.8070701018573996e-06, | |
| "loss": 0.4804, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.8034751348112643e-06, | |
| "loss": 0.4591, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.799880167765129e-06, | |
| "loss": 0.4405, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.796285200718994e-06, | |
| "loss": 0.4297, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.7926902336728586e-06, | |
| "loss": 0.44, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.7890952666267225e-06, | |
| "loss": 0.4324, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.7855002995805872e-06, | |
| "loss": 0.4386, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.781905332534452e-06, | |
| "loss": 0.4356, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.7783103654883167e-06, | |
| "loss": 0.4316, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.7747153984421806e-06, | |
| "loss": 0.4333, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.7711204313960454e-06, | |
| "loss": 0.4479, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.76752546434991e-06, | |
| "loss": 0.4613, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.763930497303775e-06, | |
| "loss": 0.4532, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.7603355302576396e-06, | |
| "loss": 0.4495, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.7567405632115035e-06, | |
| "loss": 0.4344, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.7531455961653683e-06, | |
| "loss": 0.4476, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.749550629119233e-06, | |
| "loss": 0.4236, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.7459556620730978e-06, | |
| "loss": 0.4529, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.7423606950269625e-06, | |
| "loss": 0.4271, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.738765727980827e-06, | |
| "loss": 0.4217, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.7351707609346916e-06, | |
| "loss": 0.4328, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.7315757938885564e-06, | |
| "loss": 0.4205, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.727980826842421e-06, | |
| "loss": 0.4308, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.724385859796285e-06, | |
| "loss": 0.4433, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.7207908927501498e-06, | |
| "loss": 0.4369, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.7171959257040145e-06, | |
| "loss": 0.4207, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.7136009586578793e-06, | |
| "loss": 0.4384, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.710005991611744e-06, | |
| "loss": 0.4743, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.706411024565608e-06, | |
| "loss": 0.4644, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.7028160575194727e-06, | |
| "loss": 0.4378, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.6992210904733374e-06, | |
| "loss": 0.4325, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.695626123427202e-06, | |
| "loss": 0.4443, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.692031156381067e-06, | |
| "loss": 0.4364, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.688436189334931e-06, | |
| "loss": 0.4329, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.6848412222887956e-06, | |
| "loss": 0.453, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.6812462552426603e-06, | |
| "loss": 0.448, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.677651288196525e-06, | |
| "loss": 0.4244, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.6740563211503894e-06, | |
| "loss": 0.4537, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.670461354104254e-06, | |
| "loss": 0.4306, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.666866387058119e-06, | |
| "loss": 0.428, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.6632714200119836e-06, | |
| "loss": 0.454, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.6596764529658484e-06, | |
| "loss": 0.3971, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.6560814859197123e-06, | |
| "loss": 0.4513, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.652486518873577e-06, | |
| "loss": 0.4284, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.6488915518274418e-06, | |
| "loss": 0.4211, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.6452965847813065e-06, | |
| "loss": 0.4482, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.6417016177351713e-06, | |
| "loss": 0.4549, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.638106650689035e-06, | |
| "loss": 0.4402, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.6345116836429e-06, | |
| "loss": 0.423, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.6309167165967647e-06, | |
| "loss": 0.42, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.6273217495506294e-06, | |
| "loss": 0.4471, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.6237267825044933e-06, | |
| "loss": 0.4687, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.620131815458358e-06, | |
| "loss": 0.451, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.616536848412223e-06, | |
| "loss": 0.4157, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.6129418813660876e-06, | |
| "loss": 0.4449, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.6093469143199523e-06, | |
| "loss": 0.445, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.6057519472738167e-06, | |
| "loss": 0.4523, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.6021569802276814e-06, | |
| "loss": 0.462, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.598562013181546e-06, | |
| "loss": 0.4373, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.594967046135411e-06, | |
| "loss": 0.4382, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.591372079089275e-06, | |
| "loss": 0.4482, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.5877771120431396e-06, | |
| "loss": 0.4609, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.5841821449970043e-06, | |
| "loss": 0.419, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.580587177950869e-06, | |
| "loss": 0.4535, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.576992210904734e-06, | |
| "loss": 0.4328, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.5733972438585977e-06, | |
| "loss": 0.4232, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.5698022768124625e-06, | |
| "loss": 0.4596, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.5662073097663272e-06, | |
| "loss": 0.4169, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.562612342720192e-06, | |
| "loss": 0.4692, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.5590173756740567e-06, | |
| "loss": 0.4453, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.5554224086279206e-06, | |
| "loss": 0.4391, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.5518274415817854e-06, | |
| "loss": 0.4285, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.54823247453565e-06, | |
| "loss": 0.4266, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.544637507489515e-06, | |
| "loss": 0.4496, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.541042540443379e-06, | |
| "loss": 0.452, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.537447573397244e-06, | |
| "loss": 0.4218, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.5338526063511087e-06, | |
| "loss": 0.4009, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.5302576393049734e-06, | |
| "loss": 0.418, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.526662672258838e-06, | |
| "loss": 0.4651, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.523067705212702e-06, | |
| "loss": 0.4366, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.519472738166567e-06, | |
| "loss": 0.4584, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.5158777711204316e-06, | |
| "loss": 0.4282, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.5122828040742963e-06, | |
| "loss": 0.433, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.508687837028161e-06, | |
| "loss": 0.4204, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.505092869982025e-06, | |
| "loss": 0.4501, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.5014979029358897e-06, | |
| "loss": 0.4307, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.4979029358897545e-06, | |
| "loss": 0.4593, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.4943079688436192e-06, | |
| "loss": 0.4526, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.490713001797483e-06, | |
| "loss": 0.4635, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.487118034751348e-06, | |
| "loss": 0.4358, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.4835230677052126e-06, | |
| "loss": 0.4252, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.4799281006590774e-06, | |
| "loss": 0.4103, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.476333133612942e-06, | |
| "loss": 0.4201, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.4727381665668065e-06, | |
| "loss": 0.4323, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.4691431995206712e-06, | |
| "loss": 0.4063, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.465548232474536e-06, | |
| "loss": 0.4259, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.4619532654284007e-06, | |
| "loss": 0.4419, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.458358298382265e-06, | |
| "loss": 0.4255, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.4547633313361294e-06, | |
| "loss": 0.4533, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.451168364289994e-06, | |
| "loss": 0.4193, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.447573397243859e-06, | |
| "loss": 0.4448, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.4439784301977236e-06, | |
| "loss": 0.4458, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.4403834631515875e-06, | |
| "loss": 0.4514, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.4367884961054523e-06, | |
| "loss": 0.4044, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.433193529059317e-06, | |
| "loss": 0.4463, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.4295985620131818e-06, | |
| "loss": 0.465, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.4260035949670465e-06, | |
| "loss": 0.4374, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.4224086279209104e-06, | |
| "loss": 0.4212, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.418813660874775e-06, | |
| "loss": 0.4151, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.41521869382864e-06, | |
| "loss": 0.4382, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.4116237267825047e-06, | |
| "loss": 0.4348, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.408028759736369e-06, | |
| "loss": 0.4324, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.4044337926902337e-06, | |
| "loss": 0.4362, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.4008388256440985e-06, | |
| "loss": 0.4633, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.3972438585979632e-06, | |
| "loss": 0.4246, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.3936488915518276e-06, | |
| "loss": 0.4244, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.390053924505692e-06, | |
| "loss": 0.4398, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.3864589574595566e-06, | |
| "loss": 0.4388, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.3828639904134214e-06, | |
| "loss": 0.4394, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.379269023367286e-06, | |
| "loss": 0.4569, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.375674056321151e-06, | |
| "loss": 0.4484, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.372079089275015e-06, | |
| "loss": 0.4724, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.3684841222288795e-06, | |
| "loss": 0.4403, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.3648891551827443e-06, | |
| "loss": 0.4307, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.361294188136609e-06, | |
| "loss": 0.4482, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.357699221090473e-06, | |
| "loss": 0.425, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.3541042540443377e-06, | |
| "loss": 0.4546, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.3505092869982024e-06, | |
| "loss": 0.4571, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.346914319952067e-06, | |
| "loss": 0.4193, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.343319352905932e-06, | |
| "loss": 0.4275, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.3397243858597963e-06, | |
| "loss": 0.415, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.336129418813661e-06, | |
| "loss": 0.4439, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.3325344517675258e-06, | |
| "loss": 0.4265, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.3289394847213905e-06, | |
| "loss": 0.4312, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.325344517675255e-06, | |
| "loss": 0.4605, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.321749550629119e-06, | |
| "loss": 0.414, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.318154583582984e-06, | |
| "loss": 0.421, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.3145596165368487e-06, | |
| "loss": 0.4203, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.3109646494907134e-06, | |
| "loss": 0.4166, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.3073696824445773e-06, | |
| "loss": 0.4008, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.303774715398442e-06, | |
| "loss": 0.4218, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.300179748352307e-06, | |
| "loss": 0.4511, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.2965847813061716e-06, | |
| "loss": 0.4303, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.2929898142600363e-06, | |
| "loss": 0.4198, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.2893948472139002e-06, | |
| "loss": 0.4332, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.285799880167765e-06, | |
| "loss": 0.4274, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.2822049131216297e-06, | |
| "loss": 0.4326, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.2786099460754945e-06, | |
| "loss": 0.447, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.2750149790293592e-06, | |
| "loss": 0.4371, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.2714200119832236e-06, | |
| "loss": 0.4341, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.2678250449370883e-06, | |
| "loss": 0.4243, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.264230077890953e-06, | |
| "loss": 0.4411, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.2606351108448174e-06, | |
| "loss": 0.4238, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.2570401437986817e-06, | |
| "loss": 0.4431, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.2534451767525465e-06, | |
| "loss": 0.4546, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.249850209706411e-06, | |
| "loss": 0.4407, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.246255242660276e-06, | |
| "loss": 0.4449, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.2426602756141407e-06, | |
| "loss": 0.4329, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.2390653085680046e-06, | |
| "loss": 0.4413, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.2354703415218694e-06, | |
| "loss": 0.4098, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.231875374475734e-06, | |
| "loss": 0.4288, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.228280407429599e-06, | |
| "loss": 0.4207, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.2246854403834628e-06, | |
| "loss": 0.4335, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.2210904733373275e-06, | |
| "loss": 0.445, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.2174955062911923e-06, | |
| "loss": 0.4138, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.213900539245057e-06, | |
| "loss": 0.4626, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.2103055721989218e-06, | |
| "loss": 0.4419, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.206710605152786e-06, | |
| "loss": 0.4493, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.203115638106651e-06, | |
| "loss": 0.4337, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.1995206710605156e-06, | |
| "loss": 0.446, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.19592570401438e-06, | |
| "loss": 0.4176, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.1923307369682447e-06, | |
| "loss": 0.4635, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.188735769922109e-06, | |
| "loss": 0.4326, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.1851408028759737e-06, | |
| "loss": 0.4454, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.1815458358298385e-06, | |
| "loss": 0.4368, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.1779508687837032e-06, | |
| "loss": 0.4485, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.174355901737567e-06, | |
| "loss": 0.4172, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.170760934691432e-06, | |
| "loss": 0.4466, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.1671659676452966e-06, | |
| "loss": 0.4371, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.1635710005991614e-06, | |
| "loss": 0.4421, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.159976033553026e-06, | |
| "loss": 0.4436, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.15638106650689e-06, | |
| "loss": 0.434, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.1527860994607548e-06, | |
| "loss": 0.4344, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.1491911324146195e-06, | |
| "loss": 0.4567, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.1455961653684843e-06, | |
| "loss": 0.422, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.142001198322349e-06, | |
| "loss": 0.4215, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.1384062312762134e-06, | |
| "loss": 0.4335, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.134811264230078e-06, | |
| "loss": 0.4697, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.1312162971839424e-06, | |
| "loss": 0.4409, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.127621330137807e-06, | |
| "loss": 0.4363, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.1240263630916715e-06, | |
| "loss": 0.4324, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.1204313960455363e-06, | |
| "loss": 0.4359, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.116836428999401e-06, | |
| "loss": 0.4419, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.1132414619532658e-06, | |
| "loss": 0.4174, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.1096464949071305e-06, | |
| "loss": 0.4297, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.1060515278609944e-06, | |
| "loss": 0.4319, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.102456560814859e-06, | |
| "loss": 0.4445, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.098861593768724e-06, | |
| "loss": 0.4538, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.0952666267225887e-06, | |
| "loss": 0.4253, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.0916716596764534e-06, | |
| "loss": 0.4406, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.0880766926303173e-06, | |
| "loss": 0.4385, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.084481725584182e-06, | |
| "loss": 0.4495, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.080886758538047e-06, | |
| "loss": 0.4377, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.0772917914919116e-06, | |
| "loss": 0.4066, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.073696824445776e-06, | |
| "loss": 0.4284, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.0701018573996406e-06, | |
| "loss": 0.4379, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.0665068903535054e-06, | |
| "loss": 0.4316, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.0629119233073697e-06, | |
| "loss": 0.4492, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.0593169562612345e-06, | |
| "loss": 0.4297, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.0557219892150988e-06, | |
| "loss": 0.4677, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 3.0521270221689635e-06, | |
| "loss": 0.4503, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.0485320551228283e-06, | |
| "loss": 0.429, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.044937088076693e-06, | |
| "loss": 0.4376, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.041342121030557e-06, | |
| "loss": 0.4188, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.0377471539844217e-06, | |
| "loss": 0.4184, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.0341521869382864e-06, | |
| "loss": 0.4492, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 3.030557219892151e-06, | |
| "loss": 0.4393, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 3.026962252846016e-06, | |
| "loss": 0.4182, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 3.02336728579988e-06, | |
| "loss": 0.4378, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 3.0197723187537446e-06, | |
| "loss": 0.4379, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 3.0161773517076093e-06, | |
| "loss": 0.4315, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 3.012582384661474e-06, | |
| "loss": 0.4263, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.008987417615339e-06, | |
| "loss": 0.436, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.005392450569203e-06, | |
| "loss": 0.4307, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.001797483523068e-06, | |
| "loss": 0.4567, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.9982025164769322e-06, | |
| "loss": 0.447, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.994607549430797e-06, | |
| "loss": 0.4342, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.9910125823846617e-06, | |
| "loss": 0.4591, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.987417615338526e-06, | |
| "loss": 0.4358, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.983822648292391e-06, | |
| "loss": 0.4298, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.9802276812462556e-06, | |
| "loss": 0.4525, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.97663271420012e-06, | |
| "loss": 0.444, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.9730377471539846e-06, | |
| "loss": 0.4298, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.969442780107849e-06, | |
| "loss": 0.4185, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.9658478130617137e-06, | |
| "loss": 0.4573, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.962252846015578e-06, | |
| "loss": 0.4381, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.958657878969443e-06, | |
| "loss": 0.4286, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.9550629119233075e-06, | |
| "loss": 0.4482, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.951467944877172e-06, | |
| "loss": 0.4016, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.9478729778310366e-06, | |
| "loss": 0.4517, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.944278010784901e-06, | |
| "loss": 0.4178, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.9406830437387657e-06, | |
| "loss": 0.4494, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.9370880766926304e-06, | |
| "loss": 0.4177, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.9334931096464948e-06, | |
| "loss": 0.4359, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.9298981426003595e-06, | |
| "loss": 0.4301, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.9263031755542243e-06, | |
| "loss": 0.4461, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.922708208508089e-06, | |
| "loss": 0.4204, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.9191132414619533e-06, | |
| "loss": 0.4363, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.915518274415818e-06, | |
| "loss": 0.4376, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.9119233073696824e-06, | |
| "loss": 0.441, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.908328340323547e-06, | |
| "loss": 0.4361, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.904733373277412e-06, | |
| "loss": 0.4508, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.9011384062312762e-06, | |
| "loss": 0.4315, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.897543439185141e-06, | |
| "loss": 0.4249, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.8939484721390053e-06, | |
| "loss": 0.4602, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.89035350509287e-06, | |
| "loss": 0.4386, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.8867585380467344e-06, | |
| "loss": 0.466, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.883163571000599e-06, | |
| "loss": 0.4574, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.879568603954464e-06, | |
| "loss": 0.4115, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.8759736369083282e-06, | |
| "loss": 0.4395, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.872378669862193e-06, | |
| "loss": 0.4206, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.8687837028160577e-06, | |
| "loss": 0.4274, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.865188735769922e-06, | |
| "loss": 0.4431, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.861593768723787e-06, | |
| "loss": 0.4285, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.8579988016776515e-06, | |
| "loss": 0.4259, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.8544038346315163e-06, | |
| "loss": 0.4314, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.8508088675853806e-06, | |
| "loss": 0.4092, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.8472139005392454e-06, | |
| "loss": 0.3943, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.8436189334931097e-06, | |
| "loss": 0.4184, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.8400239664469744e-06, | |
| "loss": 0.4348, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.8364289994008388e-06, | |
| "loss": 0.4588, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.8328340323547035e-06, | |
| "loss": 0.4242, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.8292390653085683e-06, | |
| "loss": 0.4279, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.8256440982624326e-06, | |
| "loss": 0.4616, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.8220491312162973e-06, | |
| "loss": 0.4153, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.8184541641701617e-06, | |
| "loss": 0.4445, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.8148591971240264e-06, | |
| "loss": 0.4062, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.8112642300778907e-06, | |
| "loss": 0.4313, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.8076692630317555e-06, | |
| "loss": 0.4519, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.8040742959856202e-06, | |
| "loss": 0.4149, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.8004793289394846e-06, | |
| "loss": 0.4223, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.7968843618933493e-06, | |
| "loss": 0.4504, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.793289394847214e-06, | |
| "loss": 0.4193, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.789694427801079e-06, | |
| "loss": 0.4321, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.786099460754943e-06, | |
| "loss": 0.411, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.782504493708808e-06, | |
| "loss": 0.4197, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.7789095266626722e-06, | |
| "loss": 0.4369, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.775314559616537e-06, | |
| "loss": 0.4394, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.7717195925704017e-06, | |
| "loss": 0.4422, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.768124625524266e-06, | |
| "loss": 0.4452, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.764529658478131e-06, | |
| "loss": 0.404, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.760934691431995e-06, | |
| "loss": 0.4237, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.75733972438586e-06, | |
| "loss": 0.4315, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.753744757339724e-06, | |
| "loss": 0.4305, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.750149790293589e-06, | |
| "loss": 0.4167, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.7465548232474537e-06, | |
| "loss": 0.4142, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.742959856201318e-06, | |
| "loss": 0.4341, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.7393648891551828e-06, | |
| "loss": 0.4471, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.735769922109047e-06, | |
| "loss": 0.4426, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.732174955062912e-06, | |
| "loss": 0.42, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.7285799880167766e-06, | |
| "loss": 0.4405, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.7249850209706414e-06, | |
| "loss": 0.458, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.721390053924506e-06, | |
| "loss": 0.4027, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.7177950868783704e-06, | |
| "loss": 0.4239, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.714200119832235e-06, | |
| "loss": 0.4504, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.7106051527860995e-06, | |
| "loss": 0.4296, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.7070101857399643e-06, | |
| "loss": 0.4171, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.7034152186938286e-06, | |
| "loss": 0.4219, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.6998202516476933e-06, | |
| "loss": 0.4244, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.696225284601558e-06, | |
| "loss": 0.4472, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.6926303175554224e-06, | |
| "loss": 0.4226, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.689035350509287e-06, | |
| "loss": 0.4333, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.6854403834631515e-06, | |
| "loss": 0.4545, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.6818454164170162e-06, | |
| "loss": 0.4249, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.6782504493708806e-06, | |
| "loss": 0.419, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.6746554823247453e-06, | |
| "loss": 0.445, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.67106051527861e-06, | |
| "loss": 0.4191, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.6674655482324744e-06, | |
| "loss": 0.4176, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.663870581186339e-06, | |
| "loss": 0.4372, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.660275614140204e-06, | |
| "loss": 0.4208, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.6566806470940686e-06, | |
| "loss": 0.4406, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.653085680047933e-06, | |
| "loss": 0.4349, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.6494907130017977e-06, | |
| "loss": 0.4418, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.6458957459556625e-06, | |
| "loss": 0.4566, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.6423007789095268e-06, | |
| "loss": 0.4342, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.6387058118633915e-06, | |
| "loss": 0.4462, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.635110844817256e-06, | |
| "loss": 0.4175, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.6315158777711206e-06, | |
| "loss": 0.4361, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.627920910724985e-06, | |
| "loss": 0.4373, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.6243259436788497e-06, | |
| "loss": 0.4544, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.620730976632714e-06, | |
| "loss": 0.4169, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.6171360095865788e-06, | |
| "loss": 0.4206, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.6135410425404435e-06, | |
| "loss": 0.4599, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.609946075494308e-06, | |
| "loss": 0.4569, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.6063511084481726e-06, | |
| "loss": 0.4592, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.602756141402037e-06, | |
| "loss": 0.4607, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.5991611743559017e-06, | |
| "loss": 0.4314, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.5955662073097664e-06, | |
| "loss": 0.4433, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.591971240263631e-06, | |
| "loss": 0.4599, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.588376273217496e-06, | |
| "loss": 0.4108, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.5847813061713602e-06, | |
| "loss": 0.4203, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.581186339125225e-06, | |
| "loss": 0.425, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.5775913720790893e-06, | |
| "loss": 0.4184, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.573996405032954e-06, | |
| "loss": 0.4594, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.5704014379868184e-06, | |
| "loss": 0.4294, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.566806470940683e-06, | |
| "loss": 0.45, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.563211503894548e-06, | |
| "loss": 0.4245, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.559616536848412e-06, | |
| "loss": 0.4216, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.556021569802277e-06, | |
| "loss": 0.4354, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.5524266027561413e-06, | |
| "loss": 0.4369, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.548831635710006e-06, | |
| "loss": 0.4141, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.5452366686638704e-06, | |
| "loss": 0.4199, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.541641701617735e-06, | |
| "loss": 0.4488, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.5380467345716e-06, | |
| "loss": 0.4315, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.534451767525464e-06, | |
| "loss": 0.4111, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.530856800479329e-06, | |
| "loss": 0.4549, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.5272618334331937e-06, | |
| "loss": 0.4448, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.5236668663870584e-06, | |
| "loss": 0.4353, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.5200718993409228e-06, | |
| "loss": 0.4429, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.5164769322947875e-06, | |
| "loss": 0.4385, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.5128819652486523e-06, | |
| "loss": 0.4174, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.5092869982025166e-06, | |
| "loss": 0.4346, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.5056920311563813e-06, | |
| "loss": 0.4402, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.5020970641102457e-06, | |
| "loss": 0.4221, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4985020970641104e-06, | |
| "loss": 0.4262, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4949071300179747e-06, | |
| "loss": 0.4569, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4913121629718395e-06, | |
| "loss": 0.4329, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4877171959257042e-06, | |
| "loss": 0.4469, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4841222288795686e-06, | |
| "loss": 0.4293, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.4805272618334333e-06, | |
| "loss": 0.4149, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.4769322947872976e-06, | |
| "loss": 0.4442, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.4733373277411624e-06, | |
| "loss": 0.4226, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.4697423606950267e-06, | |
| "loss": 0.4367, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.4661473936488915e-06, | |
| "loss": 0.4437, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.4625524266027562e-06, | |
| "loss": 0.4152, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.458957459556621e-06, | |
| "loss": 0.4091, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.4553624925104857e-06, | |
| "loss": 0.4392, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.45176752546435e-06, | |
| "loss": 0.41, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.4481725584182148e-06, | |
| "loss": 0.4304, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.444577591372079e-06, | |
| "loss": 0.4131, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.440982624325944e-06, | |
| "loss": 0.4437, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.437387657279808e-06, | |
| "loss": 0.4396, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.433792690233673e-06, | |
| "loss": 0.4476, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.4301977231875377e-06, | |
| "loss": 0.4501, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.426602756141402e-06, | |
| "loss": 0.4013, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.4230077890952668e-06, | |
| "loss": 0.4307, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.419412822049131e-06, | |
| "loss": 0.4385, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.415817855002996e-06, | |
| "loss": 0.4619, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.41222288795686e-06, | |
| "loss": 0.4358, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.408627920910725e-06, | |
| "loss": 0.415, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.4050329538645897e-06, | |
| "loss": 0.4417, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.401437986818454e-06, | |
| "loss": 0.4156, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.3978430197723187e-06, | |
| "loss": 0.4565, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.3942480527261835e-06, | |
| "loss": 0.4655, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.3906530856800482e-06, | |
| "loss": 0.4228, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.3870581186339126e-06, | |
| "loss": 0.4586, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.3834631515877773e-06, | |
| "loss": 0.4115, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.379868184541642e-06, | |
| "loss": 0.4084, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.3762732174955064e-06, | |
| "loss": 0.4154, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.372678250449371e-06, | |
| "loss": 0.4508, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.3690832834032355e-06, | |
| "loss": 0.4086, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.3654883163571002e-06, | |
| "loss": 0.4345, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.3618933493109645e-06, | |
| "loss": 0.4329, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.3582983822648293e-06, | |
| "loss": 0.4297, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.354703415218694e-06, | |
| "loss": 0.4206, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.3511084481725584e-06, | |
| "loss": 0.4353, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.347513481126423e-06, | |
| "loss": 0.4364, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.3439185140802874e-06, | |
| "loss": 0.4508, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.340323547034152e-06, | |
| "loss": 0.4358, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.3367285799880165e-06, | |
| "loss": 0.4328, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.3331336129418813e-06, | |
| "loss": 0.4037, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.329538645895746e-06, | |
| "loss": 0.4546, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.3259436788496108e-06, | |
| "loss": 0.4081, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.322348711803475e-06, | |
| "loss": 0.4143, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.31875374475734e-06, | |
| "loss": 0.4628, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.3151587777112046e-06, | |
| "loss": 0.4386, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.311563810665069e-06, | |
| "loss": 0.447, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.3079688436189337e-06, | |
| "loss": 0.4417, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.3043738765727984e-06, | |
| "loss": 0.4447, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.3007789095266627e-06, | |
| "loss": 0.4128, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.2971839424805275e-06, | |
| "loss": 0.4185, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.293588975434392e-06, | |
| "loss": 0.4128, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.2899940083882566e-06, | |
| "loss": 0.4246, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.286399041342121e-06, | |
| "loss": 0.4238, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.2828040742959856e-06, | |
| "loss": 0.4529, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.2792091072498504e-06, | |
| "loss": 0.4176, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.2756141402037147e-06, | |
| "loss": 0.4357, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.2720191731575795e-06, | |
| "loss": 0.4264, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.268424206111444e-06, | |
| "loss": 0.4423, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.2648292390653085e-06, | |
| "loss": 0.4301, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.2612342720191733e-06, | |
| "loss": 0.4354, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.2576393049730376e-06, | |
| "loss": 0.413, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.2540443379269024e-06, | |
| "loss": 0.4537, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.250449370880767e-06, | |
| "loss": 0.4369, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.246854403834632e-06, | |
| "loss": 0.4405, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.243259436788496e-06, | |
| "loss": 0.4033, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.239664469742361e-06, | |
| "loss": 0.4321, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.2360695026962253e-06, | |
| "loss": 0.438, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.23247453565009e-06, | |
| "loss": 0.4152, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.2288795686039543e-06, | |
| "loss": 0.4133, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.225284601557819e-06, | |
| "loss": 0.4164, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.221689634511684e-06, | |
| "loss": 0.4213, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.218094667465548e-06, | |
| "loss": 0.448, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.214499700419413e-06, | |
| "loss": 0.4263, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.2109047333732773e-06, | |
| "loss": 0.4185, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.207309766327142e-06, | |
| "loss": 0.4319, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.2037147992810063e-06, | |
| "loss": 0.4258, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.200119832234871e-06, | |
| "loss": 0.4294, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.196524865188736e-06, | |
| "loss": 0.4465, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.1929298981426006e-06, | |
| "loss": 0.4403, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.189334931096465e-06, | |
| "loss": 0.4191, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.1857399640503297e-06, | |
| "loss": 0.4182, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.1821449970041944e-06, | |
| "loss": 0.4339, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.1785500299580587e-06, | |
| "loss": 0.4275, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.1749550629119235e-06, | |
| "loss": 0.4383, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.1713600958657882e-06, | |
| "loss": 0.4496, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.1677651288196526e-06, | |
| "loss": 0.432, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.1641701617735173e-06, | |
| "loss": 0.4521, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.1605751947273816e-06, | |
| "loss": 0.4338, | |
| "step": 1089 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.1569802276812464e-06, | |
| "loss": 0.4299, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.1533852606351107e-06, | |
| "loss": 0.4585, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.1497902935889755e-06, | |
| "loss": 0.4616, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.14619532654284e-06, | |
| "loss": 0.4336, | |
| "step": 1093 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.1426003594967045e-06, | |
| "loss": 0.4517, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.1390053924505693e-06, | |
| "loss": 0.4471, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.1354104254044336e-06, | |
| "loss": 0.4218, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.1318154583582984e-06, | |
| "loss": 0.4151, | |
| "step": 1097 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.128220491312163e-06, | |
| "loss": 0.4289, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.1246255242660274e-06, | |
| "loss": 0.4509, | |
| "step": 1099 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.121030557219892e-06, | |
| "loss": 0.4298, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.117435590173757e-06, | |
| "loss": 0.4184, | |
| "step": 1101 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.1138406231276217e-06, | |
| "loss": 0.4283, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.110245656081486e-06, | |
| "loss": 0.4343, | |
| "step": 1103 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.1066506890353508e-06, | |
| "loss": 0.4377, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.103055721989215e-06, | |
| "loss": 0.4462, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.09946075494308e-06, | |
| "loss": 0.4375, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.0958657878969446e-06, | |
| "loss": 0.4504, | |
| "step": 1107 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.092270820850809e-06, | |
| "loss": 0.4319, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.0886758538046737e-06, | |
| "loss": 0.4208, | |
| "step": 1109 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.085080886758538e-06, | |
| "loss": 0.4281, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.0814859197124027e-06, | |
| "loss": 0.4326, | |
| "step": 1111 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.077890952666267e-06, | |
| "loss": 0.4224, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.074295985620132e-06, | |
| "loss": 0.4078, | |
| "step": 1113 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.070701018573996e-06, | |
| "loss": 0.4374, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.067106051527861e-06, | |
| "loss": 0.4485, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.0635110844817256e-06, | |
| "loss": 0.4208, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.05991611743559e-06, | |
| "loss": 0.4208, | |
| "step": 1117 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.0563211503894547e-06, | |
| "loss": 0.457, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.0527261833433195e-06, | |
| "loss": 0.4436, | |
| "step": 1119 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.049131216297184e-06, | |
| "loss": 0.4491, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.0455362492510485e-06, | |
| "loss": 0.4123, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.0419412822049133e-06, | |
| "loss": 0.4235, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.038346315158778e-06, | |
| "loss": 0.444, | |
| "step": 1123 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.0347513481126424e-06, | |
| "loss": 0.4422, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.031156381066507e-06, | |
| "loss": 0.4411, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.0275614140203714e-06, | |
| "loss": 0.407, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.023966446974236e-06, | |
| "loss": 0.4073, | |
| "step": 1127 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 2.0203714799281005e-06, | |
| "loss": 0.3811, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.0167765128819653e-06, | |
| "loss": 0.383, | |
| "step": 1129 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.01318154583583e-06, | |
| "loss": 0.3999, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.0095865787896943e-06, | |
| "loss": 0.3781, | |
| "step": 1131 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.005991611743559e-06, | |
| "loss": 0.3927, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 2.0023966446974234e-06, | |
| "loss": 0.3926, | |
| "step": 1133 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 1.998801677651288e-06, | |
| "loss": 0.3843, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.9952067106051525e-06, | |
| "loss": 0.3792, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.9916117435590172e-06, | |
| "loss": 0.3766, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.988016776512882e-06, | |
| "loss": 0.3713, | |
| "step": 1137 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.9844218094667467e-06, | |
| "loss": 0.4266, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.9808268424206115e-06, | |
| "loss": 0.3931, | |
| "step": 1139 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.977231875374476e-06, | |
| "loss": 0.3851, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.9736369083283406e-06, | |
| "loss": 0.3972, | |
| "step": 1141 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.970041941282205e-06, | |
| "loss": 0.3478, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.9664469742360696e-06, | |
| "loss": 0.3962, | |
| "step": 1143 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.9628520071899344e-06, | |
| "loss": 0.4133, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.9592570401437987e-06, | |
| "loss": 0.3934, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.9556620730976635e-06, | |
| "loss": 0.3785, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.9520671060515278e-06, | |
| "loss": 0.395, | |
| "step": 1147 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.9484721390053925e-06, | |
| "loss": 0.4045, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.944877171959257e-06, | |
| "loss": 0.3676, | |
| "step": 1149 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.9412822049131216e-06, | |
| "loss": 0.3877, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.9376872378669864e-06, | |
| "loss": 0.3615, | |
| "step": 1151 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.9340922708208507e-06, | |
| "loss": 0.4009, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.9304973037747154e-06, | |
| "loss": 0.3842, | |
| "step": 1153 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.9269023367285798e-06, | |
| "loss": 0.3672, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.9233073696824445e-06, | |
| "loss": 0.3904, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.9197124026363093e-06, | |
| "loss": 0.3921, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.916117435590174e-06, | |
| "loss": 0.3778, | |
| "step": 1157 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.9125224685440388e-06, | |
| "loss": 0.3769, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.908927501497903e-06, | |
| "loss": 0.3934, | |
| "step": 1159 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.9053325344517676e-06, | |
| "loss": 0.3738, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.9017375674056322e-06, | |
| "loss": 0.3788, | |
| "step": 1161 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.898142600359497e-06, | |
| "loss": 0.412, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.8945476333133612e-06, | |
| "loss": 0.3828, | |
| "step": 1163 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.890952666267226e-06, | |
| "loss": 0.3634, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.8873576992210903e-06, | |
| "loss": 0.3984, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.883762732174955e-06, | |
| "loss": 0.3851, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.8801677651288198e-06, | |
| "loss": 0.4053, | |
| "step": 1167 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.8765727980826841e-06, | |
| "loss": 0.3748, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.8729778310365489e-06, | |
| "loss": 0.3987, | |
| "step": 1169 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.8693828639904134e-06, | |
| "loss": 0.3535, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.8657878969442782e-06, | |
| "loss": 0.3797, | |
| "step": 1171 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.8621929298981425e-06, | |
| "loss": 0.3585, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.8585979628520073e-06, | |
| "loss": 0.4021, | |
| "step": 1173 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.855002995805872e-06, | |
| "loss": 0.3767, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.8514080287597363e-06, | |
| "loss": 0.3591, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.847813061713601e-06, | |
| "loss": 0.3696, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.8442180946674654e-06, | |
| "loss": 0.3778, | |
| "step": 1177 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.8406231276213302e-06, | |
| "loss": 0.3843, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.8370281605751947e-06, | |
| "loss": 0.4078, | |
| "step": 1179 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.8334331935290594e-06, | |
| "loss": 0.3697, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.8298382264829242e-06, | |
| "loss": 0.3688, | |
| "step": 1181 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.8262432594367885e-06, | |
| "loss": 0.3795, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.8226482923906533e-06, | |
| "loss": 0.3862, | |
| "step": 1183 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.8190533253445176e-06, | |
| "loss": 0.3943, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.8154583582983823e-06, | |
| "loss": 0.3758, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.8118633912522467e-06, | |
| "loss": 0.3771, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.8082684242061114e-06, | |
| "loss": 0.3867, | |
| "step": 1187 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.8046734571599762e-06, | |
| "loss": 0.3993, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.8010784901138407e-06, | |
| "loss": 0.3802, | |
| "step": 1189 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.7974835230677055e-06, | |
| "loss": 0.4057, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7938885560215698e-06, | |
| "loss": 0.369, | |
| "step": 1191 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7902935889754345e-06, | |
| "loss": 0.3552, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7866986219292989e-06, | |
| "loss": 0.3775, | |
| "step": 1193 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7831036548831636e-06, | |
| "loss": 0.3756, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7795086878370284e-06, | |
| "loss": 0.3768, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7759137207908927e-06, | |
| "loss": 0.3957, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.7723187537447574e-06, | |
| "loss": 0.4085, | |
| "step": 1197 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.768723786698622e-06, | |
| "loss": 0.348, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.7651288196524867e-06, | |
| "loss": 0.3953, | |
| "step": 1199 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.761533852606351e-06, | |
| "loss": 0.378, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.7579388855602158e-06, | |
| "loss": 0.3601, | |
| "step": 1201 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.7543439185140805e-06, | |
| "loss": 0.3766, | |
| "step": 1202 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.7507489514679449e-06, | |
| "loss": 0.3893, | |
| "step": 1203 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.7471539844218096e-06, | |
| "loss": 0.3732, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.743559017375674e-06, | |
| "loss": 0.3958, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.7399640503295387e-06, | |
| "loss": 0.3883, | |
| "step": 1206 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.7363690832834032e-06, | |
| "loss": 0.4191, | |
| "step": 1207 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.732774116237268e-06, | |
| "loss": 0.4085, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.7291791491911325e-06, | |
| "loss": 0.3633, | |
| "step": 1209 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.725584182144997e-06, | |
| "loss": 0.4038, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.7219892150988618e-06, | |
| "loss": 0.3516, | |
| "step": 1211 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.7183942480527261e-06, | |
| "loss": 0.3748, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.7147992810065909e-06, | |
| "loss": 0.386, | |
| "step": 1213 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.7112043139604552e-06, | |
| "loss": 0.3697, | |
| "step": 1214 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.70760934691432e-06, | |
| "loss": 0.3897, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.7040143798681845e-06, | |
| "loss": 0.3766, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.7004194128220492e-06, | |
| "loss": 0.3784, | |
| "step": 1217 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.6968244457759138e-06, | |
| "loss": 0.3998, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.6932294787297783e-06, | |
| "loss": 0.3884, | |
| "step": 1219 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.689634511683643e-06, | |
| "loss": 0.3845, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.6860395446375074e-06, | |
| "loss": 0.4006, | |
| "step": 1221 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.6824445775913721e-06, | |
| "loss": 0.3698, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.6788496105452365e-06, | |
| "loss": 0.3667, | |
| "step": 1223 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.6752546434991012e-06, | |
| "loss": 0.3762, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.671659676452966e-06, | |
| "loss": 0.3845, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6680647094068305e-06, | |
| "loss": 0.3782, | |
| "step": 1226 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6644697423606953e-06, | |
| "loss": 0.381, | |
| "step": 1227 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6608747753145596e-06, | |
| "loss": 0.3852, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6572798082684243e-06, | |
| "loss": 0.409, | |
| "step": 1229 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6536848412222887e-06, | |
| "loss": 0.4039, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.6500898741761534e-06, | |
| "loss": 0.3915, | |
| "step": 1231 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.6464949071300182e-06, | |
| "loss": 0.3991, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.6428999400838825e-06, | |
| "loss": 0.386, | |
| "step": 1233 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.6393049730377472e-06, | |
| "loss": 0.3792, | |
| "step": 1234 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.6357100059916118e-06, | |
| "loss": 0.3914, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.6321150389454765e-06, | |
| "loss": 0.3955, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.6285200718993409e-06, | |
| "loss": 0.3958, | |
| "step": 1237 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.6249251048532056e-06, | |
| "loss": 0.3759, | |
| "step": 1238 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.6213301378070704e-06, | |
| "loss": 0.3946, | |
| "step": 1239 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.6177351707609347e-06, | |
| "loss": 0.3975, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.6141402037147994e-06, | |
| "loss": 0.3819, | |
| "step": 1241 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.6105452366686638e-06, | |
| "loss": 0.3853, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.6069502696225285e-06, | |
| "loss": 0.3692, | |
| "step": 1243 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.603355302576393e-06, | |
| "loss": 0.3661, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.5997603355302578e-06, | |
| "loss": 0.378, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.5961653684841223e-06, | |
| "loss": 0.4056, | |
| "step": 1246 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.5925704014379869e-06, | |
| "loss": 0.3921, | |
| "step": 1247 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.5889754343918516e-06, | |
| "loss": 0.366, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.585380467345716e-06, | |
| "loss": 0.374, | |
| "step": 1249 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.5817855002995807e-06, | |
| "loss": 0.3851, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.578190533253445e-06, | |
| "loss": 0.4058, | |
| "step": 1251 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.5745955662073098e-06, | |
| "loss": 0.3931, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5710005991611745e-06, | |
| "loss": 0.3914, | |
| "step": 1253 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.567405632115039e-06, | |
| "loss": 0.3927, | |
| "step": 1254 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5638106650689036e-06, | |
| "loss": 0.3705, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5602156980227681e-06, | |
| "loss": 0.4131, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5566207309766329e-06, | |
| "loss": 0.3803, | |
| "step": 1257 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5530257639304972e-06, | |
| "loss": 0.3561, | |
| "step": 1258 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.549430796884362e-06, | |
| "loss": 0.3786, | |
| "step": 1259 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.5458358298382267e-06, | |
| "loss": 0.3689, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.542240862792091e-06, | |
| "loss": 0.3865, | |
| "step": 1261 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.5386458957459558e-06, | |
| "loss": 0.3607, | |
| "step": 1262 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.5350509286998203e-06, | |
| "loss": 0.375, | |
| "step": 1263 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.5314559616536849e-06, | |
| "loss": 0.3895, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.5278609946075494e-06, | |
| "loss": 0.3832, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.5242660275614141e-06, | |
| "loss": 0.3954, | |
| "step": 1266 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.5206710605152785e-06, | |
| "loss": 0.3985, | |
| "step": 1267 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.5170760934691432e-06, | |
| "loss": 0.3944, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.513481126423008e-06, | |
| "loss": 0.365, | |
| "step": 1269 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.5098861593768723e-06, | |
| "loss": 0.3923, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.506291192330737e-06, | |
| "loss": 0.4005, | |
| "step": 1271 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.5026962252846016e-06, | |
| "loss": 0.3782, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.4991012582384661e-06, | |
| "loss": 0.3849, | |
| "step": 1273 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.4955062911923309e-06, | |
| "loss": 0.3749, | |
| "step": 1274 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.4919113241461954e-06, | |
| "loss": 0.3955, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.48831635710006e-06, | |
| "loss": 0.3781, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.4847213900539245e-06, | |
| "loss": 0.383, | |
| "step": 1277 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.481126423007789e-06, | |
| "loss": 0.379, | |
| "step": 1278 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.4775314559616538e-06, | |
| "loss": 0.3646, | |
| "step": 1279 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.4739364889155183e-06, | |
| "loss": 0.4073, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.4703415218693828e-06, | |
| "loss": 0.3815, | |
| "step": 1281 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.4667465548232474e-06, | |
| "loss": 0.3898, | |
| "step": 1282 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.4631515877771121e-06, | |
| "loss": 0.383, | |
| "step": 1283 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.4595566207309767e-06, | |
| "loss": 0.3815, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.4559616536848412e-06, | |
| "loss": 0.3785, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.452366686638706e-06, | |
| "loss": 0.3936, | |
| "step": 1286 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.4487717195925705e-06, | |
| "loss": 0.389, | |
| "step": 1287 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.445176752546435e-06, | |
| "loss": 0.3477, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.4415817855002996e-06, | |
| "loss": 0.3881, | |
| "step": 1289 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.4379868184541641e-06, | |
| "loss": 0.3872, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.4343918514080289e-06, | |
| "loss": 0.3953, | |
| "step": 1291 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.4307968843618934e-06, | |
| "loss": 0.3643, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.4272019173157581e-06, | |
| "loss": 0.398, | |
| "step": 1293 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.4236069502696227e-06, | |
| "loss": 0.3761, | |
| "step": 1294 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.4200119832234872e-06, | |
| "loss": 0.3822, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.4164170161773518e-06, | |
| "loss": 0.3847, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.4128220491312163e-06, | |
| "loss": 0.3913, | |
| "step": 1297 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.4092270820850808e-06, | |
| "loss": 0.3911, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.4056321150389454e-06, | |
| "loss": 0.4136, | |
| "step": 1299 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.4020371479928101e-06, | |
| "loss": 0.3746, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.3984421809466747e-06, | |
| "loss": 0.3588, | |
| "step": 1301 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.3948472139005394e-06, | |
| "loss": 0.3817, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.391252246854404e-06, | |
| "loss": 0.3898, | |
| "step": 1303 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3876572798082685e-06, | |
| "loss": 0.4131, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.384062312762133e-06, | |
| "loss": 0.3932, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3804673457159976e-06, | |
| "loss": 0.3614, | |
| "step": 1306 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.376872378669862e-06, | |
| "loss": 0.375, | |
| "step": 1307 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3732774116237268e-06, | |
| "loss": 0.3691, | |
| "step": 1308 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.3696824445775914e-06, | |
| "loss": 0.379, | |
| "step": 1309 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.366087477531456e-06, | |
| "loss": 0.3837, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.3624925104853207e-06, | |
| "loss": 0.3866, | |
| "step": 1311 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.3588975434391852e-06, | |
| "loss": 0.3718, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.3553025763930498e-06, | |
| "loss": 0.3915, | |
| "step": 1313 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.3517076093469143e-06, | |
| "loss": 0.4006, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.348112642300779e-06, | |
| "loss": 0.3784, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.3445176752546436e-06, | |
| "loss": 0.3819, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.3409227082085081e-06, | |
| "loss": 0.3975, | |
| "step": 1317 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.3373277411623727e-06, | |
| "loss": 0.4042, | |
| "step": 1318 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.3337327741162372e-06, | |
| "loss": 0.3867, | |
| "step": 1319 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.330137807070102e-06, | |
| "loss": 0.389, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.3265428400239665e-06, | |
| "loss": 0.384, | |
| "step": 1321 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.3229478729778312e-06, | |
| "loss": 0.3988, | |
| "step": 1322 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.3193529059316958e-06, | |
| "loss": 0.3656, | |
| "step": 1323 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.3157579388855603e-06, | |
| "loss": 0.3843, | |
| "step": 1324 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.3121629718394248e-06, | |
| "loss": 0.3816, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.3085680047932894e-06, | |
| "loss": 0.3821, | |
| "step": 1326 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.304973037747154e-06, | |
| "loss": 0.3755, | |
| "step": 1327 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.3013780707010185e-06, | |
| "loss": 0.3694, | |
| "step": 1328 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.2977831036548832e-06, | |
| "loss": 0.3647, | |
| "step": 1329 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.294188136608748e-06, | |
| "loss": 0.3685, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.2905931695626125e-06, | |
| "loss": 0.3939, | |
| "step": 1331 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.286998202516477e-06, | |
| "loss": 0.3821, | |
| "step": 1332 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.2834032354703416e-06, | |
| "loss": 0.4043, | |
| "step": 1333 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.279808268424206e-06, | |
| "loss": 0.3787, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.2762133013780706e-06, | |
| "loss": 0.3687, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.2726183343319352e-06, | |
| "loss": 0.3749, | |
| "step": 1336 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.2690233672858e-06, | |
| "loss": 0.3897, | |
| "step": 1337 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.2654284002396645e-06, | |
| "loss": 0.3888, | |
| "step": 1338 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.2618334331935292e-06, | |
| "loss": 0.3519, | |
| "step": 1339 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.2582384661473938e-06, | |
| "loss": 0.3786, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.2546434991012583e-06, | |
| "loss": 0.4022, | |
| "step": 1341 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.2510485320551228e-06, | |
| "loss": 0.3579, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.2474535650089874e-06, | |
| "loss": 0.3645, | |
| "step": 1343 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.2438585979628521e-06, | |
| "loss": 0.371, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.2402636309167167e-06, | |
| "loss": 0.3872, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.2366686638705812e-06, | |
| "loss": 0.4014, | |
| "step": 1346 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.2330736968244457e-06, | |
| "loss": 0.367, | |
| "step": 1347 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.2294787297783105e-06, | |
| "loss": 0.3742, | |
| "step": 1348 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.225883762732175e-06, | |
| "loss": 0.3634, | |
| "step": 1349 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.2222887956860396e-06, | |
| "loss": 0.3724, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.218693828639904e-06, | |
| "loss": 0.367, | |
| "step": 1351 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.2150988615937688e-06, | |
| "loss": 0.3546, | |
| "step": 1352 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.2115038945476334e-06, | |
| "loss": 0.3696, | |
| "step": 1353 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.207908927501498e-06, | |
| "loss": 0.3987, | |
| "step": 1354 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.2043139604553625e-06, | |
| "loss": 0.3611, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.200718993409227e-06, | |
| "loss": 0.3766, | |
| "step": 1356 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.1971240263630917e-06, | |
| "loss": 0.3805, | |
| "step": 1357 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.1935290593169563e-06, | |
| "loss": 0.3751, | |
| "step": 1358 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.189934092270821e-06, | |
| "loss": 0.3884, | |
| "step": 1359 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1863391252246856e-06, | |
| "loss": 0.3831, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1827441581785501e-06, | |
| "loss": 0.3678, | |
| "step": 1361 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1791491911324146e-06, | |
| "loss": 0.4019, | |
| "step": 1362 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1755542240862792e-06, | |
| "loss": 0.4072, | |
| "step": 1363 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1719592570401437e-06, | |
| "loss": 0.3665, | |
| "step": 1364 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1683642899940083e-06, | |
| "loss": 0.3896, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.164769322947873e-06, | |
| "loss": 0.3716, | |
| "step": 1366 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.1611743559017375e-06, | |
| "loss": 0.4011, | |
| "step": 1367 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.1575793888556023e-06, | |
| "loss": 0.3941, | |
| "step": 1368 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.1539844218094668e-06, | |
| "loss": 0.3912, | |
| "step": 1369 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.1503894547633314e-06, | |
| "loss": 0.3958, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.146794487717196e-06, | |
| "loss": 0.3927, | |
| "step": 1371 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.1431995206710604e-06, | |
| "loss": 0.3925, | |
| "step": 1372 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.1396045536249252e-06, | |
| "loss": 0.3736, | |
| "step": 1373 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.1360095865787897e-06, | |
| "loss": 0.3956, | |
| "step": 1374 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.1324146195326543e-06, | |
| "loss": 0.3702, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.1288196524865188e-06, | |
| "loss": 0.3729, | |
| "step": 1376 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.1252246854403836e-06, | |
| "loss": 0.3713, | |
| "step": 1377 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.121629718394248e-06, | |
| "loss": 0.4047, | |
| "step": 1378 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.1180347513481126e-06, | |
| "loss": 0.378, | |
| "step": 1379 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.1144397843019772e-06, | |
| "loss": 0.4085, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.110844817255842e-06, | |
| "loss": 0.3505, | |
| "step": 1381 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.1072498502097065e-06, | |
| "loss": 0.3884, | |
| "step": 1382 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.103654883163571e-06, | |
| "loss": 0.373, | |
| "step": 1383 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.1000599161174355e-06, | |
| "loss": 0.3997, | |
| "step": 1384 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.0964649490713003e-06, | |
| "loss": 0.3655, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.0928699820251648e-06, | |
| "loss": 0.3943, | |
| "step": 1386 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.0892750149790294e-06, | |
| "loss": 0.3814, | |
| "step": 1387 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.0856800479328941e-06, | |
| "loss": 0.3955, | |
| "step": 1388 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.0820850808867587e-06, | |
| "loss": 0.3905, | |
| "step": 1389 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.0784901138406232e-06, | |
| "loss": 0.3803, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.0748951467944877e-06, | |
| "loss": 0.3784, | |
| "step": 1391 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.0713001797483523e-06, | |
| "loss": 0.3857, | |
| "step": 1392 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.0677052127022168e-06, | |
| "loss": 0.3759, | |
| "step": 1393 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.0641102456560816e-06, | |
| "loss": 0.389, | |
| "step": 1394 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.060515278609946e-06, | |
| "loss": 0.406, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.0569203115638108e-06, | |
| "loss": 0.4044, | |
| "step": 1396 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.0533253445176754e-06, | |
| "loss": 0.3742, | |
| "step": 1397 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.04973037747154e-06, | |
| "loss": 0.3801, | |
| "step": 1398 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.0461354104254045e-06, | |
| "loss": 0.362, | |
| "step": 1399 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.042540443379269e-06, | |
| "loss": 0.383, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.0389454763331335e-06, | |
| "loss": 0.3742, | |
| "step": 1401 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.035350509286998e-06, | |
| "loss": 0.4001, | |
| "step": 1402 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.0317555422408628e-06, | |
| "loss": 0.3656, | |
| "step": 1403 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.0281605751947274e-06, | |
| "loss": 0.3913, | |
| "step": 1404 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.024565608148592e-06, | |
| "loss": 0.3671, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.0209706411024566e-06, | |
| "loss": 0.3938, | |
| "step": 1406 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.0173756740563212e-06, | |
| "loss": 0.3503, | |
| "step": 1407 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.0137807070101857e-06, | |
| "loss": 0.4179, | |
| "step": 1408 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.0101857399640503e-06, | |
| "loss": 0.3798, | |
| "step": 1409 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.006590772917915e-06, | |
| "loss": 0.3855, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.0029958058717795e-06, | |
| "loss": 0.3968, | |
| "step": 1411 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.99400838825644e-07, | |
| "loss": 0.3928, | |
| "step": 1412 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.958058717795086e-07, | |
| "loss": 0.3869, | |
| "step": 1413 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.922109047333734e-07, | |
| "loss": 0.3977, | |
| "step": 1414 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.88615937687238e-07, | |
| "loss": 0.3768, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.850209706411024e-07, | |
| "loss": 0.3931, | |
| "step": 1416 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.814260035949672e-07, | |
| "loss": 0.3752, | |
| "step": 1417 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.778310365488317e-07, | |
| "loss": 0.3668, | |
| "step": 1418 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.742360695026963e-07, | |
| "loss": 0.3889, | |
| "step": 1419 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.706411024565608e-07, | |
| "loss": 0.3622, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.670461354104253e-07, | |
| "loss": 0.3974, | |
| "step": 1421 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.634511683642899e-07, | |
| "loss": 0.3881, | |
| "step": 1422 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.598562013181546e-07, | |
| "loss": 0.3845, | |
| "step": 1423 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.562612342720194e-07, | |
| "loss": 0.4073, | |
| "step": 1424 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.526662672258838e-07, | |
| "loss": 0.3742, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.490713001797485e-07, | |
| "loss": 0.3811, | |
| "step": 1426 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.45476333133613e-07, | |
| "loss": 0.3758, | |
| "step": 1427 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 9.418813660874775e-07, | |
| "loss": 0.3981, | |
| "step": 1428 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 9.382863990413421e-07, | |
| "loss": 0.3845, | |
| "step": 1429 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 9.346914319952067e-07, | |
| "loss": 0.3893, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 9.310964649490713e-07, | |
| "loss": 0.355, | |
| "step": 1431 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 9.27501497902936e-07, | |
| "loss": 0.3973, | |
| "step": 1432 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 9.239065308568005e-07, | |
| "loss": 0.3876, | |
| "step": 1433 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 9.203115638106651e-07, | |
| "loss": 0.3628, | |
| "step": 1434 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 9.167165967645297e-07, | |
| "loss": 0.3559, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 9.131216297183943e-07, | |
| "loss": 0.3947, | |
| "step": 1436 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 9.095266626722588e-07, | |
| "loss": 0.371, | |
| "step": 1437 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 9.059316956261233e-07, | |
| "loss": 0.3721, | |
| "step": 1438 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 9.023367285799881e-07, | |
| "loss": 0.3789, | |
| "step": 1439 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.987417615338527e-07, | |
| "loss": 0.3991, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.951467944877173e-07, | |
| "loss": 0.3944, | |
| "step": 1441 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.915518274415818e-07, | |
| "loss": 0.3903, | |
| "step": 1442 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.879568603954463e-07, | |
| "loss": 0.3844, | |
| "step": 1443 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.84361893349311e-07, | |
| "loss": 0.3725, | |
| "step": 1444 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.807669263031755e-07, | |
| "loss": 0.396, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.771719592570403e-07, | |
| "loss": 0.3962, | |
| "step": 1446 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.735769922109048e-07, | |
| "loss": 0.3999, | |
| "step": 1447 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.699820251647693e-07, | |
| "loss": 0.3708, | |
| "step": 1448 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.66387058118634e-07, | |
| "loss": 0.373, | |
| "step": 1449 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.627920910724985e-07, | |
| "loss": 0.388, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.591971240263631e-07, | |
| "loss": 0.3472, | |
| "step": 1451 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.556021569802276e-07, | |
| "loss": 0.3789, | |
| "step": 1452 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.520071899340922e-07, | |
| "loss": 0.3786, | |
| "step": 1453 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.484122228879569e-07, | |
| "loss": 0.3902, | |
| "step": 1454 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.448172558418215e-07, | |
| "loss": 0.4055, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.412222887956861e-07, | |
| "loss": 0.368, | |
| "step": 1456 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.376273217495506e-07, | |
| "loss": 0.4093, | |
| "step": 1457 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.340323547034153e-07, | |
| "loss": 0.3868, | |
| "step": 1458 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.304373876572798e-07, | |
| "loss": 0.3895, | |
| "step": 1459 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.268424206111443e-07, | |
| "loss": 0.3981, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.232474535650091e-07, | |
| "loss": 0.3935, | |
| "step": 1461 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.196524865188736e-07, | |
| "loss": 0.3512, | |
| "step": 1462 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.160575194727383e-07, | |
| "loss": 0.4029, | |
| "step": 1463 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.124625524266028e-07, | |
| "loss": 0.358, | |
| "step": 1464 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.088675853804673e-07, | |
| "loss": 0.3713, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.052726183343319e-07, | |
| "loss": 0.3837, | |
| "step": 1466 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.016776512881965e-07, | |
| "loss": 0.3939, | |
| "step": 1467 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 7.980826842420612e-07, | |
| "loss": 0.3887, | |
| "step": 1468 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 7.944877171959258e-07, | |
| "loss": 0.3964, | |
| "step": 1469 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 7.908927501497903e-07, | |
| "loss": 0.4059, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 7.872977831036549e-07, | |
| "loss": 0.4006, | |
| "step": 1471 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 7.837028160575195e-07, | |
| "loss": 0.3745, | |
| "step": 1472 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.801078490113841e-07, | |
| "loss": 0.3708, | |
| "step": 1473 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.765128819652486e-07, | |
| "loss": 0.387, | |
| "step": 1474 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.729179149191134e-07, | |
| "loss": 0.3714, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.693229478729779e-07, | |
| "loss": 0.3772, | |
| "step": 1476 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.657279808268424e-07, | |
| "loss": 0.398, | |
| "step": 1477 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.621330137807071e-07, | |
| "loss": 0.3733, | |
| "step": 1478 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.585380467345716e-07, | |
| "loss": 0.381, | |
| "step": 1479 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.549430796884361e-07, | |
| "loss": 0.3771, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.513481126423008e-07, | |
| "loss": 0.4006, | |
| "step": 1481 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.477531455961654e-07, | |
| "loss": 0.3774, | |
| "step": 1482 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.4415817855003e-07, | |
| "loss": 0.3845, | |
| "step": 1483 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.405632115038945e-07, | |
| "loss": 0.3757, | |
| "step": 1484 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.369682444577592e-07, | |
| "loss": 0.3898, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.333732774116237e-07, | |
| "loss": 0.3608, | |
| "step": 1486 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.297783103654883e-07, | |
| "loss": 0.3815, | |
| "step": 1487 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.26183343319353e-07, | |
| "loss": 0.3798, | |
| "step": 1488 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.225883762732175e-07, | |
| "loss": 0.3696, | |
| "step": 1489 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.189934092270821e-07, | |
| "loss": 0.3855, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.153984421809467e-07, | |
| "loss": 0.3798, | |
| "step": 1491 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.118034751348113e-07, | |
| "loss": 0.3957, | |
| "step": 1492 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.082085080886759e-07, | |
| "loss": 0.4118, | |
| "step": 1493 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.046135410425404e-07, | |
| "loss": 0.4045, | |
| "step": 1494 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 7.010185739964051e-07, | |
| "loss": 0.3813, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.974236069502697e-07, | |
| "loss": 0.3962, | |
| "step": 1496 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.938286399041342e-07, | |
| "loss": 0.3944, | |
| "step": 1497 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.902336728579988e-07, | |
| "loss": 0.412, | |
| "step": 1498 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.866387058118634e-07, | |
| "loss": 0.3843, | |
| "step": 1499 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.83043738765728e-07, | |
| "loss": 0.3546, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.794487717195926e-07, | |
| "loss": 0.4036, | |
| "step": 1501 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.758538046734571e-07, | |
| "loss": 0.3988, | |
| "step": 1502 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.722588376273218e-07, | |
| "loss": 0.3828, | |
| "step": 1503 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.686638705811863e-07, | |
| "loss": 0.3736, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.65068903535051e-07, | |
| "loss": 0.4018, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.614739364889156e-07, | |
| "loss": 0.373, | |
| "step": 1506 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.578789694427802e-07, | |
| "loss": 0.3866, | |
| "step": 1507 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.542840023966447e-07, | |
| "loss": 0.3727, | |
| "step": 1508 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.506890353505092e-07, | |
| "loss": 0.3769, | |
| "step": 1509 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.47094068304374e-07, | |
| "loss": 0.3846, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.434991012582385e-07, | |
| "loss": 0.3853, | |
| "step": 1511 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.39904134212103e-07, | |
| "loss": 0.3675, | |
| "step": 1512 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.363091671659676e-07, | |
| "loss": 0.3713, | |
| "step": 1513 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.327142001198322e-07, | |
| "loss": 0.3716, | |
| "step": 1514 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.291192330736969e-07, | |
| "loss": 0.3979, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.255242660275614e-07, | |
| "loss": 0.4059, | |
| "step": 1516 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.219292989814261e-07, | |
| "loss": 0.3897, | |
| "step": 1517 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 6.183343319352906e-07, | |
| "loss": 0.3806, | |
| "step": 1518 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 6.147393648891552e-07, | |
| "loss": 0.3926, | |
| "step": 1519 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 6.111443978430198e-07, | |
| "loss": 0.3624, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 6.075494307968844e-07, | |
| "loss": 0.3777, | |
| "step": 1521 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 6.03954463750749e-07, | |
| "loss": 0.3932, | |
| "step": 1522 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 6.003594967046135e-07, | |
| "loss": 0.404, | |
| "step": 1523 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.967645296584781e-07, | |
| "loss": 0.3878, | |
| "step": 1524 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.931695626123428e-07, | |
| "loss": 0.356, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.895745955662073e-07, | |
| "loss": 0.3679, | |
| "step": 1526 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.859796285200719e-07, | |
| "loss": 0.3812, | |
| "step": 1527 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.823846614739365e-07, | |
| "loss": 0.3858, | |
| "step": 1528 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.787896944278011e-07, | |
| "loss": 0.3832, | |
| "step": 1529 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.751947273816657e-07, | |
| "loss": 0.3496, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.715997603355302e-07, | |
| "loss": 0.3881, | |
| "step": 1531 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.680047932893949e-07, | |
| "loss": 0.3938, | |
| "step": 1532 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.644098262432594e-07, | |
| "loss": 0.3606, | |
| "step": 1533 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.60814859197124e-07, | |
| "loss": 0.4045, | |
| "step": 1534 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 5.572198921509886e-07, | |
| "loss": 0.3665, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 5.536249251048532e-07, | |
| "loss": 0.3811, | |
| "step": 1536 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 5.500299580587178e-07, | |
| "loss": 0.368, | |
| "step": 1537 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 5.464349910125824e-07, | |
| "loss": 0.3735, | |
| "step": 1538 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 5.428400239664471e-07, | |
| "loss": 0.3576, | |
| "step": 1539 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.392450569203116e-07, | |
| "loss": 0.3745, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.356500898741761e-07, | |
| "loss": 0.3751, | |
| "step": 1541 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.320551228280408e-07, | |
| "loss": 0.4006, | |
| "step": 1542 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.284601557819054e-07, | |
| "loss": 0.3867, | |
| "step": 1543 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.2486518873577e-07, | |
| "loss": 0.3694, | |
| "step": 1544 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.212702216896345e-07, | |
| "loss": 0.382, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 5.17675254643499e-07, | |
| "loss": 0.3742, | |
| "step": 1546 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 5.140802875973637e-07, | |
| "loss": 0.374, | |
| "step": 1547 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 5.104853205512283e-07, | |
| "loss": 0.376, | |
| "step": 1548 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 5.068903535050929e-07, | |
| "loss": 0.3588, | |
| "step": 1549 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 5.032953864589575e-07, | |
| "loss": 0.3821, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.99700419412822e-07, | |
| "loss": 0.4012, | |
| "step": 1551 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.961054523666867e-07, | |
| "loss": 0.3735, | |
| "step": 1552 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.925104853205512e-07, | |
| "loss": 0.3929, | |
| "step": 1553 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.889155182744159e-07, | |
| "loss": 0.3576, | |
| "step": 1554 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.853205512282804e-07, | |
| "loss": 0.383, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.817255841821449e-07, | |
| "loss": 0.404, | |
| "step": 1556 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.781306171360097e-07, | |
| "loss": 0.3946, | |
| "step": 1557 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.7453565008987423e-07, | |
| "loss": 0.3718, | |
| "step": 1558 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.7094068304373877e-07, | |
| "loss": 0.3895, | |
| "step": 1559 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.6734571599760336e-07, | |
| "loss": 0.3924, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.63750748951468e-07, | |
| "loss": 0.3806, | |
| "step": 1561 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.6015578190533254e-07, | |
| "loss": 0.3618, | |
| "step": 1562 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.5656081485919713e-07, | |
| "loss": 0.3918, | |
| "step": 1563 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.5296584781306167e-07, | |
| "loss": 0.367, | |
| "step": 1564 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.4937088076692636e-07, | |
| "loss": 0.381, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.457759137207909e-07, | |
| "loss": 0.3848, | |
| "step": 1566 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.421809466746555e-07, | |
| "loss": 0.373, | |
| "step": 1567 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.3858597962852014e-07, | |
| "loss": 0.3728, | |
| "step": 1568 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.349910125823847e-07, | |
| "loss": 0.3906, | |
| "step": 1569 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.3139604553624926e-07, | |
| "loss": 0.3942, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.278010784901138e-07, | |
| "loss": 0.3747, | |
| "step": 1571 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.2420611144397845e-07, | |
| "loss": 0.3779, | |
| "step": 1572 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.2061114439784304e-07, | |
| "loss": 0.3658, | |
| "step": 1573 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.1701617735170763e-07, | |
| "loss": 0.3689, | |
| "step": 1574 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.1342121030557217e-07, | |
| "loss": 0.3799, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.098262432594368e-07, | |
| "loss": 0.3565, | |
| "step": 1576 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.062312762133014e-07, | |
| "loss": 0.3823, | |
| "step": 1577 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.0263630916716594e-07, | |
| "loss": 0.4153, | |
| "step": 1578 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 3.990413421210306e-07, | |
| "loss": 0.3827, | |
| "step": 1579 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.9544637507489517e-07, | |
| "loss": 0.3529, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.9185140802875976e-07, | |
| "loss": 0.378, | |
| "step": 1581 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.882564409826243e-07, | |
| "loss": 0.3787, | |
| "step": 1582 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.8466147393648894e-07, | |
| "loss": 0.3647, | |
| "step": 1583 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.8106650689035354e-07, | |
| "loss": 0.3693, | |
| "step": 1584 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.774715398442181e-07, | |
| "loss": 0.4008, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.738765727980827e-07, | |
| "loss": 0.3486, | |
| "step": 1586 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.7028160575194726e-07, | |
| "loss": 0.3874, | |
| "step": 1587 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.6668663870581185e-07, | |
| "loss": 0.3915, | |
| "step": 1588 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.630916716596765e-07, | |
| "loss": 0.411, | |
| "step": 1589 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.5949670461354103e-07, | |
| "loss": 0.3738, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.5590173756740567e-07, | |
| "loss": 0.4087, | |
| "step": 1591 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.523067705212702e-07, | |
| "loss": 0.401, | |
| "step": 1592 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.4871180347513485e-07, | |
| "loss": 0.377, | |
| "step": 1593 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.451168364289994e-07, | |
| "loss": 0.3858, | |
| "step": 1594 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.41521869382864e-07, | |
| "loss": 0.3789, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.3792690233672857e-07, | |
| "loss": 0.3937, | |
| "step": 1596 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.3433193529059316e-07, | |
| "loss": 0.4194, | |
| "step": 1597 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.307369682444578e-07, | |
| "loss": 0.3739, | |
| "step": 1598 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.2714200119832234e-07, | |
| "loss": 0.3842, | |
| "step": 1599 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.23547034152187e-07, | |
| "loss": 0.4039, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.199520671060515e-07, | |
| "loss": 0.3839, | |
| "step": 1601 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 3.163571000599161e-07, | |
| "loss": 0.3709, | |
| "step": 1602 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 3.127621330137807e-07, | |
| "loss": 0.4059, | |
| "step": 1603 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 3.091671659676453e-07, | |
| "loss": 0.3762, | |
| "step": 1604 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 3.055721989215099e-07, | |
| "loss": 0.3775, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 3.019772318753745e-07, | |
| "loss": 0.382, | |
| "step": 1606 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 2.9838226482923907e-07, | |
| "loss": 0.3826, | |
| "step": 1607 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.9478729778310366e-07, | |
| "loss": 0.3742, | |
| "step": 1608 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.9119233073696825e-07, | |
| "loss": 0.3682, | |
| "step": 1609 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.8759736369083284e-07, | |
| "loss": 0.366, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.8400239664469743e-07, | |
| "loss": 0.388, | |
| "step": 1611 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.80407429598562e-07, | |
| "loss": 0.3833, | |
| "step": 1612 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.768124625524266e-07, | |
| "loss": 0.3726, | |
| "step": 1613 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.732174955062912e-07, | |
| "loss": 0.3845, | |
| "step": 1614 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.696225284601558e-07, | |
| "loss": 0.3998, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.660275614140204e-07, | |
| "loss": 0.3613, | |
| "step": 1616 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.62432594367885e-07, | |
| "loss": 0.3775, | |
| "step": 1617 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.588376273217495e-07, | |
| "loss": 0.3851, | |
| "step": 1618 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.5524266027561416e-07, | |
| "loss": 0.3801, | |
| "step": 1619 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.5164769322947875e-07, | |
| "loss": 0.3662, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.4805272618334334e-07, | |
| "loss": 0.3879, | |
| "step": 1621 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.4445775913720793e-07, | |
| "loss": 0.3524, | |
| "step": 1622 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.4086279209107247e-07, | |
| "loss": 0.3745, | |
| "step": 1623 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.3726782504493711e-07, | |
| "loss": 0.383, | |
| "step": 1624 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 2.3367285799880168e-07, | |
| "loss": 0.3772, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 2.3007789095266627e-07, | |
| "loss": 0.4023, | |
| "step": 1626 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 2.2648292390653083e-07, | |
| "loss": 0.403, | |
| "step": 1627 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 2.2288795686039545e-07, | |
| "loss": 0.377, | |
| "step": 1628 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 2.1929298981426007e-07, | |
| "loss": 0.3639, | |
| "step": 1629 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 2.1569802276812463e-07, | |
| "loss": 0.3633, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 2.1210305572198922e-07, | |
| "loss": 0.3836, | |
| "step": 1631 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 2.0850808867585381e-07, | |
| "loss": 0.3753, | |
| "step": 1632 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 2.049131216297184e-07, | |
| "loss": 0.3853, | |
| "step": 1633 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 2.0131815458358297e-07, | |
| "loss": 0.3856, | |
| "step": 1634 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 1.9772318753744759e-07, | |
| "loss": 0.3768, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.9412822049131215e-07, | |
| "loss": 0.391, | |
| "step": 1636 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.9053325344517677e-07, | |
| "loss": 0.3993, | |
| "step": 1637 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.8693828639904136e-07, | |
| "loss": 0.3922, | |
| "step": 1638 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.8334331935290592e-07, | |
| "loss": 0.3827, | |
| "step": 1639 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.7974835230677051e-07, | |
| "loss": 0.3501, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.761533852606351e-07, | |
| "loss": 0.3802, | |
| "step": 1641 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.725584182144997e-07, | |
| "loss": 0.382, | |
| "step": 1642 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.6896345116836429e-07, | |
| "loss": 0.397, | |
| "step": 1643 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.653684841222289e-07, | |
| "loss": 0.3547, | |
| "step": 1644 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.617735170760935e-07, | |
| "loss": 0.4033, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.5817855002995806e-07, | |
| "loss": 0.3999, | |
| "step": 1646 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.5458358298382265e-07, | |
| "loss": 0.3769, | |
| "step": 1647 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.5098861593768724e-07, | |
| "loss": 0.379, | |
| "step": 1648 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.4739364889155183e-07, | |
| "loss": 0.3906, | |
| "step": 1649 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.4379868184541642e-07, | |
| "loss": 0.3913, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.40203714799281e-07, | |
| "loss": 0.3885, | |
| "step": 1651 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.366087477531456e-07, | |
| "loss": 0.3714, | |
| "step": 1652 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.330137807070102e-07, | |
| "loss": 0.3876, | |
| "step": 1653 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.2941881366087476e-07, | |
| "loss": 0.3796, | |
| "step": 1654 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.2582384661473938e-07, | |
| "loss": 0.3756, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.2222887956860397e-07, | |
| "loss": 0.3915, | |
| "step": 1656 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.1863391252246856e-07, | |
| "loss": 0.3719, | |
| "step": 1657 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.1503894547633313e-07, | |
| "loss": 0.3815, | |
| "step": 1658 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.1144397843019773e-07, | |
| "loss": 0.3925, | |
| "step": 1659 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.0784901138406232e-07, | |
| "loss": 0.3526, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.0425404433792691e-07, | |
| "loss": 0.3902, | |
| "step": 1661 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.0065907729179148e-07, | |
| "loss": 0.4026, | |
| "step": 1662 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 9.706411024565608e-08, | |
| "loss": 0.3939, | |
| "step": 1663 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 9.346914319952068e-08, | |
| "loss": 0.3857, | |
| "step": 1664 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 8.987417615338526e-08, | |
| "loss": 0.3996, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 8.627920910724985e-08, | |
| "loss": 0.3682, | |
| "step": 1666 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 8.268424206111445e-08, | |
| "loss": 0.3898, | |
| "step": 1667 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 7.908927501497903e-08, | |
| "loss": 0.359, | |
| "step": 1668 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 7.549430796884362e-08, | |
| "loss": 0.3919, | |
| "step": 1669 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 7.189934092270821e-08, | |
| "loss": 0.3933, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 6.83043738765728e-08, | |
| "loss": 0.3699, | |
| "step": 1671 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 6.470940683043738e-08, | |
| "loss": 0.4004, | |
| "step": 1672 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 6.111443978430198e-08, | |
| "loss": 0.3964, | |
| "step": 1673 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 5.751947273816657e-08, | |
| "loss": 0.3836, | |
| "step": 1674 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 5.392450569203116e-08, | |
| "loss": 0.4036, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 5.032953864589574e-08, | |
| "loss": 0.3688, | |
| "step": 1676 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.673457159976034e-08, | |
| "loss": 0.3764, | |
| "step": 1677 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.3139604553624924e-08, | |
| "loss": 0.391, | |
| "step": 1678 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.9544637507489515e-08, | |
| "loss": 0.3851, | |
| "step": 1679 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.5949670461354105e-08, | |
| "loss": 0.3747, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 3.235470341521869e-08, | |
| "loss": 0.3838, | |
| "step": 1681 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.8759736369083284e-08, | |
| "loss": 0.3735, | |
| "step": 1682 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.516476932294787e-08, | |
| "loss": 0.3779, | |
| "step": 1683 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.1569802276812462e-08, | |
| "loss": 0.3876, | |
| "step": 1684 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 1.7974835230677053e-08, | |
| "loss": 0.4162, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 1.4379868184541642e-08, | |
| "loss": 0.4096, | |
| "step": 1686 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 1.0784901138406231e-08, | |
| "loss": 0.4114, | |
| "step": 1687 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 7.189934092270821e-09, | |
| "loss": 0.377, | |
| "step": 1688 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 3.5949670461354105e-09, | |
| "loss": 0.38, | |
| "step": 1689 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 1689, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "total_flos": 1.4167476619953832e+19, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |