| { | |
| "best_metric": 1.7139549255371094, | |
| "best_model_checkpoint": "outputs/checkpoint-500", | |
| "epoch": 0.7483629560336763, | |
| "eval_steps": 25, | |
| "global_step": 500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0014967259120673526, | |
| "grad_norm": 2.434373378753662, | |
| "learning_rate": 4e-05, | |
| "loss": 2.0037, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.002993451824134705, | |
| "grad_norm": 2.795464038848877, | |
| "learning_rate": 8e-05, | |
| "loss": 1.9814, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.004490177736202058, | |
| "grad_norm": 2.0998575687408447, | |
| "learning_rate": 0.00012, | |
| "loss": 2.0044, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.00598690364826941, | |
| "grad_norm": 4.470895290374756, | |
| "learning_rate": 0.00016, | |
| "loss": 1.8696, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.007483629560336763, | |
| "grad_norm": 1.6047176122665405, | |
| "learning_rate": 0.0002, | |
| "loss": 1.8021, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.008980355472404116, | |
| "grad_norm": 1.9450230598449707, | |
| "learning_rate": 0.0001996638655462185, | |
| "loss": 1.8558, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.01047708138447147, | |
| "grad_norm": 1.4420437812805176, | |
| "learning_rate": 0.00019932773109243698, | |
| "loss": 1.9079, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.01197380729653882, | |
| "grad_norm": 2.428009033203125, | |
| "learning_rate": 0.00019899159663865548, | |
| "loss": 1.7173, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.013470533208606174, | |
| "grad_norm": 1.6478683948516846, | |
| "learning_rate": 0.00019865546218487395, | |
| "loss": 1.7631, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.014967259120673527, | |
| "grad_norm": 1.6283013820648193, | |
| "learning_rate": 0.00019831932773109245, | |
| "loss": 1.9375, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01646398503274088, | |
| "grad_norm": 1.7736356258392334, | |
| "learning_rate": 0.00019798319327731095, | |
| "loss": 1.7921, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.017960710944808233, | |
| "grad_norm": 1.3353335857391357, | |
| "learning_rate": 0.00019764705882352942, | |
| "loss": 1.8699, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.019457436856875586, | |
| "grad_norm": 1.5582761764526367, | |
| "learning_rate": 0.00019731092436974792, | |
| "loss": 1.8015, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.02095416276894294, | |
| "grad_norm": 1.3032814264297485, | |
| "learning_rate": 0.00019697478991596642, | |
| "loss": 1.8513, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.02245088868101029, | |
| "grad_norm": 1.3724833726882935, | |
| "learning_rate": 0.00019663865546218486, | |
| "loss": 1.8694, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.02394761459307764, | |
| "grad_norm": 1.715156078338623, | |
| "learning_rate": 0.00019630252100840336, | |
| "loss": 1.7997, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.025444340505144995, | |
| "grad_norm": 1.9989070892333984, | |
| "learning_rate": 0.00019596638655462186, | |
| "loss": 1.7037, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.026941066417212348, | |
| "grad_norm": 1.6255011558532715, | |
| "learning_rate": 0.00019563025210084033, | |
| "loss": 1.8382, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.0284377923292797, | |
| "grad_norm": 1.305870532989502, | |
| "learning_rate": 0.00019529411764705883, | |
| "loss": 1.8095, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.029934518241347054, | |
| "grad_norm": 3.40390944480896, | |
| "learning_rate": 0.0001949579831932773, | |
| "loss": 1.6112, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0314312441534144, | |
| "grad_norm": 1.1514052152633667, | |
| "learning_rate": 0.0001946218487394958, | |
| "loss": 1.9725, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.03292797006548176, | |
| "grad_norm": 1.9419797658920288, | |
| "learning_rate": 0.0001942857142857143, | |
| "loss": 1.8492, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.03442469597754911, | |
| "grad_norm": 1.830913782119751, | |
| "learning_rate": 0.00019394957983193278, | |
| "loss": 1.7852, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.035921421889616466, | |
| "grad_norm": 1.345436453819275, | |
| "learning_rate": 0.00019361344537815127, | |
| "loss": 1.8929, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.037418147801683815, | |
| "grad_norm": 1.1833657026290894, | |
| "learning_rate": 0.00019327731092436975, | |
| "loss": 1.8822, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.037418147801683815, | |
| "eval_loss": 1.7638475894927979, | |
| "eval_runtime": 13.3387, | |
| "eval_samples_per_second": 4.798, | |
| "eval_steps_per_second": 2.399, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.03891487371375117, | |
| "grad_norm": 1.4092592000961304, | |
| "learning_rate": 0.00019294117647058825, | |
| "loss": 1.954, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.04041159962581852, | |
| "grad_norm": 1.201281189918518, | |
| "learning_rate": 0.00019260504201680674, | |
| "loss": 1.9364, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.04190832553788588, | |
| "grad_norm": 1.407148838043213, | |
| "learning_rate": 0.00019226890756302522, | |
| "loss": 1.7673, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.04340505144995323, | |
| "grad_norm": 1.3781392574310303, | |
| "learning_rate": 0.00019193277310924372, | |
| "loss": 1.81, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.04490177736202058, | |
| "grad_norm": 1.4952391386032104, | |
| "learning_rate": 0.00019159663865546221, | |
| "loss": 1.8216, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.046398503274087934, | |
| "grad_norm": 1.5127140283584595, | |
| "learning_rate": 0.0001912605042016807, | |
| "loss": 1.8316, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.04789522918615528, | |
| "grad_norm": 1.3208520412445068, | |
| "learning_rate": 0.00019092436974789919, | |
| "loss": 1.7302, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.04939195509822264, | |
| "grad_norm": 1.3473477363586426, | |
| "learning_rate": 0.00019058823529411766, | |
| "loss": 1.9661, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.05088868101028999, | |
| "grad_norm": 1.201379418373108, | |
| "learning_rate": 0.00019025210084033613, | |
| "loss": 1.8531, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.052385406922357346, | |
| "grad_norm": 1.32240891456604, | |
| "learning_rate": 0.00018991596638655463, | |
| "loss": 1.7485, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.053882132834424695, | |
| "grad_norm": 1.3222694396972656, | |
| "learning_rate": 0.0001895798319327731, | |
| "loss": 1.7939, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.05537885874649205, | |
| "grad_norm": 1.1342493295669556, | |
| "learning_rate": 0.0001892436974789916, | |
| "loss": 1.855, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.0568755846585594, | |
| "grad_norm": 1.4912521839141846, | |
| "learning_rate": 0.0001889075630252101, | |
| "loss": 1.8394, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.05837231057062675, | |
| "grad_norm": 1.4635943174362183, | |
| "learning_rate": 0.00018857142857142857, | |
| "loss": 1.7669, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.05986903648269411, | |
| "grad_norm": 1.4757208824157715, | |
| "learning_rate": 0.00018823529411764707, | |
| "loss": 1.7842, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.06136576239476146, | |
| "grad_norm": 1.5162277221679688, | |
| "learning_rate": 0.00018789915966386554, | |
| "loss": 1.8873, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.0628624883068288, | |
| "grad_norm": 1.3085792064666748, | |
| "learning_rate": 0.00018756302521008404, | |
| "loss": 1.9716, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.06435921421889616, | |
| "grad_norm": 1.1214959621429443, | |
| "learning_rate": 0.00018722689075630254, | |
| "loss": 1.8991, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.06585594013096352, | |
| "grad_norm": 1.1944588422775269, | |
| "learning_rate": 0.000186890756302521, | |
| "loss": 1.8894, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.06735266604303088, | |
| "grad_norm": 1.509717345237732, | |
| "learning_rate": 0.0001865546218487395, | |
| "loss": 1.8035, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.06884939195509822, | |
| "grad_norm": 1.3220465183258057, | |
| "learning_rate": 0.000186218487394958, | |
| "loss": 1.8673, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.07034611786716558, | |
| "grad_norm": 1.3592686653137207, | |
| "learning_rate": 0.00018588235294117648, | |
| "loss": 1.8396, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.07184284377923293, | |
| "grad_norm": 1.3568888902664185, | |
| "learning_rate": 0.00018554621848739498, | |
| "loss": 1.6562, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.07333956969130027, | |
| "grad_norm": 1.1371209621429443, | |
| "learning_rate": 0.00018521008403361345, | |
| "loss": 1.9563, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.07483629560336763, | |
| "grad_norm": 1.234221339225769, | |
| "learning_rate": 0.00018487394957983195, | |
| "loss": 1.7702, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.07483629560336763, | |
| "eval_loss": 1.7585577964782715, | |
| "eval_runtime": 9.906, | |
| "eval_samples_per_second": 6.461, | |
| "eval_steps_per_second": 3.23, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.07633302151543499, | |
| "grad_norm": 1.4679025411605835, | |
| "learning_rate": 0.00018453781512605045, | |
| "loss": 1.7903, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.07782974742750234, | |
| "grad_norm": 1.6469783782958984, | |
| "learning_rate": 0.0001842016806722689, | |
| "loss": 1.9121, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.07932647333956969, | |
| "grad_norm": 1.0950040817260742, | |
| "learning_rate": 0.0001838655462184874, | |
| "loss": 1.8015, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.08082319925163704, | |
| "grad_norm": 1.4614354372024536, | |
| "learning_rate": 0.0001835294117647059, | |
| "loss": 1.8627, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.0823199251637044, | |
| "grad_norm": 1.0772849321365356, | |
| "learning_rate": 0.00018319327731092437, | |
| "loss": 1.8474, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.08381665107577176, | |
| "grad_norm": 0.8980317115783691, | |
| "learning_rate": 0.00018285714285714286, | |
| "loss": 1.9381, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.0853133769878391, | |
| "grad_norm": 1.028698205947876, | |
| "learning_rate": 0.00018252100840336134, | |
| "loss": 1.8726, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.08681010289990646, | |
| "grad_norm": 1.2643156051635742, | |
| "learning_rate": 0.00018218487394957984, | |
| "loss": 1.937, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.08830682881197381, | |
| "grad_norm": 1.0845692157745361, | |
| "learning_rate": 0.00018184873949579833, | |
| "loss": 1.9708, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.08980355472404115, | |
| "grad_norm": 1.2025495767593384, | |
| "learning_rate": 0.0001815126050420168, | |
| "loss": 1.8674, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.09130028063610851, | |
| "grad_norm": 1.2060717344284058, | |
| "learning_rate": 0.0001811764705882353, | |
| "loss": 1.8268, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.09279700654817587, | |
| "grad_norm": 1.3296293020248413, | |
| "learning_rate": 0.0001808403361344538, | |
| "loss": 1.6956, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.09429373246024322, | |
| "grad_norm": 1.2353034019470215, | |
| "learning_rate": 0.00018050420168067228, | |
| "loss": 1.9816, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.09579045837231057, | |
| "grad_norm": 1.5975768566131592, | |
| "learning_rate": 0.00018016806722689078, | |
| "loss": 1.7846, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.09728718428437792, | |
| "grad_norm": 1.2220622301101685, | |
| "learning_rate": 0.00017983193277310925, | |
| "loss": 1.7895, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.09878391019644528, | |
| "grad_norm": 1.2025718688964844, | |
| "learning_rate": 0.00017949579831932775, | |
| "loss": 1.9242, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.10028063610851262, | |
| "grad_norm": 3.2830123901367188, | |
| "learning_rate": 0.00017915966386554625, | |
| "loss": 1.7076, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.10177736202057998, | |
| "grad_norm": 1.5499017238616943, | |
| "learning_rate": 0.00017882352941176472, | |
| "loss": 1.7964, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.10327408793264733, | |
| "grad_norm": 1.4630420207977295, | |
| "learning_rate": 0.00017848739495798322, | |
| "loss": 1.8724, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.10477081384471469, | |
| "grad_norm": 1.4005722999572754, | |
| "learning_rate": 0.0001781512605042017, | |
| "loss": 1.6889, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.10626753975678203, | |
| "grad_norm": 1.114207148551941, | |
| "learning_rate": 0.00017781512605042016, | |
| "loss": 1.8272, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.10776426566884939, | |
| "grad_norm": 1.4557619094848633, | |
| "learning_rate": 0.00017747899159663866, | |
| "loss": 1.6877, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.10926099158091675, | |
| "grad_norm": 1.4767951965332031, | |
| "learning_rate": 0.00017714285714285713, | |
| "loss": 1.8667, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.1107577174929841, | |
| "grad_norm": 1.3078974485397339, | |
| "learning_rate": 0.00017680672268907563, | |
| "loss": 1.9319, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.11225444340505145, | |
| "grad_norm": 1.1861608028411865, | |
| "learning_rate": 0.00017647058823529413, | |
| "loss": 1.9233, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.11225444340505145, | |
| "eval_loss": 1.7611756324768066, | |
| "eval_runtime": 9.9015, | |
| "eval_samples_per_second": 6.464, | |
| "eval_steps_per_second": 3.232, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.1137511693171188, | |
| "grad_norm": 1.1504981517791748, | |
| "learning_rate": 0.0001761344537815126, | |
| "loss": 1.8044, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.11524789522918616, | |
| "grad_norm": 1.3776837587356567, | |
| "learning_rate": 0.0001757983193277311, | |
| "loss": 1.725, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.1167446211412535, | |
| "grad_norm": 1.3975869417190552, | |
| "learning_rate": 0.0001754621848739496, | |
| "loss": 1.7935, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.11824134705332086, | |
| "grad_norm": 1.3506461381912231, | |
| "learning_rate": 0.00017512605042016807, | |
| "loss": 1.7342, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.11973807296538821, | |
| "grad_norm": 1.1317209005355835, | |
| "learning_rate": 0.00017478991596638657, | |
| "loss": 1.8149, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.12123479887745557, | |
| "grad_norm": 1.2540264129638672, | |
| "learning_rate": 0.00017445378151260504, | |
| "loss": 1.84, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.12273152478952291, | |
| "grad_norm": 1.23360276222229, | |
| "learning_rate": 0.00017411764705882354, | |
| "loss": 1.7623, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.12422825070159027, | |
| "grad_norm": 1.0347758531570435, | |
| "learning_rate": 0.00017378151260504204, | |
| "loss": 1.7381, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.1257249766136576, | |
| "grad_norm": 1.4501961469650269, | |
| "learning_rate": 0.0001734453781512605, | |
| "loss": 1.7075, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.12722170252572498, | |
| "grad_norm": 1.0509997606277466, | |
| "learning_rate": 0.000173109243697479, | |
| "loss": 1.7295, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.12871842843779233, | |
| "grad_norm": 1.2986621856689453, | |
| "learning_rate": 0.00017277310924369748, | |
| "loss": 1.7988, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.13021515434985967, | |
| "grad_norm": 1.1701687574386597, | |
| "learning_rate": 0.00017243697478991598, | |
| "loss": 1.6763, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.13171188026192704, | |
| "grad_norm": 1.2512173652648926, | |
| "learning_rate": 0.00017210084033613448, | |
| "loss": 1.6641, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.13320860617399438, | |
| "grad_norm": 1.658525824546814, | |
| "learning_rate": 0.00017176470588235293, | |
| "loss": 1.6849, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.13470533208606175, | |
| "grad_norm": 1.5465582609176636, | |
| "learning_rate": 0.00017142857142857143, | |
| "loss": 1.6439, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.1362020579981291, | |
| "grad_norm": 1.3289684057235718, | |
| "learning_rate": 0.00017109243697478992, | |
| "loss": 1.7429, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.13769878391019644, | |
| "grad_norm": 1.3123184442520142, | |
| "learning_rate": 0.0001707563025210084, | |
| "loss": 1.6429, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.1391955098222638, | |
| "grad_norm": 1.385330319404602, | |
| "learning_rate": 0.0001704201680672269, | |
| "loss": 1.8257, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.14069223573433115, | |
| "grad_norm": 1.3719394207000732, | |
| "learning_rate": 0.0001700840336134454, | |
| "loss": 1.7493, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.1421889616463985, | |
| "grad_norm": 1.468948245048523, | |
| "learning_rate": 0.00016974789915966387, | |
| "loss": 1.8626, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.14368568755846586, | |
| "grad_norm": 1.2705055475234985, | |
| "learning_rate": 0.00016941176470588237, | |
| "loss": 1.79, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.1451824134705332, | |
| "grad_norm": 1.0876643657684326, | |
| "learning_rate": 0.00016907563025210084, | |
| "loss": 1.9631, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.14667913938260055, | |
| "grad_norm": 1.1760327816009521, | |
| "learning_rate": 0.00016873949579831934, | |
| "loss": 1.91, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.14817586529466792, | |
| "grad_norm": 1.0915436744689941, | |
| "learning_rate": 0.00016840336134453784, | |
| "loss": 1.8369, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.14967259120673526, | |
| "grad_norm": 1.4619494676589966, | |
| "learning_rate": 0.0001680672268907563, | |
| "loss": 1.7074, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.14967259120673526, | |
| "eval_loss": 1.7622296810150146, | |
| "eval_runtime": 9.9165, | |
| "eval_samples_per_second": 6.454, | |
| "eval_steps_per_second": 3.227, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.15116931711880263, | |
| "grad_norm": 1.147395372390747, | |
| "learning_rate": 0.0001677310924369748, | |
| "loss": 1.7815, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.15266604303086997, | |
| "grad_norm": 1.1551228761672974, | |
| "learning_rate": 0.00016739495798319328, | |
| "loss": 1.9444, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.15416276894293732, | |
| "grad_norm": 1.2220309972763062, | |
| "learning_rate": 0.00016705882352941178, | |
| "loss": 1.8735, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.1556594948550047, | |
| "grad_norm": 1.315051555633545, | |
| "learning_rate": 0.00016672268907563028, | |
| "loss": 1.7284, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.15715622076707203, | |
| "grad_norm": 1.2493054866790771, | |
| "learning_rate": 0.00016638655462184875, | |
| "loss": 1.7598, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.15865294667913937, | |
| "grad_norm": 1.0625994205474854, | |
| "learning_rate": 0.00016605042016806725, | |
| "loss": 1.8334, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.16014967259120674, | |
| "grad_norm": 1.284947395324707, | |
| "learning_rate": 0.00016571428571428575, | |
| "loss": 1.7762, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.16164639850327409, | |
| "grad_norm": 1.27797269821167, | |
| "learning_rate": 0.0001653781512605042, | |
| "loss": 1.7898, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.16314312441534143, | |
| "grad_norm": 1.4802685976028442, | |
| "learning_rate": 0.0001650420168067227, | |
| "loss": 1.7004, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.1646398503274088, | |
| "grad_norm": 0.973327100276947, | |
| "learning_rate": 0.0001647058823529412, | |
| "loss": 1.9218, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.16613657623947614, | |
| "grad_norm": 1.3942281007766724, | |
| "learning_rate": 0.00016436974789915966, | |
| "loss": 1.9183, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.1676333021515435, | |
| "grad_norm": 1.2495373487472534, | |
| "learning_rate": 0.00016403361344537816, | |
| "loss": 1.8296, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.16913002806361085, | |
| "grad_norm": 1.2634400129318237, | |
| "learning_rate": 0.00016369747899159663, | |
| "loss": 1.9474, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.1706267539756782, | |
| "grad_norm": 1.2135545015335083, | |
| "learning_rate": 0.00016336134453781513, | |
| "loss": 1.787, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.17212347988774557, | |
| "grad_norm": 1.0599427223205566, | |
| "learning_rate": 0.00016302521008403363, | |
| "loss": 1.9359, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.1736202057998129, | |
| "grad_norm": 1.1955755949020386, | |
| "learning_rate": 0.0001626890756302521, | |
| "loss": 1.7344, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.17511693171188025, | |
| "grad_norm": 1.3276002407073975, | |
| "learning_rate": 0.0001623529411764706, | |
| "loss": 1.8934, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.17661365762394762, | |
| "grad_norm": 1.4872647523880005, | |
| "learning_rate": 0.00016201680672268907, | |
| "loss": 1.8229, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.17811038353601497, | |
| "grad_norm": 1.377747893333435, | |
| "learning_rate": 0.00016168067226890757, | |
| "loss": 1.7041, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.1796071094480823, | |
| "grad_norm": 1.087159276008606, | |
| "learning_rate": 0.00016134453781512607, | |
| "loss": 1.8932, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.18110383536014968, | |
| "grad_norm": 1.299407720565796, | |
| "learning_rate": 0.00016100840336134454, | |
| "loss": 1.8673, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.18260056127221702, | |
| "grad_norm": 1.172582983970642, | |
| "learning_rate": 0.00016067226890756304, | |
| "loss": 1.7616, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.1840972871842844, | |
| "grad_norm": 1.4097166061401367, | |
| "learning_rate": 0.00016033613445378154, | |
| "loss": 1.6969, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.18559401309635173, | |
| "grad_norm": 1.0662322044372559, | |
| "learning_rate": 0.00016, | |
| "loss": 1.8128, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.18709073900841908, | |
| "grad_norm": 1.2918486595153809, | |
| "learning_rate": 0.0001596638655462185, | |
| "loss": 1.8538, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.18709073900841908, | |
| "eval_loss": 1.762688159942627, | |
| "eval_runtime": 9.9256, | |
| "eval_samples_per_second": 6.448, | |
| "eval_steps_per_second": 3.224, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.18858746492048645, | |
| "grad_norm": 1.4962085485458374, | |
| "learning_rate": 0.00015932773109243698, | |
| "loss": 1.732, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.1900841908325538, | |
| "grad_norm": 1.1726781129837036, | |
| "learning_rate": 0.00015899159663865546, | |
| "loss": 1.8208, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.19158091674462113, | |
| "grad_norm": 1.1145118474960327, | |
| "learning_rate": 0.00015865546218487396, | |
| "loss": 1.8918, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.1930776426566885, | |
| "grad_norm": 1.2448960542678833, | |
| "learning_rate": 0.00015831932773109243, | |
| "loss": 1.7441, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.19457436856875585, | |
| "grad_norm": 1.1530712842941284, | |
| "learning_rate": 0.00015798319327731093, | |
| "loss": 1.8817, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.1960710944808232, | |
| "grad_norm": 0.9994822144508362, | |
| "learning_rate": 0.00015764705882352943, | |
| "loss": 1.8655, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.19756782039289056, | |
| "grad_norm": 1.475071668624878, | |
| "learning_rate": 0.0001573109243697479, | |
| "loss": 1.5336, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.1990645463049579, | |
| "grad_norm": 1.0761913061141968, | |
| "learning_rate": 0.0001569747899159664, | |
| "loss": 1.9806, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.20056127221702524, | |
| "grad_norm": 1.1356831789016724, | |
| "learning_rate": 0.00015663865546218487, | |
| "loss": 1.7876, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.20205799812909261, | |
| "grad_norm": 1.2572017908096313, | |
| "learning_rate": 0.00015630252100840337, | |
| "loss": 1.7748, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.20355472404115996, | |
| "grad_norm": 1.2248603105545044, | |
| "learning_rate": 0.00015596638655462187, | |
| "loss": 1.8696, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.20505144995322733, | |
| "grad_norm": 1.1236392259597778, | |
| "learning_rate": 0.00015563025210084034, | |
| "loss": 1.8774, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.20654817586529467, | |
| "grad_norm": 1.3141965866088867, | |
| "learning_rate": 0.00015529411764705884, | |
| "loss": 1.8864, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.208044901777362, | |
| "grad_norm": 1.364126443862915, | |
| "learning_rate": 0.00015495798319327734, | |
| "loss": 1.6608, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.20954162768942938, | |
| "grad_norm": 1.0413249731063843, | |
| "learning_rate": 0.0001546218487394958, | |
| "loss": 1.8559, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.21103835360149673, | |
| "grad_norm": 1.0397676229476929, | |
| "learning_rate": 0.0001542857142857143, | |
| "loss": 1.8886, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.21253507951356407, | |
| "grad_norm": 1.072859525680542, | |
| "learning_rate": 0.00015394957983193278, | |
| "loss": 1.8294, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.21403180542563144, | |
| "grad_norm": 1.3657593727111816, | |
| "learning_rate": 0.00015361344537815128, | |
| "loss": 1.6844, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.21552853133769878, | |
| "grad_norm": 1.4864728450775146, | |
| "learning_rate": 0.00015327731092436978, | |
| "loss": 1.7364, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.21702525724976612, | |
| "grad_norm": 1.2342066764831543, | |
| "learning_rate": 0.00015294117647058822, | |
| "loss": 1.8441, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.2185219831618335, | |
| "grad_norm": 1.788312315940857, | |
| "learning_rate": 0.00015260504201680672, | |
| "loss": 1.7228, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.22001870907390084, | |
| "grad_norm": 1.1817758083343506, | |
| "learning_rate": 0.00015226890756302522, | |
| "loss": 1.7602, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.2215154349859682, | |
| "grad_norm": 1.0265668630599976, | |
| "learning_rate": 0.0001519327731092437, | |
| "loss": 1.8208, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.22301216089803555, | |
| "grad_norm": 1.2950278520584106, | |
| "learning_rate": 0.0001515966386554622, | |
| "loss": 1.7958, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.2245088868101029, | |
| "grad_norm": 1.21755850315094, | |
| "learning_rate": 0.00015126050420168066, | |
| "loss": 1.9051, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.2245088868101029, | |
| "eval_loss": 1.760790467262268, | |
| "eval_runtime": 9.9578, | |
| "eval_samples_per_second": 6.427, | |
| "eval_steps_per_second": 3.214, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.22600561272217026, | |
| "grad_norm": 1.0923712253570557, | |
| "learning_rate": 0.00015092436974789916, | |
| "loss": 1.7108, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.2275023386342376, | |
| "grad_norm": 1.222659945487976, | |
| "learning_rate": 0.00015058823529411766, | |
| "loss": 1.8221, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.22899906454630495, | |
| "grad_norm": 1.2633992433547974, | |
| "learning_rate": 0.00015025210084033613, | |
| "loss": 1.7576, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.23049579045837232, | |
| "grad_norm": 1.3663132190704346, | |
| "learning_rate": 0.00014991596638655463, | |
| "loss": 1.7411, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.23199251637043966, | |
| "grad_norm": 1.0865604877471924, | |
| "learning_rate": 0.00014957983193277313, | |
| "loss": 1.7677, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.233489242282507, | |
| "grad_norm": 1.0313267707824707, | |
| "learning_rate": 0.0001492436974789916, | |
| "loss": 1.8802, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.23498596819457437, | |
| "grad_norm": 1.300527811050415, | |
| "learning_rate": 0.0001489075630252101, | |
| "loss": 1.7062, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.23648269410664172, | |
| "grad_norm": 1.1446460485458374, | |
| "learning_rate": 0.00014857142857142857, | |
| "loss": 1.821, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.2379794200187091, | |
| "grad_norm": 1.1422364711761475, | |
| "learning_rate": 0.00014823529411764707, | |
| "loss": 1.9409, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.23947614593077643, | |
| "grad_norm": 1.2949453592300415, | |
| "learning_rate": 0.00014789915966386557, | |
| "loss": 1.9585, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.24097287184284377, | |
| "grad_norm": 1.0998990535736084, | |
| "learning_rate": 0.00014756302521008404, | |
| "loss": 1.823, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.24246959775491114, | |
| "grad_norm": 1.3279107809066772, | |
| "learning_rate": 0.00014722689075630254, | |
| "loss": 1.7261, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.24396632366697849, | |
| "grad_norm": 1.1471811532974243, | |
| "learning_rate": 0.00014689075630252101, | |
| "loss": 1.8507, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.24546304957904583, | |
| "grad_norm": 1.0583767890930176, | |
| "learning_rate": 0.0001465546218487395, | |
| "loss": 1.9237, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.2469597754911132, | |
| "grad_norm": 1.3685698509216309, | |
| "learning_rate": 0.00014621848739495799, | |
| "loss": 1.8113, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.24845650140318054, | |
| "grad_norm": 1.5403902530670166, | |
| "learning_rate": 0.00014588235294117646, | |
| "loss": 1.797, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.24995322731524788, | |
| "grad_norm": 1.0701065063476562, | |
| "learning_rate": 0.00014554621848739496, | |
| "loss": 1.8527, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.2514499532273152, | |
| "grad_norm": 1.0490641593933105, | |
| "learning_rate": 0.00014521008403361346, | |
| "loss": 1.9737, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.2529466791393826, | |
| "grad_norm": 1.1226781606674194, | |
| "learning_rate": 0.00014487394957983193, | |
| "loss": 1.8817, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.25444340505144997, | |
| "grad_norm": 1.2346230745315552, | |
| "learning_rate": 0.00014453781512605043, | |
| "loss": 1.873, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.2559401309635173, | |
| "grad_norm": 0.9842538237571716, | |
| "learning_rate": 0.00014420168067226893, | |
| "loss": 1.8042, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.25743685687558465, | |
| "grad_norm": 1.217533826828003, | |
| "learning_rate": 0.0001438655462184874, | |
| "loss": 1.7716, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.258933582787652, | |
| "grad_norm": 1.1946215629577637, | |
| "learning_rate": 0.0001435294117647059, | |
| "loss": 1.8272, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.26043030869971934, | |
| "grad_norm": 1.1696230173110962, | |
| "learning_rate": 0.00014319327731092437, | |
| "loss": 1.801, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.26192703461178674, | |
| "grad_norm": 1.2546944618225098, | |
| "learning_rate": 0.00014285714285714287, | |
| "loss": 1.7625, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.26192703461178674, | |
| "eval_loss": 1.7579066753387451, | |
| "eval_runtime": 9.9267, | |
| "eval_samples_per_second": 6.447, | |
| "eval_steps_per_second": 3.224, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.2634237605238541, | |
| "grad_norm": 1.136839509010315, | |
| "learning_rate": 0.00014252100840336137, | |
| "loss": 1.8288, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.2649204864359214, | |
| "grad_norm": 1.2978945970535278, | |
| "learning_rate": 0.00014218487394957984, | |
| "loss": 1.6907, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.26641721234798876, | |
| "grad_norm": 1.340311050415039, | |
| "learning_rate": 0.00014184873949579834, | |
| "loss": 1.8468, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.2679139382600561, | |
| "grad_norm": 1.236281156539917, | |
| "learning_rate": 0.0001415126050420168, | |
| "loss": 1.834, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.2694106641721235, | |
| "grad_norm": 1.2765839099884033, | |
| "learning_rate": 0.0001411764705882353, | |
| "loss": 1.8231, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.27090739008419085, | |
| "grad_norm": 1.9402660131454468, | |
| "learning_rate": 0.0001408403361344538, | |
| "loss": 1.8436, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.2724041159962582, | |
| "grad_norm": 1.6397343873977661, | |
| "learning_rate": 0.00014050420168067225, | |
| "loss": 1.8738, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.27390084190832553, | |
| "grad_norm": 1.2269023656845093, | |
| "learning_rate": 0.00014016806722689075, | |
| "loss": 1.724, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.2753975678203929, | |
| "grad_norm": 1.0990972518920898, | |
| "learning_rate": 0.00013983193277310925, | |
| "loss": 1.805, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.2768942937324602, | |
| "grad_norm": 1.2719955444335938, | |
| "learning_rate": 0.00013949579831932772, | |
| "loss": 1.8694, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.2783910196445276, | |
| "grad_norm": 1.6118435859680176, | |
| "learning_rate": 0.00013915966386554622, | |
| "loss": 1.7453, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.27988774555659496, | |
| "grad_norm": 1.2474150657653809, | |
| "learning_rate": 0.00013882352941176472, | |
| "loss": 1.7035, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.2813844714686623, | |
| "grad_norm": 0.929045557975769, | |
| "learning_rate": 0.0001384873949579832, | |
| "loss": 1.4875, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.28288119738072964, | |
| "grad_norm": 1.1837550401687622, | |
| "learning_rate": 0.0001381512605042017, | |
| "loss": 1.8196, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.284377923292797, | |
| "grad_norm": 1.171769142150879, | |
| "learning_rate": 0.00013781512605042016, | |
| "loss": 1.8209, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.2858746492048644, | |
| "grad_norm": 1.0087103843688965, | |
| "learning_rate": 0.00013747899159663866, | |
| "loss": 1.8493, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.2873713751169317, | |
| "grad_norm": 1.0094703435897827, | |
| "learning_rate": 0.00013714285714285716, | |
| "loss": 1.8036, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.28886810102899907, | |
| "grad_norm": 1.2591369152069092, | |
| "learning_rate": 0.00013680672268907563, | |
| "loss": 1.6686, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.2903648269410664, | |
| "grad_norm": 1.1550267934799194, | |
| "learning_rate": 0.00013647058823529413, | |
| "loss": 1.7808, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.29186155285313375, | |
| "grad_norm": 1.575493335723877, | |
| "learning_rate": 0.0001361344537815126, | |
| "loss": 1.6981, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.2933582787652011, | |
| "grad_norm": 1.0852630138397217, | |
| "learning_rate": 0.0001357983193277311, | |
| "loss": 1.913, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.2948550046772685, | |
| "grad_norm": 1.4476265907287598, | |
| "learning_rate": 0.0001354621848739496, | |
| "loss": 1.8064, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.29635173058933584, | |
| "grad_norm": 1.1281249523162842, | |
| "learning_rate": 0.00013512605042016807, | |
| "loss": 1.9073, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.2978484565014032, | |
| "grad_norm": 1.011376142501831, | |
| "learning_rate": 0.00013478991596638657, | |
| "loss": 1.9038, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.2993451824134705, | |
| "grad_norm": 1.017004370689392, | |
| "learning_rate": 0.00013445378151260507, | |
| "loss": 1.842, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.2993451824134705, | |
| "eval_loss": 1.751793384552002, | |
| "eval_runtime": 9.9274, | |
| "eval_samples_per_second": 6.447, | |
| "eval_steps_per_second": 3.223, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.30084190832553787, | |
| "grad_norm": 1.280287504196167, | |
| "learning_rate": 0.00013411764705882352, | |
| "loss": 1.8463, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.30233863423760526, | |
| "grad_norm": 1.071548581123352, | |
| "learning_rate": 0.00013378151260504202, | |
| "loss": 1.936, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.3038353601496726, | |
| "grad_norm": 1.0402483940124512, | |
| "learning_rate": 0.00013344537815126052, | |
| "loss": 1.8919, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.30533208606173995, | |
| "grad_norm": 1.404093623161316, | |
| "learning_rate": 0.000133109243697479, | |
| "loss": 1.7126, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.3068288119738073, | |
| "grad_norm": 1.2647768259048462, | |
| "learning_rate": 0.0001327731092436975, | |
| "loss": 1.8001, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.30832553788587463, | |
| "grad_norm": 1.1771318912506104, | |
| "learning_rate": 0.00013243697478991596, | |
| "loss": 1.8045, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.309822263797942, | |
| "grad_norm": 1.1957385540008545, | |
| "learning_rate": 0.00013210084033613446, | |
| "loss": 1.7873, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.3113189897100094, | |
| "grad_norm": 1.4250292778015137, | |
| "learning_rate": 0.00013176470588235296, | |
| "loss": 1.5988, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.3128157156220767, | |
| "grad_norm": 1.1845389604568481, | |
| "learning_rate": 0.00013142857142857143, | |
| "loss": 1.7846, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.31431244153414406, | |
| "grad_norm": 1.080325722694397, | |
| "learning_rate": 0.00013109243697478993, | |
| "loss": 1.9488, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.3158091674462114, | |
| "grad_norm": 1.2887210845947266, | |
| "learning_rate": 0.0001307563025210084, | |
| "loss": 1.8574, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.31730589335827875, | |
| "grad_norm": 1.1449840068817139, | |
| "learning_rate": 0.0001304201680672269, | |
| "loss": 1.8628, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.31880261927034614, | |
| "grad_norm": 1.0899405479431152, | |
| "learning_rate": 0.0001300840336134454, | |
| "loss": 1.7998, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.3202993451824135, | |
| "grad_norm": 1.1760225296020508, | |
| "learning_rate": 0.00012974789915966387, | |
| "loss": 1.7833, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.32179607109448083, | |
| "grad_norm": 1.2030748128890991, | |
| "learning_rate": 0.00012941176470588237, | |
| "loss": 1.8504, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.32329279700654817, | |
| "grad_norm": 1.0713863372802734, | |
| "learning_rate": 0.00012907563025210087, | |
| "loss": 1.9432, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.3247895229186155, | |
| "grad_norm": 1.1058694124221802, | |
| "learning_rate": 0.00012873949579831934, | |
| "loss": 1.727, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.32628624883068286, | |
| "grad_norm": 1.1129230260849, | |
| "learning_rate": 0.00012840336134453784, | |
| "loss": 1.9422, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.32778297474275025, | |
| "grad_norm": 0.9841461181640625, | |
| "learning_rate": 0.0001280672268907563, | |
| "loss": 1.8916, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.3292797006548176, | |
| "grad_norm": 1.3188430070877075, | |
| "learning_rate": 0.00012773109243697478, | |
| "loss": 1.7102, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.33077642656688494, | |
| "grad_norm": 1.1673728227615356, | |
| "learning_rate": 0.00012739495798319328, | |
| "loss": 1.8466, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.3322731524789523, | |
| "grad_norm": 1.170258641242981, | |
| "learning_rate": 0.00012705882352941175, | |
| "loss": 1.8602, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.3337698783910196, | |
| "grad_norm": 1.1097817420959473, | |
| "learning_rate": 0.00012672268907563025, | |
| "loss": 1.912, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.335266604303087, | |
| "grad_norm": 1.0673434734344482, | |
| "learning_rate": 0.00012638655462184875, | |
| "loss": 1.808, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.33676333021515437, | |
| "grad_norm": 1.311902642250061, | |
| "learning_rate": 0.00012605042016806722, | |
| "loss": 1.8527, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.33676333021515437, | |
| "eval_loss": 1.7518868446350098, | |
| "eval_runtime": 9.8911, | |
| "eval_samples_per_second": 6.47, | |
| "eval_steps_per_second": 3.235, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.3382600561272217, | |
| "grad_norm": 1.0087488889694214, | |
| "learning_rate": 0.00012571428571428572, | |
| "loss": 1.9122, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.33975678203928905, | |
| "grad_norm": 1.2666288614273071, | |
| "learning_rate": 0.0001253781512605042, | |
| "loss": 1.7594, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.3412535079513564, | |
| "grad_norm": 1.1287747621536255, | |
| "learning_rate": 0.0001250420168067227, | |
| "loss": 1.8053, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.34275023386342374, | |
| "grad_norm": 1.206766128540039, | |
| "learning_rate": 0.0001247058823529412, | |
| "loss": 1.9256, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.34424695977549113, | |
| "grad_norm": 1.4732266664505005, | |
| "learning_rate": 0.00012436974789915966, | |
| "loss": 1.794, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.3457436856875585, | |
| "grad_norm": 1.1438567638397217, | |
| "learning_rate": 0.00012403361344537816, | |
| "loss": 1.8223, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.3472404115996258, | |
| "grad_norm": 1.1886340379714966, | |
| "learning_rate": 0.00012369747899159666, | |
| "loss": 1.7702, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.34873713751169316, | |
| "grad_norm": 1.4150899648666382, | |
| "learning_rate": 0.00012336134453781513, | |
| "loss": 1.7094, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.3502338634237605, | |
| "grad_norm": 1.446444034576416, | |
| "learning_rate": 0.00012302521008403363, | |
| "loss": 1.7575, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.3517305893358279, | |
| "grad_norm": 1.254396677017212, | |
| "learning_rate": 0.0001226890756302521, | |
| "loss": 1.6884, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.35322731524789525, | |
| "grad_norm": 1.2610015869140625, | |
| "learning_rate": 0.0001223529411764706, | |
| "loss": 1.8701, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.3547240411599626, | |
| "grad_norm": 0.932133138179779, | |
| "learning_rate": 0.00012201680672268909, | |
| "loss": 1.8865, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.35622076707202993, | |
| "grad_norm": 1.4056602716445923, | |
| "learning_rate": 0.00012168067226890756, | |
| "loss": 1.788, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.3577174929840973, | |
| "grad_norm": 1.2466380596160889, | |
| "learning_rate": 0.00012134453781512605, | |
| "loss": 1.8152, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.3592142188961646, | |
| "grad_norm": 1.1610888242721558, | |
| "learning_rate": 0.00012100840336134453, | |
| "loss": 1.8782, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.360710944808232, | |
| "grad_norm": 1.1852293014526367, | |
| "learning_rate": 0.00012067226890756302, | |
| "loss": 1.7942, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.36220767072029936, | |
| "grad_norm": 1.1470458507537842, | |
| "learning_rate": 0.00012033613445378152, | |
| "loss": 1.9395, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.3637043966323667, | |
| "grad_norm": 1.1640129089355469, | |
| "learning_rate": 0.00012, | |
| "loss": 1.8221, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.36520112254443404, | |
| "grad_norm": 1.7140247821807861, | |
| "learning_rate": 0.00011966386554621849, | |
| "loss": 1.8398, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.3666978484565014, | |
| "grad_norm": 1.1535826921463013, | |
| "learning_rate": 0.00011932773109243697, | |
| "loss": 1.7175, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.3681945743685688, | |
| "grad_norm": 1.1240558624267578, | |
| "learning_rate": 0.00011899159663865547, | |
| "loss": 1.8434, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.3696913002806361, | |
| "grad_norm": 1.2826379537582397, | |
| "learning_rate": 0.00011865546218487396, | |
| "loss": 1.7932, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.37118802619270347, | |
| "grad_norm": 1.423509955406189, | |
| "learning_rate": 0.00011831932773109244, | |
| "loss": 1.851, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.3726847521047708, | |
| "grad_norm": 1.0698450803756714, | |
| "learning_rate": 0.00011798319327731093, | |
| "loss": 1.8628, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.37418147801683815, | |
| "grad_norm": 1.2364152669906616, | |
| "learning_rate": 0.00011764705882352942, | |
| "loss": 1.8172, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.37418147801683815, | |
| "eval_loss": 1.7512738704681396, | |
| "eval_runtime": 9.9241, | |
| "eval_samples_per_second": 6.449, | |
| "eval_steps_per_second": 3.224, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.3756782039289055, | |
| "grad_norm": 1.073832392692566, | |
| "learning_rate": 0.00011731092436974791, | |
| "loss": 1.7081, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.3771749298409729, | |
| "grad_norm": 1.0931236743927002, | |
| "learning_rate": 0.0001169747899159664, | |
| "loss": 1.8104, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.37867165575304024, | |
| "grad_norm": 1.0120686292648315, | |
| "learning_rate": 0.00011663865546218489, | |
| "loss": 1.9482, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.3801683816651076, | |
| "grad_norm": 1.493310570716858, | |
| "learning_rate": 0.00011630252100840337, | |
| "loss": 1.7734, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.3816651075771749, | |
| "grad_norm": 1.1117216348648071, | |
| "learning_rate": 0.00011596638655462187, | |
| "loss": 1.8592, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.38316183348924227, | |
| "grad_norm": 1.1169261932373047, | |
| "learning_rate": 0.00011563025210084036, | |
| "loss": 1.7698, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.3846585594013096, | |
| "grad_norm": 1.1613709926605225, | |
| "learning_rate": 0.00011529411764705881, | |
| "loss": 1.8207, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.386155285313377, | |
| "grad_norm": 1.2825706005096436, | |
| "learning_rate": 0.00011495798319327731, | |
| "loss": 1.822, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.38765201122544435, | |
| "grad_norm": 0.9785429835319519, | |
| "learning_rate": 0.0001146218487394958, | |
| "loss": 1.7244, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.3891487371375117, | |
| "grad_norm": 1.177465558052063, | |
| "learning_rate": 0.00011428571428571428, | |
| "loss": 1.8048, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.39064546304957903, | |
| "grad_norm": 1.3732517957687378, | |
| "learning_rate": 0.00011394957983193277, | |
| "loss": 1.752, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.3921421889616464, | |
| "grad_norm": 1.217416763305664, | |
| "learning_rate": 0.00011361344537815127, | |
| "loss": 1.8383, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.3936389148737138, | |
| "grad_norm": 1.1581103801727295, | |
| "learning_rate": 0.00011327731092436975, | |
| "loss": 1.7979, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.3951356407857811, | |
| "grad_norm": 1.0889670848846436, | |
| "learning_rate": 0.00011294117647058824, | |
| "loss": 1.8704, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.39663236669784846, | |
| "grad_norm": 1.3366667032241821, | |
| "learning_rate": 0.00011260504201680672, | |
| "loss": 1.8054, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.3981290926099158, | |
| "grad_norm": 1.2124860286712646, | |
| "learning_rate": 0.00011226890756302521, | |
| "loss": 1.8366, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.39962581852198314, | |
| "grad_norm": 1.3980365991592407, | |
| "learning_rate": 0.00011193277310924371, | |
| "loss": 1.7174, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.4011225444340505, | |
| "grad_norm": 1.1408543586730957, | |
| "learning_rate": 0.0001115966386554622, | |
| "loss": 1.7571, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.4026192703461179, | |
| "grad_norm": 1.1843181848526, | |
| "learning_rate": 0.00011126050420168068, | |
| "loss": 1.8332, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.40411599625818523, | |
| "grad_norm": 1.4148800373077393, | |
| "learning_rate": 0.00011092436974789917, | |
| "loss": 1.9739, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.40561272217025257, | |
| "grad_norm": 1.0253487825393677, | |
| "learning_rate": 0.00011058823529411766, | |
| "loss": 1.8726, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.4071094480823199, | |
| "grad_norm": 1.3067514896392822, | |
| "learning_rate": 0.00011025210084033615, | |
| "loss": 1.6941, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.40860617399438726, | |
| "grad_norm": 1.0671433210372925, | |
| "learning_rate": 0.00010991596638655464, | |
| "loss": 1.7697, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.41010289990645465, | |
| "grad_norm": 1.3519949913024902, | |
| "learning_rate": 0.00010957983193277312, | |
| "loss": 1.7855, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.411599625818522, | |
| "grad_norm": 1.076204538345337, | |
| "learning_rate": 0.00010924369747899159, | |
| "loss": 1.9365, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.411599625818522, | |
| "eval_loss": 1.7478337287902832, | |
| "eval_runtime": 9.905, | |
| "eval_samples_per_second": 6.461, | |
| "eval_steps_per_second": 3.231, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.41309635173058934, | |
| "grad_norm": 1.2591997385025024, | |
| "learning_rate": 0.00010890756302521008, | |
| "loss": 1.7309, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.4145930776426567, | |
| "grad_norm": 1.340928554534912, | |
| "learning_rate": 0.00010857142857142856, | |
| "loss": 1.7348, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.416089803554724, | |
| "grad_norm": 0.9326046109199524, | |
| "learning_rate": 0.00010823529411764706, | |
| "loss": 1.8101, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.41758652946679137, | |
| "grad_norm": 1.3383642435073853, | |
| "learning_rate": 0.00010789915966386555, | |
| "loss": 1.7476, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.41908325537885877, | |
| "grad_norm": 0.9598699808120728, | |
| "learning_rate": 0.00010756302521008403, | |
| "loss": 1.6919, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.4205799812909261, | |
| "grad_norm": 1.019420862197876, | |
| "learning_rate": 0.00010722689075630252, | |
| "loss": 1.8148, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.42207670720299345, | |
| "grad_norm": 1.142913579940796, | |
| "learning_rate": 0.000106890756302521, | |
| "loss": 1.8095, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.4235734331150608, | |
| "grad_norm": 1.3397822380065918, | |
| "learning_rate": 0.0001065546218487395, | |
| "loss": 1.7827, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.42507015902712814, | |
| "grad_norm": 1.3161386251449585, | |
| "learning_rate": 0.00010621848739495799, | |
| "loss": 1.8575, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.42656688493919553, | |
| "grad_norm": 1.1929869651794434, | |
| "learning_rate": 0.00010588235294117647, | |
| "loss": 1.79, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.4280636108512629, | |
| "grad_norm": 1.2960401773452759, | |
| "learning_rate": 0.00010554621848739496, | |
| "loss": 1.7203, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.4295603367633302, | |
| "grad_norm": 1.117655873298645, | |
| "learning_rate": 0.00010521008403361346, | |
| "loss": 1.8556, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.43105706267539756, | |
| "grad_norm": 1.117436408996582, | |
| "learning_rate": 0.00010487394957983194, | |
| "loss": 1.7695, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.4325537885874649, | |
| "grad_norm": 1.0667588710784912, | |
| "learning_rate": 0.00010453781512605043, | |
| "loss": 1.8246, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.43405051449953225, | |
| "grad_norm": 1.0113589763641357, | |
| "learning_rate": 0.00010420168067226892, | |
| "loss": 1.9029, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.43554724041159965, | |
| "grad_norm": 1.0438803434371948, | |
| "learning_rate": 0.00010386554621848741, | |
| "loss": 1.8053, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.437043966323667, | |
| "grad_norm": 2.1361262798309326, | |
| "learning_rate": 0.0001035294117647059, | |
| "loss": 1.7788, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.43854069223573433, | |
| "grad_norm": 1.2499916553497314, | |
| "learning_rate": 0.00010319327731092439, | |
| "loss": 1.745, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.4400374181478017, | |
| "grad_norm": 1.10703444480896, | |
| "learning_rate": 0.00010285714285714286, | |
| "loss": 1.8723, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.441534144059869, | |
| "grad_norm": 0.9382535815238953, | |
| "learning_rate": 0.00010252100840336134, | |
| "loss": 1.8814, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.4430308699719364, | |
| "grad_norm": 1.1860700845718384, | |
| "learning_rate": 0.00010218487394957983, | |
| "loss": 1.7975, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.44452759588400376, | |
| "grad_norm": 1.0663989782333374, | |
| "learning_rate": 0.00010184873949579831, | |
| "loss": 1.8333, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.4460243217960711, | |
| "grad_norm": 1.0662665367126465, | |
| "learning_rate": 0.0001015126050420168, | |
| "loss": 1.7244, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.44752104770813844, | |
| "grad_norm": 1.1393229961395264, | |
| "learning_rate": 0.0001011764705882353, | |
| "loss": 1.7832, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.4490177736202058, | |
| "grad_norm": 1.1080511808395386, | |
| "learning_rate": 0.00010084033613445378, | |
| "loss": 1.9043, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.4490177736202058, | |
| "eval_loss": 1.7416434288024902, | |
| "eval_runtime": 9.9207, | |
| "eval_samples_per_second": 6.451, | |
| "eval_steps_per_second": 3.226, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.4505144995322731, | |
| "grad_norm": 1.2179614305496216, | |
| "learning_rate": 0.00010050420168067227, | |
| "loss": 1.7529, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.4520112254443405, | |
| "grad_norm": 1.0161402225494385, | |
| "learning_rate": 0.00010016806722689076, | |
| "loss": 1.9195, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.45350795135640787, | |
| "grad_norm": 0.995812714099884, | |
| "learning_rate": 9.983193277310925e-05, | |
| "loss": 1.8102, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.4550046772684752, | |
| "grad_norm": 1.1934641599655151, | |
| "learning_rate": 9.949579831932774e-05, | |
| "loss": 1.7814, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.45650140318054255, | |
| "grad_norm": 0.8843271732330322, | |
| "learning_rate": 9.915966386554623e-05, | |
| "loss": 1.6012, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.4579981290926099, | |
| "grad_norm": 1.0673537254333496, | |
| "learning_rate": 9.882352941176471e-05, | |
| "loss": 1.8338, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.4594948550046773, | |
| "grad_norm": 1.007251262664795, | |
| "learning_rate": 9.848739495798321e-05, | |
| "loss": 1.9598, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.46099158091674464, | |
| "grad_norm": 1.111372470855713, | |
| "learning_rate": 9.815126050420168e-05, | |
| "loss": 1.8086, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.462488306828812, | |
| "grad_norm": 1.2300423383712769, | |
| "learning_rate": 9.781512605042017e-05, | |
| "loss": 1.8517, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.4639850327408793, | |
| "grad_norm": 1.034952998161316, | |
| "learning_rate": 9.747899159663865e-05, | |
| "loss": 1.8743, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.46548175865294666, | |
| "grad_norm": 1.1956812143325806, | |
| "learning_rate": 9.714285714285715e-05, | |
| "loss": 1.8484, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.466978484565014, | |
| "grad_norm": 1.0862494707107544, | |
| "learning_rate": 9.680672268907564e-05, | |
| "loss": 1.8167, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.4684752104770814, | |
| "grad_norm": 1.0881495475769043, | |
| "learning_rate": 9.647058823529412e-05, | |
| "loss": 1.7105, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.46997193638914875, | |
| "grad_norm": 1.1325311660766602, | |
| "learning_rate": 9.613445378151261e-05, | |
| "loss": 1.7793, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.4714686623012161, | |
| "grad_norm": 1.0453370809555054, | |
| "learning_rate": 9.579831932773111e-05, | |
| "loss": 1.8046, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.47296538821328343, | |
| "grad_norm": 1.127502679824829, | |
| "learning_rate": 9.546218487394959e-05, | |
| "loss": 1.8062, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.4744621141253508, | |
| "grad_norm": 1.0815576314926147, | |
| "learning_rate": 9.512605042016806e-05, | |
| "loss": 1.8572, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.4759588400374182, | |
| "grad_norm": 1.1076608896255493, | |
| "learning_rate": 9.478991596638655e-05, | |
| "loss": 1.7315, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.4774555659494855, | |
| "grad_norm": 1.1549115180969238, | |
| "learning_rate": 9.445378151260505e-05, | |
| "loss": 1.749, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.47895229186155286, | |
| "grad_norm": 1.0027329921722412, | |
| "learning_rate": 9.411764705882353e-05, | |
| "loss": 1.9418, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.4804490177736202, | |
| "grad_norm": 1.1883653402328491, | |
| "learning_rate": 9.378151260504202e-05, | |
| "loss": 1.7007, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.48194574368568754, | |
| "grad_norm": 1.0235621929168701, | |
| "learning_rate": 9.34453781512605e-05, | |
| "loss": 1.7951, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.4834424695977549, | |
| "grad_norm": 1.3429903984069824, | |
| "learning_rate": 9.3109243697479e-05, | |
| "loss": 1.8396, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.4849391955098223, | |
| "grad_norm": 1.2389410734176636, | |
| "learning_rate": 9.277310924369749e-05, | |
| "loss": 1.8577, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.4864359214218896, | |
| "grad_norm": 1.1658669710159302, | |
| "learning_rate": 9.243697478991598e-05, | |
| "loss": 1.7733, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.4864359214218896, | |
| "eval_loss": 1.7379932403564453, | |
| "eval_runtime": 9.9225, | |
| "eval_samples_per_second": 6.45, | |
| "eval_steps_per_second": 3.225, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.48793264733395697, | |
| "grad_norm": 1.2423300743103027, | |
| "learning_rate": 9.210084033613445e-05, | |
| "loss": 1.7292, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.4894293732460243, | |
| "grad_norm": 1.0733031034469604, | |
| "learning_rate": 9.176470588235295e-05, | |
| "loss": 1.8459, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.49092609915809166, | |
| "grad_norm": 1.0904580354690552, | |
| "learning_rate": 9.142857142857143e-05, | |
| "loss": 1.7226, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.49242282507015905, | |
| "grad_norm": 1.271660327911377, | |
| "learning_rate": 9.109243697478992e-05, | |
| "loss": 1.7913, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.4939195509822264, | |
| "grad_norm": 1.0605442523956299, | |
| "learning_rate": 9.07563025210084e-05, | |
| "loss": 1.8367, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.49541627689429374, | |
| "grad_norm": 1.0855785608291626, | |
| "learning_rate": 9.04201680672269e-05, | |
| "loss": 1.7726, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.4969130028063611, | |
| "grad_norm": 1.0943794250488281, | |
| "learning_rate": 9.008403361344539e-05, | |
| "loss": 1.8692, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.4984097287184284, | |
| "grad_norm": 1.1238775253295898, | |
| "learning_rate": 8.974789915966387e-05, | |
| "loss": 1.6684, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.49990645463049577, | |
| "grad_norm": 1.1262332201004028, | |
| "learning_rate": 8.941176470588236e-05, | |
| "loss": 2.0474, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.5014031805425632, | |
| "grad_norm": 1.1114274263381958, | |
| "learning_rate": 8.907563025210084e-05, | |
| "loss": 1.8302, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.5028999064546305, | |
| "grad_norm": 1.2173712253570557, | |
| "learning_rate": 8.873949579831933e-05, | |
| "loss": 1.7415, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.5043966323666979, | |
| "grad_norm": 1.0768870115280151, | |
| "learning_rate": 8.840336134453782e-05, | |
| "loss": 1.8232, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.5058933582787652, | |
| "grad_norm": 1.3007467985153198, | |
| "learning_rate": 8.80672268907563e-05, | |
| "loss": 1.8447, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.5073900841908325, | |
| "grad_norm": 1.1609522104263306, | |
| "learning_rate": 8.77310924369748e-05, | |
| "loss": 1.7427, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.5088868101028999, | |
| "grad_norm": 1.3535274267196655, | |
| "learning_rate": 8.739495798319329e-05, | |
| "loss": 1.748, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.5103835360149672, | |
| "grad_norm": 1.132091999053955, | |
| "learning_rate": 8.705882352941177e-05, | |
| "loss": 1.8139, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.5118802619270346, | |
| "grad_norm": 1.1243830919265747, | |
| "learning_rate": 8.672268907563026e-05, | |
| "loss": 1.7954, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.513376987839102, | |
| "grad_norm": 1.2013453245162964, | |
| "learning_rate": 8.638655462184874e-05, | |
| "loss": 1.7759, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.5148737137511693, | |
| "grad_norm": 1.1899113655090332, | |
| "learning_rate": 8.605042016806724e-05, | |
| "loss": 1.9394, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.5163704396632367, | |
| "grad_norm": 1.1117327213287354, | |
| "learning_rate": 8.571428571428571e-05, | |
| "loss": 1.7475, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.517867165575304, | |
| "grad_norm": 0.9757189154624939, | |
| "learning_rate": 8.53781512605042e-05, | |
| "loss": 1.8084, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.5193638914873714, | |
| "grad_norm": 1.1269667148590088, | |
| "learning_rate": 8.50420168067227e-05, | |
| "loss": 1.872, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.5208606173994387, | |
| "grad_norm": 1.0271408557891846, | |
| "learning_rate": 8.470588235294118e-05, | |
| "loss": 1.8508, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.5223573433115061, | |
| "grad_norm": 1.1958681344985962, | |
| "learning_rate": 8.436974789915967e-05, | |
| "loss": 1.7609, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.5238540692235735, | |
| "grad_norm": 1.1345899105072021, | |
| "learning_rate": 8.403361344537815e-05, | |
| "loss": 1.7925, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.5238540692235735, | |
| "eval_loss": 1.7380330562591553, | |
| "eval_runtime": 9.922, | |
| "eval_samples_per_second": 6.45, | |
| "eval_steps_per_second": 3.225, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.5253507951356408, | |
| "grad_norm": 1.1461538076400757, | |
| "learning_rate": 8.369747899159664e-05, | |
| "loss": 1.7498, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.5268475210477082, | |
| "grad_norm": 1.1772956848144531, | |
| "learning_rate": 8.336134453781514e-05, | |
| "loss": 1.8021, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.5283442469597754, | |
| "grad_norm": 1.058077335357666, | |
| "learning_rate": 8.302521008403362e-05, | |
| "loss": 1.8782, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.5298409728718428, | |
| "grad_norm": 1.0576335191726685, | |
| "learning_rate": 8.26890756302521e-05, | |
| "loss": 1.917, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.5313376987839102, | |
| "grad_norm": 1.3313714265823364, | |
| "learning_rate": 8.23529411764706e-05, | |
| "loss": 1.5589, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.5328344246959775, | |
| "grad_norm": 1.1257191896438599, | |
| "learning_rate": 8.201680672268908e-05, | |
| "loss": 1.8366, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.5343311506080449, | |
| "grad_norm": 1.1489942073822021, | |
| "learning_rate": 8.168067226890757e-05, | |
| "loss": 1.7436, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.5358278765201122, | |
| "grad_norm": 1.3241493701934814, | |
| "learning_rate": 8.134453781512605e-05, | |
| "loss": 1.6859, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.5373246024321796, | |
| "grad_norm": 1.0549739599227905, | |
| "learning_rate": 8.100840336134454e-05, | |
| "loss": 1.7418, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.538821328344247, | |
| "grad_norm": 1.207139015197754, | |
| "learning_rate": 8.067226890756304e-05, | |
| "loss": 1.8467, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.5403180542563143, | |
| "grad_norm": 0.9392403364181519, | |
| "learning_rate": 8.033613445378152e-05, | |
| "loss": 1.8414, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.5418147801683817, | |
| "grad_norm": 1.0758482217788696, | |
| "learning_rate": 8e-05, | |
| "loss": 1.8437, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.543311506080449, | |
| "grad_norm": 1.3484997749328613, | |
| "learning_rate": 7.966386554621849e-05, | |
| "loss": 1.7959, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.5448082319925164, | |
| "grad_norm": 1.1805089712142944, | |
| "learning_rate": 7.932773109243698e-05, | |
| "loss": 1.9371, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.5463049579045838, | |
| "grad_norm": 1.1756744384765625, | |
| "learning_rate": 7.899159663865546e-05, | |
| "loss": 1.7701, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.5478016838166511, | |
| "grad_norm": 1.1375999450683594, | |
| "learning_rate": 7.865546218487395e-05, | |
| "loss": 1.6643, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.5492984097287185, | |
| "grad_norm": 1.069983720779419, | |
| "learning_rate": 7.831932773109243e-05, | |
| "loss": 1.8322, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.5507951356407857, | |
| "grad_norm": 1.0515434741973877, | |
| "learning_rate": 7.798319327731093e-05, | |
| "loss": 1.7744, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.5522918615528531, | |
| "grad_norm": 1.0633738040924072, | |
| "learning_rate": 7.764705882352942e-05, | |
| "loss": 1.8109, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.5537885874649204, | |
| "grad_norm": 1.359668254852295, | |
| "learning_rate": 7.73109243697479e-05, | |
| "loss": 1.7944, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.5552853133769878, | |
| "grad_norm": 1.2288023233413696, | |
| "learning_rate": 7.697478991596639e-05, | |
| "loss": 1.7918, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.5567820392890552, | |
| "grad_norm": 1.1025428771972656, | |
| "learning_rate": 7.663865546218489e-05, | |
| "loss": 1.823, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.5582787652011225, | |
| "grad_norm": 1.005429744720459, | |
| "learning_rate": 7.630252100840336e-05, | |
| "loss": 1.8222, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.5597754911131899, | |
| "grad_norm": 1.2293034791946411, | |
| "learning_rate": 7.596638655462185e-05, | |
| "loss": 1.8342, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.5612722170252572, | |
| "grad_norm": 1.0827324390411377, | |
| "learning_rate": 7.563025210084033e-05, | |
| "loss": 1.643, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.5612722170252572, | |
| "eval_loss": 1.7346194982528687, | |
| "eval_runtime": 9.9111, | |
| "eval_samples_per_second": 6.457, | |
| "eval_steps_per_second": 3.229, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.5627689429373246, | |
| "grad_norm": 1.0979185104370117, | |
| "learning_rate": 7.529411764705883e-05, | |
| "loss": 1.792, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.564265668849392, | |
| "grad_norm": 1.0999720096588135, | |
| "learning_rate": 7.495798319327732e-05, | |
| "loss": 1.75, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.5657623947614593, | |
| "grad_norm": 1.1695318222045898, | |
| "learning_rate": 7.46218487394958e-05, | |
| "loss": 1.7434, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.5672591206735267, | |
| "grad_norm": 1.452539324760437, | |
| "learning_rate": 7.428571428571429e-05, | |
| "loss": 1.8034, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.568755846585594, | |
| "grad_norm": 0.971097469329834, | |
| "learning_rate": 7.394957983193279e-05, | |
| "loss": 1.9281, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.5702525724976614, | |
| "grad_norm": 1.190000057220459, | |
| "learning_rate": 7.361344537815127e-05, | |
| "loss": 1.8709, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.5717492984097288, | |
| "grad_norm": 1.2394530773162842, | |
| "learning_rate": 7.327731092436974e-05, | |
| "loss": 1.8645, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.573246024321796, | |
| "grad_norm": 1.0058298110961914, | |
| "learning_rate": 7.294117647058823e-05, | |
| "loss": 1.7103, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.5747427502338635, | |
| "grad_norm": 0.9850262999534607, | |
| "learning_rate": 7.260504201680673e-05, | |
| "loss": 1.8576, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.5762394761459307, | |
| "grad_norm": 0.9836248159408569, | |
| "learning_rate": 7.226890756302521e-05, | |
| "loss": 1.8989, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.5777362020579981, | |
| "grad_norm": 0.9555202126502991, | |
| "learning_rate": 7.19327731092437e-05, | |
| "loss": 1.9098, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.5792329279700655, | |
| "grad_norm": 1.2133311033248901, | |
| "learning_rate": 7.159663865546218e-05, | |
| "loss": 1.719, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.5807296538821328, | |
| "grad_norm": 1.0879029035568237, | |
| "learning_rate": 7.126050420168068e-05, | |
| "loss": 1.7404, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.5822263797942002, | |
| "grad_norm": 1.1183611154556274, | |
| "learning_rate": 7.092436974789917e-05, | |
| "loss": 1.7065, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.5837231057062675, | |
| "grad_norm": 1.1441535949707031, | |
| "learning_rate": 7.058823529411765e-05, | |
| "loss": 1.876, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.5852198316183349, | |
| "grad_norm": 1.0589473247528076, | |
| "learning_rate": 7.025210084033613e-05, | |
| "loss": 1.7505, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.5867165575304022, | |
| "grad_norm": 1.149380087852478, | |
| "learning_rate": 6.991596638655463e-05, | |
| "loss": 1.7483, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.5882132834424696, | |
| "grad_norm": 1.18876314163208, | |
| "learning_rate": 6.957983193277311e-05, | |
| "loss": 1.6457, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.589710009354537, | |
| "grad_norm": 1.4560132026672363, | |
| "learning_rate": 6.92436974789916e-05, | |
| "loss": 1.747, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.5912067352666043, | |
| "grad_norm": 1.0865813493728638, | |
| "learning_rate": 6.890756302521008e-05, | |
| "loss": 1.7372, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.5927034611786717, | |
| "grad_norm": 1.1931579113006592, | |
| "learning_rate": 6.857142857142858e-05, | |
| "loss": 1.9051, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.594200187090739, | |
| "grad_norm": 1.0811692476272583, | |
| "learning_rate": 6.823529411764707e-05, | |
| "loss": 1.8224, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.5956969130028064, | |
| "grad_norm": 1.2077957391738892, | |
| "learning_rate": 6.789915966386555e-05, | |
| "loss": 1.8336, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.5971936389148738, | |
| "grad_norm": 1.0455660820007324, | |
| "learning_rate": 6.756302521008404e-05, | |
| "loss": 1.797, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.598690364826941, | |
| "grad_norm": 1.0006000995635986, | |
| "learning_rate": 6.722689075630254e-05, | |
| "loss": 1.7179, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.598690364826941, | |
| "eval_loss": 1.7308931350708008, | |
| "eval_runtime": 9.9241, | |
| "eval_samples_per_second": 6.449, | |
| "eval_steps_per_second": 3.224, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.6001870907390084, | |
| "grad_norm": 1.143531322479248, | |
| "learning_rate": 6.689075630252101e-05, | |
| "loss": 1.8215, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.6016838166510757, | |
| "grad_norm": 0.9073050618171692, | |
| "learning_rate": 6.65546218487395e-05, | |
| "loss": 1.9348, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.6031805425631431, | |
| "grad_norm": 0.994017481803894, | |
| "learning_rate": 6.621848739495798e-05, | |
| "loss": 1.5425, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.6046772684752105, | |
| "grad_norm": 1.1705002784729004, | |
| "learning_rate": 6.588235294117648e-05, | |
| "loss": 1.8536, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.6061739943872778, | |
| "grad_norm": 0.9837265610694885, | |
| "learning_rate": 6.554621848739496e-05, | |
| "loss": 1.7408, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.6076707202993452, | |
| "grad_norm": 0.9924471378326416, | |
| "learning_rate": 6.521008403361345e-05, | |
| "loss": 1.7051, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.6091674462114125, | |
| "grad_norm": 1.0559359788894653, | |
| "learning_rate": 6.487394957983193e-05, | |
| "loss": 1.9459, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.6106641721234799, | |
| "grad_norm": 1.2436074018478394, | |
| "learning_rate": 6.453781512605043e-05, | |
| "loss": 1.7981, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.6121608980355472, | |
| "grad_norm": 1.278940200805664, | |
| "learning_rate": 6.420168067226892e-05, | |
| "loss": 1.7406, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.6136576239476146, | |
| "grad_norm": 1.1248103380203247, | |
| "learning_rate": 6.386554621848739e-05, | |
| "loss": 1.7909, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.615154349859682, | |
| "grad_norm": 1.1026619672775269, | |
| "learning_rate": 6.352941176470588e-05, | |
| "loss": 1.8588, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.6166510757717493, | |
| "grad_norm": 1.1886570453643799, | |
| "learning_rate": 6.319327731092438e-05, | |
| "loss": 1.7935, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.6181478016838167, | |
| "grad_norm": 1.0327427387237549, | |
| "learning_rate": 6.285714285714286e-05, | |
| "loss": 1.8984, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.619644527595884, | |
| "grad_norm": 0.9914230108261108, | |
| "learning_rate": 6.252100840336135e-05, | |
| "loss": 1.9701, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.6211412535079514, | |
| "grad_norm": 0.9550872445106506, | |
| "learning_rate": 6.218487394957983e-05, | |
| "loss": 1.7874, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.6226379794200188, | |
| "grad_norm": 1.10657799243927, | |
| "learning_rate": 6.184873949579833e-05, | |
| "loss": 1.7691, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.624134705332086, | |
| "grad_norm": 0.956917405128479, | |
| "learning_rate": 6.151260504201682e-05, | |
| "loss": 1.7304, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.6256314312441534, | |
| "grad_norm": 1.0174245834350586, | |
| "learning_rate": 6.11764705882353e-05, | |
| "loss": 1.8446, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.6271281571562207, | |
| "grad_norm": 1.321598768234253, | |
| "learning_rate": 6.084033613445378e-05, | |
| "loss": 1.5281, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.6286248830682881, | |
| "grad_norm": 0.977022111415863, | |
| "learning_rate": 6.0504201680672267e-05, | |
| "loss": 1.8618, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.6301216089803555, | |
| "grad_norm": 1.2729384899139404, | |
| "learning_rate": 6.016806722689076e-05, | |
| "loss": 1.6538, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.6316183348924228, | |
| "grad_norm": 0.982009768486023, | |
| "learning_rate": 5.9831932773109244e-05, | |
| "loss": 1.8753, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.6331150608044902, | |
| "grad_norm": 1.1528300046920776, | |
| "learning_rate": 5.9495798319327737e-05, | |
| "loss": 1.8639, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.6346117867165575, | |
| "grad_norm": 1.096552848815918, | |
| "learning_rate": 5.915966386554622e-05, | |
| "loss": 1.7805, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.6361085126286249, | |
| "grad_norm": 0.9799396991729736, | |
| "learning_rate": 5.882352941176471e-05, | |
| "loss": 1.9097, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.6361085126286249, | |
| "eval_loss": 1.7265229225158691, | |
| "eval_runtime": 9.9192, | |
| "eval_samples_per_second": 6.452, | |
| "eval_steps_per_second": 3.226, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.6376052385406923, | |
| "grad_norm": 0.9917983412742615, | |
| "learning_rate": 5.84873949579832e-05, | |
| "loss": 1.8196, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.6391019644527596, | |
| "grad_norm": 0.9843238592147827, | |
| "learning_rate": 5.8151260504201685e-05, | |
| "loss": 1.8505, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.640598690364827, | |
| "grad_norm": 1.1680494546890259, | |
| "learning_rate": 5.781512605042018e-05, | |
| "loss": 1.7219, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.6420954162768943, | |
| "grad_norm": 1.100326657295227, | |
| "learning_rate": 5.7478991596638656e-05, | |
| "loss": 1.8383, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.6435921421889617, | |
| "grad_norm": 1.1734800338745117, | |
| "learning_rate": 5.714285714285714e-05, | |
| "loss": 1.7812, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.645088868101029, | |
| "grad_norm": 1.129654049873352, | |
| "learning_rate": 5.6806722689075634e-05, | |
| "loss": 1.6585, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.6465855940130963, | |
| "grad_norm": 0.925308108329773, | |
| "learning_rate": 5.647058823529412e-05, | |
| "loss": 2.008, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.6480823199251637, | |
| "grad_norm": 1.2162028551101685, | |
| "learning_rate": 5.6134453781512605e-05, | |
| "loss": 1.7252, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.649579045837231, | |
| "grad_norm": 1.0221952199935913, | |
| "learning_rate": 5.57983193277311e-05, | |
| "loss": 1.8121, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.6510757717492984, | |
| "grad_norm": 1.125623106956482, | |
| "learning_rate": 5.546218487394958e-05, | |
| "loss": 1.7751, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.6525724976613657, | |
| "grad_norm": 1.0839701890945435, | |
| "learning_rate": 5.5126050420168075e-05, | |
| "loss": 1.7789, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.6540692235734331, | |
| "grad_norm": 1.1495144367218018, | |
| "learning_rate": 5.478991596638656e-05, | |
| "loss": 1.7065, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.6555659494855005, | |
| "grad_norm": 1.1935160160064697, | |
| "learning_rate": 5.445378151260504e-05, | |
| "loss": 1.6623, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.6570626753975678, | |
| "grad_norm": 1.0761985778808594, | |
| "learning_rate": 5.411764705882353e-05, | |
| "loss": 1.7685, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.6585594013096352, | |
| "grad_norm": 1.2685760259628296, | |
| "learning_rate": 5.378151260504202e-05, | |
| "loss": 1.6965, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.6600561272217025, | |
| "grad_norm": 1.3311073780059814, | |
| "learning_rate": 5.34453781512605e-05, | |
| "loss": 1.7738, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.6615528531337699, | |
| "grad_norm": 1.2063039541244507, | |
| "learning_rate": 5.3109243697478995e-05, | |
| "loss": 1.8287, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.6630495790458373, | |
| "grad_norm": 0.8949533104896545, | |
| "learning_rate": 5.277310924369748e-05, | |
| "loss": 1.7777, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.6645463049579046, | |
| "grad_norm": 0.9890621900558472, | |
| "learning_rate": 5.243697478991597e-05, | |
| "loss": 1.8908, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.666043030869972, | |
| "grad_norm": 1.141076683998108, | |
| "learning_rate": 5.210084033613446e-05, | |
| "loss": 1.7678, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.6675397567820393, | |
| "grad_norm": 1.4433619976043701, | |
| "learning_rate": 5.176470588235295e-05, | |
| "loss": 1.7579, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.6690364826941066, | |
| "grad_norm": 1.088409423828125, | |
| "learning_rate": 5.142857142857143e-05, | |
| "loss": 1.7632, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.670533208606174, | |
| "grad_norm": 0.9711755514144897, | |
| "learning_rate": 5.1092436974789914e-05, | |
| "loss": 1.9189, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.6720299345182413, | |
| "grad_norm": 1.2208515405654907, | |
| "learning_rate": 5.07563025210084e-05, | |
| "loss": 1.8619, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.6735266604303087, | |
| "grad_norm": 1.202310562133789, | |
| "learning_rate": 5.042016806722689e-05, | |
| "loss": 1.7852, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.6735266604303087, | |
| "eval_loss": 1.7230315208435059, | |
| "eval_runtime": 9.9156, | |
| "eval_samples_per_second": 6.454, | |
| "eval_steps_per_second": 3.227, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.675023386342376, | |
| "grad_norm": 1.0285881757736206, | |
| "learning_rate": 5.008403361344538e-05, | |
| "loss": 1.8857, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.6765201122544434, | |
| "grad_norm": 1.0473498106002808, | |
| "learning_rate": 4.974789915966387e-05, | |
| "loss": 1.8324, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.6780168381665107, | |
| "grad_norm": 1.129944920539856, | |
| "learning_rate": 4.9411764705882355e-05, | |
| "loss": 1.9199, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.6795135640785781, | |
| "grad_norm": 0.9636753797531128, | |
| "learning_rate": 4.907563025210084e-05, | |
| "loss": 1.8929, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.6810102899906455, | |
| "grad_norm": 1.014259934425354, | |
| "learning_rate": 4.8739495798319326e-05, | |
| "loss": 1.9402, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.6825070159027128, | |
| "grad_norm": 1.033557415008545, | |
| "learning_rate": 4.840336134453782e-05, | |
| "loss": 1.7401, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.6840037418147802, | |
| "grad_norm": 1.017338752746582, | |
| "learning_rate": 4.8067226890756304e-05, | |
| "loss": 1.7473, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.6855004677268475, | |
| "grad_norm": 1.0434340238571167, | |
| "learning_rate": 4.7731092436974796e-05, | |
| "loss": 1.8009, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.6869971936389149, | |
| "grad_norm": 1.1409871578216553, | |
| "learning_rate": 4.7394957983193275e-05, | |
| "loss": 1.7357, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.6884939195509823, | |
| "grad_norm": 1.2012221813201904, | |
| "learning_rate": 4.705882352941177e-05, | |
| "loss": 1.6896, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.6899906454630496, | |
| "grad_norm": 1.079777479171753, | |
| "learning_rate": 4.672268907563025e-05, | |
| "loss": 1.8386, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.691487371375117, | |
| "grad_norm": 1.0131338834762573, | |
| "learning_rate": 4.6386554621848745e-05, | |
| "loss": 1.8507, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.6929840972871842, | |
| "grad_norm": 1.123460292816162, | |
| "learning_rate": 4.6050420168067224e-05, | |
| "loss": 1.7637, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.6944808231992516, | |
| "grad_norm": 1.4412729740142822, | |
| "learning_rate": 4.5714285714285716e-05, | |
| "loss": 1.6948, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.695977549111319, | |
| "grad_norm": 1.220767617225647, | |
| "learning_rate": 4.53781512605042e-05, | |
| "loss": 1.8524, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.6974742750233863, | |
| "grad_norm": 1.0192736387252808, | |
| "learning_rate": 4.5042016806722694e-05, | |
| "loss": 1.8267, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.6989710009354537, | |
| "grad_norm": 1.0492209196090698, | |
| "learning_rate": 4.470588235294118e-05, | |
| "loss": 1.6835, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.700467726847521, | |
| "grad_norm": 0.9897292256355286, | |
| "learning_rate": 4.4369747899159665e-05, | |
| "loss": 1.6976, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.7019644527595884, | |
| "grad_norm": 1.3974395990371704, | |
| "learning_rate": 4.403361344537815e-05, | |
| "loss": 1.7081, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.7034611786716558, | |
| "grad_norm": 1.0798640251159668, | |
| "learning_rate": 4.369747899159664e-05, | |
| "loss": 1.5884, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.7049579045837231, | |
| "grad_norm": 1.010471224784851, | |
| "learning_rate": 4.336134453781513e-05, | |
| "loss": 1.8187, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.7064546304957905, | |
| "grad_norm": 1.2067906856536865, | |
| "learning_rate": 4.302521008403362e-05, | |
| "loss": 1.7909, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.7079513564078578, | |
| "grad_norm": 0.9305792450904846, | |
| "learning_rate": 4.26890756302521e-05, | |
| "loss": 1.8915, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.7094480823199252, | |
| "grad_norm": 0.9608831405639648, | |
| "learning_rate": 4.235294117647059e-05, | |
| "loss": 1.8287, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.7109448082319925, | |
| "grad_norm": 1.055216908454895, | |
| "learning_rate": 4.201680672268908e-05, | |
| "loss": 1.8816, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.7109448082319925, | |
| "eval_loss": 1.7186332941055298, | |
| "eval_runtime": 9.9278, | |
| "eval_samples_per_second": 6.447, | |
| "eval_steps_per_second": 3.223, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.7124415341440599, | |
| "grad_norm": 1.1744225025177002, | |
| "learning_rate": 4.168067226890757e-05, | |
| "loss": 1.8259, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.7139382600561273, | |
| "grad_norm": 1.0045477151870728, | |
| "learning_rate": 4.134453781512605e-05, | |
| "loss": 1.9401, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.7154349859681945, | |
| "grad_norm": 1.0020571947097778, | |
| "learning_rate": 4.100840336134454e-05, | |
| "loss": 1.801, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.716931711880262, | |
| "grad_norm": 1.0695722103118896, | |
| "learning_rate": 4.0672268907563026e-05, | |
| "loss": 1.7671, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.7184284377923292, | |
| "grad_norm": 1.0022177696228027, | |
| "learning_rate": 4.033613445378152e-05, | |
| "loss": 1.9396, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.7199251637043966, | |
| "grad_norm": 1.308337688446045, | |
| "learning_rate": 4e-05, | |
| "loss": 1.8176, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.721421889616464, | |
| "grad_norm": 1.133042573928833, | |
| "learning_rate": 3.966386554621849e-05, | |
| "loss": 1.726, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.7229186155285313, | |
| "grad_norm": 0.9932194352149963, | |
| "learning_rate": 3.9327731092436974e-05, | |
| "loss": 1.8504, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.7244153414405987, | |
| "grad_norm": 1.0440658330917358, | |
| "learning_rate": 3.8991596638655467e-05, | |
| "loss": 1.7443, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.725912067352666, | |
| "grad_norm": 1.0093355178833008, | |
| "learning_rate": 3.865546218487395e-05, | |
| "loss": 1.9789, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.7274087932647334, | |
| "grad_norm": 1.0096389055252075, | |
| "learning_rate": 3.8319327731092444e-05, | |
| "loss": 1.7372, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.7289055191768008, | |
| "grad_norm": 0.9682419300079346, | |
| "learning_rate": 3.798319327731092e-05, | |
| "loss": 1.8613, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.7304022450888681, | |
| "grad_norm": 1.174164891242981, | |
| "learning_rate": 3.7647058823529415e-05, | |
| "loss": 1.6717, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.7318989710009355, | |
| "grad_norm": 1.0377211570739746, | |
| "learning_rate": 3.73109243697479e-05, | |
| "loss": 1.8994, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.7333956969130028, | |
| "grad_norm": 1.137771725654602, | |
| "learning_rate": 3.697478991596639e-05, | |
| "loss": 1.6331, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.7348924228250702, | |
| "grad_norm": 1.159590482711792, | |
| "learning_rate": 3.663865546218487e-05, | |
| "loss": 1.7866, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.7363891487371376, | |
| "grad_norm": 1.0992891788482666, | |
| "learning_rate": 3.6302521008403364e-05, | |
| "loss": 1.6787, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.7378858746492049, | |
| "grad_norm": 1.1458029747009277, | |
| "learning_rate": 3.596638655462185e-05, | |
| "loss": 1.8032, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.7393826005612723, | |
| "grad_norm": 1.0241297483444214, | |
| "learning_rate": 3.563025210084034e-05, | |
| "loss": 1.8382, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.7408793264733395, | |
| "grad_norm": 1.0229034423828125, | |
| "learning_rate": 3.529411764705883e-05, | |
| "loss": 1.6115, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.7423760523854069, | |
| "grad_norm": 1.0173096656799316, | |
| "learning_rate": 3.495798319327731e-05, | |
| "loss": 1.719, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.7438727782974742, | |
| "grad_norm": 1.0528109073638916, | |
| "learning_rate": 3.46218487394958e-05, | |
| "loss": 1.8406, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.7453695042095416, | |
| "grad_norm": 1.131954312324524, | |
| "learning_rate": 3.428571428571429e-05, | |
| "loss": 1.7171, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.746866230121609, | |
| "grad_norm": 1.2483571767807007, | |
| "learning_rate": 3.3949579831932776e-05, | |
| "loss": 1.6404, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.7483629560336763, | |
| "grad_norm": 1.1184672117233276, | |
| "learning_rate": 3.361344537815127e-05, | |
| "loss": 1.7858, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.7483629560336763, | |
| "eval_loss": 1.7139549255371094, | |
| "eval_runtime": 9.9252, | |
| "eval_samples_per_second": 6.448, | |
| "eval_steps_per_second": 3.224, | |
| "step": 500 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 600, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 50, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.418709572519199e+17, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |