{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.996102883865939, "eval_steps": 500, "global_step": 1923, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001558846453624318, "grad_norm": 52.22184914143673, "learning_rate": 2.590673575129534e-07, "loss": 11.3056, "step": 1 }, { "epoch": 0.003117692907248636, "grad_norm": 55.3801908988395, "learning_rate": 5.181347150259068e-07, "loss": 11.1148, "step": 2 }, { "epoch": 0.004676539360872954, "grad_norm": 53.76354651632375, "learning_rate": 7.772020725388602e-07, "loss": 11.215, "step": 3 }, { "epoch": 0.006235385814497272, "grad_norm": 52.06232793344078, "learning_rate": 1.0362694300518136e-06, "loss": 11.2145, "step": 4 }, { "epoch": 0.00779423226812159, "grad_norm": 55.06385321064369, "learning_rate": 1.2953367875647669e-06, "loss": 11.1779, "step": 5 }, { "epoch": 0.009353078721745909, "grad_norm": 53.1236773756813, "learning_rate": 1.5544041450777204e-06, "loss": 11.1812, "step": 6 }, { "epoch": 0.010911925175370226, "grad_norm": 58.28685190281901, "learning_rate": 1.8134715025906736e-06, "loss": 10.864, "step": 7 }, { "epoch": 0.012470771628994544, "grad_norm": 57.49860454762129, "learning_rate": 2.0725388601036273e-06, "loss": 10.8997, "step": 8 }, { "epoch": 0.014029618082618862, "grad_norm": 61.33406437121332, "learning_rate": 2.3316062176165805e-06, "loss": 10.7659, "step": 9 }, { "epoch": 0.01558846453624318, "grad_norm": 83.51848150551032, "learning_rate": 2.5906735751295338e-06, "loss": 9.7158, "step": 10 }, { "epoch": 0.0171473109898675, "grad_norm": 87.40268379033651, "learning_rate": 2.8497409326424875e-06, "loss": 9.3576, "step": 11 }, { "epoch": 0.018706157443491817, "grad_norm": 94.86364555239567, "learning_rate": 3.1088082901554407e-06, "loss": 9.0476, "step": 12 }, { "epoch": 0.020265003897116135, "grad_norm": 66.58727850770262, "learning_rate": 3.367875647668394e-06, "loss": 3.7998, "step": 13 }, { "epoch": 0.021823850350740453, "grad_norm": 60.23863297900514, "learning_rate": 3.626943005181347e-06, "loss": 3.5631, "step": 14 }, { "epoch": 0.02338269680436477, "grad_norm": 42.715651899746156, "learning_rate": 3.886010362694301e-06, "loss": 2.8496, "step": 15 }, { "epoch": 0.02494154325798909, "grad_norm": 35.86036496654402, "learning_rate": 4.1450777202072546e-06, "loss": 2.5679, "step": 16 }, { "epoch": 0.026500389711613406, "grad_norm": 8.204333077091338, "learning_rate": 4.404145077720207e-06, "loss": 1.4646, "step": 17 }, { "epoch": 0.028059236165237724, "grad_norm": 5.413485907710649, "learning_rate": 4.663212435233161e-06, "loss": 1.3395, "step": 18 }, { "epoch": 0.029618082618862042, "grad_norm": 4.181557631031653, "learning_rate": 4.922279792746115e-06, "loss": 1.282, "step": 19 }, { "epoch": 0.03117692907248636, "grad_norm": 3.3106118750979614, "learning_rate": 5.1813471502590676e-06, "loss": 1.192, "step": 20 }, { "epoch": 0.03273577552611068, "grad_norm": 2.5377335055220023, "learning_rate": 5.440414507772021e-06, "loss": 1.1684, "step": 21 }, { "epoch": 0.034294621979735, "grad_norm": 2.1309505059833436, "learning_rate": 5.699481865284975e-06, "loss": 1.0969, "step": 22 }, { "epoch": 0.03585346843335931, "grad_norm": 3.8622301532964998, "learning_rate": 5.958549222797928e-06, "loss": 1.0732, "step": 23 }, { "epoch": 0.037412314886983634, "grad_norm": 47.58759142573856, "learning_rate": 6.217616580310881e-06, "loss": 1.0184, "step": 24 }, { "epoch": 0.03897116134060795, "grad_norm": 16.958376825756968, "learning_rate": 6.476683937823834e-06, "loss": 0.9784, "step": 25 }, { "epoch": 0.04053000779423227, "grad_norm": 1.7724891024655978, "learning_rate": 6.735751295336788e-06, "loss": 0.9303, "step": 26 }, { "epoch": 0.042088854247856584, "grad_norm": 1.2517023402408312, "learning_rate": 6.994818652849741e-06, "loss": 0.8952, "step": 27 }, { "epoch": 0.043647700701480906, "grad_norm": 1.0169924063509375, "learning_rate": 7.253886010362694e-06, "loss": 0.8697, "step": 28 }, { "epoch": 0.04520654715510522, "grad_norm": 0.8637449789891917, "learning_rate": 7.512953367875648e-06, "loss": 0.8588, "step": 29 }, { "epoch": 0.04676539360872954, "grad_norm": 0.8722633450269738, "learning_rate": 7.772020725388602e-06, "loss": 0.8327, "step": 30 }, { "epoch": 0.048324240062353856, "grad_norm": 0.7979319477336262, "learning_rate": 8.031088082901555e-06, "loss": 0.815, "step": 31 }, { "epoch": 0.04988308651597818, "grad_norm": 0.7636414965508945, "learning_rate": 8.290155440414509e-06, "loss": 0.7699, "step": 32 }, { "epoch": 0.05144193296960249, "grad_norm": 0.6723426478607053, "learning_rate": 8.549222797927462e-06, "loss": 0.7777, "step": 33 }, { "epoch": 0.05300077942322681, "grad_norm": 0.671680033034605, "learning_rate": 8.808290155440415e-06, "loss": 0.7429, "step": 34 }, { "epoch": 0.05455962587685113, "grad_norm": 0.6367482766420466, "learning_rate": 9.06735751295337e-06, "loss": 0.7442, "step": 35 }, { "epoch": 0.05611847233047545, "grad_norm": 0.5985756418410207, "learning_rate": 9.326424870466322e-06, "loss": 0.7354, "step": 36 }, { "epoch": 0.05767731878409977, "grad_norm": 0.5965999524800757, "learning_rate": 9.585492227979275e-06, "loss": 0.7226, "step": 37 }, { "epoch": 0.059236165237724084, "grad_norm": 0.5390447035986822, "learning_rate": 9.84455958549223e-06, "loss": 0.6732, "step": 38 }, { "epoch": 0.060795011691348405, "grad_norm": 0.4740676330023613, "learning_rate": 1.0103626943005182e-05, "loss": 0.6923, "step": 39 }, { "epoch": 0.06235385814497272, "grad_norm": 0.4870908612880822, "learning_rate": 1.0362694300518135e-05, "loss": 0.6898, "step": 40 }, { "epoch": 0.06391270459859703, "grad_norm": 0.49048808676105743, "learning_rate": 1.062176165803109e-05, "loss": 0.6817, "step": 41 }, { "epoch": 0.06547155105222136, "grad_norm": 0.4775360211701637, "learning_rate": 1.0880829015544042e-05, "loss": 0.6806, "step": 42 }, { "epoch": 0.06703039750584568, "grad_norm": 0.540847742943483, "learning_rate": 1.1139896373056995e-05, "loss": 0.6548, "step": 43 }, { "epoch": 0.06858924395947, "grad_norm": 0.4168099062285217, "learning_rate": 1.139896373056995e-05, "loss": 0.6584, "step": 44 }, { "epoch": 0.0701480904130943, "grad_norm": 0.4163477313182357, "learning_rate": 1.1658031088082903e-05, "loss": 0.6511, "step": 45 }, { "epoch": 0.07170693686671863, "grad_norm": 0.4425510534250163, "learning_rate": 1.1917098445595855e-05, "loss": 0.6432, "step": 46 }, { "epoch": 0.07326578332034295, "grad_norm": 0.3769848201392934, "learning_rate": 1.2176165803108808e-05, "loss": 0.6004, "step": 47 }, { "epoch": 0.07482462977396727, "grad_norm": 0.36307003967305024, "learning_rate": 1.2435233160621763e-05, "loss": 0.6339, "step": 48 }, { "epoch": 0.07638347622759158, "grad_norm": 0.36365771900218874, "learning_rate": 1.2694300518134716e-05, "loss": 0.6573, "step": 49 }, { "epoch": 0.0779423226812159, "grad_norm": 0.3554573096821652, "learning_rate": 1.2953367875647668e-05, "loss": 0.6606, "step": 50 }, { "epoch": 0.07950116913484022, "grad_norm": 0.4045859236452954, "learning_rate": 1.3212435233160623e-05, "loss": 0.6475, "step": 51 }, { "epoch": 0.08106001558846454, "grad_norm": 0.32890043969215965, "learning_rate": 1.3471502590673576e-05, "loss": 0.6148, "step": 52 }, { "epoch": 0.08261886204208885, "grad_norm": 0.3083895203558219, "learning_rate": 1.3730569948186529e-05, "loss": 0.6347, "step": 53 }, { "epoch": 0.08417770849571317, "grad_norm": 0.33340100324523925, "learning_rate": 1.3989637305699481e-05, "loss": 0.6165, "step": 54 }, { "epoch": 0.08573655494933749, "grad_norm": 0.29517015584889045, "learning_rate": 1.4248704663212436e-05, "loss": 0.5802, "step": 55 }, { "epoch": 0.08729540140296181, "grad_norm": 0.28430427065168046, "learning_rate": 1.4507772020725389e-05, "loss": 0.5895, "step": 56 }, { "epoch": 0.08885424785658613, "grad_norm": 0.3338831939267004, "learning_rate": 1.4766839378238342e-05, "loss": 0.5715, "step": 57 }, { "epoch": 0.09041309431021044, "grad_norm": 0.3093980059650245, "learning_rate": 1.5025906735751296e-05, "loss": 0.5847, "step": 58 }, { "epoch": 0.09197194076383476, "grad_norm": 0.2948901838276863, "learning_rate": 1.528497409326425e-05, "loss": 0.5995, "step": 59 }, { "epoch": 0.09353078721745908, "grad_norm": 0.31762615759493373, "learning_rate": 1.5544041450777204e-05, "loss": 0.6043, "step": 60 }, { "epoch": 0.0950896336710834, "grad_norm": 0.29491512166284634, "learning_rate": 1.5803108808290158e-05, "loss": 0.5992, "step": 61 }, { "epoch": 0.09664848012470771, "grad_norm": 0.29897052685770187, "learning_rate": 1.606217616580311e-05, "loss": 0.5931, "step": 62 }, { "epoch": 0.09820732657833203, "grad_norm": 0.30055023786228835, "learning_rate": 1.6321243523316064e-05, "loss": 0.5849, "step": 63 }, { "epoch": 0.09976617303195635, "grad_norm": 0.26046107559568865, "learning_rate": 1.6580310880829018e-05, "loss": 0.5512, "step": 64 }, { "epoch": 0.10132501948558068, "grad_norm": 0.30065721102538445, "learning_rate": 1.683937823834197e-05, "loss": 0.5814, "step": 65 }, { "epoch": 0.10288386593920498, "grad_norm": 0.29570037471107435, "learning_rate": 1.7098445595854924e-05, "loss": 0.5889, "step": 66 }, { "epoch": 0.1044427123928293, "grad_norm": 0.2675735266669228, "learning_rate": 1.735751295336788e-05, "loss": 0.5442, "step": 67 }, { "epoch": 0.10600155884645363, "grad_norm": 0.3100468880818413, "learning_rate": 1.761658031088083e-05, "loss": 0.5776, "step": 68 }, { "epoch": 0.10756040530007795, "grad_norm": 0.27731104939451146, "learning_rate": 1.7875647668393784e-05, "loss": 0.5644, "step": 69 }, { "epoch": 0.10911925175370225, "grad_norm": 0.3214790758034934, "learning_rate": 1.813471502590674e-05, "loss": 0.5361, "step": 70 }, { "epoch": 0.11067809820732658, "grad_norm": 0.29991490625376055, "learning_rate": 1.839378238341969e-05, "loss": 0.5567, "step": 71 }, { "epoch": 0.1122369446609509, "grad_norm": 0.2826573502857747, "learning_rate": 1.8652849740932644e-05, "loss": 0.5901, "step": 72 }, { "epoch": 0.11379579111457522, "grad_norm": 0.30249730284978976, "learning_rate": 1.89119170984456e-05, "loss": 0.5699, "step": 73 }, { "epoch": 0.11535463756819954, "grad_norm": 0.32336196262393, "learning_rate": 1.917098445595855e-05, "loss": 0.5842, "step": 74 }, { "epoch": 0.11691348402182385, "grad_norm": 0.2940197733300188, "learning_rate": 1.9430051813471504e-05, "loss": 0.56, "step": 75 }, { "epoch": 0.11847233047544817, "grad_norm": 0.27234922818526347, "learning_rate": 1.968911917098446e-05, "loss": 0.5596, "step": 76 }, { "epoch": 0.12003117692907249, "grad_norm": 0.31604251280714674, "learning_rate": 1.994818652849741e-05, "loss": 0.5638, "step": 77 }, { "epoch": 0.12159002338269681, "grad_norm": 0.2640909764215779, "learning_rate": 2.0207253886010365e-05, "loss": 0.5357, "step": 78 }, { "epoch": 0.12314886983632112, "grad_norm": 0.3325125655849214, "learning_rate": 2.046632124352332e-05, "loss": 0.5634, "step": 79 }, { "epoch": 0.12470771628994544, "grad_norm": 0.2728623091830791, "learning_rate": 2.072538860103627e-05, "loss": 0.567, "step": 80 }, { "epoch": 0.12626656274356976, "grad_norm": 0.3120421349580941, "learning_rate": 2.0984455958549225e-05, "loss": 0.5428, "step": 81 }, { "epoch": 0.12782540919719407, "grad_norm": 0.2771731681198735, "learning_rate": 2.124352331606218e-05, "loss": 0.5526, "step": 82 }, { "epoch": 0.1293842556508184, "grad_norm": 0.3027775941751013, "learning_rate": 2.150259067357513e-05, "loss": 0.5432, "step": 83 }, { "epoch": 0.1309431021044427, "grad_norm": 0.3063900417637035, "learning_rate": 2.1761658031088085e-05, "loss": 0.5633, "step": 84 }, { "epoch": 0.13250194855806702, "grad_norm": 0.29978974426938987, "learning_rate": 2.202072538860104e-05, "loss": 0.5521, "step": 85 }, { "epoch": 0.13406079501169135, "grad_norm": 0.3066667325804301, "learning_rate": 2.227979274611399e-05, "loss": 0.5576, "step": 86 }, { "epoch": 0.13561964146531566, "grad_norm": 0.3257258293999494, "learning_rate": 2.2538860103626945e-05, "loss": 0.5409, "step": 87 }, { "epoch": 0.13717848791894, "grad_norm": 0.32964630331063166, "learning_rate": 2.27979274611399e-05, "loss": 0.5524, "step": 88 }, { "epoch": 0.1387373343725643, "grad_norm": 0.3390745114943544, "learning_rate": 2.305699481865285e-05, "loss": 0.5539, "step": 89 }, { "epoch": 0.1402961808261886, "grad_norm": 0.296430130882943, "learning_rate": 2.3316062176165805e-05, "loss": 0.5519, "step": 90 }, { "epoch": 0.14185502727981295, "grad_norm": 0.3159884102421548, "learning_rate": 2.3575129533678756e-05, "loss": 0.5265, "step": 91 }, { "epoch": 0.14341387373343725, "grad_norm": 0.3136672635617097, "learning_rate": 2.383419689119171e-05, "loss": 0.5331, "step": 92 }, { "epoch": 0.1449727201870616, "grad_norm": 0.3203194252949439, "learning_rate": 2.4093264248704665e-05, "loss": 0.5472, "step": 93 }, { "epoch": 0.1465315666406859, "grad_norm": 0.34456319750216174, "learning_rate": 2.4352331606217617e-05, "loss": 0.5575, "step": 94 }, { "epoch": 0.1480904130943102, "grad_norm": 0.3228736119075391, "learning_rate": 2.461139896373057e-05, "loss": 0.5293, "step": 95 }, { "epoch": 0.14964925954793454, "grad_norm": 0.34523710171851074, "learning_rate": 2.4870466321243526e-05, "loss": 0.5287, "step": 96 }, { "epoch": 0.15120810600155885, "grad_norm": 0.3667004301126633, "learning_rate": 2.5129533678756477e-05, "loss": 0.5385, "step": 97 }, { "epoch": 0.15276695245518315, "grad_norm": 0.36137195700706454, "learning_rate": 2.538860103626943e-05, "loss": 0.5182, "step": 98 }, { "epoch": 0.1543257989088075, "grad_norm": 0.27897710920322044, "learning_rate": 2.5647668393782386e-05, "loss": 0.5301, "step": 99 }, { "epoch": 0.1558846453624318, "grad_norm": 0.28832759123856166, "learning_rate": 2.5906735751295337e-05, "loss": 0.5118, "step": 100 }, { "epoch": 0.15744349181605613, "grad_norm": 0.34377322242370634, "learning_rate": 2.616580310880829e-05, "loss": 0.532, "step": 101 }, { "epoch": 0.15900233826968044, "grad_norm": 0.3036711410586048, "learning_rate": 2.6424870466321246e-05, "loss": 0.5165, "step": 102 }, { "epoch": 0.16056118472330475, "grad_norm": 0.36103093391647273, "learning_rate": 2.6683937823834197e-05, "loss": 0.5287, "step": 103 }, { "epoch": 0.16212003117692908, "grad_norm": 0.29337030826817656, "learning_rate": 2.694300518134715e-05, "loss": 0.528, "step": 104 }, { "epoch": 0.1636788776305534, "grad_norm": 0.33805107509606247, "learning_rate": 2.7202072538860106e-05, "loss": 0.5354, "step": 105 }, { "epoch": 0.1652377240841777, "grad_norm": 0.31771328671144855, "learning_rate": 2.7461139896373057e-05, "loss": 0.5282, "step": 106 }, { "epoch": 0.16679657053780203, "grad_norm": 0.28313518361082185, "learning_rate": 2.7720207253886012e-05, "loss": 0.5086, "step": 107 }, { "epoch": 0.16835541699142634, "grad_norm": 0.34173304219782913, "learning_rate": 2.7979274611398963e-05, "loss": 0.5266, "step": 108 }, { "epoch": 0.16991426344505067, "grad_norm": 0.2895019297309633, "learning_rate": 2.8238341968911917e-05, "loss": 0.5305, "step": 109 }, { "epoch": 0.17147310989867498, "grad_norm": 0.3271405004182382, "learning_rate": 2.8497409326424872e-05, "loss": 0.5412, "step": 110 }, { "epoch": 0.1730319563522993, "grad_norm": 0.3254108424111249, "learning_rate": 2.8756476683937827e-05, "loss": 0.5267, "step": 111 }, { "epoch": 0.17459080280592362, "grad_norm": 0.43636064754159903, "learning_rate": 2.9015544041450778e-05, "loss": 0.546, "step": 112 }, { "epoch": 0.17614964925954793, "grad_norm": 0.35895062463773536, "learning_rate": 2.9274611398963732e-05, "loss": 0.5379, "step": 113 }, { "epoch": 0.17770849571317227, "grad_norm": 0.48008348852035654, "learning_rate": 2.9533678756476683e-05, "loss": 0.5251, "step": 114 }, { "epoch": 0.17926734216679657, "grad_norm": 0.40042199319895483, "learning_rate": 2.9792746113989638e-05, "loss": 0.5043, "step": 115 }, { "epoch": 0.18082618862042088, "grad_norm": 0.44332954681106973, "learning_rate": 3.0051813471502592e-05, "loss": 0.5135, "step": 116 }, { "epoch": 0.18238503507404522, "grad_norm": 0.43275195875835987, "learning_rate": 3.0310880829015547e-05, "loss": 0.5196, "step": 117 }, { "epoch": 0.18394388152766952, "grad_norm": 0.3939079245195568, "learning_rate": 3.05699481865285e-05, "loss": 0.5261, "step": 118 }, { "epoch": 0.18550272798129383, "grad_norm": 0.45003761498432127, "learning_rate": 3.082901554404145e-05, "loss": 0.5285, "step": 119 }, { "epoch": 0.18706157443491817, "grad_norm": 0.38744988267232733, "learning_rate": 3.108808290155441e-05, "loss": 0.5183, "step": 120 }, { "epoch": 0.18862042088854247, "grad_norm": 0.42417740462219883, "learning_rate": 3.134715025906736e-05, "loss": 0.5366, "step": 121 }, { "epoch": 0.1901792673421668, "grad_norm": 0.38878746316666607, "learning_rate": 3.1606217616580316e-05, "loss": 0.5405, "step": 122 }, { "epoch": 0.19173811379579112, "grad_norm": 0.39470480511600553, "learning_rate": 3.186528497409327e-05, "loss": 0.5177, "step": 123 }, { "epoch": 0.19329696024941542, "grad_norm": 0.3425139995303132, "learning_rate": 3.212435233160622e-05, "loss": 0.5278, "step": 124 }, { "epoch": 0.19485580670303976, "grad_norm": 0.3855003673988402, "learning_rate": 3.238341968911917e-05, "loss": 0.5303, "step": 125 }, { "epoch": 0.19641465315666407, "grad_norm": 0.3708329304695599, "learning_rate": 3.264248704663213e-05, "loss": 0.5308, "step": 126 }, { "epoch": 0.1979734996102884, "grad_norm": 0.42197888435438696, "learning_rate": 3.290155440414508e-05, "loss": 0.5199, "step": 127 }, { "epoch": 0.1995323460639127, "grad_norm": 0.39698217513178863, "learning_rate": 3.3160621761658036e-05, "loss": 0.506, "step": 128 }, { "epoch": 0.20109119251753702, "grad_norm": 0.4043130382762473, "learning_rate": 3.341968911917099e-05, "loss": 0.5065, "step": 129 }, { "epoch": 0.20265003897116135, "grad_norm": 0.34367190967592626, "learning_rate": 3.367875647668394e-05, "loss": 0.509, "step": 130 }, { "epoch": 0.20420888542478566, "grad_norm": 0.49458742478228435, "learning_rate": 3.393782383419689e-05, "loss": 0.5057, "step": 131 }, { "epoch": 0.20576773187840997, "grad_norm": 0.3946670126655184, "learning_rate": 3.419689119170985e-05, "loss": 0.5074, "step": 132 }, { "epoch": 0.2073265783320343, "grad_norm": 0.4590062123156051, "learning_rate": 3.44559585492228e-05, "loss": 0.5196, "step": 133 }, { "epoch": 0.2088854247856586, "grad_norm": 0.41794015093559495, "learning_rate": 3.471502590673576e-05, "loss": 0.5293, "step": 134 }, { "epoch": 0.21044427123928294, "grad_norm": 0.3976262150889235, "learning_rate": 3.497409326424871e-05, "loss": 0.4958, "step": 135 }, { "epoch": 0.21200311769290725, "grad_norm": 0.3801765387110536, "learning_rate": 3.523316062176166e-05, "loss": 0.5283, "step": 136 }, { "epoch": 0.21356196414653156, "grad_norm": 0.4393098932539108, "learning_rate": 3.549222797927461e-05, "loss": 0.5106, "step": 137 }, { "epoch": 0.2151208106001559, "grad_norm": 0.37730905366487805, "learning_rate": 3.575129533678757e-05, "loss": 0.5004, "step": 138 }, { "epoch": 0.2166796570537802, "grad_norm": 0.47264900412277805, "learning_rate": 3.601036269430052e-05, "loss": 0.5107, "step": 139 }, { "epoch": 0.2182385035074045, "grad_norm": 0.4572211602618532, "learning_rate": 3.626943005181348e-05, "loss": 0.5258, "step": 140 }, { "epoch": 0.21979734996102884, "grad_norm": 0.566903195196336, "learning_rate": 3.652849740932643e-05, "loss": 0.4926, "step": 141 }, { "epoch": 0.22135619641465315, "grad_norm": 0.40493119123467375, "learning_rate": 3.678756476683938e-05, "loss": 0.5167, "step": 142 }, { "epoch": 0.22291504286827749, "grad_norm": 0.5800300099928885, "learning_rate": 3.704663212435233e-05, "loss": 0.4888, "step": 143 }, { "epoch": 0.2244738893219018, "grad_norm": 0.49267006079774145, "learning_rate": 3.730569948186529e-05, "loss": 0.4931, "step": 144 }, { "epoch": 0.2260327357755261, "grad_norm": 0.5509681396457901, "learning_rate": 3.756476683937824e-05, "loss": 0.515, "step": 145 }, { "epoch": 0.22759158222915044, "grad_norm": 0.7109027605533598, "learning_rate": 3.78238341968912e-05, "loss": 0.4932, "step": 146 }, { "epoch": 0.22915042868277474, "grad_norm": 0.35899713776420916, "learning_rate": 3.808290155440415e-05, "loss": 0.4893, "step": 147 }, { "epoch": 0.23070927513639908, "grad_norm": 0.5392148285010113, "learning_rate": 3.83419689119171e-05, "loss": 0.4865, "step": 148 }, { "epoch": 0.23226812159002339, "grad_norm": 0.4200166849207286, "learning_rate": 3.860103626943005e-05, "loss": 0.516, "step": 149 }, { "epoch": 0.2338269680436477, "grad_norm": 0.4668106441613405, "learning_rate": 3.886010362694301e-05, "loss": 0.5104, "step": 150 }, { "epoch": 0.23538581449727203, "grad_norm": 0.388638095128956, "learning_rate": 3.911917098445596e-05, "loss": 0.5251, "step": 151 }, { "epoch": 0.23694466095089634, "grad_norm": 0.3927756772978556, "learning_rate": 3.937823834196892e-05, "loss": 0.4898, "step": 152 }, { "epoch": 0.23850350740452064, "grad_norm": 0.4474650309815615, "learning_rate": 3.963730569948187e-05, "loss": 0.5002, "step": 153 }, { "epoch": 0.24006235385814498, "grad_norm": 0.40465455399123634, "learning_rate": 3.989637305699482e-05, "loss": 0.4805, "step": 154 }, { "epoch": 0.24162120031176929, "grad_norm": 0.46329230044256475, "learning_rate": 4.015544041450777e-05, "loss": 0.5148, "step": 155 }, { "epoch": 0.24318004676539362, "grad_norm": 0.4221898620021292, "learning_rate": 4.041450777202073e-05, "loss": 0.5087, "step": 156 }, { "epoch": 0.24473889321901793, "grad_norm": 0.41405351180679445, "learning_rate": 4.067357512953368e-05, "loss": 0.5206, "step": 157 }, { "epoch": 0.24629773967264224, "grad_norm": 0.3718428876018395, "learning_rate": 4.093264248704664e-05, "loss": 0.5018, "step": 158 }, { "epoch": 0.24785658612626657, "grad_norm": 0.4108732557264086, "learning_rate": 4.119170984455959e-05, "loss": 0.4721, "step": 159 }, { "epoch": 0.24941543257989088, "grad_norm": 0.5196664214895078, "learning_rate": 4.145077720207254e-05, "loss": 0.501, "step": 160 }, { "epoch": 0.2509742790335152, "grad_norm": 0.34375475907972514, "learning_rate": 4.170984455958549e-05, "loss": 0.4944, "step": 161 }, { "epoch": 0.2525331254871395, "grad_norm": 0.5064449794040312, "learning_rate": 4.196891191709845e-05, "loss": 0.5029, "step": 162 }, { "epoch": 0.2540919719407638, "grad_norm": 0.4647723748878803, "learning_rate": 4.22279792746114e-05, "loss": 0.5024, "step": 163 }, { "epoch": 0.25565081839438814, "grad_norm": 0.4006925999521958, "learning_rate": 4.248704663212436e-05, "loss": 0.4655, "step": 164 }, { "epoch": 0.2572096648480125, "grad_norm": 0.45646986120048183, "learning_rate": 4.274611398963731e-05, "loss": 0.5059, "step": 165 }, { "epoch": 0.2587685113016368, "grad_norm": 0.4337454689280252, "learning_rate": 4.300518134715026e-05, "loss": 0.4847, "step": 166 }, { "epoch": 0.2603273577552611, "grad_norm": 0.4818011779118697, "learning_rate": 4.326424870466321e-05, "loss": 0.5039, "step": 167 }, { "epoch": 0.2618862042088854, "grad_norm": 0.4004895532370599, "learning_rate": 4.352331606217617e-05, "loss": 0.4997, "step": 168 }, { "epoch": 0.2634450506625097, "grad_norm": 0.4776133221437833, "learning_rate": 4.378238341968912e-05, "loss": 0.4747, "step": 169 }, { "epoch": 0.26500389711613404, "grad_norm": 0.3542482573234714, "learning_rate": 4.404145077720208e-05, "loss": 0.4995, "step": 170 }, { "epoch": 0.2665627435697584, "grad_norm": 0.47546134775525584, "learning_rate": 4.430051813471503e-05, "loss": 0.5021, "step": 171 }, { "epoch": 0.2681215900233827, "grad_norm": 0.37668384089912865, "learning_rate": 4.455958549222798e-05, "loss": 0.4847, "step": 172 }, { "epoch": 0.269680436477007, "grad_norm": 0.5250094003553915, "learning_rate": 4.481865284974093e-05, "loss": 0.4965, "step": 173 }, { "epoch": 0.2712392829306313, "grad_norm": 0.5118060286220544, "learning_rate": 4.507772020725389e-05, "loss": 0.5002, "step": 174 }, { "epoch": 0.2727981293842556, "grad_norm": 0.5439924404821223, "learning_rate": 4.533678756476684e-05, "loss": 0.4935, "step": 175 }, { "epoch": 0.27435697583788, "grad_norm": 0.46651416281963, "learning_rate": 4.55958549222798e-05, "loss": 0.4936, "step": 176 }, { "epoch": 0.2759158222915043, "grad_norm": 0.6025736893384493, "learning_rate": 4.585492227979275e-05, "loss": 0.4936, "step": 177 }, { "epoch": 0.2774746687451286, "grad_norm": 0.49477813313240576, "learning_rate": 4.61139896373057e-05, "loss": 0.5014, "step": 178 }, { "epoch": 0.2790335151987529, "grad_norm": 0.38453152894256626, "learning_rate": 4.637305699481865e-05, "loss": 0.5026, "step": 179 }, { "epoch": 0.2805923616523772, "grad_norm": 0.6693985713554009, "learning_rate": 4.663212435233161e-05, "loss": 0.4966, "step": 180 }, { "epoch": 0.2821512081060016, "grad_norm": 0.49890072866209745, "learning_rate": 4.689119170984456e-05, "loss": 0.4951, "step": 181 }, { "epoch": 0.2837100545596259, "grad_norm": 0.4986049703889912, "learning_rate": 4.715025906735751e-05, "loss": 0.4848, "step": 182 }, { "epoch": 0.2852689010132502, "grad_norm": 0.7123183153704133, "learning_rate": 4.740932642487047e-05, "loss": 0.5087, "step": 183 }, { "epoch": 0.2868277474668745, "grad_norm": 0.5101437582246441, "learning_rate": 4.766839378238342e-05, "loss": 0.496, "step": 184 }, { "epoch": 0.2883865939204988, "grad_norm": 0.5462740650609419, "learning_rate": 4.792746113989637e-05, "loss": 0.4891, "step": 185 }, { "epoch": 0.2899454403741232, "grad_norm": 0.6739608210631324, "learning_rate": 4.818652849740933e-05, "loss": 0.4809, "step": 186 }, { "epoch": 0.2915042868277475, "grad_norm": 0.36454838249845395, "learning_rate": 4.844559585492228e-05, "loss": 0.4826, "step": 187 }, { "epoch": 0.2930631332813718, "grad_norm": 0.8180118832431814, "learning_rate": 4.870466321243523e-05, "loss": 0.5007, "step": 188 }, { "epoch": 0.2946219797349961, "grad_norm": 0.8582599255383014, "learning_rate": 4.896373056994819e-05, "loss": 0.5101, "step": 189 }, { "epoch": 0.2961808261886204, "grad_norm": 0.5339198441824459, "learning_rate": 4.922279792746114e-05, "loss": 0.4755, "step": 190 }, { "epoch": 0.2977396726422447, "grad_norm": 0.5038691245377587, "learning_rate": 4.948186528497409e-05, "loss": 0.4987, "step": 191 }, { "epoch": 0.2992985190958691, "grad_norm": 0.7512130026001776, "learning_rate": 4.974093264248705e-05, "loss": 0.5168, "step": 192 }, { "epoch": 0.3008573655494934, "grad_norm": 0.5965449485798306, "learning_rate": 5e-05, "loss": 0.4825, "step": 193 }, { "epoch": 0.3024162120031177, "grad_norm": 0.5874775355232069, "learning_rate": 4.9971098265895956e-05, "loss": 0.484, "step": 194 }, { "epoch": 0.303975058456742, "grad_norm": 0.6794164991276913, "learning_rate": 4.994219653179191e-05, "loss": 0.4956, "step": 195 }, { "epoch": 0.3055339049103663, "grad_norm": 0.5324211441774743, "learning_rate": 4.991329479768786e-05, "loss": 0.469, "step": 196 }, { "epoch": 0.30709275136399067, "grad_norm": 0.7381982819903451, "learning_rate": 4.9884393063583816e-05, "loss": 0.4859, "step": 197 }, { "epoch": 0.308651597817615, "grad_norm": 0.7316455279398204, "learning_rate": 4.985549132947977e-05, "loss": 0.4962, "step": 198 }, { "epoch": 0.3102104442712393, "grad_norm": 0.5513173476365167, "learning_rate": 4.982658959537572e-05, "loss": 0.4973, "step": 199 }, { "epoch": 0.3117692907248636, "grad_norm": 0.6961187825800337, "learning_rate": 4.979768786127168e-05, "loss": 0.504, "step": 200 }, { "epoch": 0.3133281371784879, "grad_norm": 0.6675059607265288, "learning_rate": 4.976878612716763e-05, "loss": 0.4818, "step": 201 }, { "epoch": 0.31488698363211226, "grad_norm": 0.5775315130950791, "learning_rate": 4.9739884393063583e-05, "loss": 0.5108, "step": 202 }, { "epoch": 0.31644583008573657, "grad_norm": 0.5493567678427346, "learning_rate": 4.971098265895954e-05, "loss": 0.4933, "step": 203 }, { "epoch": 0.3180046765393609, "grad_norm": 0.5788493631561638, "learning_rate": 4.96820809248555e-05, "loss": 0.4921, "step": 204 }, { "epoch": 0.3195635229929852, "grad_norm": 0.5111462707368106, "learning_rate": 4.965317919075145e-05, "loss": 0.4617, "step": 205 }, { "epoch": 0.3211223694466095, "grad_norm": 0.5647593330505578, "learning_rate": 4.96242774566474e-05, "loss": 0.4953, "step": 206 }, { "epoch": 0.32268121590023385, "grad_norm": 0.4207725153414849, "learning_rate": 4.959537572254335e-05, "loss": 0.5115, "step": 207 }, { "epoch": 0.32424006235385816, "grad_norm": 0.4624044434848398, "learning_rate": 4.956647398843931e-05, "loss": 0.4722, "step": 208 }, { "epoch": 0.32579890880748247, "grad_norm": 0.4376386673543866, "learning_rate": 4.9537572254335264e-05, "loss": 0.4832, "step": 209 }, { "epoch": 0.3273577552611068, "grad_norm": 0.531865499974187, "learning_rate": 4.950867052023122e-05, "loss": 0.5055, "step": 210 }, { "epoch": 0.3289166017147311, "grad_norm": 0.5050602424671378, "learning_rate": 4.9479768786127164e-05, "loss": 0.4697, "step": 211 }, { "epoch": 0.3304754481683554, "grad_norm": 0.38168626861209426, "learning_rate": 4.9450867052023125e-05, "loss": 0.4907, "step": 212 }, { "epoch": 0.33203429462197975, "grad_norm": 0.5281790174146758, "learning_rate": 4.942196531791908e-05, "loss": 0.4876, "step": 213 }, { "epoch": 0.33359314107560406, "grad_norm": 0.5297198653114203, "learning_rate": 4.939306358381503e-05, "loss": 0.4913, "step": 214 }, { "epoch": 0.33515198752922837, "grad_norm": 0.45152753639301624, "learning_rate": 4.9364161849710985e-05, "loss": 0.4731, "step": 215 }, { "epoch": 0.3367108339828527, "grad_norm": 0.3683086256111994, "learning_rate": 4.933526011560694e-05, "loss": 0.4779, "step": 216 }, { "epoch": 0.338269680436477, "grad_norm": 0.4739643840797579, "learning_rate": 4.930635838150289e-05, "loss": 0.4992, "step": 217 }, { "epoch": 0.33982852689010135, "grad_norm": 0.4068837265657986, "learning_rate": 4.9277456647398845e-05, "loss": 0.4996, "step": 218 }, { "epoch": 0.34138737334372565, "grad_norm": 0.4885011148922096, "learning_rate": 4.92485549132948e-05, "loss": 0.4865, "step": 219 }, { "epoch": 0.34294621979734996, "grad_norm": 0.493814836908405, "learning_rate": 4.921965317919075e-05, "loss": 0.4731, "step": 220 }, { "epoch": 0.34450506625097427, "grad_norm": 0.46892553233657863, "learning_rate": 4.9190751445086706e-05, "loss": 0.4699, "step": 221 }, { "epoch": 0.3460639127045986, "grad_norm": 0.464609243189492, "learning_rate": 4.916184971098266e-05, "loss": 0.478, "step": 222 }, { "epoch": 0.34762275915822294, "grad_norm": 0.37152543938557336, "learning_rate": 4.913294797687861e-05, "loss": 0.4694, "step": 223 }, { "epoch": 0.34918160561184725, "grad_norm": 0.5025252675976578, "learning_rate": 4.910404624277457e-05, "loss": 0.4715, "step": 224 }, { "epoch": 0.35074045206547155, "grad_norm": 0.3402466722282753, "learning_rate": 4.9075144508670526e-05, "loss": 0.463, "step": 225 }, { "epoch": 0.35229929851909586, "grad_norm": 0.3965061942369021, "learning_rate": 4.904624277456647e-05, "loss": 0.4792, "step": 226 }, { "epoch": 0.35385814497272017, "grad_norm": 0.4206144149046459, "learning_rate": 4.9017341040462426e-05, "loss": 0.4901, "step": 227 }, { "epoch": 0.35541699142634453, "grad_norm": 0.4286051420798536, "learning_rate": 4.8988439306358387e-05, "loss": 0.4742, "step": 228 }, { "epoch": 0.35697583787996884, "grad_norm": 0.47141158644523007, "learning_rate": 4.895953757225434e-05, "loss": 0.4968, "step": 229 }, { "epoch": 0.35853468433359315, "grad_norm": 0.3415005284976459, "learning_rate": 4.8930635838150293e-05, "loss": 0.4568, "step": 230 }, { "epoch": 0.36009353078721745, "grad_norm": 0.48287718381612604, "learning_rate": 4.890173410404624e-05, "loss": 0.4993, "step": 231 }, { "epoch": 0.36165237724084176, "grad_norm": 0.5803189730929726, "learning_rate": 4.88728323699422e-05, "loss": 0.4653, "step": 232 }, { "epoch": 0.36321122369446607, "grad_norm": 0.3907951855741137, "learning_rate": 4.8843930635838154e-05, "loss": 0.4794, "step": 233 }, { "epoch": 0.36477007014809043, "grad_norm": 0.5028194362808669, "learning_rate": 4.881502890173411e-05, "loss": 0.4942, "step": 234 }, { "epoch": 0.36632891660171474, "grad_norm": 0.4945111081672912, "learning_rate": 4.878612716763006e-05, "loss": 0.4771, "step": 235 }, { "epoch": 0.36788776305533905, "grad_norm": 0.43827490144476094, "learning_rate": 4.8757225433526014e-05, "loss": 0.459, "step": 236 }, { "epoch": 0.36944660950896335, "grad_norm": 0.5256559700101235, "learning_rate": 4.872832369942197e-05, "loss": 0.4696, "step": 237 }, { "epoch": 0.37100545596258766, "grad_norm": 0.5899114320019979, "learning_rate": 4.869942196531792e-05, "loss": 0.4769, "step": 238 }, { "epoch": 0.372564302416212, "grad_norm": 0.4326187078710821, "learning_rate": 4.8670520231213874e-05, "loss": 0.4761, "step": 239 }, { "epoch": 0.37412314886983633, "grad_norm": 0.4922313590375816, "learning_rate": 4.864161849710983e-05, "loss": 0.4871, "step": 240 }, { "epoch": 0.37568199532346064, "grad_norm": 0.4403169878610757, "learning_rate": 4.861271676300578e-05, "loss": 0.4918, "step": 241 }, { "epoch": 0.37724084177708495, "grad_norm": 0.39128925993823416, "learning_rate": 4.8583815028901735e-05, "loss": 0.4695, "step": 242 }, { "epoch": 0.37879968823070925, "grad_norm": 0.4947169815070576, "learning_rate": 4.855491329479769e-05, "loss": 0.4785, "step": 243 }, { "epoch": 0.3803585346843336, "grad_norm": 0.36561636850124596, "learning_rate": 4.852601156069365e-05, "loss": 0.4687, "step": 244 }, { "epoch": 0.3819173811379579, "grad_norm": 0.5740065574334354, "learning_rate": 4.8497109826589595e-05, "loss": 0.4777, "step": 245 }, { "epoch": 0.38347622759158223, "grad_norm": 0.37164208100522045, "learning_rate": 4.846820809248555e-05, "loss": 0.4769, "step": 246 }, { "epoch": 0.38503507404520654, "grad_norm": 0.45893063215148555, "learning_rate": 4.84393063583815e-05, "loss": 0.4754, "step": 247 }, { "epoch": 0.38659392049883085, "grad_norm": 0.4592126499581869, "learning_rate": 4.841040462427746e-05, "loss": 0.4667, "step": 248 }, { "epoch": 0.3881527669524552, "grad_norm": 0.42130274935477224, "learning_rate": 4.8381502890173416e-05, "loss": 0.4727, "step": 249 }, { "epoch": 0.3897116134060795, "grad_norm": 0.5562234140465554, "learning_rate": 4.835260115606937e-05, "loss": 0.5142, "step": 250 }, { "epoch": 0.3912704598597038, "grad_norm": 0.35335610698932396, "learning_rate": 4.8323699421965316e-05, "loss": 0.4726, "step": 251 }, { "epoch": 0.39282930631332813, "grad_norm": 0.5355853337204556, "learning_rate": 4.8294797687861276e-05, "loss": 0.468, "step": 252 }, { "epoch": 0.39438815276695244, "grad_norm": 0.3940457770663424, "learning_rate": 4.826589595375723e-05, "loss": 0.4795, "step": 253 }, { "epoch": 0.3959469992205768, "grad_norm": 0.48308987476155735, "learning_rate": 4.823699421965318e-05, "loss": 0.5007, "step": 254 }, { "epoch": 0.3975058456742011, "grad_norm": 0.529695325269085, "learning_rate": 4.8208092485549136e-05, "loss": 0.4928, "step": 255 }, { "epoch": 0.3990646921278254, "grad_norm": 0.46950723041884723, "learning_rate": 4.817919075144509e-05, "loss": 0.4637, "step": 256 }, { "epoch": 0.4006235385814497, "grad_norm": 0.5659006167244565, "learning_rate": 4.815028901734104e-05, "loss": 0.4719, "step": 257 }, { "epoch": 0.40218238503507403, "grad_norm": 0.39896571715761564, "learning_rate": 4.8121387283237e-05, "loss": 0.4847, "step": 258 }, { "epoch": 0.40374123148869834, "grad_norm": 0.5485795723559732, "learning_rate": 4.809248554913295e-05, "loss": 0.4782, "step": 259 }, { "epoch": 0.4053000779423227, "grad_norm": 0.3776722335922268, "learning_rate": 4.8063583815028904e-05, "loss": 0.4728, "step": 260 }, { "epoch": 0.406858924395947, "grad_norm": 0.42769770440042826, "learning_rate": 4.803468208092486e-05, "loss": 0.4718, "step": 261 }, { "epoch": 0.4084177708495713, "grad_norm": 0.38140213686869234, "learning_rate": 4.800578034682081e-05, "loss": 0.4798, "step": 262 }, { "epoch": 0.4099766173031956, "grad_norm": 0.3693912240886004, "learning_rate": 4.7976878612716764e-05, "loss": 0.4677, "step": 263 }, { "epoch": 0.41153546375681993, "grad_norm": 0.540507172297762, "learning_rate": 4.7947976878612724e-05, "loss": 0.4783, "step": 264 }, { "epoch": 0.4130943102104443, "grad_norm": 0.38027657459985387, "learning_rate": 4.791907514450867e-05, "loss": 0.4977, "step": 265 }, { "epoch": 0.4146531566640686, "grad_norm": 0.5498501393924234, "learning_rate": 4.7890173410404624e-05, "loss": 0.5207, "step": 266 }, { "epoch": 0.4162120031176929, "grad_norm": 0.44454634948662736, "learning_rate": 4.786127167630058e-05, "loss": 0.449, "step": 267 }, { "epoch": 0.4177708495713172, "grad_norm": 0.5112086728008995, "learning_rate": 4.783236994219654e-05, "loss": 0.4803, "step": 268 }, { "epoch": 0.4193296960249415, "grad_norm": 0.6274330925404005, "learning_rate": 4.780346820809249e-05, "loss": 0.477, "step": 269 }, { "epoch": 0.4208885424785659, "grad_norm": 0.45606522557073254, "learning_rate": 4.777456647398844e-05, "loss": 0.4794, "step": 270 }, { "epoch": 0.4224473889321902, "grad_norm": 0.5630490443389089, "learning_rate": 4.774566473988439e-05, "loss": 0.48, "step": 271 }, { "epoch": 0.4240062353858145, "grad_norm": 0.5538678824248618, "learning_rate": 4.771676300578035e-05, "loss": 0.4972, "step": 272 }, { "epoch": 0.4255650818394388, "grad_norm": 0.5075041955013312, "learning_rate": 4.7687861271676305e-05, "loss": 0.4935, "step": 273 }, { "epoch": 0.4271239282930631, "grad_norm": 0.5416003321516639, "learning_rate": 4.765895953757226e-05, "loss": 0.4732, "step": 274 }, { "epoch": 0.4286827747466875, "grad_norm": 0.4993962101559131, "learning_rate": 4.763005780346821e-05, "loss": 0.477, "step": 275 }, { "epoch": 0.4302416212003118, "grad_norm": 0.5254171958882733, "learning_rate": 4.7601156069364166e-05, "loss": 0.4899, "step": 276 }, { "epoch": 0.4318004676539361, "grad_norm": 0.5484868681986969, "learning_rate": 4.757225433526012e-05, "loss": 0.4608, "step": 277 }, { "epoch": 0.4333593141075604, "grad_norm": 0.6405428710095884, "learning_rate": 4.754335260115607e-05, "loss": 0.4816, "step": 278 }, { "epoch": 0.4349181605611847, "grad_norm": 0.48377922320779415, "learning_rate": 4.7514450867052026e-05, "loss": 0.4838, "step": 279 }, { "epoch": 0.436477007014809, "grad_norm": 0.7085436078292358, "learning_rate": 4.748554913294798e-05, "loss": 0.4718, "step": 280 }, { "epoch": 0.4380358534684334, "grad_norm": 0.4043009368444505, "learning_rate": 4.745664739884393e-05, "loss": 0.4442, "step": 281 }, { "epoch": 0.4395946999220577, "grad_norm": 0.5866137208959042, "learning_rate": 4.7427745664739886e-05, "loss": 0.461, "step": 282 }, { "epoch": 0.441153546375682, "grad_norm": 0.5810371524784574, "learning_rate": 4.739884393063584e-05, "loss": 0.4839, "step": 283 }, { "epoch": 0.4427123928293063, "grad_norm": 0.438670614683413, "learning_rate": 4.73699421965318e-05, "loss": 0.4615, "step": 284 }, { "epoch": 0.4442712392829306, "grad_norm": 0.6406824783627434, "learning_rate": 4.7341040462427747e-05, "loss": 0.4611, "step": 285 }, { "epoch": 0.44583008573655497, "grad_norm": 0.529477326835477, "learning_rate": 4.73121387283237e-05, "loss": 0.4788, "step": 286 }, { "epoch": 0.4473889321901793, "grad_norm": 0.4838077152970437, "learning_rate": 4.7283236994219653e-05, "loss": 0.4727, "step": 287 }, { "epoch": 0.4489477786438036, "grad_norm": 0.47119490593405255, "learning_rate": 4.7254335260115614e-05, "loss": 0.473, "step": 288 }, { "epoch": 0.4505066250974279, "grad_norm": 0.4366369786625669, "learning_rate": 4.722543352601157e-05, "loss": 0.4526, "step": 289 }, { "epoch": 0.4520654715510522, "grad_norm": 0.4281616340000569, "learning_rate": 4.7196531791907514e-05, "loss": 0.482, "step": 290 }, { "epoch": 0.45362431800467656, "grad_norm": 0.5820798503771535, "learning_rate": 4.716763005780347e-05, "loss": 0.4716, "step": 291 }, { "epoch": 0.45518316445830087, "grad_norm": 0.45736556696474184, "learning_rate": 4.713872832369942e-05, "loss": 0.4569, "step": 292 }, { "epoch": 0.4567420109119252, "grad_norm": 0.5568676859423672, "learning_rate": 4.710982658959538e-05, "loss": 0.4738, "step": 293 }, { "epoch": 0.4583008573655495, "grad_norm": 0.6028680330451817, "learning_rate": 4.7080924855491334e-05, "loss": 0.476, "step": 294 }, { "epoch": 0.4598597038191738, "grad_norm": 0.47393834722502826, "learning_rate": 4.705202312138728e-05, "loss": 0.4896, "step": 295 }, { "epoch": 0.46141855027279816, "grad_norm": 0.5313704646163404, "learning_rate": 4.7023121387283234e-05, "loss": 0.4805, "step": 296 }, { "epoch": 0.46297739672642246, "grad_norm": 0.5051367616146644, "learning_rate": 4.6994219653179195e-05, "loss": 0.4569, "step": 297 }, { "epoch": 0.46453624318004677, "grad_norm": 0.3831029585767748, "learning_rate": 4.696531791907515e-05, "loss": 0.4744, "step": 298 }, { "epoch": 0.4660950896336711, "grad_norm": 0.521461479028099, "learning_rate": 4.69364161849711e-05, "loss": 0.4483, "step": 299 }, { "epoch": 0.4676539360872954, "grad_norm": 0.38896958455232095, "learning_rate": 4.690751445086705e-05, "loss": 0.4498, "step": 300 }, { "epoch": 0.4692127825409197, "grad_norm": 0.39085029504274005, "learning_rate": 4.687861271676301e-05, "loss": 0.4671, "step": 301 }, { "epoch": 0.47077162899454406, "grad_norm": 0.4788937605695569, "learning_rate": 4.684971098265896e-05, "loss": 0.476, "step": 302 }, { "epoch": 0.47233047544816836, "grad_norm": 0.34601184680325847, "learning_rate": 4.6820809248554915e-05, "loss": 0.4638, "step": 303 }, { "epoch": 0.47388932190179267, "grad_norm": 0.43700546489204795, "learning_rate": 4.679190751445087e-05, "loss": 0.4814, "step": 304 }, { "epoch": 0.475448168355417, "grad_norm": 0.3838353985965739, "learning_rate": 4.676300578034682e-05, "loss": 0.451, "step": 305 }, { "epoch": 0.4770070148090413, "grad_norm": 0.4177639814145164, "learning_rate": 4.6734104046242776e-05, "loss": 0.4758, "step": 306 }, { "epoch": 0.47856586126266565, "grad_norm": 0.3922398233693093, "learning_rate": 4.670520231213873e-05, "loss": 0.4657, "step": 307 }, { "epoch": 0.48012470771628996, "grad_norm": 0.3757724195621053, "learning_rate": 4.667630057803468e-05, "loss": 0.4793, "step": 308 }, { "epoch": 0.48168355416991426, "grad_norm": 0.3414831861075698, "learning_rate": 4.664739884393064e-05, "loss": 0.4636, "step": 309 }, { "epoch": 0.48324240062353857, "grad_norm": 0.41270011789282307, "learning_rate": 4.661849710982659e-05, "loss": 0.4508, "step": 310 }, { "epoch": 0.4848012470771629, "grad_norm": 0.40583792569372235, "learning_rate": 4.658959537572254e-05, "loss": 0.4849, "step": 311 }, { "epoch": 0.48636009353078724, "grad_norm": 0.44731637705198873, "learning_rate": 4.6560693641618496e-05, "loss": 0.4956, "step": 312 }, { "epoch": 0.48791893998441155, "grad_norm": 0.39572009537399316, "learning_rate": 4.653179190751446e-05, "loss": 0.4752, "step": 313 }, { "epoch": 0.48947778643803586, "grad_norm": 0.39239985837197133, "learning_rate": 4.650289017341041e-05, "loss": 0.4448, "step": 314 }, { "epoch": 0.49103663289166016, "grad_norm": 0.47254520649722037, "learning_rate": 4.647398843930636e-05, "loss": 0.4424, "step": 315 }, { "epoch": 0.49259547934528447, "grad_norm": 0.3655766693795544, "learning_rate": 4.644508670520231e-05, "loss": 0.4585, "step": 316 }, { "epoch": 0.49415432579890883, "grad_norm": 0.5024262207840765, "learning_rate": 4.641618497109827e-05, "loss": 0.4711, "step": 317 }, { "epoch": 0.49571317225253314, "grad_norm": 0.41554970235113214, "learning_rate": 4.6387283236994224e-05, "loss": 0.4606, "step": 318 }, { "epoch": 0.49727201870615745, "grad_norm": 0.4597741099913984, "learning_rate": 4.635838150289018e-05, "loss": 0.4527, "step": 319 }, { "epoch": 0.49883086515978176, "grad_norm": 0.4425235361753541, "learning_rate": 4.6329479768786124e-05, "loss": 0.4769, "step": 320 }, { "epoch": 0.5003897116134061, "grad_norm": 0.38187283401074307, "learning_rate": 4.6300578034682084e-05, "loss": 0.4594, "step": 321 }, { "epoch": 0.5019485580670304, "grad_norm": 0.3797910266033653, "learning_rate": 4.627167630057804e-05, "loss": 0.462, "step": 322 }, { "epoch": 0.5035074045206547, "grad_norm": 0.39511922436954366, "learning_rate": 4.624277456647399e-05, "loss": 0.4703, "step": 323 }, { "epoch": 0.505066250974279, "grad_norm": 0.4061473825369823, "learning_rate": 4.6213872832369945e-05, "loss": 0.4497, "step": 324 }, { "epoch": 0.5066250974279034, "grad_norm": 0.36187280391895393, "learning_rate": 4.61849710982659e-05, "loss": 0.4536, "step": 325 }, { "epoch": 0.5081839438815277, "grad_norm": 0.4400520460188613, "learning_rate": 4.615606936416185e-05, "loss": 0.4608, "step": 326 }, { "epoch": 0.509742790335152, "grad_norm": 0.3663575693429021, "learning_rate": 4.6127167630057805e-05, "loss": 0.4626, "step": 327 }, { "epoch": 0.5113016367887763, "grad_norm": 0.4766184153960843, "learning_rate": 4.609826589595376e-05, "loss": 0.4773, "step": 328 }, { "epoch": 0.5128604832424006, "grad_norm": 0.45206345178046475, "learning_rate": 4.606936416184972e-05, "loss": 0.4469, "step": 329 }, { "epoch": 0.514419329696025, "grad_norm": 0.39026528576069053, "learning_rate": 4.6040462427745665e-05, "loss": 0.4604, "step": 330 }, { "epoch": 0.5159781761496492, "grad_norm": 0.5361204288029048, "learning_rate": 4.601156069364162e-05, "loss": 0.4613, "step": 331 }, { "epoch": 0.5175370226032736, "grad_norm": 0.43012235247767305, "learning_rate": 4.598265895953757e-05, "loss": 0.4579, "step": 332 }, { "epoch": 0.5190958690568979, "grad_norm": 0.527243694253446, "learning_rate": 4.595375722543353e-05, "loss": 0.4602, "step": 333 }, { "epoch": 0.5206547155105222, "grad_norm": 0.4179764821236328, "learning_rate": 4.5924855491329486e-05, "loss": 0.4585, "step": 334 }, { "epoch": 0.5222135619641465, "grad_norm": 0.4510997406406127, "learning_rate": 4.589595375722543e-05, "loss": 0.4636, "step": 335 }, { "epoch": 0.5237724084177708, "grad_norm": 0.3187587883976425, "learning_rate": 4.5867052023121386e-05, "loss": 0.4612, "step": 336 }, { "epoch": 0.5253312548713952, "grad_norm": 0.46286631819853313, "learning_rate": 4.5838150289017346e-05, "loss": 0.4455, "step": 337 }, { "epoch": 0.5268901013250195, "grad_norm": 0.43393995615442665, "learning_rate": 4.58092485549133e-05, "loss": 0.481, "step": 338 }, { "epoch": 0.5284489477786438, "grad_norm": 0.5297318671699833, "learning_rate": 4.578034682080925e-05, "loss": 0.464, "step": 339 }, { "epoch": 0.5300077942322681, "grad_norm": 0.42416355410040585, "learning_rate": 4.57514450867052e-05, "loss": 0.4644, "step": 340 }, { "epoch": 0.5315666406858924, "grad_norm": 0.3712827662390051, "learning_rate": 4.572254335260116e-05, "loss": 0.4675, "step": 341 }, { "epoch": 0.5331254871395168, "grad_norm": 0.39762495205102294, "learning_rate": 4.569364161849711e-05, "loss": 0.469, "step": 342 }, { "epoch": 0.534684333593141, "grad_norm": 0.3254212965910093, "learning_rate": 4.566473988439307e-05, "loss": 0.4534, "step": 343 }, { "epoch": 0.5362431800467654, "grad_norm": 0.4848016190354553, "learning_rate": 4.563583815028902e-05, "loss": 0.4616, "step": 344 }, { "epoch": 0.5378020265003897, "grad_norm": 0.31661190588364, "learning_rate": 4.5606936416184974e-05, "loss": 0.4568, "step": 345 }, { "epoch": 0.539360872954014, "grad_norm": 0.4676445060911043, "learning_rate": 4.557803468208093e-05, "loss": 0.472, "step": 346 }, { "epoch": 0.5409197194076384, "grad_norm": 0.3160669772415641, "learning_rate": 4.554913294797688e-05, "loss": 0.444, "step": 347 }, { "epoch": 0.5424785658612626, "grad_norm": 0.4323013049552334, "learning_rate": 4.5520231213872834e-05, "loss": 0.4392, "step": 348 }, { "epoch": 0.544037412314887, "grad_norm": 0.4523674768546962, "learning_rate": 4.549132947976879e-05, "loss": 0.4682, "step": 349 }, { "epoch": 0.5455962587685113, "grad_norm": 0.3575280341982122, "learning_rate": 4.546242774566474e-05, "loss": 0.4483, "step": 350 }, { "epoch": 0.5471551052221356, "grad_norm": 0.4418729569254676, "learning_rate": 4.5433526011560694e-05, "loss": 0.4489, "step": 351 }, { "epoch": 0.54871395167576, "grad_norm": 0.3280270922434657, "learning_rate": 4.540462427745665e-05, "loss": 0.4616, "step": 352 }, { "epoch": 0.5502727981293842, "grad_norm": 0.4873805692806027, "learning_rate": 4.537572254335261e-05, "loss": 0.4802, "step": 353 }, { "epoch": 0.5518316445830086, "grad_norm": 0.36573476600102434, "learning_rate": 4.5346820809248555e-05, "loss": 0.475, "step": 354 }, { "epoch": 0.5533904910366328, "grad_norm": 0.41050124588561904, "learning_rate": 4.531791907514451e-05, "loss": 0.4535, "step": 355 }, { "epoch": 0.5549493374902572, "grad_norm": 0.3620545249321528, "learning_rate": 4.528901734104046e-05, "loss": 0.4428, "step": 356 }, { "epoch": 0.5565081839438816, "grad_norm": 0.467231893264037, "learning_rate": 4.526011560693642e-05, "loss": 0.4566, "step": 357 }, { "epoch": 0.5580670303975058, "grad_norm": 0.4030396125824727, "learning_rate": 4.5231213872832375e-05, "loss": 0.4717, "step": 358 }, { "epoch": 0.5596258768511302, "grad_norm": 0.34726035881210865, "learning_rate": 4.520231213872833e-05, "loss": 0.4497, "step": 359 }, { "epoch": 0.5611847233047544, "grad_norm": 0.39289264377072153, "learning_rate": 4.5173410404624275e-05, "loss": 0.4318, "step": 360 }, { "epoch": 0.5627435697583788, "grad_norm": 0.3890869555190299, "learning_rate": 4.5144508670520236e-05, "loss": 0.4569, "step": 361 }, { "epoch": 0.5643024162120032, "grad_norm": 0.39781174379277523, "learning_rate": 4.511560693641619e-05, "loss": 0.4638, "step": 362 }, { "epoch": 0.5658612626656274, "grad_norm": 0.36372440090872343, "learning_rate": 4.508670520231214e-05, "loss": 0.441, "step": 363 }, { "epoch": 0.5674201091192518, "grad_norm": 0.4284734591617738, "learning_rate": 4.5057803468208096e-05, "loss": 0.4745, "step": 364 }, { "epoch": 0.568978955572876, "grad_norm": 0.3838382539954402, "learning_rate": 4.502890173410405e-05, "loss": 0.4538, "step": 365 }, { "epoch": 0.5705378020265004, "grad_norm": 0.4837580518409696, "learning_rate": 4.5e-05, "loss": 0.4291, "step": 366 }, { "epoch": 0.5720966484801248, "grad_norm": 0.3807661371797664, "learning_rate": 4.4971098265895956e-05, "loss": 0.4681, "step": 367 }, { "epoch": 0.573655494933749, "grad_norm": 0.4409972991204447, "learning_rate": 4.494219653179191e-05, "loss": 0.4425, "step": 368 }, { "epoch": 0.5752143413873734, "grad_norm": 0.371972546132041, "learning_rate": 4.491329479768786e-05, "loss": 0.4543, "step": 369 }, { "epoch": 0.5767731878409976, "grad_norm": 0.37683721417276733, "learning_rate": 4.488439306358382e-05, "loss": 0.4575, "step": 370 }, { "epoch": 0.578332034294622, "grad_norm": 0.45855850398477666, "learning_rate": 4.485549132947977e-05, "loss": 0.4847, "step": 371 }, { "epoch": 0.5798908807482464, "grad_norm": 0.4223420944336656, "learning_rate": 4.4826589595375724e-05, "loss": 0.4649, "step": 372 }, { "epoch": 0.5814497272018706, "grad_norm": 0.4597887227283376, "learning_rate": 4.4797687861271684e-05, "loss": 0.4828, "step": 373 }, { "epoch": 0.583008573655495, "grad_norm": 0.4998303192690309, "learning_rate": 4.476878612716763e-05, "loss": 0.452, "step": 374 }, { "epoch": 0.5845674201091192, "grad_norm": 0.3756104589899762, "learning_rate": 4.4739884393063584e-05, "loss": 0.4411, "step": 375 }, { "epoch": 0.5861262665627436, "grad_norm": 0.4224811548039192, "learning_rate": 4.471098265895954e-05, "loss": 0.4564, "step": 376 }, { "epoch": 0.5876851130163678, "grad_norm": 0.4569199095589889, "learning_rate": 4.46820809248555e-05, "loss": 0.4664, "step": 377 }, { "epoch": 0.5892439594699922, "grad_norm": 0.3608347158374211, "learning_rate": 4.465317919075145e-05, "loss": 0.4734, "step": 378 }, { "epoch": 0.5908028059236166, "grad_norm": 0.4552865050320062, "learning_rate": 4.46242774566474e-05, "loss": 0.4723, "step": 379 }, { "epoch": 0.5923616523772408, "grad_norm": 0.3539125767256611, "learning_rate": 4.459537572254335e-05, "loss": 0.4762, "step": 380 }, { "epoch": 0.5939204988308652, "grad_norm": 0.3880352014434992, "learning_rate": 4.456647398843931e-05, "loss": 0.4388, "step": 381 }, { "epoch": 0.5954793452844894, "grad_norm": 0.37410095368321145, "learning_rate": 4.4537572254335265e-05, "loss": 0.4533, "step": 382 }, { "epoch": 0.5970381917381138, "grad_norm": 0.3133270680014226, "learning_rate": 4.450867052023122e-05, "loss": 0.4525, "step": 383 }, { "epoch": 0.5985970381917382, "grad_norm": 0.41920980039661443, "learning_rate": 4.447976878612717e-05, "loss": 0.4333, "step": 384 }, { "epoch": 0.6001558846453624, "grad_norm": 0.3774275243895145, "learning_rate": 4.4450867052023125e-05, "loss": 0.4459, "step": 385 }, { "epoch": 0.6017147310989868, "grad_norm": 0.369753586583569, "learning_rate": 4.442196531791908e-05, "loss": 0.4628, "step": 386 }, { "epoch": 0.603273577552611, "grad_norm": 0.43933813534803284, "learning_rate": 4.439306358381503e-05, "loss": 0.4866, "step": 387 }, { "epoch": 0.6048324240062354, "grad_norm": 0.36877162976354194, "learning_rate": 4.4364161849710985e-05, "loss": 0.4489, "step": 388 }, { "epoch": 0.6063912704598597, "grad_norm": 0.36797218493960576, "learning_rate": 4.433526011560694e-05, "loss": 0.4681, "step": 389 }, { "epoch": 0.607950116913484, "grad_norm": 0.3598201801544274, "learning_rate": 4.430635838150289e-05, "loss": 0.4451, "step": 390 }, { "epoch": 0.6095089633671084, "grad_norm": 0.37415526691302536, "learning_rate": 4.4277456647398846e-05, "loss": 0.4649, "step": 391 }, { "epoch": 0.6110678098207326, "grad_norm": 0.3217861980835781, "learning_rate": 4.42485549132948e-05, "loss": 0.4579, "step": 392 }, { "epoch": 0.612626656274357, "grad_norm": 0.4455765725783138, "learning_rate": 4.421965317919075e-05, "loss": 0.4603, "step": 393 }, { "epoch": 0.6141855027279813, "grad_norm": 0.37219414271910367, "learning_rate": 4.4190751445086706e-05, "loss": 0.4548, "step": 394 }, { "epoch": 0.6157443491816056, "grad_norm": 0.45080978894718876, "learning_rate": 4.416184971098266e-05, "loss": 0.4504, "step": 395 }, { "epoch": 0.61730319563523, "grad_norm": 0.4358441423072049, "learning_rate": 4.413294797687861e-05, "loss": 0.4446, "step": 396 }, { "epoch": 0.6188620420888542, "grad_norm": 0.37738329645269036, "learning_rate": 4.4104046242774566e-05, "loss": 0.4436, "step": 397 }, { "epoch": 0.6204208885424786, "grad_norm": 0.4204646254215696, "learning_rate": 4.407514450867053e-05, "loss": 0.4512, "step": 398 }, { "epoch": 0.6219797349961029, "grad_norm": 0.39901225883937147, "learning_rate": 4.404624277456647e-05, "loss": 0.463, "step": 399 }, { "epoch": 0.6235385814497272, "grad_norm": 0.320265409159274, "learning_rate": 4.401734104046243e-05, "loss": 0.4329, "step": 400 }, { "epoch": 0.6250974279033515, "grad_norm": 0.49013995892880263, "learning_rate": 4.398843930635838e-05, "loss": 0.4542, "step": 401 }, { "epoch": 0.6266562743569758, "grad_norm": 0.3510400983584625, "learning_rate": 4.395953757225434e-05, "loss": 0.4589, "step": 402 }, { "epoch": 0.6282151208106002, "grad_norm": 0.4402933914634823, "learning_rate": 4.3930635838150294e-05, "loss": 0.4593, "step": 403 }, { "epoch": 0.6297739672642245, "grad_norm": 0.47377011043879347, "learning_rate": 4.390173410404624e-05, "loss": 0.4675, "step": 404 }, { "epoch": 0.6313328137178488, "grad_norm": 0.37916907989218646, "learning_rate": 4.3872832369942194e-05, "loss": 0.4412, "step": 405 }, { "epoch": 0.6328916601714731, "grad_norm": 0.35621367384404673, "learning_rate": 4.3843930635838154e-05, "loss": 0.4618, "step": 406 }, { "epoch": 0.6344505066250974, "grad_norm": 0.3834554558658467, "learning_rate": 4.381502890173411e-05, "loss": 0.4591, "step": 407 }, { "epoch": 0.6360093530787218, "grad_norm": 0.3273576509892712, "learning_rate": 4.378612716763006e-05, "loss": 0.4307, "step": 408 }, { "epoch": 0.6375681995323461, "grad_norm": 0.3260428886276838, "learning_rate": 4.3757225433526015e-05, "loss": 0.4315, "step": 409 }, { "epoch": 0.6391270459859704, "grad_norm": 0.3828677152566117, "learning_rate": 4.372832369942197e-05, "loss": 0.431, "step": 410 }, { "epoch": 0.6406858924395947, "grad_norm": 0.3428782387615178, "learning_rate": 4.369942196531792e-05, "loss": 0.4535, "step": 411 }, { "epoch": 0.642244738893219, "grad_norm": 0.3568689766352524, "learning_rate": 4.3670520231213875e-05, "loss": 0.4471, "step": 412 }, { "epoch": 0.6438035853468433, "grad_norm": 0.3200198497931682, "learning_rate": 4.364161849710983e-05, "loss": 0.4675, "step": 413 }, { "epoch": 0.6453624318004677, "grad_norm": 0.3439904016830579, "learning_rate": 4.361271676300578e-05, "loss": 0.4434, "step": 414 }, { "epoch": 0.646921278254092, "grad_norm": 0.3910285780339108, "learning_rate": 4.3583815028901735e-05, "loss": 0.4291, "step": 415 }, { "epoch": 0.6484801247077163, "grad_norm": 0.35633702898935654, "learning_rate": 4.355491329479769e-05, "loss": 0.4465, "step": 416 }, { "epoch": 0.6500389711613406, "grad_norm": 0.4280187254199856, "learning_rate": 4.352601156069364e-05, "loss": 0.4467, "step": 417 }, { "epoch": 0.6515978176149649, "grad_norm": 0.3860941239745589, "learning_rate": 4.34971098265896e-05, "loss": 0.4443, "step": 418 }, { "epoch": 0.6531566640685893, "grad_norm": 0.44940034037874554, "learning_rate": 4.346820809248555e-05, "loss": 0.48, "step": 419 }, { "epoch": 0.6547155105222136, "grad_norm": 0.4625797028642205, "learning_rate": 4.34393063583815e-05, "loss": 0.4568, "step": 420 }, { "epoch": 0.6562743569758379, "grad_norm": 0.36493835250776474, "learning_rate": 4.3410404624277456e-05, "loss": 0.4622, "step": 421 }, { "epoch": 0.6578332034294622, "grad_norm": 0.4451390677090527, "learning_rate": 4.3381502890173416e-05, "loss": 0.4399, "step": 422 }, { "epoch": 0.6593920498830865, "grad_norm": 0.40509517400153716, "learning_rate": 4.335260115606937e-05, "loss": 0.4548, "step": 423 }, { "epoch": 0.6609508963367108, "grad_norm": 0.4745439842667765, "learning_rate": 4.3323699421965316e-05, "loss": 0.4409, "step": 424 }, { "epoch": 0.6625097427903351, "grad_norm": 0.40589601301353223, "learning_rate": 4.329479768786127e-05, "loss": 0.4386, "step": 425 }, { "epoch": 0.6640685892439595, "grad_norm": 0.410367648733596, "learning_rate": 4.326589595375723e-05, "loss": 0.4359, "step": 426 }, { "epoch": 0.6656274356975838, "grad_norm": 0.39329540338329594, "learning_rate": 4.3236994219653183e-05, "loss": 0.4503, "step": 427 }, { "epoch": 0.6671862821512081, "grad_norm": 0.4110086370635056, "learning_rate": 4.320809248554914e-05, "loss": 0.4447, "step": 428 }, { "epoch": 0.6687451286048324, "grad_norm": 0.47772877851625145, "learning_rate": 4.3179190751445084e-05, "loss": 0.4687, "step": 429 }, { "epoch": 0.6703039750584567, "grad_norm": 0.3866102449562328, "learning_rate": 4.3150289017341044e-05, "loss": 0.4393, "step": 430 }, { "epoch": 0.6718628215120811, "grad_norm": 0.45739240597886627, "learning_rate": 4.3121387283237e-05, "loss": 0.4668, "step": 431 }, { "epoch": 0.6734216679657054, "grad_norm": 0.4314476402559539, "learning_rate": 4.309248554913295e-05, "loss": 0.4649, "step": 432 }, { "epoch": 0.6749805144193297, "grad_norm": 0.45634509121859473, "learning_rate": 4.3063583815028904e-05, "loss": 0.4441, "step": 433 }, { "epoch": 0.676539360872954, "grad_norm": 0.4811305736485706, "learning_rate": 4.303468208092486e-05, "loss": 0.4529, "step": 434 }, { "epoch": 0.6780982073265783, "grad_norm": 0.39207564033387, "learning_rate": 4.300578034682081e-05, "loss": 0.4588, "step": 435 }, { "epoch": 0.6796570537802027, "grad_norm": 0.4658600522478725, "learning_rate": 4.2976878612716764e-05, "loss": 0.4379, "step": 436 }, { "epoch": 0.6812159002338269, "grad_norm": 0.41119750867971183, "learning_rate": 4.294797687861272e-05, "loss": 0.4418, "step": 437 }, { "epoch": 0.6827747466874513, "grad_norm": 0.4230535920977332, "learning_rate": 4.291907514450868e-05, "loss": 0.4536, "step": 438 }, { "epoch": 0.6843335931410756, "grad_norm": 0.5092465838859024, "learning_rate": 4.2890173410404625e-05, "loss": 0.4584, "step": 439 }, { "epoch": 0.6858924395946999, "grad_norm": 0.3601271310299026, "learning_rate": 4.286127167630058e-05, "loss": 0.4368, "step": 440 }, { "epoch": 0.6874512860483243, "grad_norm": 0.5033625655714722, "learning_rate": 4.283236994219653e-05, "loss": 0.4484, "step": 441 }, { "epoch": 0.6890101325019485, "grad_norm": 0.3994249544123872, "learning_rate": 4.280346820809249e-05, "loss": 0.4418, "step": 442 }, { "epoch": 0.6905689789555729, "grad_norm": 0.45539241489062715, "learning_rate": 4.2774566473988445e-05, "loss": 0.4473, "step": 443 }, { "epoch": 0.6921278254091972, "grad_norm": 0.4417751638358608, "learning_rate": 4.274566473988439e-05, "loss": 0.4496, "step": 444 }, { "epoch": 0.6936866718628215, "grad_norm": 0.40882073328312935, "learning_rate": 4.2716763005780345e-05, "loss": 0.4622, "step": 445 }, { "epoch": 0.6952455183164459, "grad_norm": 0.39444852000587594, "learning_rate": 4.2687861271676306e-05, "loss": 0.4769, "step": 446 }, { "epoch": 0.6968043647700701, "grad_norm": 0.37669038630165647, "learning_rate": 4.265895953757226e-05, "loss": 0.4463, "step": 447 }, { "epoch": 0.6983632112236945, "grad_norm": 0.4180560176644135, "learning_rate": 4.263005780346821e-05, "loss": 0.4502, "step": 448 }, { "epoch": 0.6999220576773187, "grad_norm": 0.3606230739957135, "learning_rate": 4.260115606936416e-05, "loss": 0.4686, "step": 449 }, { "epoch": 0.7014809041309431, "grad_norm": 0.4763884079373346, "learning_rate": 4.257225433526012e-05, "loss": 0.4465, "step": 450 }, { "epoch": 0.7030397505845675, "grad_norm": 0.36030780349125946, "learning_rate": 4.254335260115607e-05, "loss": 0.4475, "step": 451 }, { "epoch": 0.7045985970381917, "grad_norm": 0.4451176525342345, "learning_rate": 4.2514450867052026e-05, "loss": 0.4268, "step": 452 }, { "epoch": 0.7061574434918161, "grad_norm": 0.33875598529236145, "learning_rate": 4.248554913294798e-05, "loss": 0.4499, "step": 453 }, { "epoch": 0.7077162899454403, "grad_norm": 0.5071769434422747, "learning_rate": 4.245664739884393e-05, "loss": 0.4537, "step": 454 }, { "epoch": 0.7092751363990647, "grad_norm": 0.3488438534196029, "learning_rate": 4.242774566473989e-05, "loss": 0.4623, "step": 455 }, { "epoch": 0.7108339828526891, "grad_norm": 0.4454725104594618, "learning_rate": 4.239884393063584e-05, "loss": 0.4516, "step": 456 }, { "epoch": 0.7123928293063133, "grad_norm": 0.3583891995078908, "learning_rate": 4.2369942196531794e-05, "loss": 0.4581, "step": 457 }, { "epoch": 0.7139516757599377, "grad_norm": 0.36757937011684644, "learning_rate": 4.234104046242775e-05, "loss": 0.4505, "step": 458 }, { "epoch": 0.7155105222135619, "grad_norm": 0.39437787323637835, "learning_rate": 4.23121387283237e-05, "loss": 0.4555, "step": 459 }, { "epoch": 0.7170693686671863, "grad_norm": 0.3526260112578523, "learning_rate": 4.2283236994219654e-05, "loss": 0.4462, "step": 460 }, { "epoch": 0.7186282151208107, "grad_norm": 0.3610517367016722, "learning_rate": 4.225433526011561e-05, "loss": 0.4439, "step": 461 }, { "epoch": 0.7201870615744349, "grad_norm": 0.33046116660611863, "learning_rate": 4.222543352601157e-05, "loss": 0.4553, "step": 462 }, { "epoch": 0.7217459080280593, "grad_norm": 0.3479382421778036, "learning_rate": 4.2196531791907514e-05, "loss": 0.4615, "step": 463 }, { "epoch": 0.7233047544816835, "grad_norm": 0.3563546046165941, "learning_rate": 4.216763005780347e-05, "loss": 0.4502, "step": 464 }, { "epoch": 0.7248636009353079, "grad_norm": 0.33104574079840804, "learning_rate": 4.213872832369942e-05, "loss": 0.4522, "step": 465 }, { "epoch": 0.7264224473889321, "grad_norm": 0.3615689785424514, "learning_rate": 4.210982658959538e-05, "loss": 0.4456, "step": 466 }, { "epoch": 0.7279812938425565, "grad_norm": 0.33245430385705754, "learning_rate": 4.2080924855491335e-05, "loss": 0.451, "step": 467 }, { "epoch": 0.7295401402961809, "grad_norm": 0.31570139553177473, "learning_rate": 4.205202312138729e-05, "loss": 0.4387, "step": 468 }, { "epoch": 0.7310989867498051, "grad_norm": 0.38612290372770497, "learning_rate": 4.2023121387283235e-05, "loss": 0.4391, "step": 469 }, { "epoch": 0.7326578332034295, "grad_norm": 0.4573667515885595, "learning_rate": 4.1994219653179195e-05, "loss": 0.4689, "step": 470 }, { "epoch": 0.7342166796570537, "grad_norm": 0.3807562323188079, "learning_rate": 4.196531791907515e-05, "loss": 0.4632, "step": 471 }, { "epoch": 0.7357755261106781, "grad_norm": 0.3525392715340891, "learning_rate": 4.19364161849711e-05, "loss": 0.4381, "step": 472 }, { "epoch": 0.7373343725643025, "grad_norm": 0.3782760064170116, "learning_rate": 4.1907514450867055e-05, "loss": 0.4477, "step": 473 }, { "epoch": 0.7388932190179267, "grad_norm": 0.41115398176210943, "learning_rate": 4.187861271676301e-05, "loss": 0.4602, "step": 474 }, { "epoch": 0.7404520654715511, "grad_norm": 0.32624650584367754, "learning_rate": 4.184971098265896e-05, "loss": 0.4569, "step": 475 }, { "epoch": 0.7420109119251753, "grad_norm": 0.378893068450349, "learning_rate": 4.1820809248554916e-05, "loss": 0.4489, "step": 476 }, { "epoch": 0.7435697583787997, "grad_norm": 0.3829792535253302, "learning_rate": 4.179190751445087e-05, "loss": 0.4466, "step": 477 }, { "epoch": 0.745128604832424, "grad_norm": 0.33863233675349563, "learning_rate": 4.176300578034682e-05, "loss": 0.4383, "step": 478 }, { "epoch": 0.7466874512860483, "grad_norm": 0.40164735104812754, "learning_rate": 4.1734104046242776e-05, "loss": 0.4203, "step": 479 }, { "epoch": 0.7482462977396727, "grad_norm": 0.35580789823430714, "learning_rate": 4.170520231213873e-05, "loss": 0.4744, "step": 480 }, { "epoch": 0.7498051441932969, "grad_norm": 0.31246598578051804, "learning_rate": 4.167630057803468e-05, "loss": 0.4219, "step": 481 }, { "epoch": 0.7513639906469213, "grad_norm": 0.329237795782184, "learning_rate": 4.1647398843930636e-05, "loss": 0.4462, "step": 482 }, { "epoch": 0.7529228371005456, "grad_norm": 0.32134761939251955, "learning_rate": 4.161849710982659e-05, "loss": 0.4533, "step": 483 }, { "epoch": 0.7544816835541699, "grad_norm": 0.34349515426855415, "learning_rate": 4.1589595375722543e-05, "loss": 0.4381, "step": 484 }, { "epoch": 0.7560405300077943, "grad_norm": 0.3706250923736234, "learning_rate": 4.15606936416185e-05, "loss": 0.4406, "step": 485 }, { "epoch": 0.7575993764614185, "grad_norm": 0.30867896526600563, "learning_rate": 4.153179190751445e-05, "loss": 0.4328, "step": 486 }, { "epoch": 0.7591582229150429, "grad_norm": 0.3756083443390548, "learning_rate": 4.150289017341041e-05, "loss": 0.4593, "step": 487 }, { "epoch": 0.7607170693686672, "grad_norm": 0.3359590028525238, "learning_rate": 4.147398843930636e-05, "loss": 0.4452, "step": 488 }, { "epoch": 0.7622759158222915, "grad_norm": 0.3805120865629876, "learning_rate": 4.144508670520231e-05, "loss": 0.448, "step": 489 }, { "epoch": 0.7638347622759158, "grad_norm": 0.36856409628466563, "learning_rate": 4.1416184971098264e-05, "loss": 0.4361, "step": 490 }, { "epoch": 0.7653936087295401, "grad_norm": 0.3130789506998712, "learning_rate": 4.1387283236994224e-05, "loss": 0.4422, "step": 491 }, { "epoch": 0.7669524551831645, "grad_norm": 0.3420964053109533, "learning_rate": 4.135838150289018e-05, "loss": 0.44, "step": 492 }, { "epoch": 0.7685113016367888, "grad_norm": 0.3761751528633877, "learning_rate": 4.132947976878613e-05, "loss": 0.4512, "step": 493 }, { "epoch": 0.7700701480904131, "grad_norm": 0.3217947372011256, "learning_rate": 4.130057803468208e-05, "loss": 0.4542, "step": 494 }, { "epoch": 0.7716289945440374, "grad_norm": 0.36411196358092746, "learning_rate": 4.127167630057804e-05, "loss": 0.4453, "step": 495 }, { "epoch": 0.7731878409976617, "grad_norm": 0.36062847920871005, "learning_rate": 4.124277456647399e-05, "loss": 0.432, "step": 496 }, { "epoch": 0.774746687451286, "grad_norm": 0.31936162560311365, "learning_rate": 4.1213872832369945e-05, "loss": 0.4477, "step": 497 }, { "epoch": 0.7763055339049104, "grad_norm": 0.3761636426405107, "learning_rate": 4.11849710982659e-05, "loss": 0.4353, "step": 498 }, { "epoch": 0.7778643803585347, "grad_norm": 0.32932231835478515, "learning_rate": 4.115606936416185e-05, "loss": 0.4467, "step": 499 }, { "epoch": 0.779423226812159, "grad_norm": 0.38070437317284483, "learning_rate": 4.1127167630057805e-05, "loss": 0.4395, "step": 500 }, { "epoch": 0.7809820732657833, "grad_norm": 0.3246837506466452, "learning_rate": 4.109826589595376e-05, "loss": 0.4447, "step": 501 }, { "epoch": 0.7825409197194076, "grad_norm": 0.34953071175085054, "learning_rate": 4.106936416184971e-05, "loss": 0.4333, "step": 502 }, { "epoch": 0.784099766173032, "grad_norm": 0.37305833228096014, "learning_rate": 4.1040462427745666e-05, "loss": 0.4314, "step": 503 }, { "epoch": 0.7856586126266563, "grad_norm": 0.3472419499146332, "learning_rate": 4.101156069364162e-05, "loss": 0.4363, "step": 504 }, { "epoch": 0.7872174590802806, "grad_norm": 0.38866158249187055, "learning_rate": 4.098265895953757e-05, "loss": 0.4429, "step": 505 }, { "epoch": 0.7887763055339049, "grad_norm": 0.330824609948283, "learning_rate": 4.0953757225433526e-05, "loss": 0.4442, "step": 506 }, { "epoch": 0.7903351519875292, "grad_norm": 0.39663219339611827, "learning_rate": 4.0924855491329486e-05, "loss": 0.4434, "step": 507 }, { "epoch": 0.7918939984411536, "grad_norm": 0.3619324831325355, "learning_rate": 4.089595375722543e-05, "loss": 0.4416, "step": 508 }, { "epoch": 0.7934528448947779, "grad_norm": 0.35246995122654845, "learning_rate": 4.0867052023121386e-05, "loss": 0.4265, "step": 509 }, { "epoch": 0.7950116913484022, "grad_norm": 0.3612965001810029, "learning_rate": 4.083815028901734e-05, "loss": 0.433, "step": 510 }, { "epoch": 0.7965705378020265, "grad_norm": 0.38138895157014485, "learning_rate": 4.08092485549133e-05, "loss": 0.4341, "step": 511 }, { "epoch": 0.7981293842556508, "grad_norm": 0.3449144364761348, "learning_rate": 4.0780346820809253e-05, "loss": 0.457, "step": 512 }, { "epoch": 0.7996882307092751, "grad_norm": 0.4015780437671286, "learning_rate": 4.07514450867052e-05, "loss": 0.43, "step": 513 }, { "epoch": 0.8012470771628994, "grad_norm": 0.31072059103382504, "learning_rate": 4.0722543352601154e-05, "loss": 0.4424, "step": 514 }, { "epoch": 0.8028059236165238, "grad_norm": 0.3758869241826082, "learning_rate": 4.0693641618497114e-05, "loss": 0.4568, "step": 515 }, { "epoch": 0.8043647700701481, "grad_norm": 0.3383461562057852, "learning_rate": 4.066473988439307e-05, "loss": 0.4513, "step": 516 }, { "epoch": 0.8059236165237724, "grad_norm": 0.35697829825213634, "learning_rate": 4.063583815028902e-05, "loss": 0.4512, "step": 517 }, { "epoch": 0.8074824629773967, "grad_norm": 0.3930071983033836, "learning_rate": 4.0606936416184974e-05, "loss": 0.4386, "step": 518 }, { "epoch": 0.809041309431021, "grad_norm": 0.34936814110167913, "learning_rate": 4.057803468208093e-05, "loss": 0.4249, "step": 519 }, { "epoch": 0.8106001558846454, "grad_norm": 0.3477156813236279, "learning_rate": 4.054913294797688e-05, "loss": 0.4421, "step": 520 }, { "epoch": 0.8121590023382697, "grad_norm": 0.3771551756284452, "learning_rate": 4.0520231213872834e-05, "loss": 0.4422, "step": 521 }, { "epoch": 0.813717848791894, "grad_norm": 0.3915428314100825, "learning_rate": 4.049132947976879e-05, "loss": 0.4489, "step": 522 }, { "epoch": 0.8152766952455183, "grad_norm": 0.3438944943929536, "learning_rate": 4.046242774566474e-05, "loss": 0.4403, "step": 523 }, { "epoch": 0.8168355416991426, "grad_norm": 0.41104576558641404, "learning_rate": 4.0433526011560695e-05, "loss": 0.4606, "step": 524 }, { "epoch": 0.818394388152767, "grad_norm": 0.37712289992365605, "learning_rate": 4.040462427745665e-05, "loss": 0.4325, "step": 525 }, { "epoch": 0.8199532346063912, "grad_norm": 0.3008825974501281, "learning_rate": 4.03757225433526e-05, "loss": 0.4348, "step": 526 }, { "epoch": 0.8215120810600156, "grad_norm": 0.3701724452827897, "learning_rate": 4.034682080924856e-05, "loss": 0.444, "step": 527 }, { "epoch": 0.8230709275136399, "grad_norm": 0.4810485926905224, "learning_rate": 4.031791907514451e-05, "loss": 0.4608, "step": 528 }, { "epoch": 0.8246297739672642, "grad_norm": 0.3549215989456713, "learning_rate": 4.028901734104046e-05, "loss": 0.4539, "step": 529 }, { "epoch": 0.8261886204208886, "grad_norm": 0.37028937018443525, "learning_rate": 4.0260115606936415e-05, "loss": 0.4485, "step": 530 }, { "epoch": 0.8277474668745128, "grad_norm": 0.4323084487068229, "learning_rate": 4.0231213872832376e-05, "loss": 0.4622, "step": 531 }, { "epoch": 0.8293063133281372, "grad_norm": 0.37099996438121063, "learning_rate": 4.020231213872833e-05, "loss": 0.4389, "step": 532 }, { "epoch": 0.8308651597817615, "grad_norm": 0.4485476616523096, "learning_rate": 4.0173410404624276e-05, "loss": 0.458, "step": 533 }, { "epoch": 0.8324240062353858, "grad_norm": 0.47843701934513816, "learning_rate": 4.014450867052023e-05, "loss": 0.4313, "step": 534 }, { "epoch": 0.8339828526890102, "grad_norm": 0.39097267691711673, "learning_rate": 4.011560693641619e-05, "loss": 0.457, "step": 535 }, { "epoch": 0.8355416991426344, "grad_norm": 0.44705834158808144, "learning_rate": 4.008670520231214e-05, "loss": 0.4411, "step": 536 }, { "epoch": 0.8371005455962588, "grad_norm": 0.39009820610398244, "learning_rate": 4.0057803468208096e-05, "loss": 0.4666, "step": 537 }, { "epoch": 0.838659392049883, "grad_norm": 0.40676152233104423, "learning_rate": 4.002890173410404e-05, "loss": 0.4464, "step": 538 }, { "epoch": 0.8402182385035074, "grad_norm": 0.38980475885919486, "learning_rate": 4e-05, "loss": 0.4287, "step": 539 }, { "epoch": 0.8417770849571318, "grad_norm": 0.4155553631446244, "learning_rate": 3.997109826589596e-05, "loss": 0.4592, "step": 540 }, { "epoch": 0.843335931410756, "grad_norm": 0.4683782615674356, "learning_rate": 3.994219653179191e-05, "loss": 0.4434, "step": 541 }, { "epoch": 0.8448947778643804, "grad_norm": 0.4270570741139377, "learning_rate": 3.9913294797687864e-05, "loss": 0.4471, "step": 542 }, { "epoch": 0.8464536243180046, "grad_norm": 0.3328227432824558, "learning_rate": 3.988439306358382e-05, "loss": 0.4342, "step": 543 }, { "epoch": 0.848012470771629, "grad_norm": 0.383218659320037, "learning_rate": 3.985549132947977e-05, "loss": 0.4418, "step": 544 }, { "epoch": 0.8495713172252534, "grad_norm": 0.41491251710244376, "learning_rate": 3.9826589595375724e-05, "loss": 0.4562, "step": 545 }, { "epoch": 0.8511301636788776, "grad_norm": 0.4581749465782902, "learning_rate": 3.979768786127168e-05, "loss": 0.4543, "step": 546 }, { "epoch": 0.852689010132502, "grad_norm": 0.39063683965855545, "learning_rate": 3.976878612716764e-05, "loss": 0.4403, "step": 547 }, { "epoch": 0.8542478565861262, "grad_norm": 0.4827779522726547, "learning_rate": 3.9739884393063584e-05, "loss": 0.4371, "step": 548 }, { "epoch": 0.8558067030397506, "grad_norm": 0.4619643532977416, "learning_rate": 3.971098265895954e-05, "loss": 0.4481, "step": 549 }, { "epoch": 0.857365549493375, "grad_norm": 0.38833379166687454, "learning_rate": 3.968208092485549e-05, "loss": 0.4298, "step": 550 }, { "epoch": 0.8589243959469992, "grad_norm": 0.47500496297331074, "learning_rate": 3.965317919075145e-05, "loss": 0.4242, "step": 551 }, { "epoch": 0.8604832424006236, "grad_norm": 0.43102309998132354, "learning_rate": 3.9624277456647405e-05, "loss": 0.4417, "step": 552 }, { "epoch": 0.8620420888542478, "grad_norm": 0.5161140253561636, "learning_rate": 3.959537572254335e-05, "loss": 0.4497, "step": 553 }, { "epoch": 0.8636009353078722, "grad_norm": 0.5456731985809448, "learning_rate": 3.9566473988439305e-05, "loss": 0.4179, "step": 554 }, { "epoch": 0.8651597817614964, "grad_norm": 0.4207485633634238, "learning_rate": 3.9537572254335265e-05, "loss": 0.4427, "step": 555 }, { "epoch": 0.8667186282151208, "grad_norm": 0.5759488662135421, "learning_rate": 3.950867052023122e-05, "loss": 0.4477, "step": 556 }, { "epoch": 0.8682774746687452, "grad_norm": 0.4021704450074726, "learning_rate": 3.947976878612717e-05, "loss": 0.4463, "step": 557 }, { "epoch": 0.8698363211223694, "grad_norm": 0.42177083348273137, "learning_rate": 3.945086705202312e-05, "loss": 0.4598, "step": 558 }, { "epoch": 0.8713951675759938, "grad_norm": 0.4701952158517483, "learning_rate": 3.942196531791908e-05, "loss": 0.4394, "step": 559 }, { "epoch": 0.872954014029618, "grad_norm": 0.44080844830392035, "learning_rate": 3.939306358381503e-05, "loss": 0.4493, "step": 560 }, { "epoch": 0.8745128604832424, "grad_norm": 0.3977306328304877, "learning_rate": 3.9364161849710986e-05, "loss": 0.4524, "step": 561 }, { "epoch": 0.8760717069368668, "grad_norm": 0.43018670105627926, "learning_rate": 3.933526011560694e-05, "loss": 0.4267, "step": 562 }, { "epoch": 0.877630553390491, "grad_norm": 0.4268384770791736, "learning_rate": 3.930635838150289e-05, "loss": 0.4607, "step": 563 }, { "epoch": 0.8791893998441154, "grad_norm": 0.38783009673051294, "learning_rate": 3.9277456647398846e-05, "loss": 0.4476, "step": 564 }, { "epoch": 0.8807482462977396, "grad_norm": 0.35687832758423854, "learning_rate": 3.92485549132948e-05, "loss": 0.435, "step": 565 }, { "epoch": 0.882307092751364, "grad_norm": 0.43764488183272643, "learning_rate": 3.921965317919075e-05, "loss": 0.4466, "step": 566 }, { "epoch": 0.8838659392049883, "grad_norm": 0.40005960295573123, "learning_rate": 3.9190751445086707e-05, "loss": 0.4288, "step": 567 }, { "epoch": 0.8854247856586126, "grad_norm": 0.46289540585836514, "learning_rate": 3.916184971098266e-05, "loss": 0.4545, "step": 568 }, { "epoch": 0.886983632112237, "grad_norm": 0.3824987032253999, "learning_rate": 3.9132947976878613e-05, "loss": 0.4414, "step": 569 }, { "epoch": 0.8885424785658612, "grad_norm": 0.47993501337835254, "learning_rate": 3.910404624277457e-05, "loss": 0.4411, "step": 570 }, { "epoch": 0.8901013250194856, "grad_norm": 0.3194097378426562, "learning_rate": 3.907514450867053e-05, "loss": 0.4532, "step": 571 }, { "epoch": 0.8916601714731099, "grad_norm": 0.5232202230976029, "learning_rate": 3.904624277456648e-05, "loss": 0.423, "step": 572 }, { "epoch": 0.8932190179267342, "grad_norm": 0.3126690343426845, "learning_rate": 3.901734104046243e-05, "loss": 0.4333, "step": 573 }, { "epoch": 0.8947778643803586, "grad_norm": 0.45718182082215625, "learning_rate": 3.898843930635838e-05, "loss": 0.4579, "step": 574 }, { "epoch": 0.8963367108339828, "grad_norm": 0.3578668564536035, "learning_rate": 3.895953757225434e-05, "loss": 0.4365, "step": 575 }, { "epoch": 0.8978955572876072, "grad_norm": 0.39528003293735503, "learning_rate": 3.8930635838150294e-05, "loss": 0.4373, "step": 576 }, { "epoch": 0.8994544037412315, "grad_norm": 0.4144241684007145, "learning_rate": 3.890173410404625e-05, "loss": 0.4261, "step": 577 }, { "epoch": 0.9010132501948558, "grad_norm": 0.37062244540768974, "learning_rate": 3.8872832369942194e-05, "loss": 0.4495, "step": 578 }, { "epoch": 0.9025720966484801, "grad_norm": 0.38183987870941294, "learning_rate": 3.884393063583815e-05, "loss": 0.4338, "step": 579 }, { "epoch": 0.9041309431021044, "grad_norm": 0.39082808942712405, "learning_rate": 3.881502890173411e-05, "loss": 0.4444, "step": 580 }, { "epoch": 0.9056897895557288, "grad_norm": 0.4477498201218956, "learning_rate": 3.878612716763006e-05, "loss": 0.4519, "step": 581 }, { "epoch": 0.9072486360093531, "grad_norm": 0.44239045295490875, "learning_rate": 3.8757225433526015e-05, "loss": 0.4481, "step": 582 }, { "epoch": 0.9088074824629774, "grad_norm": 0.3680081228407661, "learning_rate": 3.872832369942196e-05, "loss": 0.44, "step": 583 }, { "epoch": 0.9103663289166017, "grad_norm": 0.49017226032397, "learning_rate": 3.869942196531792e-05, "loss": 0.4523, "step": 584 }, { "epoch": 0.911925175370226, "grad_norm": 0.36889936915401544, "learning_rate": 3.8670520231213875e-05, "loss": 0.4252, "step": 585 }, { "epoch": 0.9134840218238504, "grad_norm": 0.42106930607963194, "learning_rate": 3.864161849710983e-05, "loss": 0.4366, "step": 586 }, { "epoch": 0.9150428682774747, "grad_norm": 0.33426016889258103, "learning_rate": 3.861271676300578e-05, "loss": 0.4364, "step": 587 }, { "epoch": 0.916601714731099, "grad_norm": 0.38283532007129306, "learning_rate": 3.8583815028901736e-05, "loss": 0.4299, "step": 588 }, { "epoch": 0.9181605611847233, "grad_norm": 0.3843431490728921, "learning_rate": 3.855491329479769e-05, "loss": 0.4423, "step": 589 }, { "epoch": 0.9197194076383476, "grad_norm": 0.39391225219289805, "learning_rate": 3.852601156069364e-05, "loss": 0.4553, "step": 590 }, { "epoch": 0.921278254091972, "grad_norm": 0.3921483022893861, "learning_rate": 3.8497109826589596e-05, "loss": 0.4374, "step": 591 }, { "epoch": 0.9228371005455963, "grad_norm": 0.4586348772859173, "learning_rate": 3.846820809248555e-05, "loss": 0.4391, "step": 592 }, { "epoch": 0.9243959469992206, "grad_norm": 0.3870941468579518, "learning_rate": 3.84393063583815e-05, "loss": 0.4588, "step": 593 }, { "epoch": 0.9259547934528449, "grad_norm": 0.4121790906090321, "learning_rate": 3.8410404624277456e-05, "loss": 0.4192, "step": 594 }, { "epoch": 0.9275136399064692, "grad_norm": 0.3287491996044717, "learning_rate": 3.838150289017341e-05, "loss": 0.4376, "step": 595 }, { "epoch": 0.9290724863600935, "grad_norm": 0.4668763195586917, "learning_rate": 3.835260115606937e-05, "loss": 0.4513, "step": 596 }, { "epoch": 0.9306313328137179, "grad_norm": 0.39213339962767724, "learning_rate": 3.832369942196532e-05, "loss": 0.4349, "step": 597 }, { "epoch": 0.9321901792673422, "grad_norm": 0.42578108338194876, "learning_rate": 3.829479768786127e-05, "loss": 0.4365, "step": 598 }, { "epoch": 0.9337490257209665, "grad_norm": 0.3949037922002061, "learning_rate": 3.8265895953757224e-05, "loss": 0.4366, "step": 599 }, { "epoch": 0.9353078721745908, "grad_norm": 0.3826228697949301, "learning_rate": 3.8236994219653184e-05, "loss": 0.4426, "step": 600 }, { "epoch": 0.9368667186282151, "grad_norm": 0.3650038749614227, "learning_rate": 3.820809248554914e-05, "loss": 0.4339, "step": 601 }, { "epoch": 0.9384255650818394, "grad_norm": 0.40293156157563836, "learning_rate": 3.817919075144509e-05, "loss": 0.4744, "step": 602 }, { "epoch": 0.9399844115354637, "grad_norm": 0.40210846514825743, "learning_rate": 3.815028901734104e-05, "loss": 0.4404, "step": 603 }, { "epoch": 0.9415432579890881, "grad_norm": 0.39097895288839746, "learning_rate": 3.8121387283237e-05, "loss": 0.4401, "step": 604 }, { "epoch": 0.9431021044427124, "grad_norm": 0.3227109176868314, "learning_rate": 3.809248554913295e-05, "loss": 0.4312, "step": 605 }, { "epoch": 0.9446609508963367, "grad_norm": 0.4608466312979325, "learning_rate": 3.8063583815028905e-05, "loss": 0.4591, "step": 606 }, { "epoch": 0.946219797349961, "grad_norm": 0.34756293281666123, "learning_rate": 3.803468208092486e-05, "loss": 0.4566, "step": 607 }, { "epoch": 0.9477786438035853, "grad_norm": 0.37558636760339875, "learning_rate": 3.800578034682081e-05, "loss": 0.4205, "step": 608 }, { "epoch": 0.9493374902572097, "grad_norm": 0.4205365075958774, "learning_rate": 3.7976878612716765e-05, "loss": 0.4388, "step": 609 }, { "epoch": 0.950896336710834, "grad_norm": 0.45096900527646666, "learning_rate": 3.794797687861272e-05, "loss": 0.4316, "step": 610 }, { "epoch": 0.9524551831644583, "grad_norm": 0.3812977384877062, "learning_rate": 3.791907514450867e-05, "loss": 0.4363, "step": 611 }, { "epoch": 0.9540140296180826, "grad_norm": 0.48866253778453445, "learning_rate": 3.7890173410404625e-05, "loss": 0.428, "step": 612 }, { "epoch": 0.9555728760717069, "grad_norm": 0.40499141402747296, "learning_rate": 3.786127167630058e-05, "loss": 0.4381, "step": 613 }, { "epoch": 0.9571317225253313, "grad_norm": 0.46430129913008983, "learning_rate": 3.783236994219653e-05, "loss": 0.4429, "step": 614 }, { "epoch": 0.9586905689789555, "grad_norm": 0.34966737213764454, "learning_rate": 3.7803468208092486e-05, "loss": 0.4381, "step": 615 }, { "epoch": 0.9602494154325799, "grad_norm": 0.4138090253112475, "learning_rate": 3.7774566473988446e-05, "loss": 0.4258, "step": 616 }, { "epoch": 0.9618082618862042, "grad_norm": 0.3297979382576517, "learning_rate": 3.774566473988439e-05, "loss": 0.4413, "step": 617 }, { "epoch": 0.9633671083398285, "grad_norm": 0.3863735923761598, "learning_rate": 3.7716763005780346e-05, "loss": 0.431, "step": 618 }, { "epoch": 0.9649259547934529, "grad_norm": 0.33528959794283103, "learning_rate": 3.76878612716763e-05, "loss": 0.4377, "step": 619 }, { "epoch": 0.9664848012470771, "grad_norm": 0.3292524974297176, "learning_rate": 3.765895953757226e-05, "loss": 0.4326, "step": 620 }, { "epoch": 0.9680436477007015, "grad_norm": 0.3156599159357699, "learning_rate": 3.763005780346821e-05, "loss": 0.4236, "step": 621 }, { "epoch": 0.9696024941543258, "grad_norm": 0.349891638695045, "learning_rate": 3.760115606936416e-05, "loss": 0.4226, "step": 622 }, { "epoch": 0.9711613406079501, "grad_norm": 0.3027036773941601, "learning_rate": 3.757225433526011e-05, "loss": 0.4357, "step": 623 }, { "epoch": 0.9727201870615745, "grad_norm": 0.3418983655828593, "learning_rate": 3.754335260115607e-05, "loss": 0.4072, "step": 624 }, { "epoch": 0.9742790335151987, "grad_norm": 0.32424716914194174, "learning_rate": 3.751445086705203e-05, "loss": 0.4229, "step": 625 }, { "epoch": 0.9758378799688231, "grad_norm": 0.35397027714852347, "learning_rate": 3.748554913294798e-05, "loss": 0.443, "step": 626 }, { "epoch": 0.9773967264224473, "grad_norm": 0.35983171780105105, "learning_rate": 3.7456647398843934e-05, "loss": 0.4218, "step": 627 }, { "epoch": 0.9789555728760717, "grad_norm": 0.36410259602213924, "learning_rate": 3.742774566473989e-05, "loss": 0.4293, "step": 628 }, { "epoch": 0.9805144193296961, "grad_norm": 0.3865124494448919, "learning_rate": 3.739884393063584e-05, "loss": 0.4229, "step": 629 }, { "epoch": 0.9820732657833203, "grad_norm": 0.3646826145021347, "learning_rate": 3.7369942196531794e-05, "loss": 0.4308, "step": 630 }, { "epoch": 0.9836321122369447, "grad_norm": 0.35897147135714, "learning_rate": 3.734104046242775e-05, "loss": 0.4407, "step": 631 }, { "epoch": 0.9851909586905689, "grad_norm": 0.3391861968595338, "learning_rate": 3.73121387283237e-05, "loss": 0.4354, "step": 632 }, { "epoch": 0.9867498051441933, "grad_norm": 0.43647757995986053, "learning_rate": 3.7283236994219654e-05, "loss": 0.4274, "step": 633 }, { "epoch": 0.9883086515978177, "grad_norm": 0.3618260493378971, "learning_rate": 3.725433526011561e-05, "loss": 0.4203, "step": 634 }, { "epoch": 0.9898674980514419, "grad_norm": 0.40188293786147516, "learning_rate": 3.722543352601156e-05, "loss": 0.4306, "step": 635 }, { "epoch": 0.9914263445050663, "grad_norm": 0.4129221508285703, "learning_rate": 3.719653179190752e-05, "loss": 0.4248, "step": 636 }, { "epoch": 0.9929851909586905, "grad_norm": 0.39766604692773766, "learning_rate": 3.716763005780347e-05, "loss": 0.4574, "step": 637 }, { "epoch": 0.9945440374123149, "grad_norm": 0.5559578417632596, "learning_rate": 3.713872832369942e-05, "loss": 0.469, "step": 638 }, { "epoch": 0.9961028838659393, "grad_norm": 0.3580889923606767, "learning_rate": 3.7109826589595375e-05, "loss": 0.4091, "step": 639 }, { "epoch": 0.9976617303195635, "grad_norm": 0.5076402218866569, "learning_rate": 3.7080924855491335e-05, "loss": 0.4305, "step": 640 }, { "epoch": 0.9992205767731879, "grad_norm": 0.3789119376045301, "learning_rate": 3.705202312138729e-05, "loss": 0.4306, "step": 641 }, { "epoch": 1.0, "grad_norm": 0.3789119376045301, "learning_rate": 3.7023121387283235e-05, "loss": 0.4336, "step": 642 }, { "epoch": 1.0015588464536243, "grad_norm": 0.628371201501632, "learning_rate": 3.699421965317919e-05, "loss": 0.3748, "step": 643 }, { "epoch": 1.0031176929072487, "grad_norm": 0.49964431541022886, "learning_rate": 3.696531791907515e-05, "loss": 0.3842, "step": 644 }, { "epoch": 1.004676539360873, "grad_norm": 0.4115388078983886, "learning_rate": 3.69364161849711e-05, "loss": 0.3921, "step": 645 }, { "epoch": 1.0062353858144972, "grad_norm": 0.36847896294616717, "learning_rate": 3.6907514450867056e-05, "loss": 0.3917, "step": 646 }, { "epoch": 1.0077942322681215, "grad_norm": 0.41235306439266933, "learning_rate": 3.6878612716763e-05, "loss": 0.3661, "step": 647 }, { "epoch": 1.009353078721746, "grad_norm": 0.36043638686025053, "learning_rate": 3.684971098265896e-05, "loss": 0.3793, "step": 648 }, { "epoch": 1.0109119251753702, "grad_norm": 0.5072805021746171, "learning_rate": 3.6820809248554916e-05, "loss": 0.3731, "step": 649 }, { "epoch": 1.0124707716289945, "grad_norm": 0.3774967084990004, "learning_rate": 3.679190751445087e-05, "loss": 0.3837, "step": 650 }, { "epoch": 1.014029618082619, "grad_norm": 0.47148649785731656, "learning_rate": 3.676300578034682e-05, "loss": 0.3655, "step": 651 }, { "epoch": 1.0155884645362432, "grad_norm": 0.37322483271463347, "learning_rate": 3.6734104046242777e-05, "loss": 0.3955, "step": 652 }, { "epoch": 1.0171473109898674, "grad_norm": 0.3705138885107383, "learning_rate": 3.670520231213873e-05, "loss": 0.3753, "step": 653 }, { "epoch": 1.018706157443492, "grad_norm": 0.5310690376603894, "learning_rate": 3.6676300578034683e-05, "loss": 0.365, "step": 654 }, { "epoch": 1.0202650038971162, "grad_norm": 0.3622397954678426, "learning_rate": 3.664739884393064e-05, "loss": 0.3691, "step": 655 }, { "epoch": 1.0218238503507404, "grad_norm": 0.48489876341543203, "learning_rate": 3.66184971098266e-05, "loss": 0.3757, "step": 656 }, { "epoch": 1.0233826968043647, "grad_norm": 0.3846528970782493, "learning_rate": 3.6589595375722544e-05, "loss": 0.3766, "step": 657 }, { "epoch": 1.0249415432579891, "grad_norm": 0.4581460525204425, "learning_rate": 3.65606936416185e-05, "loss": 0.3513, "step": 658 }, { "epoch": 1.0265003897116134, "grad_norm": 0.3465258787798717, "learning_rate": 3.653179190751445e-05, "loss": 0.3624, "step": 659 }, { "epoch": 1.0280592361652376, "grad_norm": 0.336442810655498, "learning_rate": 3.650289017341041e-05, "loss": 0.3572, "step": 660 }, { "epoch": 1.0296180826188621, "grad_norm": 0.3003091419390878, "learning_rate": 3.6473988439306364e-05, "loss": 0.367, "step": 661 }, { "epoch": 1.0311769290724864, "grad_norm": 0.32146010358650395, "learning_rate": 3.644508670520231e-05, "loss": 0.3758, "step": 662 }, { "epoch": 1.0327357755261106, "grad_norm": 0.3233773848232951, "learning_rate": 3.6416184971098265e-05, "loss": 0.3602, "step": 663 }, { "epoch": 1.034294621979735, "grad_norm": 0.3511338330800115, "learning_rate": 3.6387283236994225e-05, "loss": 0.3496, "step": 664 }, { "epoch": 1.0358534684333593, "grad_norm": 0.2996717533713208, "learning_rate": 3.635838150289018e-05, "loss": 0.3462, "step": 665 }, { "epoch": 1.0374123148869836, "grad_norm": 0.32342558777316216, "learning_rate": 3.632947976878613e-05, "loss": 0.3747, "step": 666 }, { "epoch": 1.0389711613406079, "grad_norm": 0.3367570922014961, "learning_rate": 3.630057803468208e-05, "loss": 0.3614, "step": 667 }, { "epoch": 1.0405300077942323, "grad_norm": 0.3441547440118924, "learning_rate": 3.627167630057804e-05, "loss": 0.3917, "step": 668 }, { "epoch": 1.0420888542478566, "grad_norm": 0.3361181158280621, "learning_rate": 3.624277456647399e-05, "loss": 0.3832, "step": 669 }, { "epoch": 1.0436477007014808, "grad_norm": 0.3291404615432134, "learning_rate": 3.6213872832369945e-05, "loss": 0.3622, "step": 670 }, { "epoch": 1.0452065471551053, "grad_norm": 0.30771644853510227, "learning_rate": 3.61849710982659e-05, "loss": 0.3796, "step": 671 }, { "epoch": 1.0467653936087296, "grad_norm": 0.33176245015646244, "learning_rate": 3.615606936416185e-05, "loss": 0.3788, "step": 672 }, { "epoch": 1.0483242400623538, "grad_norm": 0.9925708666951382, "learning_rate": 3.6127167630057806e-05, "loss": 0.4087, "step": 673 }, { "epoch": 1.0498830865159783, "grad_norm": 0.3681749130556859, "learning_rate": 3.609826589595376e-05, "loss": 0.3647, "step": 674 }, { "epoch": 1.0514419329696025, "grad_norm": 0.3804125438000747, "learning_rate": 3.606936416184971e-05, "loss": 0.387, "step": 675 }, { "epoch": 1.0530007794232268, "grad_norm": 0.3792677722995949, "learning_rate": 3.6040462427745666e-05, "loss": 0.3806, "step": 676 }, { "epoch": 1.054559625876851, "grad_norm": 0.3546135860064871, "learning_rate": 3.601156069364162e-05, "loss": 0.3858, "step": 677 }, { "epoch": 1.0561184723304755, "grad_norm": 0.3603149324311687, "learning_rate": 3.598265895953757e-05, "loss": 0.3744, "step": 678 }, { "epoch": 1.0576773187840998, "grad_norm": 0.34480487604338456, "learning_rate": 3.5953757225433526e-05, "loss": 0.3642, "step": 679 }, { "epoch": 1.059236165237724, "grad_norm": 0.3405204986814321, "learning_rate": 3.592485549132948e-05, "loss": 0.364, "step": 680 }, { "epoch": 1.0607950116913485, "grad_norm": 0.3605005587786407, "learning_rate": 3.589595375722544e-05, "loss": 0.3619, "step": 681 }, { "epoch": 1.0623538581449727, "grad_norm": 0.4006257143609181, "learning_rate": 3.586705202312139e-05, "loss": 0.3818, "step": 682 }, { "epoch": 1.063912704598597, "grad_norm": 0.29455634443366796, "learning_rate": 3.583815028901734e-05, "loss": 0.3673, "step": 683 }, { "epoch": 1.0654715510522212, "grad_norm": 0.36050845424709793, "learning_rate": 3.5809248554913294e-05, "loss": 0.367, "step": 684 }, { "epoch": 1.0670303975058457, "grad_norm": 0.3404885010186498, "learning_rate": 3.5780346820809254e-05, "loss": 0.3931, "step": 685 }, { "epoch": 1.06858924395947, "grad_norm": 0.3209534029364564, "learning_rate": 3.575144508670521e-05, "loss": 0.371, "step": 686 }, { "epoch": 1.0701480904130942, "grad_norm": 0.33657202823646476, "learning_rate": 3.5722543352601154e-05, "loss": 0.3789, "step": 687 }, { "epoch": 1.0717069368667187, "grad_norm": 0.3056381970556524, "learning_rate": 3.569364161849711e-05, "loss": 0.3781, "step": 688 }, { "epoch": 1.073265783320343, "grad_norm": 0.2869272425111672, "learning_rate": 3.566473988439307e-05, "loss": 0.3795, "step": 689 }, { "epoch": 1.0748246297739672, "grad_norm": 0.3423224557400581, "learning_rate": 3.563583815028902e-05, "loss": 0.3569, "step": 690 }, { "epoch": 1.0763834762275917, "grad_norm": 0.3620299786614207, "learning_rate": 3.5606936416184975e-05, "loss": 0.3813, "step": 691 }, { "epoch": 1.077942322681216, "grad_norm": 0.3164910374888722, "learning_rate": 3.557803468208092e-05, "loss": 0.3548, "step": 692 }, { "epoch": 1.0795011691348402, "grad_norm": 0.32077378022129677, "learning_rate": 3.554913294797688e-05, "loss": 0.3714, "step": 693 }, { "epoch": 1.0810600155884647, "grad_norm": 0.3449220187992264, "learning_rate": 3.5520231213872835e-05, "loss": 0.363, "step": 694 }, { "epoch": 1.082618862042089, "grad_norm": 0.3342998444359934, "learning_rate": 3.549132947976879e-05, "loss": 0.3482, "step": 695 }, { "epoch": 1.0841777084957132, "grad_norm": 0.35266941801482377, "learning_rate": 3.546242774566474e-05, "loss": 0.3784, "step": 696 }, { "epoch": 1.0857365549493374, "grad_norm": 0.4488757216839682, "learning_rate": 3.5433526011560695e-05, "loss": 0.3788, "step": 697 }, { "epoch": 1.0872954014029619, "grad_norm": 0.3288290304532232, "learning_rate": 3.540462427745665e-05, "loss": 0.3812, "step": 698 }, { "epoch": 1.0888542478565861, "grad_norm": 0.4417200363539964, "learning_rate": 3.53757225433526e-05, "loss": 0.377, "step": 699 }, { "epoch": 1.0904130943102104, "grad_norm": 0.31742949476273763, "learning_rate": 3.5346820809248556e-05, "loss": 0.3639, "step": 700 }, { "epoch": 1.0919719407638349, "grad_norm": 0.3308705930780861, "learning_rate": 3.531791907514451e-05, "loss": 0.3634, "step": 701 }, { "epoch": 1.093530787217459, "grad_norm": 0.3382413793786327, "learning_rate": 3.528901734104046e-05, "loss": 0.3598, "step": 702 }, { "epoch": 1.0950896336710834, "grad_norm": 0.3621055356977116, "learning_rate": 3.5260115606936416e-05, "loss": 0.3657, "step": 703 }, { "epoch": 1.0966484801247076, "grad_norm": 0.30903567078078115, "learning_rate": 3.523121387283237e-05, "loss": 0.3746, "step": 704 }, { "epoch": 1.098207326578332, "grad_norm": 0.43168788302923716, "learning_rate": 3.520231213872833e-05, "loss": 0.3786, "step": 705 }, { "epoch": 1.0997661730319563, "grad_norm": 0.3735078603768068, "learning_rate": 3.5173410404624276e-05, "loss": 0.3554, "step": 706 }, { "epoch": 1.1013250194855806, "grad_norm": 0.3705144355540284, "learning_rate": 3.514450867052023e-05, "loss": 0.3491, "step": 707 }, { "epoch": 1.102883865939205, "grad_norm": 0.36212715939080664, "learning_rate": 3.511560693641618e-05, "loss": 0.3665, "step": 708 }, { "epoch": 1.1044427123928293, "grad_norm": 0.3844205403442548, "learning_rate": 3.508670520231214e-05, "loss": 0.369, "step": 709 }, { "epoch": 1.1060015588464536, "grad_norm": 0.3434945710583639, "learning_rate": 3.50578034682081e-05, "loss": 0.3724, "step": 710 }, { "epoch": 1.107560405300078, "grad_norm": 0.4376944680100341, "learning_rate": 3.502890173410405e-05, "loss": 0.3829, "step": 711 }, { "epoch": 1.1091192517537023, "grad_norm": 0.3162790873067193, "learning_rate": 3.5e-05, "loss": 0.3794, "step": 712 }, { "epoch": 1.1106780982073265, "grad_norm": 0.30611808166388454, "learning_rate": 3.497109826589596e-05, "loss": 0.3652, "step": 713 }, { "epoch": 1.1122369446609508, "grad_norm": 0.3100967615655972, "learning_rate": 3.494219653179191e-05, "loss": 0.3715, "step": 714 }, { "epoch": 1.1137957911145753, "grad_norm": 0.29998053632892707, "learning_rate": 3.4913294797687864e-05, "loss": 0.385, "step": 715 }, { "epoch": 1.1153546375681995, "grad_norm": 0.3531103029098729, "learning_rate": 3.488439306358382e-05, "loss": 0.3852, "step": 716 }, { "epoch": 1.1169134840218238, "grad_norm": 0.28836989657125967, "learning_rate": 3.485549132947977e-05, "loss": 0.369, "step": 717 }, { "epoch": 1.1184723304754483, "grad_norm": 0.3726105295277464, "learning_rate": 3.4826589595375724e-05, "loss": 0.3901, "step": 718 }, { "epoch": 1.1200311769290725, "grad_norm": 0.30082206554969804, "learning_rate": 3.479768786127168e-05, "loss": 0.3672, "step": 719 }, { "epoch": 1.1215900233826968, "grad_norm": 0.3229975948532406, "learning_rate": 3.476878612716763e-05, "loss": 0.3741, "step": 720 }, { "epoch": 1.1231488698363212, "grad_norm": 0.30690351827295265, "learning_rate": 3.4739884393063585e-05, "loss": 0.3612, "step": 721 }, { "epoch": 1.1247077162899455, "grad_norm": 0.3558269454615068, "learning_rate": 3.471098265895954e-05, "loss": 0.3887, "step": 722 }, { "epoch": 1.1262665627435697, "grad_norm": 0.34137813880164614, "learning_rate": 3.468208092485549e-05, "loss": 0.3835, "step": 723 }, { "epoch": 1.127825409197194, "grad_norm": 0.3060779272980073, "learning_rate": 3.4653179190751445e-05, "loss": 0.3825, "step": 724 }, { "epoch": 1.1293842556508185, "grad_norm": 0.34107259706747284, "learning_rate": 3.4624277456647405e-05, "loss": 0.3716, "step": 725 }, { "epoch": 1.1309431021044427, "grad_norm": 0.28915982893470976, "learning_rate": 3.459537572254335e-05, "loss": 0.3658, "step": 726 }, { "epoch": 1.132501948558067, "grad_norm": 0.2927164690689312, "learning_rate": 3.4566473988439305e-05, "loss": 0.3604, "step": 727 }, { "epoch": 1.1340607950116914, "grad_norm": 0.3191085857306229, "learning_rate": 3.453757225433526e-05, "loss": 0.3837, "step": 728 }, { "epoch": 1.1356196414653157, "grad_norm": 0.3079933409669794, "learning_rate": 3.450867052023122e-05, "loss": 0.374, "step": 729 }, { "epoch": 1.13717848791894, "grad_norm": 0.2904492631644862, "learning_rate": 3.447976878612717e-05, "loss": 0.3778, "step": 730 }, { "epoch": 1.1387373343725642, "grad_norm": 0.29735218541832303, "learning_rate": 3.445086705202312e-05, "loss": 0.3573, "step": 731 }, { "epoch": 1.1402961808261887, "grad_norm": 0.3152575212565156, "learning_rate": 3.442196531791907e-05, "loss": 0.373, "step": 732 }, { "epoch": 1.141855027279813, "grad_norm": 0.32039618688552635, "learning_rate": 3.439306358381503e-05, "loss": 0.3779, "step": 733 }, { "epoch": 1.1434138737334372, "grad_norm": 0.3168459279894005, "learning_rate": 3.4364161849710986e-05, "loss": 0.3612, "step": 734 }, { "epoch": 1.1449727201870616, "grad_norm": 0.3382435631010463, "learning_rate": 3.433526011560694e-05, "loss": 0.375, "step": 735 }, { "epoch": 1.146531566640686, "grad_norm": 0.3451916527844319, "learning_rate": 3.430635838150289e-05, "loss": 0.3656, "step": 736 }, { "epoch": 1.1480904130943101, "grad_norm": 0.29795458770948763, "learning_rate": 3.427745664739885e-05, "loss": 0.3517, "step": 737 }, { "epoch": 1.1496492595479346, "grad_norm": 0.3204048119901522, "learning_rate": 3.42485549132948e-05, "loss": 0.3753, "step": 738 }, { "epoch": 1.1512081060015589, "grad_norm": 0.3509132683834904, "learning_rate": 3.4219653179190754e-05, "loss": 0.3718, "step": 739 }, { "epoch": 1.1527669524551831, "grad_norm": 0.3252836365240409, "learning_rate": 3.419075144508671e-05, "loss": 0.3884, "step": 740 }, { "epoch": 1.1543257989088076, "grad_norm": 0.2938786860494216, "learning_rate": 3.416184971098266e-05, "loss": 0.3508, "step": 741 }, { "epoch": 1.1558846453624319, "grad_norm": 0.3286847180631177, "learning_rate": 3.4132947976878614e-05, "loss": 0.3887, "step": 742 }, { "epoch": 1.157443491816056, "grad_norm": 0.293702845880183, "learning_rate": 3.410404624277457e-05, "loss": 0.3625, "step": 743 }, { "epoch": 1.1590023382696804, "grad_norm": 0.3637581698851474, "learning_rate": 3.407514450867052e-05, "loss": 0.3807, "step": 744 }, { "epoch": 1.1605611847233048, "grad_norm": 0.33030454774450496, "learning_rate": 3.404624277456648e-05, "loss": 0.3625, "step": 745 }, { "epoch": 1.162120031176929, "grad_norm": 0.3413072692858561, "learning_rate": 3.401734104046243e-05, "loss": 0.3799, "step": 746 }, { "epoch": 1.1636788776305533, "grad_norm": 0.3282042861228878, "learning_rate": 3.398843930635838e-05, "loss": 0.3744, "step": 747 }, { "epoch": 1.1652377240841778, "grad_norm": 0.4051993980268251, "learning_rate": 3.3959537572254335e-05, "loss": 0.3658, "step": 748 }, { "epoch": 1.166796570537802, "grad_norm": 0.3513826597230518, "learning_rate": 3.3930635838150295e-05, "loss": 0.3641, "step": 749 }, { "epoch": 1.1683554169914263, "grad_norm": 0.35299247249790444, "learning_rate": 3.390173410404625e-05, "loss": 0.3654, "step": 750 }, { "epoch": 1.1699142634450506, "grad_norm": 0.3014386293388838, "learning_rate": 3.3872832369942195e-05, "loss": 0.3765, "step": 751 }, { "epoch": 1.171473109898675, "grad_norm": 0.3603523634257503, "learning_rate": 3.384393063583815e-05, "loss": 0.3845, "step": 752 }, { "epoch": 1.1730319563522993, "grad_norm": 0.29665311038360115, "learning_rate": 3.381502890173411e-05, "loss": 0.3797, "step": 753 }, { "epoch": 1.1745908028059235, "grad_norm": 0.2935340087785823, "learning_rate": 3.378612716763006e-05, "loss": 0.3602, "step": 754 }, { "epoch": 1.176149649259548, "grad_norm": 0.33634845355876714, "learning_rate": 3.3757225433526015e-05, "loss": 0.3796, "step": 755 }, { "epoch": 1.1777084957131723, "grad_norm": 0.33117541570348524, "learning_rate": 3.372832369942196e-05, "loss": 0.4078, "step": 756 }, { "epoch": 1.1792673421667965, "grad_norm": 0.2734330693617624, "learning_rate": 3.369942196531792e-05, "loss": 0.3731, "step": 757 }, { "epoch": 1.1808261886204208, "grad_norm": 0.35008178961287134, "learning_rate": 3.3670520231213876e-05, "loss": 0.3764, "step": 758 }, { "epoch": 1.1823850350740452, "grad_norm": 0.3094977957875163, "learning_rate": 3.364161849710983e-05, "loss": 0.3565, "step": 759 }, { "epoch": 1.1839438815276695, "grad_norm": 0.29463076564571844, "learning_rate": 3.361271676300578e-05, "loss": 0.3705, "step": 760 }, { "epoch": 1.1855027279812937, "grad_norm": 0.3164209650580971, "learning_rate": 3.3583815028901736e-05, "loss": 0.3764, "step": 761 }, { "epoch": 1.1870615744349182, "grad_norm": 0.28906263199367943, "learning_rate": 3.355491329479769e-05, "loss": 0.3889, "step": 762 }, { "epoch": 1.1886204208885425, "grad_norm": 0.3138486170125581, "learning_rate": 3.352601156069364e-05, "loss": 0.3781, "step": 763 }, { "epoch": 1.1901792673421667, "grad_norm": 0.31679716557668985, "learning_rate": 3.3497109826589596e-05, "loss": 0.3725, "step": 764 }, { "epoch": 1.1917381137957912, "grad_norm": 0.2819822969115435, "learning_rate": 3.346820809248556e-05, "loss": 0.3621, "step": 765 }, { "epoch": 1.1932969602494155, "grad_norm": 0.2900191750020783, "learning_rate": 3.34393063583815e-05, "loss": 0.3864, "step": 766 }, { "epoch": 1.1948558067030397, "grad_norm": 0.3475437886006249, "learning_rate": 3.341040462427746e-05, "loss": 0.3501, "step": 767 }, { "epoch": 1.1964146531566642, "grad_norm": 0.27289955081057704, "learning_rate": 3.338150289017341e-05, "loss": 0.3738, "step": 768 }, { "epoch": 1.1979734996102884, "grad_norm": 0.37715118746357845, "learning_rate": 3.335260115606937e-05, "loss": 0.3566, "step": 769 }, { "epoch": 1.1995323460639127, "grad_norm": 0.30510592253569463, "learning_rate": 3.3323699421965324e-05, "loss": 0.3579, "step": 770 }, { "epoch": 1.201091192517537, "grad_norm": 0.3705292948456225, "learning_rate": 3.329479768786127e-05, "loss": 0.3655, "step": 771 }, { "epoch": 1.2026500389711614, "grad_norm": 0.32470886193713566, "learning_rate": 3.3265895953757224e-05, "loss": 0.3512, "step": 772 }, { "epoch": 1.2042088854247857, "grad_norm": 0.3267984083578711, "learning_rate": 3.323699421965318e-05, "loss": 0.3559, "step": 773 }, { "epoch": 1.20576773187841, "grad_norm": 0.316676369330289, "learning_rate": 3.320809248554914e-05, "loss": 0.3504, "step": 774 }, { "epoch": 1.2073265783320344, "grad_norm": 0.3083068143586948, "learning_rate": 3.317919075144509e-05, "loss": 0.3497, "step": 775 }, { "epoch": 1.2088854247856586, "grad_norm": 0.3112543283121555, "learning_rate": 3.315028901734104e-05, "loss": 0.3626, "step": 776 }, { "epoch": 1.2104442712392829, "grad_norm": 0.2930169607949705, "learning_rate": 3.312138728323699e-05, "loss": 0.3767, "step": 777 }, { "epoch": 1.2120031176929071, "grad_norm": 0.3740917898852477, "learning_rate": 3.309248554913295e-05, "loss": 0.3623, "step": 778 }, { "epoch": 1.2135619641465316, "grad_norm": 0.2909666222558865, "learning_rate": 3.3063583815028905e-05, "loss": 0.3571, "step": 779 }, { "epoch": 1.2151208106001559, "grad_norm": 0.3349084531099002, "learning_rate": 3.303468208092486e-05, "loss": 0.3809, "step": 780 }, { "epoch": 1.2166796570537801, "grad_norm": 0.32395231786814505, "learning_rate": 3.3005780346820805e-05, "loss": 0.3701, "step": 781 }, { "epoch": 1.2182385035074046, "grad_norm": 0.35641499162371804, "learning_rate": 3.2976878612716765e-05, "loss": 0.3767, "step": 782 }, { "epoch": 1.2197973499610288, "grad_norm": 0.371466143629439, "learning_rate": 3.294797687861272e-05, "loss": 0.3716, "step": 783 }, { "epoch": 1.221356196414653, "grad_norm": 0.3246192454526884, "learning_rate": 3.291907514450867e-05, "loss": 0.3742, "step": 784 }, { "epoch": 1.2229150428682776, "grad_norm": 0.35868229918432093, "learning_rate": 3.2890173410404626e-05, "loss": 0.38, "step": 785 }, { "epoch": 1.2244738893219018, "grad_norm": 0.353282741567031, "learning_rate": 3.286127167630058e-05, "loss": 0.3621, "step": 786 }, { "epoch": 1.226032735775526, "grad_norm": 0.34468589401361616, "learning_rate": 3.283236994219653e-05, "loss": 0.3739, "step": 787 }, { "epoch": 1.2275915822291505, "grad_norm": 0.2790582877348234, "learning_rate": 3.2803468208092486e-05, "loss": 0.3759, "step": 788 }, { "epoch": 1.2291504286827748, "grad_norm": 0.455365887136131, "learning_rate": 3.277456647398844e-05, "loss": 0.371, "step": 789 }, { "epoch": 1.230709275136399, "grad_norm": 0.32167141704971974, "learning_rate": 3.27456647398844e-05, "loss": 0.3728, "step": 790 }, { "epoch": 1.2322681215900233, "grad_norm": 0.3938144396135692, "learning_rate": 3.2716763005780346e-05, "loss": 0.3613, "step": 791 }, { "epoch": 1.2338269680436478, "grad_norm": 0.35340743429724525, "learning_rate": 3.26878612716763e-05, "loss": 0.362, "step": 792 }, { "epoch": 1.235385814497272, "grad_norm": 0.3887479835449903, "learning_rate": 3.265895953757225e-05, "loss": 0.3534, "step": 793 }, { "epoch": 1.2369446609508963, "grad_norm": 0.3810148179704415, "learning_rate": 3.2630057803468213e-05, "loss": 0.3708, "step": 794 }, { "epoch": 1.2385035074045208, "grad_norm": 0.45308168447987623, "learning_rate": 3.260115606936417e-05, "loss": 0.363, "step": 795 }, { "epoch": 1.240062353858145, "grad_norm": 0.3530892057764186, "learning_rate": 3.2572254335260114e-05, "loss": 0.3748, "step": 796 }, { "epoch": 1.2416212003117693, "grad_norm": 0.4323443583689573, "learning_rate": 3.254335260115607e-05, "loss": 0.3727, "step": 797 }, { "epoch": 1.2431800467653935, "grad_norm": 0.3213425990151126, "learning_rate": 3.251445086705203e-05, "loss": 0.391, "step": 798 }, { "epoch": 1.244738893219018, "grad_norm": 0.33159050533895373, "learning_rate": 3.248554913294798e-05, "loss": 0.378, "step": 799 }, { "epoch": 1.2462977396726422, "grad_norm": 0.35918958392112355, "learning_rate": 3.2456647398843934e-05, "loss": 0.372, "step": 800 }, { "epoch": 1.2478565861262665, "grad_norm": 0.3065526765937574, "learning_rate": 3.242774566473988e-05, "loss": 0.3678, "step": 801 }, { "epoch": 1.249415432579891, "grad_norm": 0.3162819465443684, "learning_rate": 3.239884393063584e-05, "loss": 0.3554, "step": 802 }, { "epoch": 1.2509742790335152, "grad_norm": 0.3623636030150004, "learning_rate": 3.2369942196531794e-05, "loss": 0.3742, "step": 803 }, { "epoch": 1.2525331254871395, "grad_norm": 0.2955704493514888, "learning_rate": 3.234104046242775e-05, "loss": 0.3704, "step": 804 }, { "epoch": 1.2540919719407637, "grad_norm": 0.36701054340991607, "learning_rate": 3.23121387283237e-05, "loss": 0.3728, "step": 805 }, { "epoch": 1.2556508183943882, "grad_norm": 0.3328636144558375, "learning_rate": 3.2283236994219655e-05, "loss": 0.3816, "step": 806 }, { "epoch": 1.2572096648480124, "grad_norm": 0.3578865290981359, "learning_rate": 3.225433526011561e-05, "loss": 0.3667, "step": 807 }, { "epoch": 1.258768511301637, "grad_norm": 0.3408336723453489, "learning_rate": 3.222543352601156e-05, "loss": 0.3817, "step": 808 }, { "epoch": 1.2603273577552612, "grad_norm": 0.3216030123089057, "learning_rate": 3.2196531791907515e-05, "loss": 0.3691, "step": 809 }, { "epoch": 1.2618862042088854, "grad_norm": 0.37724572418313224, "learning_rate": 3.216763005780347e-05, "loss": 0.3824, "step": 810 }, { "epoch": 1.2634450506625097, "grad_norm": 0.3056707439719446, "learning_rate": 3.213872832369942e-05, "loss": 0.3718, "step": 811 }, { "epoch": 1.265003897116134, "grad_norm": 0.3614010471408866, "learning_rate": 3.2109826589595375e-05, "loss": 0.378, "step": 812 }, { "epoch": 1.2665627435697584, "grad_norm": 0.30433373829385135, "learning_rate": 3.208092485549133e-05, "loss": 0.4059, "step": 813 }, { "epoch": 1.2681215900233826, "grad_norm": 0.3789679238319672, "learning_rate": 3.205202312138729e-05, "loss": 0.3739, "step": 814 }, { "epoch": 1.2696804364770071, "grad_norm": 0.31090210614669855, "learning_rate": 3.202312138728324e-05, "loss": 0.3796, "step": 815 }, { "epoch": 1.2712392829306314, "grad_norm": 0.3395673581732934, "learning_rate": 3.199421965317919e-05, "loss": 0.3842, "step": 816 }, { "epoch": 1.2727981293842556, "grad_norm": 0.3814013963708837, "learning_rate": 3.196531791907514e-05, "loss": 0.3818, "step": 817 }, { "epoch": 1.2743569758378799, "grad_norm": 0.2881312269831677, "learning_rate": 3.19364161849711e-05, "loss": 0.3755, "step": 818 }, { "epoch": 1.2759158222915044, "grad_norm": 0.3386831635441485, "learning_rate": 3.1907514450867056e-05, "loss": 0.3666, "step": 819 }, { "epoch": 1.2774746687451286, "grad_norm": 0.30697028027634937, "learning_rate": 3.187861271676301e-05, "loss": 0.3843, "step": 820 }, { "epoch": 1.2790335151987529, "grad_norm": 0.29980913900639733, "learning_rate": 3.1849710982658956e-05, "loss": 0.3666, "step": 821 }, { "epoch": 1.2805923616523773, "grad_norm": 0.3327607375144203, "learning_rate": 3.182080924855492e-05, "loss": 0.3858, "step": 822 }, { "epoch": 1.2821512081060016, "grad_norm": 0.29432860827743634, "learning_rate": 3.179190751445087e-05, "loss": 0.3614, "step": 823 }, { "epoch": 1.2837100545596258, "grad_norm": 0.3621120310764712, "learning_rate": 3.1763005780346824e-05, "loss": 0.3737, "step": 824 }, { "epoch": 1.28526890101325, "grad_norm": 0.3445821791263814, "learning_rate": 3.173410404624278e-05, "loss": 0.3672, "step": 825 }, { "epoch": 1.2868277474668746, "grad_norm": 0.2954327384154805, "learning_rate": 3.170520231213873e-05, "loss": 0.3607, "step": 826 }, { "epoch": 1.2883865939204988, "grad_norm": 0.328451959411728, "learning_rate": 3.1676300578034684e-05, "loss": 0.3702, "step": 827 }, { "epoch": 1.2899454403741233, "grad_norm": 0.2944126016476037, "learning_rate": 3.164739884393064e-05, "loss": 0.3889, "step": 828 }, { "epoch": 1.2915042868277475, "grad_norm": 0.3969777724956226, "learning_rate": 3.161849710982659e-05, "loss": 0.3547, "step": 829 }, { "epoch": 1.2930631332813718, "grad_norm": 0.2951235605441101, "learning_rate": 3.1589595375722544e-05, "loss": 0.378, "step": 830 }, { "epoch": 1.294621979734996, "grad_norm": 0.28308239600029, "learning_rate": 3.15606936416185e-05, "loss": 0.3828, "step": 831 }, { "epoch": 1.2961808261886203, "grad_norm": 0.3901401668691733, "learning_rate": 3.153179190751445e-05, "loss": 0.3842, "step": 832 }, { "epoch": 1.2977396726422448, "grad_norm": 0.3137907326756719, "learning_rate": 3.1502890173410405e-05, "loss": 0.3517, "step": 833 }, { "epoch": 1.299298519095869, "grad_norm": 0.30348942495737735, "learning_rate": 3.1473988439306365e-05, "loss": 0.3584, "step": 834 }, { "epoch": 1.3008573655494935, "grad_norm": 0.3134452203812044, "learning_rate": 3.144508670520231e-05, "loss": 0.3668, "step": 835 }, { "epoch": 1.3024162120031177, "grad_norm": 0.3330164303286052, "learning_rate": 3.1416184971098265e-05, "loss": 0.3791, "step": 836 }, { "epoch": 1.303975058456742, "grad_norm": 0.3242958410056007, "learning_rate": 3.138728323699422e-05, "loss": 0.3644, "step": 837 }, { "epoch": 1.3055339049103662, "grad_norm": 0.3372448607214704, "learning_rate": 3.135838150289018e-05, "loss": 0.3755, "step": 838 }, { "epoch": 1.3070927513639907, "grad_norm": 0.34805951436795246, "learning_rate": 3.132947976878613e-05, "loss": 0.4029, "step": 839 }, { "epoch": 1.308651597817615, "grad_norm": 0.3375892298162743, "learning_rate": 3.130057803468208e-05, "loss": 0.3688, "step": 840 }, { "epoch": 1.3102104442712392, "grad_norm": 0.3427372693638577, "learning_rate": 3.127167630057803e-05, "loss": 0.3583, "step": 841 }, { "epoch": 1.3117692907248637, "grad_norm": 0.39528017078512534, "learning_rate": 3.124277456647399e-05, "loss": 0.3815, "step": 842 }, { "epoch": 1.313328137178488, "grad_norm": 0.300797549857878, "learning_rate": 3.1213872832369946e-05, "loss": 0.3789, "step": 843 }, { "epoch": 1.3148869836321122, "grad_norm": 0.37253735153697626, "learning_rate": 3.11849710982659e-05, "loss": 0.38, "step": 844 }, { "epoch": 1.3164458300857365, "grad_norm": 0.3636396680715121, "learning_rate": 3.115606936416185e-05, "loss": 0.3584, "step": 845 }, { "epoch": 1.318004676539361, "grad_norm": 0.3225603467896192, "learning_rate": 3.1127167630057806e-05, "loss": 0.3801, "step": 846 }, { "epoch": 1.3195635229929852, "grad_norm": 0.3923444106986942, "learning_rate": 3.109826589595376e-05, "loss": 0.3838, "step": 847 }, { "epoch": 1.3211223694466094, "grad_norm": 0.29792571705489457, "learning_rate": 3.106936416184971e-05, "loss": 0.3877, "step": 848 }, { "epoch": 1.322681215900234, "grad_norm": 0.41101804688404436, "learning_rate": 3.1040462427745667e-05, "loss": 0.3711, "step": 849 }, { "epoch": 1.3242400623538582, "grad_norm": 0.31527820005688345, "learning_rate": 3.101156069364162e-05, "loss": 0.3672, "step": 850 }, { "epoch": 1.3257989088074824, "grad_norm": 0.3247891491953755, "learning_rate": 3.0982658959537573e-05, "loss": 0.3875, "step": 851 }, { "epoch": 1.3273577552611067, "grad_norm": 0.31678580564634845, "learning_rate": 3.095375722543353e-05, "loss": 0.3767, "step": 852 }, { "epoch": 1.3289166017147311, "grad_norm": 0.42289684401021704, "learning_rate": 3.092485549132948e-05, "loss": 0.3896, "step": 853 }, { "epoch": 1.3304754481683554, "grad_norm": 0.34344843446611606, "learning_rate": 3.089595375722544e-05, "loss": 0.3837, "step": 854 }, { "epoch": 1.3320342946219799, "grad_norm": 0.46291047151845677, "learning_rate": 3.086705202312139e-05, "loss": 0.3757, "step": 855 }, { "epoch": 1.3335931410756041, "grad_norm": 0.30760449980458326, "learning_rate": 3.083815028901734e-05, "loss": 0.3784, "step": 856 }, { "epoch": 1.3351519875292284, "grad_norm": 0.36618124968476307, "learning_rate": 3.0809248554913294e-05, "loss": 0.3646, "step": 857 }, { "epoch": 1.3367108339828526, "grad_norm": 0.3517633452852137, "learning_rate": 3.0780346820809254e-05, "loss": 0.3678, "step": 858 }, { "epoch": 1.3382696804364769, "grad_norm": 0.36650573052741386, "learning_rate": 3.075144508670521e-05, "loss": 0.3666, "step": 859 }, { "epoch": 1.3398285268901013, "grad_norm": 0.32309525776364156, "learning_rate": 3.0722543352601154e-05, "loss": 0.3596, "step": 860 }, { "epoch": 1.3413873733437256, "grad_norm": 0.2918364251771976, "learning_rate": 3.069364161849711e-05, "loss": 0.363, "step": 861 }, { "epoch": 1.34294621979735, "grad_norm": 0.34546628975016486, "learning_rate": 3.066473988439307e-05, "loss": 0.3553, "step": 862 }, { "epoch": 1.3445050662509743, "grad_norm": 0.2923967277989431, "learning_rate": 3.063583815028902e-05, "loss": 0.3489, "step": 863 }, { "epoch": 1.3460639127045986, "grad_norm": 0.29914296381783523, "learning_rate": 3.0606936416184975e-05, "loss": 0.3787, "step": 864 }, { "epoch": 1.3476227591582228, "grad_norm": 0.31943444292620105, "learning_rate": 3.057803468208092e-05, "loss": 0.3846, "step": 865 }, { "epoch": 1.3491816056118473, "grad_norm": 0.29039633067671383, "learning_rate": 3.0549132947976875e-05, "loss": 0.3639, "step": 866 }, { "epoch": 1.3507404520654716, "grad_norm": 0.3421893491096786, "learning_rate": 3.0520231213872835e-05, "loss": 0.3734, "step": 867 }, { "epoch": 1.3522992985190958, "grad_norm": 0.3170017953585186, "learning_rate": 3.049132947976879e-05, "loss": 0.369, "step": 868 }, { "epoch": 1.3538581449727203, "grad_norm": 0.31067447367056183, "learning_rate": 3.046242774566474e-05, "loss": 0.3879, "step": 869 }, { "epoch": 1.3554169914263445, "grad_norm": 0.3212256602786557, "learning_rate": 3.0433526011560692e-05, "loss": 0.3805, "step": 870 }, { "epoch": 1.3569758378799688, "grad_norm": 0.32052080705580777, "learning_rate": 3.040462427745665e-05, "loss": 0.3879, "step": 871 }, { "epoch": 1.358534684333593, "grad_norm": 0.27323028010679573, "learning_rate": 3.0375722543352603e-05, "loss": 0.3753, "step": 872 }, { "epoch": 1.3600935307872175, "grad_norm": 0.31472694270289026, "learning_rate": 3.0346820809248556e-05, "loss": 0.3676, "step": 873 }, { "epoch": 1.3616523772408418, "grad_norm": 0.2817828326023501, "learning_rate": 3.0317919075144506e-05, "loss": 0.3602, "step": 874 }, { "epoch": 1.363211223694466, "grad_norm": 0.31483550680330125, "learning_rate": 3.0289017341040466e-05, "loss": 0.3801, "step": 875 }, { "epoch": 1.3647700701480905, "grad_norm": 0.31361863279521, "learning_rate": 3.0260115606936416e-05, "loss": 0.3574, "step": 876 }, { "epoch": 1.3663289166017147, "grad_norm": 0.3795186696828694, "learning_rate": 3.023121387283237e-05, "loss": 0.3608, "step": 877 }, { "epoch": 1.367887763055339, "grad_norm": 0.3024681697597954, "learning_rate": 3.0202312138728323e-05, "loss": 0.3597, "step": 878 }, { "epoch": 1.3694466095089632, "grad_norm": 0.2839484436314578, "learning_rate": 3.017341040462428e-05, "loss": 0.354, "step": 879 }, { "epoch": 1.3710054559625877, "grad_norm": 0.2989185656953038, "learning_rate": 3.0144508670520234e-05, "loss": 0.3997, "step": 880 }, { "epoch": 1.372564302416212, "grad_norm": 0.3332125838479124, "learning_rate": 3.0115606936416184e-05, "loss": 0.3866, "step": 881 }, { "epoch": 1.3741231488698364, "grad_norm": 0.32196595663758937, "learning_rate": 3.0086705202312137e-05, "loss": 0.3703, "step": 882 }, { "epoch": 1.3756819953234607, "grad_norm": 0.3218885635026921, "learning_rate": 3.0057803468208097e-05, "loss": 0.3644, "step": 883 }, { "epoch": 1.377240841777085, "grad_norm": 0.3653050901483605, "learning_rate": 3.0028901734104047e-05, "loss": 0.3589, "step": 884 }, { "epoch": 1.3787996882307092, "grad_norm": 0.3686597692797899, "learning_rate": 3e-05, "loss": 0.3691, "step": 885 }, { "epoch": 1.3803585346843337, "grad_norm": 0.31536889278803354, "learning_rate": 2.9971098265895954e-05, "loss": 0.3719, "step": 886 }, { "epoch": 1.381917381137958, "grad_norm": 0.33299163893566386, "learning_rate": 2.994219653179191e-05, "loss": 0.376, "step": 887 }, { "epoch": 1.3834762275915822, "grad_norm": 0.3637154017248889, "learning_rate": 2.9913294797687864e-05, "loss": 0.3763, "step": 888 }, { "epoch": 1.3850350740452066, "grad_norm": 0.31459018536275674, "learning_rate": 2.9884393063583815e-05, "loss": 0.3719, "step": 889 }, { "epoch": 1.386593920498831, "grad_norm": 0.38863024038763483, "learning_rate": 2.9855491329479768e-05, "loss": 0.3897, "step": 890 }, { "epoch": 1.3881527669524552, "grad_norm": 0.2713739072144362, "learning_rate": 2.9826589595375725e-05, "loss": 0.3774, "step": 891 }, { "epoch": 1.3897116134060794, "grad_norm": 0.3161296579386695, "learning_rate": 2.9797687861271678e-05, "loss": 0.367, "step": 892 }, { "epoch": 1.3912704598597039, "grad_norm": 0.32321623162651764, "learning_rate": 2.9768786127167632e-05, "loss": 0.3908, "step": 893 }, { "epoch": 1.3928293063133281, "grad_norm": 0.28969209370393095, "learning_rate": 2.9739884393063582e-05, "loss": 0.3715, "step": 894 }, { "epoch": 1.3943881527669524, "grad_norm": 0.35841207954661103, "learning_rate": 2.9710982658959542e-05, "loss": 0.3714, "step": 895 }, { "epoch": 1.3959469992205769, "grad_norm": 0.3324679248139433, "learning_rate": 2.9682080924855492e-05, "loss": 0.3717, "step": 896 }, { "epoch": 1.397505845674201, "grad_norm": 0.43923218037881906, "learning_rate": 2.9653179190751446e-05, "loss": 0.3619, "step": 897 }, { "epoch": 1.3990646921278254, "grad_norm": 0.31489100441245577, "learning_rate": 2.96242774566474e-05, "loss": 0.3523, "step": 898 }, { "epoch": 1.4006235385814496, "grad_norm": 0.3292376441511829, "learning_rate": 2.9595375722543356e-05, "loss": 0.3661, "step": 899 }, { "epoch": 1.402182385035074, "grad_norm": 0.2988144575593582, "learning_rate": 2.956647398843931e-05, "loss": 0.3749, "step": 900 }, { "epoch": 1.4037412314886983, "grad_norm": 0.41020723022893746, "learning_rate": 2.953757225433526e-05, "loss": 0.366, "step": 901 }, { "epoch": 1.4053000779423228, "grad_norm": 0.310023862219356, "learning_rate": 2.9508670520231213e-05, "loss": 0.3884, "step": 902 }, { "epoch": 1.406858924395947, "grad_norm": 0.3454601431005727, "learning_rate": 2.947976878612717e-05, "loss": 0.3634, "step": 903 }, { "epoch": 1.4084177708495713, "grad_norm": 0.34356963535549423, "learning_rate": 2.9450867052023123e-05, "loss": 0.3588, "step": 904 }, { "epoch": 1.4099766173031956, "grad_norm": 0.3283997995226211, "learning_rate": 2.9421965317919076e-05, "loss": 0.3637, "step": 905 }, { "epoch": 1.4115354637568198, "grad_norm": 0.3554228840104719, "learning_rate": 2.9393063583815027e-05, "loss": 0.343, "step": 906 }, { "epoch": 1.4130943102104443, "grad_norm": 0.36071645685077747, "learning_rate": 2.9364161849710987e-05, "loss": 0.3596, "step": 907 }, { "epoch": 1.4146531566640685, "grad_norm": 0.3112359765865261, "learning_rate": 2.9335260115606937e-05, "loss": 0.3526, "step": 908 }, { "epoch": 1.416212003117693, "grad_norm": 0.34359020030558024, "learning_rate": 2.930635838150289e-05, "loss": 0.3844, "step": 909 }, { "epoch": 1.4177708495713173, "grad_norm": 0.3008878572757736, "learning_rate": 2.9277456647398844e-05, "loss": 0.3772, "step": 910 }, { "epoch": 1.4193296960249415, "grad_norm": 0.349466144708206, "learning_rate": 2.92485549132948e-05, "loss": 0.3473, "step": 911 }, { "epoch": 1.4208885424785658, "grad_norm": 0.3382202790340046, "learning_rate": 2.9219653179190754e-05, "loss": 0.3757, "step": 912 }, { "epoch": 1.4224473889321902, "grad_norm": 0.32188902141101255, "learning_rate": 2.9190751445086707e-05, "loss": 0.3712, "step": 913 }, { "epoch": 1.4240062353858145, "grad_norm": 0.29754261721596836, "learning_rate": 2.9161849710982657e-05, "loss": 0.3674, "step": 914 }, { "epoch": 1.4255650818394388, "grad_norm": 0.41389196935864453, "learning_rate": 2.9132947976878618e-05, "loss": 0.3835, "step": 915 }, { "epoch": 1.4271239282930632, "grad_norm": 0.2943029270951012, "learning_rate": 2.9104046242774568e-05, "loss": 0.359, "step": 916 }, { "epoch": 1.4286827747466875, "grad_norm": 0.28723479498577376, "learning_rate": 2.907514450867052e-05, "loss": 0.3695, "step": 917 }, { "epoch": 1.4302416212003117, "grad_norm": 0.35137358020533505, "learning_rate": 2.9046242774566475e-05, "loss": 0.3626, "step": 918 }, { "epoch": 1.431800467653936, "grad_norm": 0.3198068865556259, "learning_rate": 2.901734104046243e-05, "loss": 0.3855, "step": 919 }, { "epoch": 1.4333593141075605, "grad_norm": 0.30362892326696816, "learning_rate": 2.8988439306358385e-05, "loss": 0.3882, "step": 920 }, { "epoch": 1.4349181605611847, "grad_norm": 0.3191735883892925, "learning_rate": 2.8959537572254335e-05, "loss": 0.381, "step": 921 }, { "epoch": 1.436477007014809, "grad_norm": 0.3034703934282239, "learning_rate": 2.893063583815029e-05, "loss": 0.3756, "step": 922 }, { "epoch": 1.4380358534684334, "grad_norm": 0.32215086435520635, "learning_rate": 2.8901734104046245e-05, "loss": 0.3505, "step": 923 }, { "epoch": 1.4395946999220577, "grad_norm": 0.3038326438989388, "learning_rate": 2.88728323699422e-05, "loss": 0.3578, "step": 924 }, { "epoch": 1.441153546375682, "grad_norm": 0.32683148429237635, "learning_rate": 2.8843930635838152e-05, "loss": 0.358, "step": 925 }, { "epoch": 1.4427123928293062, "grad_norm": 0.3682668995271643, "learning_rate": 2.8815028901734102e-05, "loss": 0.3693, "step": 926 }, { "epoch": 1.4442712392829307, "grad_norm": 0.2743528467933646, "learning_rate": 2.8786127167630062e-05, "loss": 0.366, "step": 927 }, { "epoch": 1.445830085736555, "grad_norm": 0.3966807244623201, "learning_rate": 2.8757225433526013e-05, "loss": 0.386, "step": 928 }, { "epoch": 1.4473889321901794, "grad_norm": 0.32237328497617107, "learning_rate": 2.8728323699421966e-05, "loss": 0.3659, "step": 929 }, { "epoch": 1.4489477786438036, "grad_norm": 0.31054470894310937, "learning_rate": 2.869942196531792e-05, "loss": 0.3512, "step": 930 }, { "epoch": 1.450506625097428, "grad_norm": 0.43587348795942593, "learning_rate": 2.8670520231213876e-05, "loss": 0.3673, "step": 931 }, { "epoch": 1.4520654715510521, "grad_norm": 0.2889720776490319, "learning_rate": 2.864161849710983e-05, "loss": 0.3834, "step": 932 }, { "epoch": 1.4536243180046766, "grad_norm": 0.41243764294739627, "learning_rate": 2.861271676300578e-05, "loss": 0.3718, "step": 933 }, { "epoch": 1.4551831644583009, "grad_norm": 0.3609717686516069, "learning_rate": 2.8583815028901733e-05, "loss": 0.3666, "step": 934 }, { "epoch": 1.4567420109119251, "grad_norm": 0.3210455895295045, "learning_rate": 2.855491329479769e-05, "loss": 0.3512, "step": 935 }, { "epoch": 1.4583008573655496, "grad_norm": 0.36374442891877706, "learning_rate": 2.8526011560693643e-05, "loss": 0.3789, "step": 936 }, { "epoch": 1.4598597038191738, "grad_norm": 0.302661164884665, "learning_rate": 2.8497109826589597e-05, "loss": 0.3849, "step": 937 }, { "epoch": 1.461418550272798, "grad_norm": 0.2922986064549059, "learning_rate": 2.846820809248555e-05, "loss": 0.3799, "step": 938 }, { "epoch": 1.4629773967264224, "grad_norm": 0.3106937568151618, "learning_rate": 2.8439306358381507e-05, "loss": 0.3934, "step": 939 }, { "epoch": 1.4645362431800468, "grad_norm": 0.3894062278825112, "learning_rate": 2.841040462427746e-05, "loss": 0.366, "step": 940 }, { "epoch": 1.466095089633671, "grad_norm": 0.3094641338569837, "learning_rate": 2.838150289017341e-05, "loss": 0.3667, "step": 941 }, { "epoch": 1.4676539360872953, "grad_norm": 0.2993216786356107, "learning_rate": 2.8352601156069364e-05, "loss": 0.3778, "step": 942 }, { "epoch": 1.4692127825409198, "grad_norm": 0.290935367620464, "learning_rate": 2.832369942196532e-05, "loss": 0.3705, "step": 943 }, { "epoch": 1.470771628994544, "grad_norm": 0.3157318710559096, "learning_rate": 2.8294797687861274e-05, "loss": 0.3595, "step": 944 }, { "epoch": 1.4723304754481683, "grad_norm": 0.2890101359777515, "learning_rate": 2.8265895953757228e-05, "loss": 0.3773, "step": 945 }, { "epoch": 1.4738893219017926, "grad_norm": 0.3502615396853657, "learning_rate": 2.8236994219653178e-05, "loss": 0.3605, "step": 946 }, { "epoch": 1.475448168355417, "grad_norm": 0.34510628572007923, "learning_rate": 2.8208092485549138e-05, "loss": 0.3567, "step": 947 }, { "epoch": 1.4770070148090413, "grad_norm": 0.30987576563699276, "learning_rate": 2.8179190751445088e-05, "loss": 0.3876, "step": 948 }, { "epoch": 1.4785658612626658, "grad_norm": 0.38023748103458505, "learning_rate": 2.815028901734104e-05, "loss": 0.3692, "step": 949 }, { "epoch": 1.48012470771629, "grad_norm": 0.30194848904733007, "learning_rate": 2.8121387283236995e-05, "loss": 0.3639, "step": 950 }, { "epoch": 1.4816835541699143, "grad_norm": 0.36247398033072276, "learning_rate": 2.8092485549132952e-05, "loss": 0.3647, "step": 951 }, { "epoch": 1.4832424006235385, "grad_norm": 0.3225418884761915, "learning_rate": 2.8063583815028905e-05, "loss": 0.3728, "step": 952 }, { "epoch": 1.4848012470771628, "grad_norm": 0.33143640861599943, "learning_rate": 2.8034682080924855e-05, "loss": 0.3786, "step": 953 }, { "epoch": 1.4863600935307872, "grad_norm": 0.37859874084366085, "learning_rate": 2.800578034682081e-05, "loss": 0.3676, "step": 954 }, { "epoch": 1.4879189399844115, "grad_norm": 0.376572202242789, "learning_rate": 2.7976878612716766e-05, "loss": 0.3741, "step": 955 }, { "epoch": 1.489477786438036, "grad_norm": 0.33305139704253905, "learning_rate": 2.794797687861272e-05, "loss": 0.3676, "step": 956 }, { "epoch": 1.4910366328916602, "grad_norm": 0.5151154703791708, "learning_rate": 2.7919075144508673e-05, "loss": 0.3712, "step": 957 }, { "epoch": 1.4925954793452845, "grad_norm": 0.289014937897802, "learning_rate": 2.7890173410404623e-05, "loss": 0.3522, "step": 958 }, { "epoch": 1.4941543257989087, "grad_norm": 0.4050759220300978, "learning_rate": 2.7861271676300583e-05, "loss": 0.3695, "step": 959 }, { "epoch": 1.4957131722525332, "grad_norm": 0.4162682771540926, "learning_rate": 2.7832369942196533e-05, "loss": 0.3864, "step": 960 }, { "epoch": 1.4972720187061574, "grad_norm": 0.3267115752346114, "learning_rate": 2.7803468208092486e-05, "loss": 0.398, "step": 961 }, { "epoch": 1.4988308651597817, "grad_norm": 0.46629831946057043, "learning_rate": 2.777456647398844e-05, "loss": 0.3667, "step": 962 }, { "epoch": 1.5003897116134062, "grad_norm": 0.3126837897461622, "learning_rate": 2.7745664739884393e-05, "loss": 0.3903, "step": 963 }, { "epoch": 1.5019485580670304, "grad_norm": 0.3025185534818139, "learning_rate": 2.771676300578035e-05, "loss": 0.4, "step": 964 }, { "epoch": 1.5035074045206547, "grad_norm": 0.4008151783936587, "learning_rate": 2.7687861271676304e-05, "loss": 0.3839, "step": 965 }, { "epoch": 1.505066250974279, "grad_norm": 0.26063162497897896, "learning_rate": 2.7658959537572254e-05, "loss": 0.361, "step": 966 }, { "epoch": 1.5066250974279034, "grad_norm": 0.2888143269712675, "learning_rate": 2.7630057803468207e-05, "loss": 0.365, "step": 967 }, { "epoch": 1.5081839438815277, "grad_norm": 0.3227561551129935, "learning_rate": 2.7601156069364164e-05, "loss": 0.3867, "step": 968 }, { "epoch": 1.5097427903351521, "grad_norm": 0.29415161701639214, "learning_rate": 2.7572254335260117e-05, "loss": 0.3766, "step": 969 }, { "epoch": 1.5113016367887764, "grad_norm": 0.30622334937549417, "learning_rate": 2.754335260115607e-05, "loss": 0.362, "step": 970 }, { "epoch": 1.5128604832424006, "grad_norm": 0.29765107975268507, "learning_rate": 2.751445086705202e-05, "loss": 0.3691, "step": 971 }, { "epoch": 1.5144193296960249, "grad_norm": 0.28887375199419957, "learning_rate": 2.748554913294798e-05, "loss": 0.376, "step": 972 }, { "epoch": 1.5159781761496491, "grad_norm": 0.30491806233389507, "learning_rate": 2.745664739884393e-05, "loss": 0.3767, "step": 973 }, { "epoch": 1.5175370226032736, "grad_norm": 0.2645879031934002, "learning_rate": 2.7427745664739885e-05, "loss": 0.3812, "step": 974 }, { "epoch": 1.5190958690568979, "grad_norm": 0.28565659154478995, "learning_rate": 2.7398843930635838e-05, "loss": 0.3785, "step": 975 }, { "epoch": 1.5206547155105223, "grad_norm": 0.28811620089387, "learning_rate": 2.7369942196531795e-05, "loss": 0.3537, "step": 976 }, { "epoch": 1.5222135619641466, "grad_norm": 0.2577090443673922, "learning_rate": 2.734104046242775e-05, "loss": 0.355, "step": 977 }, { "epoch": 1.5237724084177708, "grad_norm": 0.2624546459669209, "learning_rate": 2.73121387283237e-05, "loss": 0.3665, "step": 978 }, { "epoch": 1.525331254871395, "grad_norm": 0.29221496113302586, "learning_rate": 2.7283236994219652e-05, "loss": 0.3683, "step": 979 }, { "epoch": 1.5268901013250193, "grad_norm": 0.2680361810204351, "learning_rate": 2.725433526011561e-05, "loss": 0.3699, "step": 980 }, { "epoch": 1.5284489477786438, "grad_norm": 0.3207752203807436, "learning_rate": 2.7225433526011562e-05, "loss": 0.3741, "step": 981 }, { "epoch": 1.530007794232268, "grad_norm": 0.282031401318972, "learning_rate": 2.7196531791907516e-05, "loss": 0.3853, "step": 982 }, { "epoch": 1.5315666406858925, "grad_norm": 0.27982574114644004, "learning_rate": 2.7167630057803466e-05, "loss": 0.3832, "step": 983 }, { "epoch": 1.5331254871395168, "grad_norm": 0.2804241252514004, "learning_rate": 2.7138728323699426e-05, "loss": 0.3593, "step": 984 }, { "epoch": 1.534684333593141, "grad_norm": 0.2628827070596599, "learning_rate": 2.7109826589595376e-05, "loss": 0.3669, "step": 985 }, { "epoch": 1.5362431800467653, "grad_norm": 0.29615449790999526, "learning_rate": 2.708092485549133e-05, "loss": 0.3559, "step": 986 }, { "epoch": 1.5378020265003896, "grad_norm": 0.28549992629405135, "learning_rate": 2.7052023121387283e-05, "loss": 0.3859, "step": 987 }, { "epoch": 1.539360872954014, "grad_norm": 0.34007692401926715, "learning_rate": 2.702312138728324e-05, "loss": 0.3715, "step": 988 }, { "epoch": 1.5409197194076385, "grad_norm": 0.40294192708679744, "learning_rate": 2.6994219653179193e-05, "loss": 0.3821, "step": 989 }, { "epoch": 1.5424785658612628, "grad_norm": 0.265218381447006, "learning_rate": 2.6965317919075143e-05, "loss": 0.369, "step": 990 }, { "epoch": 1.544037412314887, "grad_norm": 0.37011390060826654, "learning_rate": 2.6936416184971097e-05, "loss": 0.3643, "step": 991 }, { "epoch": 1.5455962587685113, "grad_norm": 0.31157711298908103, "learning_rate": 2.6907514450867057e-05, "loss": 0.3688, "step": 992 }, { "epoch": 1.5471551052221355, "grad_norm": 0.30367205453380675, "learning_rate": 2.6878612716763007e-05, "loss": 0.362, "step": 993 }, { "epoch": 1.54871395167576, "grad_norm": 0.40306120286710007, "learning_rate": 2.684971098265896e-05, "loss": 0.3679, "step": 994 }, { "epoch": 1.5502727981293842, "grad_norm": 0.3044105968804601, "learning_rate": 2.6820809248554914e-05, "loss": 0.3719, "step": 995 }, { "epoch": 1.5518316445830087, "grad_norm": 0.3384424306040368, "learning_rate": 2.679190751445087e-05, "loss": 0.3669, "step": 996 }, { "epoch": 1.553390491036633, "grad_norm": 0.2879848276257589, "learning_rate": 2.6763005780346824e-05, "loss": 0.3698, "step": 997 }, { "epoch": 1.5549493374902572, "grad_norm": 0.3215450830190318, "learning_rate": 2.6734104046242774e-05, "loss": 0.3687, "step": 998 }, { "epoch": 1.5565081839438815, "grad_norm": 0.2820349630596858, "learning_rate": 2.6705202312138728e-05, "loss": 0.3665, "step": 999 }, { "epoch": 1.5580670303975057, "grad_norm": 0.3606469331809056, "learning_rate": 2.6676300578034684e-05, "loss": 0.351, "step": 1000 }, { "epoch": 1.5596258768511302, "grad_norm": 0.3028843154170325, "learning_rate": 2.6647398843930638e-05, "loss": 0.3736, "step": 1001 }, { "epoch": 1.5611847233047544, "grad_norm": 0.3174824654872977, "learning_rate": 2.661849710982659e-05, "loss": 0.3629, "step": 1002 }, { "epoch": 1.562743569758379, "grad_norm": 0.34721718744253455, "learning_rate": 2.658959537572254e-05, "loss": 0.379, "step": 1003 }, { "epoch": 1.5643024162120032, "grad_norm": 0.2978398739891535, "learning_rate": 2.65606936416185e-05, "loss": 0.3683, "step": 1004 }, { "epoch": 1.5658612626656274, "grad_norm": 0.39334684334782377, "learning_rate": 2.653179190751445e-05, "loss": 0.3569, "step": 1005 }, { "epoch": 1.5674201091192517, "grad_norm": 0.2717962848929705, "learning_rate": 2.6502890173410405e-05, "loss": 0.3561, "step": 1006 }, { "epoch": 1.568978955572876, "grad_norm": 0.3736251973751076, "learning_rate": 2.647398843930636e-05, "loss": 0.3669, "step": 1007 }, { "epoch": 1.5705378020265004, "grad_norm": 0.29899355625093155, "learning_rate": 2.6445086705202315e-05, "loss": 0.3604, "step": 1008 }, { "epoch": 1.5720966484801249, "grad_norm": 0.38876556071538754, "learning_rate": 2.641618497109827e-05, "loss": 0.3605, "step": 1009 }, { "epoch": 1.5736554949337491, "grad_norm": 0.325884443622183, "learning_rate": 2.638728323699422e-05, "loss": 0.3858, "step": 1010 }, { "epoch": 1.5752143413873734, "grad_norm": 0.3267871100560373, "learning_rate": 2.6358381502890172e-05, "loss": 0.3504, "step": 1011 }, { "epoch": 1.5767731878409976, "grad_norm": 0.33526147496951697, "learning_rate": 2.632947976878613e-05, "loss": 0.3559, "step": 1012 }, { "epoch": 1.5783320342946219, "grad_norm": 0.2944668930671466, "learning_rate": 2.6300578034682083e-05, "loss": 0.3755, "step": 1013 }, { "epoch": 1.5798908807482464, "grad_norm": 0.3138451746470972, "learning_rate": 2.6271676300578036e-05, "loss": 0.3643, "step": 1014 }, { "epoch": 1.5814497272018706, "grad_norm": 0.26776940795700943, "learning_rate": 2.6242774566473986e-05, "loss": 0.3466, "step": 1015 }, { "epoch": 1.583008573655495, "grad_norm": 0.26784179313482087, "learning_rate": 2.6213872832369946e-05, "loss": 0.3486, "step": 1016 }, { "epoch": 1.5845674201091193, "grad_norm": 0.38607190770403105, "learning_rate": 2.6184971098265896e-05, "loss": 0.353, "step": 1017 }, { "epoch": 1.5861262665627436, "grad_norm": 0.3065474352566968, "learning_rate": 2.615606936416185e-05, "loss": 0.3589, "step": 1018 }, { "epoch": 1.5876851130163678, "grad_norm": 0.2832505582830171, "learning_rate": 2.6127167630057803e-05, "loss": 0.3678, "step": 1019 }, { "epoch": 1.589243959469992, "grad_norm": 0.36412176076022207, "learning_rate": 2.609826589595376e-05, "loss": 0.3798, "step": 1020 }, { "epoch": 1.5908028059236166, "grad_norm": 0.28000224882500524, "learning_rate": 2.6069364161849714e-05, "loss": 0.3497, "step": 1021 }, { "epoch": 1.5923616523772408, "grad_norm": 1.8214222618798932, "learning_rate": 2.6040462427745667e-05, "loss": 0.4108, "step": 1022 }, { "epoch": 1.5939204988308653, "grad_norm": 0.44865333521920003, "learning_rate": 2.6011560693641617e-05, "loss": 0.3772, "step": 1023 }, { "epoch": 1.5954793452844895, "grad_norm": 0.25598641910972036, "learning_rate": 2.5982658959537577e-05, "loss": 0.3657, "step": 1024 }, { "epoch": 1.5970381917381138, "grad_norm": 0.33942270049254686, "learning_rate": 2.5953757225433527e-05, "loss": 0.3995, "step": 1025 }, { "epoch": 1.598597038191738, "grad_norm": 0.3657115461703349, "learning_rate": 2.592485549132948e-05, "loss": 0.3941, "step": 1026 }, { "epoch": 1.6001558846453623, "grad_norm": 0.28531237255153097, "learning_rate": 2.5895953757225434e-05, "loss": 0.3806, "step": 1027 }, { "epoch": 1.6017147310989868, "grad_norm": 0.35592372312382153, "learning_rate": 2.586705202312139e-05, "loss": 0.3781, "step": 1028 }, { "epoch": 1.603273577552611, "grad_norm": 0.29099766422855, "learning_rate": 2.5838150289017344e-05, "loss": 0.3638, "step": 1029 }, { "epoch": 1.6048324240062355, "grad_norm": 0.3174734446542449, "learning_rate": 2.5809248554913295e-05, "loss": 0.3435, "step": 1030 }, { "epoch": 1.6063912704598597, "grad_norm": 0.4203285741115712, "learning_rate": 2.5780346820809248e-05, "loss": 0.3788, "step": 1031 }, { "epoch": 1.607950116913484, "grad_norm": 0.32251687798177475, "learning_rate": 2.5751445086705205e-05, "loss": 0.3632, "step": 1032 }, { "epoch": 1.6095089633671082, "grad_norm": 0.3531877282131029, "learning_rate": 2.5722543352601158e-05, "loss": 0.366, "step": 1033 }, { "epoch": 1.6110678098207325, "grad_norm": 0.34051504848522857, "learning_rate": 2.5693641618497112e-05, "loss": 0.3591, "step": 1034 }, { "epoch": 1.612626656274357, "grad_norm": 0.3072876546153309, "learning_rate": 2.5664739884393062e-05, "loss": 0.3807, "step": 1035 }, { "epoch": 1.6141855027279814, "grad_norm": 0.3229272762875624, "learning_rate": 2.5635838150289022e-05, "loss": 0.3707, "step": 1036 }, { "epoch": 1.6157443491816057, "grad_norm": 0.3309237615111176, "learning_rate": 2.5606936416184972e-05, "loss": 0.3612, "step": 1037 }, { "epoch": 1.61730319563523, "grad_norm": 0.3111215230340758, "learning_rate": 2.5578034682080925e-05, "loss": 0.3803, "step": 1038 }, { "epoch": 1.6188620420888542, "grad_norm": 0.32805176015176535, "learning_rate": 2.554913294797688e-05, "loss": 0.3695, "step": 1039 }, { "epoch": 1.6204208885424785, "grad_norm": 0.2835660539290897, "learning_rate": 2.5520231213872836e-05, "loss": 0.3807, "step": 1040 }, { "epoch": 1.621979734996103, "grad_norm": 0.3032998483914643, "learning_rate": 2.549132947976879e-05, "loss": 0.3695, "step": 1041 }, { "epoch": 1.6235385814497272, "grad_norm": 0.27617969523468905, "learning_rate": 2.546242774566474e-05, "loss": 0.3723, "step": 1042 }, { "epoch": 1.6250974279033517, "grad_norm": 0.32157009299636713, "learning_rate": 2.5433526011560693e-05, "loss": 0.3572, "step": 1043 }, { "epoch": 1.626656274356976, "grad_norm": 0.2842521187908904, "learning_rate": 2.540462427745665e-05, "loss": 0.3656, "step": 1044 }, { "epoch": 1.6282151208106002, "grad_norm": 0.3288780675289122, "learning_rate": 2.5375722543352603e-05, "loss": 0.385, "step": 1045 }, { "epoch": 1.6297739672642244, "grad_norm": 0.32678871102435747, "learning_rate": 2.5346820809248556e-05, "loss": 0.3844, "step": 1046 }, { "epoch": 1.6313328137178487, "grad_norm": 0.39119637546888175, "learning_rate": 2.531791907514451e-05, "loss": 0.3542, "step": 1047 }, { "epoch": 1.6328916601714731, "grad_norm": 0.288052529248597, "learning_rate": 2.5289017341040467e-05, "loss": 0.3572, "step": 1048 }, { "epoch": 1.6344505066250974, "grad_norm": 0.3487713869931321, "learning_rate": 2.526011560693642e-05, "loss": 0.3542, "step": 1049 }, { "epoch": 1.6360093530787219, "grad_norm": 0.26513201445656637, "learning_rate": 2.523121387283237e-05, "loss": 0.3731, "step": 1050 }, { "epoch": 1.6375681995323461, "grad_norm": 0.32844205151640604, "learning_rate": 2.5202312138728324e-05, "loss": 0.3691, "step": 1051 }, { "epoch": 1.6391270459859704, "grad_norm": 0.30490163955005906, "learning_rate": 2.517341040462428e-05, "loss": 0.3813, "step": 1052 }, { "epoch": 1.6406858924395946, "grad_norm": 0.4555429114132785, "learning_rate": 2.5144508670520234e-05, "loss": 0.3818, "step": 1053 }, { "epoch": 1.6422447388932189, "grad_norm": 0.29026515692355687, "learning_rate": 2.5115606936416187e-05, "loss": 0.3626, "step": 1054 }, { "epoch": 1.6438035853468433, "grad_norm": 0.39179865186142615, "learning_rate": 2.5086705202312137e-05, "loss": 0.3502, "step": 1055 }, { "epoch": 1.6453624318004678, "grad_norm": 0.32433692268192194, "learning_rate": 2.5057803468208098e-05, "loss": 0.3781, "step": 1056 }, { "epoch": 1.646921278254092, "grad_norm": 0.33527613308127024, "learning_rate": 2.5028901734104048e-05, "loss": 0.3619, "step": 1057 }, { "epoch": 1.6484801247077163, "grad_norm": 0.3159961287138508, "learning_rate": 2.5e-05, "loss": 0.3632, "step": 1058 }, { "epoch": 1.6500389711613406, "grad_norm": 0.31940038082116945, "learning_rate": 2.4971098265895955e-05, "loss": 0.3633, "step": 1059 }, { "epoch": 1.6515978176149648, "grad_norm": 0.3303935987168502, "learning_rate": 2.4942196531791908e-05, "loss": 0.3745, "step": 1060 }, { "epoch": 1.6531566640685893, "grad_norm": 0.34507486262396847, "learning_rate": 2.491329479768786e-05, "loss": 0.3507, "step": 1061 }, { "epoch": 1.6547155105222136, "grad_norm": 0.2898972340965827, "learning_rate": 2.4884393063583815e-05, "loss": 0.3583, "step": 1062 }, { "epoch": 1.656274356975838, "grad_norm": 0.29250387172577985, "learning_rate": 2.485549132947977e-05, "loss": 0.3754, "step": 1063 }, { "epoch": 1.6578332034294623, "grad_norm": 0.3144587373068325, "learning_rate": 2.4826589595375725e-05, "loss": 0.3675, "step": 1064 }, { "epoch": 1.6593920498830865, "grad_norm": 0.28878104517546993, "learning_rate": 2.4797687861271675e-05, "loss": 0.3535, "step": 1065 }, { "epoch": 1.6609508963367108, "grad_norm": 0.2915446431012894, "learning_rate": 2.4768786127167632e-05, "loss": 0.3461, "step": 1066 }, { "epoch": 1.662509742790335, "grad_norm": 0.3180916504846977, "learning_rate": 2.4739884393063582e-05, "loss": 0.3482, "step": 1067 }, { "epoch": 1.6640685892439595, "grad_norm": 0.35825594962772983, "learning_rate": 2.471098265895954e-05, "loss": 0.3822, "step": 1068 }, { "epoch": 1.6656274356975838, "grad_norm": 0.2716382477039464, "learning_rate": 2.4682080924855492e-05, "loss": 0.3712, "step": 1069 }, { "epoch": 1.6671862821512082, "grad_norm": 0.3832340579480356, "learning_rate": 2.4653179190751446e-05, "loss": 0.3744, "step": 1070 }, { "epoch": 1.6687451286048325, "grad_norm": 0.30678572957271294, "learning_rate": 2.46242774566474e-05, "loss": 0.3579, "step": 1071 }, { "epoch": 1.6703039750584567, "grad_norm": 0.2843857794151713, "learning_rate": 2.4595375722543353e-05, "loss": 0.3826, "step": 1072 }, { "epoch": 1.671862821512081, "grad_norm": 0.34064253973495867, "learning_rate": 2.4566473988439306e-05, "loss": 0.3761, "step": 1073 }, { "epoch": 1.6734216679657052, "grad_norm": 0.29119739353949486, "learning_rate": 2.4537572254335263e-05, "loss": 0.3755, "step": 1074 }, { "epoch": 1.6749805144193297, "grad_norm": 0.2816956584902144, "learning_rate": 2.4508670520231213e-05, "loss": 0.3741, "step": 1075 }, { "epoch": 1.676539360872954, "grad_norm": 0.3111411062453413, "learning_rate": 2.447976878612717e-05, "loss": 0.3825, "step": 1076 }, { "epoch": 1.6780982073265784, "grad_norm": 0.2726926618664558, "learning_rate": 2.445086705202312e-05, "loss": 0.3879, "step": 1077 }, { "epoch": 1.6796570537802027, "grad_norm": 0.2834300086794794, "learning_rate": 2.4421965317919077e-05, "loss": 0.3705, "step": 1078 }, { "epoch": 1.681215900233827, "grad_norm": 0.31189081369621346, "learning_rate": 2.439306358381503e-05, "loss": 0.3663, "step": 1079 }, { "epoch": 1.6827747466874512, "grad_norm": 0.2829807568606232, "learning_rate": 2.4364161849710984e-05, "loss": 0.3626, "step": 1080 }, { "epoch": 1.6843335931410754, "grad_norm": 0.32489614588339644, "learning_rate": 2.4335260115606937e-05, "loss": 0.3733, "step": 1081 }, { "epoch": 1.6858924395947, "grad_norm": 0.3379997934308957, "learning_rate": 2.430635838150289e-05, "loss": 0.3843, "step": 1082 }, { "epoch": 1.6874512860483244, "grad_norm": 0.2730933075226193, "learning_rate": 2.4277456647398844e-05, "loss": 0.3825, "step": 1083 }, { "epoch": 1.6890101325019486, "grad_norm": 0.3404758440705323, "learning_rate": 2.4248554913294798e-05, "loss": 0.3692, "step": 1084 }, { "epoch": 1.690568978955573, "grad_norm": 0.2678166339094811, "learning_rate": 2.421965317919075e-05, "loss": 0.358, "step": 1085 }, { "epoch": 1.6921278254091972, "grad_norm": 0.3179761438396162, "learning_rate": 2.4190751445086708e-05, "loss": 0.3615, "step": 1086 }, { "epoch": 1.6936866718628214, "grad_norm": 0.30230071697163113, "learning_rate": 2.4161849710982658e-05, "loss": 0.3778, "step": 1087 }, { "epoch": 1.6952455183164459, "grad_norm": 0.291898879189332, "learning_rate": 2.4132947976878615e-05, "loss": 0.373, "step": 1088 }, { "epoch": 1.6968043647700701, "grad_norm": 0.2503981788139211, "learning_rate": 2.4104046242774568e-05, "loss": 0.3568, "step": 1089 }, { "epoch": 1.6983632112236946, "grad_norm": 0.28111682299046925, "learning_rate": 2.407514450867052e-05, "loss": 0.352, "step": 1090 }, { "epoch": 1.6999220576773189, "grad_norm": 0.2448550241175499, "learning_rate": 2.4046242774566475e-05, "loss": 0.3657, "step": 1091 }, { "epoch": 1.701480904130943, "grad_norm": 0.3065398908103042, "learning_rate": 2.401734104046243e-05, "loss": 0.3632, "step": 1092 }, { "epoch": 1.7030397505845674, "grad_norm": 0.2822693214761903, "learning_rate": 2.3988439306358382e-05, "loss": 0.401, "step": 1093 }, { "epoch": 1.7045985970381916, "grad_norm": 0.29360065434429594, "learning_rate": 2.3959537572254335e-05, "loss": 0.3682, "step": 1094 }, { "epoch": 1.706157443491816, "grad_norm": 0.24618104109851893, "learning_rate": 2.393063583815029e-05, "loss": 0.3653, "step": 1095 }, { "epoch": 1.7077162899454403, "grad_norm": 0.27658727281555634, "learning_rate": 2.3901734104046246e-05, "loss": 0.3703, "step": 1096 }, { "epoch": 1.7092751363990648, "grad_norm": 0.2777982792931725, "learning_rate": 2.3872832369942196e-05, "loss": 0.382, "step": 1097 }, { "epoch": 1.710833982852689, "grad_norm": 0.30238858038905553, "learning_rate": 2.3843930635838153e-05, "loss": 0.3646, "step": 1098 }, { "epoch": 1.7123928293063133, "grad_norm": 0.27806470539498046, "learning_rate": 2.3815028901734106e-05, "loss": 0.3745, "step": 1099 }, { "epoch": 1.7139516757599376, "grad_norm": 0.29767876896518725, "learning_rate": 2.378612716763006e-05, "loss": 0.3758, "step": 1100 }, { "epoch": 1.7155105222135618, "grad_norm": 0.2829737040629793, "learning_rate": 2.3757225433526013e-05, "loss": 0.3666, "step": 1101 }, { "epoch": 1.7170693686671863, "grad_norm": 0.3578161092071975, "learning_rate": 2.3728323699421966e-05, "loss": 0.3886, "step": 1102 }, { "epoch": 1.7186282151208108, "grad_norm": 0.29906079097318256, "learning_rate": 2.369942196531792e-05, "loss": 0.3671, "step": 1103 }, { "epoch": 1.720187061574435, "grad_norm": 0.28648360330595535, "learning_rate": 2.3670520231213873e-05, "loss": 0.3552, "step": 1104 }, { "epoch": 1.7217459080280593, "grad_norm": 0.2690083195767578, "learning_rate": 2.3641618497109827e-05, "loss": 0.3642, "step": 1105 }, { "epoch": 1.7233047544816835, "grad_norm": 0.3412710940527978, "learning_rate": 2.3612716763005784e-05, "loss": 0.3851, "step": 1106 }, { "epoch": 1.7248636009353078, "grad_norm": 0.2945093059468343, "learning_rate": 2.3583815028901734e-05, "loss": 0.3623, "step": 1107 }, { "epoch": 1.726422447388932, "grad_norm": 0.2745278267359284, "learning_rate": 2.355491329479769e-05, "loss": 0.3573, "step": 1108 }, { "epoch": 1.7279812938425565, "grad_norm": 0.2718225903508495, "learning_rate": 2.352601156069364e-05, "loss": 0.3641, "step": 1109 }, { "epoch": 1.729540140296181, "grad_norm": 0.27774676623694455, "learning_rate": 2.3497109826589597e-05, "loss": 0.3621, "step": 1110 }, { "epoch": 1.7310989867498052, "grad_norm": 0.26445890337068356, "learning_rate": 2.346820809248555e-05, "loss": 0.3616, "step": 1111 }, { "epoch": 1.7326578332034295, "grad_norm": 0.34740235201216224, "learning_rate": 2.3439306358381504e-05, "loss": 0.3549, "step": 1112 }, { "epoch": 1.7342166796570537, "grad_norm": 0.27846442379370767, "learning_rate": 2.3410404624277458e-05, "loss": 0.3582, "step": 1113 }, { "epoch": 1.735775526110678, "grad_norm": 0.3157341078739885, "learning_rate": 2.338150289017341e-05, "loss": 0.3777, "step": 1114 }, { "epoch": 1.7373343725643025, "grad_norm": 0.27171759230918247, "learning_rate": 2.3352601156069365e-05, "loss": 0.3752, "step": 1115 }, { "epoch": 1.7388932190179267, "grad_norm": 0.3016475132282448, "learning_rate": 2.332369942196532e-05, "loss": 0.3815, "step": 1116 }, { "epoch": 1.7404520654715512, "grad_norm": 0.28434294314632325, "learning_rate": 2.329479768786127e-05, "loss": 0.3771, "step": 1117 }, { "epoch": 1.7420109119251754, "grad_norm": 0.3013627601659192, "learning_rate": 2.326589595375723e-05, "loss": 0.3668, "step": 1118 }, { "epoch": 1.7435697583787997, "grad_norm": 0.2907093179335693, "learning_rate": 2.323699421965318e-05, "loss": 0.3939, "step": 1119 }, { "epoch": 1.745128604832424, "grad_norm": 0.2727965148056477, "learning_rate": 2.3208092485549135e-05, "loss": 0.3892, "step": 1120 }, { "epoch": 1.7466874512860482, "grad_norm": 0.2993429686182799, "learning_rate": 2.317919075144509e-05, "loss": 0.3519, "step": 1121 }, { "epoch": 1.7482462977396727, "grad_norm": 0.2693671384061593, "learning_rate": 2.3150289017341042e-05, "loss": 0.3549, "step": 1122 }, { "epoch": 1.749805144193297, "grad_norm": 0.3209589746292965, "learning_rate": 2.3121387283236996e-05, "loss": 0.375, "step": 1123 }, { "epoch": 1.7513639906469214, "grad_norm": 0.24312926167519938, "learning_rate": 2.309248554913295e-05, "loss": 0.3499, "step": 1124 }, { "epoch": 1.7529228371005456, "grad_norm": 0.3104273862594975, "learning_rate": 2.3063583815028902e-05, "loss": 0.389, "step": 1125 }, { "epoch": 1.75448168355417, "grad_norm": 0.2752541588528496, "learning_rate": 2.303468208092486e-05, "loss": 0.3646, "step": 1126 }, { "epoch": 1.7560405300077941, "grad_norm": 0.27374708814696136, "learning_rate": 2.300578034682081e-05, "loss": 0.3576, "step": 1127 }, { "epoch": 1.7575993764614184, "grad_norm": 0.31492934426283914, "learning_rate": 2.2976878612716766e-05, "loss": 0.3872, "step": 1128 }, { "epoch": 1.7591582229150429, "grad_norm": 0.28256761554526544, "learning_rate": 2.2947976878612716e-05, "loss": 0.3659, "step": 1129 }, { "epoch": 1.7607170693686673, "grad_norm": 0.2543705987468934, "learning_rate": 2.2919075144508673e-05, "loss": 0.3607, "step": 1130 }, { "epoch": 1.7622759158222916, "grad_norm": 0.298933402272287, "learning_rate": 2.2890173410404627e-05, "loss": 0.3683, "step": 1131 }, { "epoch": 1.7638347622759158, "grad_norm": 0.311816355725973, "learning_rate": 2.286127167630058e-05, "loss": 0.3597, "step": 1132 }, { "epoch": 1.76539360872954, "grad_norm": 0.3067218252259407, "learning_rate": 2.2832369942196533e-05, "loss": 0.3786, "step": 1133 }, { "epoch": 1.7669524551831643, "grad_norm": 0.2861975440862343, "learning_rate": 2.2803468208092487e-05, "loss": 0.3731, "step": 1134 }, { "epoch": 1.7685113016367888, "grad_norm": 0.316953066205446, "learning_rate": 2.277456647398844e-05, "loss": 0.3518, "step": 1135 }, { "epoch": 1.770070148090413, "grad_norm": 0.2708691409104658, "learning_rate": 2.2745664739884394e-05, "loss": 0.373, "step": 1136 }, { "epoch": 1.7716289945440375, "grad_norm": 0.26797311923183315, "learning_rate": 2.2716763005780347e-05, "loss": 0.3665, "step": 1137 }, { "epoch": 1.7731878409976618, "grad_norm": 0.3099497339980096, "learning_rate": 2.2687861271676304e-05, "loss": 0.3951, "step": 1138 }, { "epoch": 1.774746687451286, "grad_norm": 0.34054924476007226, "learning_rate": 2.2658959537572254e-05, "loss": 0.3594, "step": 1139 }, { "epoch": 1.7763055339049103, "grad_norm": 0.25223898072450013, "learning_rate": 2.263005780346821e-05, "loss": 0.361, "step": 1140 }, { "epoch": 1.7778643803585346, "grad_norm": 0.29284462313911747, "learning_rate": 2.2601156069364164e-05, "loss": 0.3575, "step": 1141 }, { "epoch": 1.779423226812159, "grad_norm": 0.2960130777592313, "learning_rate": 2.2572254335260118e-05, "loss": 0.3715, "step": 1142 }, { "epoch": 1.7809820732657833, "grad_norm": 0.26708445021070565, "learning_rate": 2.254335260115607e-05, "loss": 0.3685, "step": 1143 }, { "epoch": 1.7825409197194078, "grad_norm": 0.23850951752764862, "learning_rate": 2.2514450867052025e-05, "loss": 0.3528, "step": 1144 }, { "epoch": 1.784099766173032, "grad_norm": 0.28725957396773283, "learning_rate": 2.2485549132947978e-05, "loss": 0.3575, "step": 1145 }, { "epoch": 1.7856586126266563, "grad_norm": 0.27977866068073226, "learning_rate": 2.245664739884393e-05, "loss": 0.3605, "step": 1146 }, { "epoch": 1.7872174590802805, "grad_norm": 0.27572937920147894, "learning_rate": 2.2427745664739885e-05, "loss": 0.3797, "step": 1147 }, { "epoch": 1.7887763055339048, "grad_norm": 0.3044698433558107, "learning_rate": 2.2398843930635842e-05, "loss": 0.3561, "step": 1148 }, { "epoch": 1.7903351519875292, "grad_norm": 0.3034254736946843, "learning_rate": 2.2369942196531792e-05, "loss": 0.3729, "step": 1149 }, { "epoch": 1.7918939984411537, "grad_norm": 0.29184309753197535, "learning_rate": 2.234104046242775e-05, "loss": 0.3602, "step": 1150 }, { "epoch": 1.793452844894778, "grad_norm": 0.2752185880115312, "learning_rate": 2.23121387283237e-05, "loss": 0.3705, "step": 1151 }, { "epoch": 1.7950116913484022, "grad_norm": 0.2809903038728107, "learning_rate": 2.2283236994219656e-05, "loss": 0.3545, "step": 1152 }, { "epoch": 1.7965705378020265, "grad_norm": 0.3185758597066887, "learning_rate": 2.225433526011561e-05, "loss": 0.3775, "step": 1153 }, { "epoch": 1.7981293842556507, "grad_norm": 0.32522837307739677, "learning_rate": 2.2225433526011563e-05, "loss": 0.3762, "step": 1154 }, { "epoch": 1.799688230709275, "grad_norm": 0.29829669655375685, "learning_rate": 2.2196531791907516e-05, "loss": 0.3431, "step": 1155 }, { "epoch": 1.8012470771628994, "grad_norm": 0.2923880518129942, "learning_rate": 2.216763005780347e-05, "loss": 0.3625, "step": 1156 }, { "epoch": 1.802805923616524, "grad_norm": 0.31193429426470687, "learning_rate": 2.2138728323699423e-05, "loss": 0.3687, "step": 1157 }, { "epoch": 1.8043647700701482, "grad_norm": 0.3028214816745831, "learning_rate": 2.2109826589595376e-05, "loss": 0.3835, "step": 1158 }, { "epoch": 1.8059236165237724, "grad_norm": 0.35099054840606936, "learning_rate": 2.208092485549133e-05, "loss": 0.3785, "step": 1159 }, { "epoch": 1.8074824629773967, "grad_norm": 0.3153451160744472, "learning_rate": 2.2052023121387283e-05, "loss": 0.367, "step": 1160 }, { "epoch": 1.809041309431021, "grad_norm": 0.3253785936574865, "learning_rate": 2.2023121387283237e-05, "loss": 0.3877, "step": 1161 }, { "epoch": 1.8106001558846454, "grad_norm": 0.30760323974470033, "learning_rate": 2.199421965317919e-05, "loss": 0.3643, "step": 1162 }, { "epoch": 1.8121590023382697, "grad_norm": 0.3598054851684999, "learning_rate": 2.1965317919075147e-05, "loss": 0.3586, "step": 1163 }, { "epoch": 1.8137178487918941, "grad_norm": 0.26304147816154894, "learning_rate": 2.1936416184971097e-05, "loss": 0.3621, "step": 1164 }, { "epoch": 1.8152766952455184, "grad_norm": 0.3322641073371267, "learning_rate": 2.1907514450867054e-05, "loss": 0.3787, "step": 1165 }, { "epoch": 1.8168355416991426, "grad_norm": 0.2972445280530977, "learning_rate": 2.1878612716763007e-05, "loss": 0.3623, "step": 1166 }, { "epoch": 1.8183943881527669, "grad_norm": 0.29860325366761714, "learning_rate": 2.184971098265896e-05, "loss": 0.3693, "step": 1167 }, { "epoch": 1.8199532346063911, "grad_norm": 0.31932300688288, "learning_rate": 2.1820809248554914e-05, "loss": 0.3687, "step": 1168 }, { "epoch": 1.8215120810600156, "grad_norm": 0.25058398882971866, "learning_rate": 2.1791907514450868e-05, "loss": 0.3688, "step": 1169 }, { "epoch": 1.8230709275136399, "grad_norm": 0.30451618236243205, "learning_rate": 2.176300578034682e-05, "loss": 0.3604, "step": 1170 }, { "epoch": 1.8246297739672643, "grad_norm": 0.30692079739121414, "learning_rate": 2.1734104046242775e-05, "loss": 0.3515, "step": 1171 }, { "epoch": 1.8261886204208886, "grad_norm": 0.2807778315885177, "learning_rate": 2.1705202312138728e-05, "loss": 0.3509, "step": 1172 }, { "epoch": 1.8277474668745128, "grad_norm": 0.2905244115992112, "learning_rate": 2.1676300578034685e-05, "loss": 0.3661, "step": 1173 }, { "epoch": 1.829306313328137, "grad_norm": 0.3105261582464057, "learning_rate": 2.1647398843930635e-05, "loss": 0.372, "step": 1174 }, { "epoch": 1.8308651597817613, "grad_norm": 0.3095524114344275, "learning_rate": 2.1618497109826592e-05, "loss": 0.3683, "step": 1175 }, { "epoch": 1.8324240062353858, "grad_norm": 0.3224707852323462, "learning_rate": 2.1589595375722542e-05, "loss": 0.3941, "step": 1176 }, { "epoch": 1.8339828526890103, "grad_norm": 0.37265316733469883, "learning_rate": 2.15606936416185e-05, "loss": 0.3729, "step": 1177 }, { "epoch": 1.8355416991426345, "grad_norm": 0.3484805036172097, "learning_rate": 2.1531791907514452e-05, "loss": 0.3695, "step": 1178 }, { "epoch": 1.8371005455962588, "grad_norm": 0.3062185230582236, "learning_rate": 2.1502890173410405e-05, "loss": 0.3722, "step": 1179 }, { "epoch": 1.838659392049883, "grad_norm": 0.3514017844930539, "learning_rate": 2.147398843930636e-05, "loss": 0.3716, "step": 1180 }, { "epoch": 1.8402182385035073, "grad_norm": 0.29370679610653583, "learning_rate": 2.1445086705202312e-05, "loss": 0.3511, "step": 1181 }, { "epoch": 1.8417770849571318, "grad_norm": 0.3650311060921678, "learning_rate": 2.1416184971098266e-05, "loss": 0.3746, "step": 1182 }, { "epoch": 1.843335931410756, "grad_norm": 0.28361922423446095, "learning_rate": 2.1387283236994223e-05, "loss": 0.354, "step": 1183 }, { "epoch": 1.8448947778643805, "grad_norm": 0.2915388311697381, "learning_rate": 2.1358381502890173e-05, "loss": 0.36, "step": 1184 }, { "epoch": 1.8464536243180047, "grad_norm": 0.33038140947084105, "learning_rate": 2.132947976878613e-05, "loss": 0.3554, "step": 1185 }, { "epoch": 1.848012470771629, "grad_norm": 0.419071588239709, "learning_rate": 2.130057803468208e-05, "loss": 0.3834, "step": 1186 }, { "epoch": 1.8495713172252533, "grad_norm": 0.3514728435923063, "learning_rate": 2.1271676300578036e-05, "loss": 0.3627, "step": 1187 }, { "epoch": 1.8511301636788775, "grad_norm": 0.35707340267403825, "learning_rate": 2.124277456647399e-05, "loss": 0.3874, "step": 1188 }, { "epoch": 1.852689010132502, "grad_norm": 0.37099787424788894, "learning_rate": 2.1213872832369943e-05, "loss": 0.3726, "step": 1189 }, { "epoch": 1.8542478565861262, "grad_norm": 0.330071756650617, "learning_rate": 2.1184971098265897e-05, "loss": 0.3715, "step": 1190 }, { "epoch": 1.8558067030397507, "grad_norm": 0.37948140477016307, "learning_rate": 2.115606936416185e-05, "loss": 0.3561, "step": 1191 }, { "epoch": 1.857365549493375, "grad_norm": 0.27396995420812964, "learning_rate": 2.1127167630057804e-05, "loss": 0.38, "step": 1192 }, { "epoch": 1.8589243959469992, "grad_norm": 0.35094585580789595, "learning_rate": 2.1098265895953757e-05, "loss": 0.3646, "step": 1193 }, { "epoch": 1.8604832424006235, "grad_norm": 0.3047789152628225, "learning_rate": 2.106936416184971e-05, "loss": 0.3778, "step": 1194 }, { "epoch": 1.8620420888542477, "grad_norm": 0.33234879657650934, "learning_rate": 2.1040462427745667e-05, "loss": 0.3713, "step": 1195 }, { "epoch": 1.8636009353078722, "grad_norm": 0.32766164310333334, "learning_rate": 2.1011560693641617e-05, "loss": 0.3736, "step": 1196 }, { "epoch": 1.8651597817614964, "grad_norm": 0.2759861630025658, "learning_rate": 2.0982658959537574e-05, "loss": 0.3796, "step": 1197 }, { "epoch": 1.866718628215121, "grad_norm": 0.32521839552334825, "learning_rate": 2.0953757225433528e-05, "loss": 0.362, "step": 1198 }, { "epoch": 1.8682774746687452, "grad_norm": 0.2931283862783075, "learning_rate": 2.092485549132948e-05, "loss": 0.3502, "step": 1199 }, { "epoch": 1.8698363211223694, "grad_norm": 0.2991343395943882, "learning_rate": 2.0895953757225435e-05, "loss": 0.3714, "step": 1200 }, { "epoch": 1.8713951675759937, "grad_norm": 0.2932605788220294, "learning_rate": 2.0867052023121388e-05, "loss": 0.3819, "step": 1201 }, { "epoch": 1.872954014029618, "grad_norm": 0.30785885245721334, "learning_rate": 2.083815028901734e-05, "loss": 0.3471, "step": 1202 }, { "epoch": 1.8745128604832424, "grad_norm": 0.32014124433096525, "learning_rate": 2.0809248554913295e-05, "loss": 0.3561, "step": 1203 }, { "epoch": 1.8760717069368669, "grad_norm": 0.3235541208435597, "learning_rate": 2.078034682080925e-05, "loss": 0.3866, "step": 1204 }, { "epoch": 1.8776305533904911, "grad_norm": 0.30949299738115854, "learning_rate": 2.0751445086705205e-05, "loss": 0.3639, "step": 1205 }, { "epoch": 1.8791893998441154, "grad_norm": 0.2817005798429569, "learning_rate": 2.0722543352601155e-05, "loss": 0.3597, "step": 1206 }, { "epoch": 1.8807482462977396, "grad_norm": 0.30497938827386945, "learning_rate": 2.0693641618497112e-05, "loss": 0.383, "step": 1207 }, { "epoch": 1.8823070927513639, "grad_norm": 0.25436146038693197, "learning_rate": 2.0664739884393066e-05, "loss": 0.3692, "step": 1208 }, { "epoch": 1.8838659392049883, "grad_norm": 0.3056404459036848, "learning_rate": 2.063583815028902e-05, "loss": 0.3704, "step": 1209 }, { "epoch": 1.8854247856586126, "grad_norm": 0.26875488415904514, "learning_rate": 2.0606936416184972e-05, "loss": 0.3594, "step": 1210 }, { "epoch": 1.886983632112237, "grad_norm": 0.3511619874207967, "learning_rate": 2.0578034682080926e-05, "loss": 0.3798, "step": 1211 }, { "epoch": 1.8885424785658613, "grad_norm": 0.26158426756531344, "learning_rate": 2.054913294797688e-05, "loss": 0.3856, "step": 1212 }, { "epoch": 1.8901013250194856, "grad_norm": 0.3122928786540747, "learning_rate": 2.0520231213872833e-05, "loss": 0.3677, "step": 1213 }, { "epoch": 1.8916601714731098, "grad_norm": 0.2684224099385814, "learning_rate": 2.0491329479768786e-05, "loss": 0.3534, "step": 1214 }, { "epoch": 1.893219017926734, "grad_norm": 0.2571315731383385, "learning_rate": 2.0462427745664743e-05, "loss": 0.3722, "step": 1215 }, { "epoch": 1.8947778643803586, "grad_norm": 0.25588546444512905, "learning_rate": 2.0433526011560693e-05, "loss": 0.3681, "step": 1216 }, { "epoch": 1.8963367108339828, "grad_norm": 0.3187680200863735, "learning_rate": 2.040462427745665e-05, "loss": 0.3634, "step": 1217 }, { "epoch": 1.8978955572876073, "grad_norm": 0.2611861721574427, "learning_rate": 2.03757225433526e-05, "loss": 0.3879, "step": 1218 }, { "epoch": 1.8994544037412315, "grad_norm": 0.30654388499949137, "learning_rate": 2.0346820809248557e-05, "loss": 0.363, "step": 1219 }, { "epoch": 1.9010132501948558, "grad_norm": 0.32726906906242276, "learning_rate": 2.031791907514451e-05, "loss": 0.369, "step": 1220 }, { "epoch": 1.90257209664848, "grad_norm": 0.28441311387650836, "learning_rate": 2.0289017341040464e-05, "loss": 0.3559, "step": 1221 }, { "epoch": 1.9041309431021043, "grad_norm": 0.2657998910454343, "learning_rate": 2.0260115606936417e-05, "loss": 0.3794, "step": 1222 }, { "epoch": 1.9056897895557288, "grad_norm": 0.31053785670596307, "learning_rate": 2.023121387283237e-05, "loss": 0.3658, "step": 1223 }, { "epoch": 1.9072486360093532, "grad_norm": 0.30798708262091334, "learning_rate": 2.0202312138728324e-05, "loss": 0.3603, "step": 1224 }, { "epoch": 1.9088074824629775, "grad_norm": 0.2523008465363555, "learning_rate": 2.017341040462428e-05, "loss": 0.3724, "step": 1225 }, { "epoch": 1.9103663289166017, "grad_norm": 0.2763374659384913, "learning_rate": 2.014450867052023e-05, "loss": 0.3548, "step": 1226 }, { "epoch": 1.911925175370226, "grad_norm": 0.24964103816057576, "learning_rate": 2.0115606936416188e-05, "loss": 0.3694, "step": 1227 }, { "epoch": 1.9134840218238502, "grad_norm": 0.28584599942242367, "learning_rate": 2.0086705202312138e-05, "loss": 0.3644, "step": 1228 }, { "epoch": 1.9150428682774747, "grad_norm": 0.253224234190821, "learning_rate": 2.0057803468208095e-05, "loss": 0.3651, "step": 1229 }, { "epoch": 1.916601714731099, "grad_norm": 0.2853460997991044, "learning_rate": 2.0028901734104048e-05, "loss": 0.3865, "step": 1230 }, { "epoch": 1.9181605611847234, "grad_norm": 0.2914953043711197, "learning_rate": 2e-05, "loss": 0.3705, "step": 1231 }, { "epoch": 1.9197194076383477, "grad_norm": 0.2703628033375394, "learning_rate": 1.9971098265895955e-05, "loss": 0.35, "step": 1232 }, { "epoch": 1.921278254091972, "grad_norm": 0.3403993339455389, "learning_rate": 1.994219653179191e-05, "loss": 0.367, "step": 1233 }, { "epoch": 1.9228371005455962, "grad_norm": 0.2403088129031985, "learning_rate": 1.9913294797687862e-05, "loss": 0.3554, "step": 1234 }, { "epoch": 1.9243959469992205, "grad_norm": 0.2811341273215568, "learning_rate": 1.988439306358382e-05, "loss": 0.3617, "step": 1235 }, { "epoch": 1.925954793452845, "grad_norm": 0.2878197944500182, "learning_rate": 1.985549132947977e-05, "loss": 0.3574, "step": 1236 }, { "epoch": 1.9275136399064692, "grad_norm": 0.2774310293587101, "learning_rate": 1.9826589595375726e-05, "loss": 0.3748, "step": 1237 }, { "epoch": 1.9290724863600937, "grad_norm": 0.3115605490343856, "learning_rate": 1.9797687861271676e-05, "loss": 0.353, "step": 1238 }, { "epoch": 1.930631332813718, "grad_norm": 0.25590953954693557, "learning_rate": 1.9768786127167633e-05, "loss": 0.3662, "step": 1239 }, { "epoch": 1.9321901792673422, "grad_norm": 0.2555007229781606, "learning_rate": 1.9739884393063586e-05, "loss": 0.3494, "step": 1240 }, { "epoch": 1.9337490257209664, "grad_norm": 0.2691775864628584, "learning_rate": 1.971098265895954e-05, "loss": 0.3695, "step": 1241 }, { "epoch": 1.9353078721745907, "grad_norm": 0.27904072318741047, "learning_rate": 1.9682080924855493e-05, "loss": 0.3553, "step": 1242 }, { "epoch": 1.9368667186282151, "grad_norm": 0.2940422073737695, "learning_rate": 1.9653179190751446e-05, "loss": 0.3732, "step": 1243 }, { "epoch": 1.9384255650818394, "grad_norm": 0.28761825900324317, "learning_rate": 1.96242774566474e-05, "loss": 0.3739, "step": 1244 }, { "epoch": 1.9399844115354639, "grad_norm": 0.30744221511987735, "learning_rate": 1.9595375722543353e-05, "loss": 0.3598, "step": 1245 }, { "epoch": 1.9415432579890881, "grad_norm": 0.2669547151465641, "learning_rate": 1.9566473988439307e-05, "loss": 0.3625, "step": 1246 }, { "epoch": 1.9431021044427124, "grad_norm": 0.3163867104214392, "learning_rate": 1.9537572254335264e-05, "loss": 0.3527, "step": 1247 }, { "epoch": 1.9446609508963366, "grad_norm": 0.2724508920694651, "learning_rate": 1.9508670520231214e-05, "loss": 0.349, "step": 1248 }, { "epoch": 1.9462197973499609, "grad_norm": 0.31987516157242796, "learning_rate": 1.947976878612717e-05, "loss": 0.3578, "step": 1249 }, { "epoch": 1.9477786438035853, "grad_norm": 0.27194855512008154, "learning_rate": 1.9450867052023124e-05, "loss": 0.3717, "step": 1250 }, { "epoch": 1.9493374902572098, "grad_norm": 0.42931537141861814, "learning_rate": 1.9421965317919074e-05, "loss": 0.3762, "step": 1251 }, { "epoch": 1.950896336710834, "grad_norm": 0.263639511648232, "learning_rate": 1.939306358381503e-05, "loss": 0.37, "step": 1252 }, { "epoch": 1.9524551831644583, "grad_norm": 0.3046013215628072, "learning_rate": 1.936416184971098e-05, "loss": 0.3657, "step": 1253 }, { "epoch": 1.9540140296180826, "grad_norm": 0.27822048356909573, "learning_rate": 1.9335260115606938e-05, "loss": 0.3702, "step": 1254 }, { "epoch": 1.9555728760717068, "grad_norm": 0.2503963888170451, "learning_rate": 1.930635838150289e-05, "loss": 0.3533, "step": 1255 }, { "epoch": 1.9571317225253313, "grad_norm": 0.3608440008147506, "learning_rate": 1.9277456647398845e-05, "loss": 0.3617, "step": 1256 }, { "epoch": 1.9586905689789555, "grad_norm": 0.2618137124636811, "learning_rate": 1.9248554913294798e-05, "loss": 0.3536, "step": 1257 }, { "epoch": 1.96024941543258, "grad_norm": 0.31638615142182935, "learning_rate": 1.921965317919075e-05, "loss": 0.3646, "step": 1258 }, { "epoch": 1.9618082618862043, "grad_norm": 0.2902981949466536, "learning_rate": 1.9190751445086705e-05, "loss": 0.3576, "step": 1259 }, { "epoch": 1.9633671083398285, "grad_norm": 0.24681016366367486, "learning_rate": 1.916184971098266e-05, "loss": 0.37, "step": 1260 }, { "epoch": 1.9649259547934528, "grad_norm": 0.30886035269615025, "learning_rate": 1.9132947976878612e-05, "loss": 0.3735, "step": 1261 }, { "epoch": 1.966484801247077, "grad_norm": 0.3261707681560328, "learning_rate": 1.910404624277457e-05, "loss": 0.3802, "step": 1262 }, { "epoch": 1.9680436477007015, "grad_norm": 0.27343056543520194, "learning_rate": 1.907514450867052e-05, "loss": 0.3689, "step": 1263 }, { "epoch": 1.9696024941543258, "grad_norm": 0.2919468083028249, "learning_rate": 1.9046242774566476e-05, "loss": 0.3599, "step": 1264 }, { "epoch": 1.9711613406079502, "grad_norm": 0.2994649798188123, "learning_rate": 1.901734104046243e-05, "loss": 0.3713, "step": 1265 }, { "epoch": 1.9727201870615745, "grad_norm": 0.3564440516448399, "learning_rate": 1.8988439306358382e-05, "loss": 0.3451, "step": 1266 }, { "epoch": 1.9742790335151987, "grad_norm": 0.28761498604656505, "learning_rate": 1.8959537572254336e-05, "loss": 0.3691, "step": 1267 }, { "epoch": 1.975837879968823, "grad_norm": 0.29458672500997857, "learning_rate": 1.893063583815029e-05, "loss": 0.3694, "step": 1268 }, { "epoch": 1.9773967264224472, "grad_norm": 0.28968659606971875, "learning_rate": 1.8901734104046243e-05, "loss": 0.3757, "step": 1269 }, { "epoch": 1.9789555728760717, "grad_norm": 0.25074130760016594, "learning_rate": 1.8872832369942196e-05, "loss": 0.3671, "step": 1270 }, { "epoch": 1.9805144193296962, "grad_norm": 0.2727775661900619, "learning_rate": 1.884393063583815e-05, "loss": 0.3763, "step": 1271 }, { "epoch": 1.9820732657833204, "grad_norm": 0.2540694971749628, "learning_rate": 1.8815028901734106e-05, "loss": 0.361, "step": 1272 }, { "epoch": 1.9836321122369447, "grad_norm": 0.26620323560623, "learning_rate": 1.8786127167630057e-05, "loss": 0.3804, "step": 1273 }, { "epoch": 1.985190958690569, "grad_norm": 0.2770753239201809, "learning_rate": 1.8757225433526013e-05, "loss": 0.3511, "step": 1274 }, { "epoch": 1.9867498051441932, "grad_norm": 0.28316981515390427, "learning_rate": 1.8728323699421967e-05, "loss": 0.3527, "step": 1275 }, { "epoch": 1.9883086515978177, "grad_norm": 0.24806656551525558, "learning_rate": 1.869942196531792e-05, "loss": 0.3888, "step": 1276 }, { "epoch": 1.989867498051442, "grad_norm": 0.2751522114384385, "learning_rate": 1.8670520231213874e-05, "loss": 0.3591, "step": 1277 }, { "epoch": 1.9914263445050664, "grad_norm": 0.29543750469737046, "learning_rate": 1.8641618497109827e-05, "loss": 0.3604, "step": 1278 }, { "epoch": 1.9929851909586906, "grad_norm": 0.27104672954069825, "learning_rate": 1.861271676300578e-05, "loss": 0.3686, "step": 1279 }, { "epoch": 1.994544037412315, "grad_norm": 0.26770020711721565, "learning_rate": 1.8583815028901734e-05, "loss": 0.3919, "step": 1280 }, { "epoch": 1.9961028838659391, "grad_norm": 0.3002772154980313, "learning_rate": 1.8554913294797688e-05, "loss": 0.378, "step": 1281 }, { "epoch": 1.9976617303195634, "grad_norm": 0.2934928063887189, "learning_rate": 1.8526011560693644e-05, "loss": 0.3851, "step": 1282 }, { "epoch": 1.9992205767731879, "grad_norm": 0.26071286159069573, "learning_rate": 1.8497109826589594e-05, "loss": 0.3661, "step": 1283 }, { "epoch": 2.0, "grad_norm": 1.2407353049559013, "learning_rate": 1.846820809248555e-05, "loss": 0.3315, "step": 1284 }, { "epoch": 2.0015588464536243, "grad_norm": 0.3019354121209205, "learning_rate": 1.84393063583815e-05, "loss": 0.3132, "step": 1285 }, { "epoch": 2.0031176929072485, "grad_norm": 0.3221654485813362, "learning_rate": 1.8410404624277458e-05, "loss": 0.3049, "step": 1286 }, { "epoch": 2.0046765393608728, "grad_norm": 0.3107212962226372, "learning_rate": 1.838150289017341e-05, "loss": 0.2962, "step": 1287 }, { "epoch": 2.0062353858144975, "grad_norm": 0.2652451153347466, "learning_rate": 1.8352601156069365e-05, "loss": 0.2848, "step": 1288 }, { "epoch": 2.0077942322681217, "grad_norm": 0.29115331241478903, "learning_rate": 1.832369942196532e-05, "loss": 0.2974, "step": 1289 }, { "epoch": 2.009353078721746, "grad_norm": 0.29937191649859674, "learning_rate": 1.8294797687861272e-05, "loss": 0.28, "step": 1290 }, { "epoch": 2.01091192517537, "grad_norm": 0.3181952793594262, "learning_rate": 1.8265895953757225e-05, "loss": 0.2795, "step": 1291 }, { "epoch": 2.0124707716289945, "grad_norm": 0.3030873579967208, "learning_rate": 1.8236994219653182e-05, "loss": 0.3004, "step": 1292 }, { "epoch": 2.0140296180826187, "grad_norm": 0.30093086334228963, "learning_rate": 1.8208092485549132e-05, "loss": 0.2987, "step": 1293 }, { "epoch": 2.015588464536243, "grad_norm": 0.31992474845831625, "learning_rate": 1.817919075144509e-05, "loss": 0.2921, "step": 1294 }, { "epoch": 2.0171473109898677, "grad_norm": 0.31971945976797334, "learning_rate": 1.815028901734104e-05, "loss": 0.2938, "step": 1295 }, { "epoch": 2.018706157443492, "grad_norm": 0.2784528999275469, "learning_rate": 1.8121387283236996e-05, "loss": 0.2969, "step": 1296 }, { "epoch": 2.020265003897116, "grad_norm": 0.28547739386324084, "learning_rate": 1.809248554913295e-05, "loss": 0.2819, "step": 1297 }, { "epoch": 2.0218238503507404, "grad_norm": 0.35064000763654246, "learning_rate": 1.8063583815028903e-05, "loss": 0.2886, "step": 1298 }, { "epoch": 2.0233826968043647, "grad_norm": 0.298377942201609, "learning_rate": 1.8034682080924856e-05, "loss": 0.2888, "step": 1299 }, { "epoch": 2.024941543257989, "grad_norm": 0.2741254018613014, "learning_rate": 1.800578034682081e-05, "loss": 0.2846, "step": 1300 }, { "epoch": 2.0265003897116136, "grad_norm": 0.3278384877642372, "learning_rate": 1.7976878612716763e-05, "loss": 0.2996, "step": 1301 }, { "epoch": 2.028059236165238, "grad_norm": 0.31062098942293614, "learning_rate": 1.794797687861272e-05, "loss": 0.3004, "step": 1302 }, { "epoch": 2.029618082618862, "grad_norm": 0.2840675944649226, "learning_rate": 1.791907514450867e-05, "loss": 0.2915, "step": 1303 }, { "epoch": 2.0311769290724864, "grad_norm": 0.3066133656454931, "learning_rate": 1.7890173410404627e-05, "loss": 0.2887, "step": 1304 }, { "epoch": 2.0327357755261106, "grad_norm": 0.2729351456670734, "learning_rate": 1.7861271676300577e-05, "loss": 0.2984, "step": 1305 }, { "epoch": 2.034294621979735, "grad_norm": 0.2636535310871996, "learning_rate": 1.7832369942196534e-05, "loss": 0.291, "step": 1306 }, { "epoch": 2.035853468433359, "grad_norm": 0.2964005719112254, "learning_rate": 1.7803468208092487e-05, "loss": 0.2878, "step": 1307 }, { "epoch": 2.037412314886984, "grad_norm": 0.2362770470391929, "learning_rate": 1.777456647398844e-05, "loss": 0.2889, "step": 1308 }, { "epoch": 2.038971161340608, "grad_norm": 0.2877360765498249, "learning_rate": 1.7745664739884394e-05, "loss": 0.306, "step": 1309 }, { "epoch": 2.0405300077942323, "grad_norm": 0.26635846301125965, "learning_rate": 1.7716763005780348e-05, "loss": 0.2835, "step": 1310 }, { "epoch": 2.0420888542478566, "grad_norm": 0.2364083261820846, "learning_rate": 1.76878612716763e-05, "loss": 0.2813, "step": 1311 }, { "epoch": 2.043647700701481, "grad_norm": 0.25071829512152555, "learning_rate": 1.7658959537572255e-05, "loss": 0.2722, "step": 1312 }, { "epoch": 2.045206547155105, "grad_norm": 0.2640068234739541, "learning_rate": 1.7630057803468208e-05, "loss": 0.3078, "step": 1313 }, { "epoch": 2.0467653936087293, "grad_norm": 0.2733838421400439, "learning_rate": 1.7601156069364165e-05, "loss": 0.2936, "step": 1314 }, { "epoch": 2.048324240062354, "grad_norm": 0.2550755795463892, "learning_rate": 1.7572254335260115e-05, "loss": 0.2819, "step": 1315 }, { "epoch": 2.0498830865159783, "grad_norm": 0.25545376074388176, "learning_rate": 1.754335260115607e-05, "loss": 0.2823, "step": 1316 }, { "epoch": 2.0514419329696025, "grad_norm": 0.22661427370917822, "learning_rate": 1.7514450867052025e-05, "loss": 0.2824, "step": 1317 }, { "epoch": 2.053000779423227, "grad_norm": 0.29774507660120497, "learning_rate": 1.748554913294798e-05, "loss": 0.2838, "step": 1318 }, { "epoch": 2.054559625876851, "grad_norm": 0.2598671788363091, "learning_rate": 1.7456647398843932e-05, "loss": 0.2903, "step": 1319 }, { "epoch": 2.0561184723304753, "grad_norm": 0.2519961479463723, "learning_rate": 1.7427745664739885e-05, "loss": 0.2571, "step": 1320 }, { "epoch": 2.0576773187841, "grad_norm": 0.25326241935804067, "learning_rate": 1.739884393063584e-05, "loss": 0.2977, "step": 1321 }, { "epoch": 2.0592361652377242, "grad_norm": 0.24057089058554867, "learning_rate": 1.7369942196531792e-05, "loss": 0.2866, "step": 1322 }, { "epoch": 2.0607950116913485, "grad_norm": 0.22295403456879342, "learning_rate": 1.7341040462427746e-05, "loss": 0.28, "step": 1323 }, { "epoch": 2.0623538581449727, "grad_norm": 0.2868118439567236, "learning_rate": 1.7312138728323703e-05, "loss": 0.3044, "step": 1324 }, { "epoch": 2.063912704598597, "grad_norm": 0.2708711985497848, "learning_rate": 1.7283236994219653e-05, "loss": 0.2823, "step": 1325 }, { "epoch": 2.0654715510522212, "grad_norm": 0.42537603462611046, "learning_rate": 1.725433526011561e-05, "loss": 0.2984, "step": 1326 }, { "epoch": 2.0670303975058455, "grad_norm": 0.25936897272013576, "learning_rate": 1.722543352601156e-05, "loss": 0.2996, "step": 1327 }, { "epoch": 2.06858924395947, "grad_norm": 0.29205530294187826, "learning_rate": 1.7196531791907516e-05, "loss": 0.2815, "step": 1328 }, { "epoch": 2.0701480904130944, "grad_norm": 0.2498440090336756, "learning_rate": 1.716763005780347e-05, "loss": 0.2908, "step": 1329 }, { "epoch": 2.0717069368667187, "grad_norm": 0.2803332912261779, "learning_rate": 1.7138728323699423e-05, "loss": 0.2919, "step": 1330 }, { "epoch": 2.073265783320343, "grad_norm": 0.2714375808140227, "learning_rate": 1.7109826589595377e-05, "loss": 0.285, "step": 1331 }, { "epoch": 2.074824629773967, "grad_norm": 0.23047443561287922, "learning_rate": 1.708092485549133e-05, "loss": 0.2982, "step": 1332 }, { "epoch": 2.0763834762275915, "grad_norm": 0.27974401775468655, "learning_rate": 1.7052023121387284e-05, "loss": 0.2816, "step": 1333 }, { "epoch": 2.0779423226812157, "grad_norm": 0.27728883446303665, "learning_rate": 1.702312138728324e-05, "loss": 0.2874, "step": 1334 }, { "epoch": 2.0795011691348404, "grad_norm": 0.22428181416464069, "learning_rate": 1.699421965317919e-05, "loss": 0.2959, "step": 1335 }, { "epoch": 2.0810600155884647, "grad_norm": 0.4070173071535244, "learning_rate": 1.6965317919075147e-05, "loss": 0.2853, "step": 1336 }, { "epoch": 2.082618862042089, "grad_norm": 0.33199185273932175, "learning_rate": 1.6936416184971097e-05, "loss": 0.2825, "step": 1337 }, { "epoch": 2.084177708495713, "grad_norm": 0.23889026155636925, "learning_rate": 1.6907514450867054e-05, "loss": 0.2878, "step": 1338 }, { "epoch": 2.0857365549493374, "grad_norm": 0.2588106933535315, "learning_rate": 1.6878612716763008e-05, "loss": 0.291, "step": 1339 }, { "epoch": 2.0872954014029617, "grad_norm": 0.2521130177780321, "learning_rate": 1.684971098265896e-05, "loss": 0.2927, "step": 1340 }, { "epoch": 2.0888542478565864, "grad_norm": 0.28556085649359164, "learning_rate": 1.6820809248554915e-05, "loss": 0.3141, "step": 1341 }, { "epoch": 2.0904130943102106, "grad_norm": 0.23613222180841584, "learning_rate": 1.6791907514450868e-05, "loss": 0.2861, "step": 1342 }, { "epoch": 2.091971940763835, "grad_norm": 0.24684587451326986, "learning_rate": 1.676300578034682e-05, "loss": 0.2853, "step": 1343 }, { "epoch": 2.093530787217459, "grad_norm": 0.23325395667135312, "learning_rate": 1.673410404624278e-05, "loss": 0.2908, "step": 1344 }, { "epoch": 2.0950896336710834, "grad_norm": 0.2594871779004086, "learning_rate": 1.670520231213873e-05, "loss": 0.3087, "step": 1345 }, { "epoch": 2.0966484801247076, "grad_norm": 0.23321943315628127, "learning_rate": 1.6676300578034685e-05, "loss": 0.2958, "step": 1346 }, { "epoch": 2.098207326578332, "grad_norm": 0.23482451299773394, "learning_rate": 1.6647398843930635e-05, "loss": 0.3037, "step": 1347 }, { "epoch": 2.0997661730319566, "grad_norm": 0.2237656119837914, "learning_rate": 1.661849710982659e-05, "loss": 0.2788, "step": 1348 }, { "epoch": 2.101325019485581, "grad_norm": 0.25698435105220185, "learning_rate": 1.6589595375722546e-05, "loss": 0.2747, "step": 1349 }, { "epoch": 2.102883865939205, "grad_norm": 0.23963767268717012, "learning_rate": 1.6560693641618496e-05, "loss": 0.292, "step": 1350 }, { "epoch": 2.1044427123928293, "grad_norm": 0.2588056696749546, "learning_rate": 1.6531791907514452e-05, "loss": 0.3035, "step": 1351 }, { "epoch": 2.1060015588464536, "grad_norm": 0.25164368263083836, "learning_rate": 1.6502890173410403e-05, "loss": 0.297, "step": 1352 }, { "epoch": 2.107560405300078, "grad_norm": 0.2570592630746674, "learning_rate": 1.647398843930636e-05, "loss": 0.2982, "step": 1353 }, { "epoch": 2.109119251753702, "grad_norm": 0.2848315307917638, "learning_rate": 1.6445086705202313e-05, "loss": 0.3007, "step": 1354 }, { "epoch": 2.1106780982073268, "grad_norm": 0.2634643134825361, "learning_rate": 1.6416184971098266e-05, "loss": 0.2818, "step": 1355 }, { "epoch": 2.112236944660951, "grad_norm": 0.25838881698032484, "learning_rate": 1.638728323699422e-05, "loss": 0.2675, "step": 1356 }, { "epoch": 2.1137957911145753, "grad_norm": 0.2438857581178678, "learning_rate": 1.6358381502890173e-05, "loss": 0.2931, "step": 1357 }, { "epoch": 2.1153546375681995, "grad_norm": 0.23647456464356711, "learning_rate": 1.6329479768786127e-05, "loss": 0.2758, "step": 1358 }, { "epoch": 2.1169134840218238, "grad_norm": 0.245819384339159, "learning_rate": 1.6300578034682083e-05, "loss": 0.2935, "step": 1359 }, { "epoch": 2.118472330475448, "grad_norm": 0.21611647939557374, "learning_rate": 1.6271676300578033e-05, "loss": 0.2761, "step": 1360 }, { "epoch": 2.1200311769290723, "grad_norm": 0.24308227386319844, "learning_rate": 1.624277456647399e-05, "loss": 0.2911, "step": 1361 }, { "epoch": 2.121590023382697, "grad_norm": 0.28166564601668537, "learning_rate": 1.621387283236994e-05, "loss": 0.291, "step": 1362 }, { "epoch": 2.1231488698363212, "grad_norm": 0.2333910024555345, "learning_rate": 1.6184971098265897e-05, "loss": 0.282, "step": 1363 }, { "epoch": 2.1247077162899455, "grad_norm": 0.24801232636156478, "learning_rate": 1.615606936416185e-05, "loss": 0.286, "step": 1364 }, { "epoch": 2.1262665627435697, "grad_norm": 0.2529976131043813, "learning_rate": 1.6127167630057804e-05, "loss": 0.2907, "step": 1365 }, { "epoch": 2.127825409197194, "grad_norm": 0.24496746678348746, "learning_rate": 1.6098265895953758e-05, "loss": 0.3096, "step": 1366 }, { "epoch": 2.1293842556508182, "grad_norm": 0.22826596743843666, "learning_rate": 1.606936416184971e-05, "loss": 0.2837, "step": 1367 }, { "epoch": 2.1309431021044425, "grad_norm": 0.22953989118587162, "learning_rate": 1.6040462427745664e-05, "loss": 0.3016, "step": 1368 }, { "epoch": 2.132501948558067, "grad_norm": 0.24938076633148093, "learning_rate": 1.601156069364162e-05, "loss": 0.2938, "step": 1369 }, { "epoch": 2.1340607950116914, "grad_norm": 0.21892102115489234, "learning_rate": 1.598265895953757e-05, "loss": 0.2815, "step": 1370 }, { "epoch": 2.1356196414653157, "grad_norm": 0.2590092923549659, "learning_rate": 1.5953757225433528e-05, "loss": 0.2954, "step": 1371 }, { "epoch": 2.13717848791894, "grad_norm": 0.24890666085552407, "learning_rate": 1.5924855491329478e-05, "loss": 0.2906, "step": 1372 }, { "epoch": 2.138737334372564, "grad_norm": 0.2266359560410207, "learning_rate": 1.5895953757225435e-05, "loss": 0.2817, "step": 1373 }, { "epoch": 2.1402961808261884, "grad_norm": 0.24930550949426433, "learning_rate": 1.586705202312139e-05, "loss": 0.2902, "step": 1374 }, { "epoch": 2.141855027279813, "grad_norm": 0.22270650770572073, "learning_rate": 1.5838150289017342e-05, "loss": 0.2963, "step": 1375 }, { "epoch": 2.1434138737334374, "grad_norm": 0.21082837821856862, "learning_rate": 1.5809248554913295e-05, "loss": 0.2823, "step": 1376 }, { "epoch": 2.1449727201870616, "grad_norm": 0.2506283104261599, "learning_rate": 1.578034682080925e-05, "loss": 0.3003, "step": 1377 }, { "epoch": 2.146531566640686, "grad_norm": 0.2389054933616818, "learning_rate": 1.5751445086705202e-05, "loss": 0.2885, "step": 1378 }, { "epoch": 2.14809041309431, "grad_norm": 0.2334569505339403, "learning_rate": 1.5722543352601156e-05, "loss": 0.3052, "step": 1379 }, { "epoch": 2.1496492595479344, "grad_norm": 0.23175590298956983, "learning_rate": 1.569364161849711e-05, "loss": 0.2739, "step": 1380 }, { "epoch": 2.1512081060015587, "grad_norm": 0.23033041297872123, "learning_rate": 1.5664739884393066e-05, "loss": 0.286, "step": 1381 }, { "epoch": 2.1527669524551833, "grad_norm": 0.2465779921606415, "learning_rate": 1.5635838150289016e-05, "loss": 0.3009, "step": 1382 }, { "epoch": 2.1543257989088076, "grad_norm": 0.21179227741678436, "learning_rate": 1.5606936416184973e-05, "loss": 0.2847, "step": 1383 }, { "epoch": 2.155884645362432, "grad_norm": 0.25428536535374435, "learning_rate": 1.5578034682080926e-05, "loss": 0.2954, "step": 1384 }, { "epoch": 2.157443491816056, "grad_norm": 0.22930157749520183, "learning_rate": 1.554913294797688e-05, "loss": 0.2818, "step": 1385 }, { "epoch": 2.1590023382696804, "grad_norm": 0.23105202022376808, "learning_rate": 1.5520231213872833e-05, "loss": 0.3053, "step": 1386 }, { "epoch": 2.1605611847233046, "grad_norm": 0.23246069929983043, "learning_rate": 1.5491329479768787e-05, "loss": 0.2803, "step": 1387 }, { "epoch": 2.1621200311769293, "grad_norm": 0.22704083693916596, "learning_rate": 1.546242774566474e-05, "loss": 0.2891, "step": 1388 }, { "epoch": 2.1636788776305536, "grad_norm": 0.2540369139980788, "learning_rate": 1.5433526011560694e-05, "loss": 0.2847, "step": 1389 }, { "epoch": 2.165237724084178, "grad_norm": 0.21737272977125724, "learning_rate": 1.5404624277456647e-05, "loss": 0.2681, "step": 1390 }, { "epoch": 2.166796570537802, "grad_norm": 0.2515481263822816, "learning_rate": 1.5375722543352604e-05, "loss": 0.2763, "step": 1391 }, { "epoch": 2.1683554169914263, "grad_norm": 0.26477321713964047, "learning_rate": 1.5346820809248554e-05, "loss": 0.2808, "step": 1392 }, { "epoch": 2.1699142634450506, "grad_norm": 0.22854862492021247, "learning_rate": 1.531791907514451e-05, "loss": 0.2943, "step": 1393 }, { "epoch": 2.171473109898675, "grad_norm": 0.27158405015797865, "learning_rate": 1.528901734104046e-05, "loss": 0.2925, "step": 1394 }, { "epoch": 2.1730319563522995, "grad_norm": 0.2347337281042923, "learning_rate": 1.5260115606936418e-05, "loss": 0.2943, "step": 1395 }, { "epoch": 2.1745908028059238, "grad_norm": 0.23975916222494428, "learning_rate": 1.523121387283237e-05, "loss": 0.2804, "step": 1396 }, { "epoch": 2.176149649259548, "grad_norm": 0.22779116644041048, "learning_rate": 1.5202312138728325e-05, "loss": 0.2847, "step": 1397 }, { "epoch": 2.1777084957131723, "grad_norm": 0.2576587147864288, "learning_rate": 1.5173410404624278e-05, "loss": 0.2996, "step": 1398 }, { "epoch": 2.1792673421667965, "grad_norm": 0.21115000700822473, "learning_rate": 1.5144508670520233e-05, "loss": 0.2787, "step": 1399 }, { "epoch": 2.1808261886204208, "grad_norm": 0.24108973804933587, "learning_rate": 1.5115606936416185e-05, "loss": 0.287, "step": 1400 }, { "epoch": 2.182385035074045, "grad_norm": 0.24288289350228465, "learning_rate": 1.508670520231214e-05, "loss": 0.2852, "step": 1401 }, { "epoch": 2.1839438815276697, "grad_norm": 0.23706474437913794, "learning_rate": 1.5057803468208092e-05, "loss": 0.2968, "step": 1402 }, { "epoch": 2.185502727981294, "grad_norm": 0.2183208225549947, "learning_rate": 1.5028901734104049e-05, "loss": 0.2865, "step": 1403 }, { "epoch": 2.187061574434918, "grad_norm": 0.24253894248682953, "learning_rate": 1.5e-05, "loss": 0.2847, "step": 1404 }, { "epoch": 2.1886204208885425, "grad_norm": 0.22598056970641814, "learning_rate": 1.4971098265895956e-05, "loss": 0.2982, "step": 1405 }, { "epoch": 2.1901792673421667, "grad_norm": 0.2525617645303132, "learning_rate": 1.4942196531791907e-05, "loss": 0.2861, "step": 1406 }, { "epoch": 2.191738113795791, "grad_norm": 0.2192754370403558, "learning_rate": 1.4913294797687862e-05, "loss": 0.2872, "step": 1407 }, { "epoch": 2.1932969602494152, "grad_norm": 0.22900250201312447, "learning_rate": 1.4884393063583816e-05, "loss": 0.2845, "step": 1408 }, { "epoch": 2.19485580670304, "grad_norm": 0.30054519236450217, "learning_rate": 1.4855491329479771e-05, "loss": 0.2943, "step": 1409 }, { "epoch": 2.196414653156664, "grad_norm": 0.24441363373477734, "learning_rate": 1.4826589595375723e-05, "loss": 0.2975, "step": 1410 }, { "epoch": 2.1979734996102884, "grad_norm": 0.2267613912682233, "learning_rate": 1.4797687861271678e-05, "loss": 0.3009, "step": 1411 }, { "epoch": 2.1995323460639127, "grad_norm": 0.21100387037469506, "learning_rate": 1.476878612716763e-05, "loss": 0.2971, "step": 1412 }, { "epoch": 2.201091192517537, "grad_norm": 0.21361711019848872, "learning_rate": 1.4739884393063585e-05, "loss": 0.2986, "step": 1413 }, { "epoch": 2.202650038971161, "grad_norm": 0.23486378364056842, "learning_rate": 1.4710982658959538e-05, "loss": 0.2983, "step": 1414 }, { "epoch": 2.2042088854247854, "grad_norm": 0.2194691571773862, "learning_rate": 1.4682080924855493e-05, "loss": 0.2805, "step": 1415 }, { "epoch": 2.20576773187841, "grad_norm": 0.22839671067672832, "learning_rate": 1.4653179190751445e-05, "loss": 0.2821, "step": 1416 }, { "epoch": 2.2073265783320344, "grad_norm": 0.22721582376446814, "learning_rate": 1.46242774566474e-05, "loss": 0.2883, "step": 1417 }, { "epoch": 2.2088854247856586, "grad_norm": 0.21273654291505045, "learning_rate": 1.4595375722543354e-05, "loss": 0.2871, "step": 1418 }, { "epoch": 2.210444271239283, "grad_norm": 0.22181506285423208, "learning_rate": 1.4566473988439309e-05, "loss": 0.2937, "step": 1419 }, { "epoch": 2.212003117692907, "grad_norm": 0.21682833270568624, "learning_rate": 1.453757225433526e-05, "loss": 0.295, "step": 1420 }, { "epoch": 2.2135619641465314, "grad_norm": 0.2544117406843311, "learning_rate": 1.4508670520231216e-05, "loss": 0.2959, "step": 1421 }, { "epoch": 2.215120810600156, "grad_norm": 0.2496480472797846, "learning_rate": 1.4479768786127167e-05, "loss": 0.2955, "step": 1422 }, { "epoch": 2.2166796570537803, "grad_norm": 0.25230712728942756, "learning_rate": 1.4450867052023123e-05, "loss": 0.2834, "step": 1423 }, { "epoch": 2.2182385035074046, "grad_norm": 0.22497012216327553, "learning_rate": 1.4421965317919076e-05, "loss": 0.292, "step": 1424 }, { "epoch": 2.219797349961029, "grad_norm": 0.2374016981159356, "learning_rate": 1.4393063583815031e-05, "loss": 0.2825, "step": 1425 }, { "epoch": 2.221356196414653, "grad_norm": 0.25465184050475065, "learning_rate": 1.4364161849710983e-05, "loss": 0.2903, "step": 1426 }, { "epoch": 2.2229150428682773, "grad_norm": 0.23870397107279917, "learning_rate": 1.4335260115606938e-05, "loss": 0.2936, "step": 1427 }, { "epoch": 2.2244738893219016, "grad_norm": 0.2318933878119134, "learning_rate": 1.430635838150289e-05, "loss": 0.2849, "step": 1428 }, { "epoch": 2.2260327357755263, "grad_norm": 0.23210851294757337, "learning_rate": 1.4277456647398845e-05, "loss": 0.2919, "step": 1429 }, { "epoch": 2.2275915822291505, "grad_norm": 0.20510591201151088, "learning_rate": 1.4248554913294798e-05, "loss": 0.291, "step": 1430 }, { "epoch": 2.229150428682775, "grad_norm": 0.2254853152020144, "learning_rate": 1.4219653179190754e-05, "loss": 0.2855, "step": 1431 }, { "epoch": 2.230709275136399, "grad_norm": 0.22024099508735234, "learning_rate": 1.4190751445086705e-05, "loss": 0.2927, "step": 1432 }, { "epoch": 2.2322681215900233, "grad_norm": 0.26043420023310093, "learning_rate": 1.416184971098266e-05, "loss": 0.3052, "step": 1433 }, { "epoch": 2.2338269680436476, "grad_norm": 0.2133056223198868, "learning_rate": 1.4132947976878614e-05, "loss": 0.2927, "step": 1434 }, { "epoch": 2.2353858144972722, "grad_norm": 0.2259340299782071, "learning_rate": 1.4104046242774569e-05, "loss": 0.2857, "step": 1435 }, { "epoch": 2.2369446609508965, "grad_norm": 0.22200399124864398, "learning_rate": 1.407514450867052e-05, "loss": 0.2833, "step": 1436 }, { "epoch": 2.2385035074045208, "grad_norm": 0.22962299042832857, "learning_rate": 1.4046242774566476e-05, "loss": 0.2799, "step": 1437 }, { "epoch": 2.240062353858145, "grad_norm": 0.21438991547193448, "learning_rate": 1.4017341040462428e-05, "loss": 0.2859, "step": 1438 }, { "epoch": 2.2416212003117693, "grad_norm": 0.22069076358373807, "learning_rate": 1.3988439306358383e-05, "loss": 0.2902, "step": 1439 }, { "epoch": 2.2431800467653935, "grad_norm": 0.21552495322148416, "learning_rate": 1.3959537572254336e-05, "loss": 0.2717, "step": 1440 }, { "epoch": 2.2447388932190178, "grad_norm": 0.22292189496523734, "learning_rate": 1.3930635838150291e-05, "loss": 0.2837, "step": 1441 }, { "epoch": 2.2462977396726425, "grad_norm": 0.2235003322526891, "learning_rate": 1.3901734104046243e-05, "loss": 0.2785, "step": 1442 }, { "epoch": 2.2478565861262667, "grad_norm": 0.21304911072284627, "learning_rate": 1.3872832369942197e-05, "loss": 0.2879, "step": 1443 }, { "epoch": 2.249415432579891, "grad_norm": 0.25029914205006026, "learning_rate": 1.3843930635838152e-05, "loss": 0.2988, "step": 1444 }, { "epoch": 2.250974279033515, "grad_norm": 0.2165276488818481, "learning_rate": 1.3815028901734104e-05, "loss": 0.272, "step": 1445 }, { "epoch": 2.2525331254871395, "grad_norm": 0.24444489582018986, "learning_rate": 1.3786127167630059e-05, "loss": 0.3014, "step": 1446 }, { "epoch": 2.2540919719407637, "grad_norm": 0.24771789944118036, "learning_rate": 1.375722543352601e-05, "loss": 0.2984, "step": 1447 }, { "epoch": 2.255650818394388, "grad_norm": 0.23706904927685987, "learning_rate": 1.3728323699421966e-05, "loss": 0.3033, "step": 1448 }, { "epoch": 2.2572096648480127, "grad_norm": 0.23835195294951741, "learning_rate": 1.3699421965317919e-05, "loss": 0.2774, "step": 1449 }, { "epoch": 2.258768511301637, "grad_norm": 0.2492838137926293, "learning_rate": 1.3670520231213874e-05, "loss": 0.2843, "step": 1450 }, { "epoch": 2.260327357755261, "grad_norm": 0.22427981907836916, "learning_rate": 1.3641618497109826e-05, "loss": 0.2862, "step": 1451 }, { "epoch": 2.2618862042088854, "grad_norm": 0.2425076666411249, "learning_rate": 1.3612716763005781e-05, "loss": 0.2908, "step": 1452 }, { "epoch": 2.2634450506625097, "grad_norm": 0.2686851435498773, "learning_rate": 1.3583815028901733e-05, "loss": 0.2853, "step": 1453 }, { "epoch": 2.265003897116134, "grad_norm": 0.23258321364166276, "learning_rate": 1.3554913294797688e-05, "loss": 0.292, "step": 1454 }, { "epoch": 2.266562743569758, "grad_norm": 0.2178252049659836, "learning_rate": 1.3526011560693641e-05, "loss": 0.2798, "step": 1455 }, { "epoch": 2.268121590023383, "grad_norm": 0.23464644401450185, "learning_rate": 1.3497109826589597e-05, "loss": 0.279, "step": 1456 }, { "epoch": 2.269680436477007, "grad_norm": 0.25551325960953597, "learning_rate": 1.3468208092485548e-05, "loss": 0.2997, "step": 1457 }, { "epoch": 2.2712392829306314, "grad_norm": 0.2352887691568654, "learning_rate": 1.3439306358381503e-05, "loss": 0.2843, "step": 1458 }, { "epoch": 2.2727981293842556, "grad_norm": 0.24160307618137192, "learning_rate": 1.3410404624277457e-05, "loss": 0.2979, "step": 1459 }, { "epoch": 2.27435697583788, "grad_norm": 0.247635002623451, "learning_rate": 1.3381502890173412e-05, "loss": 0.2819, "step": 1460 }, { "epoch": 2.275915822291504, "grad_norm": 0.2335426458405105, "learning_rate": 1.3352601156069364e-05, "loss": 0.2813, "step": 1461 }, { "epoch": 2.2774746687451284, "grad_norm": 0.22663182041299806, "learning_rate": 1.3323699421965319e-05, "loss": 0.2981, "step": 1462 }, { "epoch": 2.279033515198753, "grad_norm": 0.21171630873754166, "learning_rate": 1.329479768786127e-05, "loss": 0.2826, "step": 1463 }, { "epoch": 2.2805923616523773, "grad_norm": 0.22757568106572626, "learning_rate": 1.3265895953757226e-05, "loss": 0.2923, "step": 1464 }, { "epoch": 2.2821512081060016, "grad_norm": 0.22020055614811698, "learning_rate": 1.323699421965318e-05, "loss": 0.2807, "step": 1465 }, { "epoch": 2.283710054559626, "grad_norm": 0.23965818423420435, "learning_rate": 1.3208092485549134e-05, "loss": 0.295, "step": 1466 }, { "epoch": 2.28526890101325, "grad_norm": 0.21933495891348634, "learning_rate": 1.3179190751445086e-05, "loss": 0.2808, "step": 1467 }, { "epoch": 2.2868277474668743, "grad_norm": 0.22394396080019968, "learning_rate": 1.3150289017341041e-05, "loss": 0.2828, "step": 1468 }, { "epoch": 2.2883865939204986, "grad_norm": 0.2341244557636366, "learning_rate": 1.3121387283236993e-05, "loss": 0.2872, "step": 1469 }, { "epoch": 2.2899454403741233, "grad_norm": 0.24597918258046544, "learning_rate": 1.3092485549132948e-05, "loss": 0.2854, "step": 1470 }, { "epoch": 2.2915042868277475, "grad_norm": 0.23301804978413262, "learning_rate": 1.3063583815028902e-05, "loss": 0.3053, "step": 1471 }, { "epoch": 2.293063133281372, "grad_norm": 0.2360765813557192, "learning_rate": 1.3034682080924857e-05, "loss": 0.2831, "step": 1472 }, { "epoch": 2.294621979734996, "grad_norm": 0.2658423968018713, "learning_rate": 1.3005780346820809e-05, "loss": 0.2869, "step": 1473 }, { "epoch": 2.2961808261886203, "grad_norm": 0.25016881594307117, "learning_rate": 1.2976878612716764e-05, "loss": 0.2763, "step": 1474 }, { "epoch": 2.2977396726422445, "grad_norm": 0.2382698683460339, "learning_rate": 1.2947976878612717e-05, "loss": 0.3028, "step": 1475 }, { "epoch": 2.2992985190958692, "grad_norm": 0.24221989883612682, "learning_rate": 1.2919075144508672e-05, "loss": 0.3088, "step": 1476 }, { "epoch": 2.3008573655494935, "grad_norm": 0.25017938488171704, "learning_rate": 1.2890173410404624e-05, "loss": 0.2869, "step": 1477 }, { "epoch": 2.3024162120031177, "grad_norm": 0.23057606820343765, "learning_rate": 1.2861271676300579e-05, "loss": 0.2812, "step": 1478 }, { "epoch": 2.303975058456742, "grad_norm": 0.22851951981455979, "learning_rate": 1.2832369942196531e-05, "loss": 0.2903, "step": 1479 }, { "epoch": 2.3055339049103662, "grad_norm": 0.25023954290485484, "learning_rate": 1.2803468208092486e-05, "loss": 0.2995, "step": 1480 }, { "epoch": 2.3070927513639905, "grad_norm": 0.22267748488279954, "learning_rate": 1.277456647398844e-05, "loss": 0.2946, "step": 1481 }, { "epoch": 2.308651597817615, "grad_norm": 0.2433472732667005, "learning_rate": 1.2745664739884395e-05, "loss": 0.2805, "step": 1482 }, { "epoch": 2.3102104442712394, "grad_norm": 0.22084205335592666, "learning_rate": 1.2716763005780346e-05, "loss": 0.2718, "step": 1483 }, { "epoch": 2.3117692907248637, "grad_norm": 0.24304148596183653, "learning_rate": 1.2687861271676302e-05, "loss": 0.2988, "step": 1484 }, { "epoch": 2.313328137178488, "grad_norm": 0.2424544214432883, "learning_rate": 1.2658959537572255e-05, "loss": 0.2899, "step": 1485 }, { "epoch": 2.314886983632112, "grad_norm": 0.2328910388765159, "learning_rate": 1.263005780346821e-05, "loss": 0.3127, "step": 1486 }, { "epoch": 2.3164458300857365, "grad_norm": 0.23286533311552443, "learning_rate": 1.2601156069364162e-05, "loss": 0.2995, "step": 1487 }, { "epoch": 2.3180046765393607, "grad_norm": 0.2549382777228154, "learning_rate": 1.2572254335260117e-05, "loss": 0.2928, "step": 1488 }, { "epoch": 2.3195635229929854, "grad_norm": 0.2559163608878674, "learning_rate": 1.2543352601156069e-05, "loss": 0.3113, "step": 1489 }, { "epoch": 2.3211223694466097, "grad_norm": 0.2265685017446726, "learning_rate": 1.2514450867052024e-05, "loss": 0.295, "step": 1490 }, { "epoch": 2.322681215900234, "grad_norm": 0.22273564838191307, "learning_rate": 1.2485549132947977e-05, "loss": 0.2717, "step": 1491 }, { "epoch": 2.324240062353858, "grad_norm": 0.21674653038960545, "learning_rate": 1.245664739884393e-05, "loss": 0.2845, "step": 1492 }, { "epoch": 2.3257989088074824, "grad_norm": 0.21909928945574206, "learning_rate": 1.2427745664739884e-05, "loss": 0.284, "step": 1493 }, { "epoch": 2.3273577552611067, "grad_norm": 0.2750007906217208, "learning_rate": 1.2398843930635838e-05, "loss": 0.2862, "step": 1494 }, { "epoch": 2.328916601714731, "grad_norm": 0.21703802020708915, "learning_rate": 1.2369942196531791e-05, "loss": 0.2829, "step": 1495 }, { "epoch": 2.3304754481683556, "grad_norm": 0.22950831232595262, "learning_rate": 1.2341040462427746e-05, "loss": 0.2883, "step": 1496 }, { "epoch": 2.33203429462198, "grad_norm": 0.24426647745397967, "learning_rate": 1.23121387283237e-05, "loss": 0.2955, "step": 1497 }, { "epoch": 2.333593141075604, "grad_norm": 0.23949518800459763, "learning_rate": 1.2283236994219653e-05, "loss": 0.2771, "step": 1498 }, { "epoch": 2.3351519875292284, "grad_norm": 0.21894296552516587, "learning_rate": 1.2254335260115607e-05, "loss": 0.2899, "step": 1499 }, { "epoch": 2.3367108339828526, "grad_norm": 0.22130870442542108, "learning_rate": 1.222543352601156e-05, "loss": 0.2836, "step": 1500 }, { "epoch": 2.338269680436477, "grad_norm": 0.2659075059435682, "learning_rate": 1.2196531791907515e-05, "loss": 0.2838, "step": 1501 }, { "epoch": 2.339828526890101, "grad_norm": 0.23607600423290415, "learning_rate": 1.2167630057803469e-05, "loss": 0.2972, "step": 1502 }, { "epoch": 2.341387373343726, "grad_norm": 0.2352342145609656, "learning_rate": 1.2138728323699422e-05, "loss": 0.3003, "step": 1503 }, { "epoch": 2.34294621979735, "grad_norm": 0.23180782759300908, "learning_rate": 1.2109826589595376e-05, "loss": 0.2925, "step": 1504 }, { "epoch": 2.3445050662509743, "grad_norm": 0.2421492426629086, "learning_rate": 1.2080924855491329e-05, "loss": 0.3153, "step": 1505 }, { "epoch": 2.3460639127045986, "grad_norm": 0.220047149003109, "learning_rate": 1.2052023121387284e-05, "loss": 0.2915, "step": 1506 }, { "epoch": 2.347622759158223, "grad_norm": 0.2081371282315123, "learning_rate": 1.2023121387283238e-05, "loss": 0.2786, "step": 1507 }, { "epoch": 2.349181605611847, "grad_norm": 0.2209132364079271, "learning_rate": 1.1994219653179191e-05, "loss": 0.2919, "step": 1508 }, { "epoch": 2.3507404520654713, "grad_norm": 0.24203159776015792, "learning_rate": 1.1965317919075144e-05, "loss": 0.2944, "step": 1509 }, { "epoch": 2.352299298519096, "grad_norm": 0.21125490785458723, "learning_rate": 1.1936416184971098e-05, "loss": 0.2819, "step": 1510 }, { "epoch": 2.3538581449727203, "grad_norm": 0.21084827155820668, "learning_rate": 1.1907514450867053e-05, "loss": 0.2735, "step": 1511 }, { "epoch": 2.3554169914263445, "grad_norm": 0.2341059203258637, "learning_rate": 1.1878612716763006e-05, "loss": 0.2991, "step": 1512 }, { "epoch": 2.356975837879969, "grad_norm": 0.21296853713090008, "learning_rate": 1.184971098265896e-05, "loss": 0.2833, "step": 1513 }, { "epoch": 2.358534684333593, "grad_norm": 0.2567119485230043, "learning_rate": 1.1820809248554913e-05, "loss": 0.2947, "step": 1514 }, { "epoch": 2.3600935307872173, "grad_norm": 0.2333031479163138, "learning_rate": 1.1791907514450867e-05, "loss": 0.2864, "step": 1515 }, { "epoch": 2.3616523772408415, "grad_norm": 0.24056595254646906, "learning_rate": 1.176300578034682e-05, "loss": 0.2802, "step": 1516 }, { "epoch": 2.3632112236944662, "grad_norm": 0.23247035275872605, "learning_rate": 1.1734104046242775e-05, "loss": 0.2821, "step": 1517 }, { "epoch": 2.3647700701480905, "grad_norm": 0.23551357842178408, "learning_rate": 1.1705202312138729e-05, "loss": 0.2799, "step": 1518 }, { "epoch": 2.3663289166017147, "grad_norm": 0.22199252143905965, "learning_rate": 1.1676300578034682e-05, "loss": 0.2921, "step": 1519 }, { "epoch": 2.367887763055339, "grad_norm": 0.2557073653528457, "learning_rate": 1.1647398843930636e-05, "loss": 0.2819, "step": 1520 }, { "epoch": 2.3694466095089632, "grad_norm": 0.23177632971310883, "learning_rate": 1.161849710982659e-05, "loss": 0.3, "step": 1521 }, { "epoch": 2.3710054559625875, "grad_norm": 0.2141949825685525, "learning_rate": 1.1589595375722544e-05, "loss": 0.2804, "step": 1522 }, { "epoch": 2.372564302416212, "grad_norm": 0.22456594982131964, "learning_rate": 1.1560693641618498e-05, "loss": 0.2919, "step": 1523 }, { "epoch": 2.3741231488698364, "grad_norm": 0.2328422263169891, "learning_rate": 1.1531791907514451e-05, "loss": 0.298, "step": 1524 }, { "epoch": 2.3756819953234607, "grad_norm": 0.23696530509869898, "learning_rate": 1.1502890173410405e-05, "loss": 0.2826, "step": 1525 }, { "epoch": 2.377240841777085, "grad_norm": 0.23471588381502256, "learning_rate": 1.1473988439306358e-05, "loss": 0.2954, "step": 1526 }, { "epoch": 2.378799688230709, "grad_norm": 0.24324817997977416, "learning_rate": 1.1445086705202313e-05, "loss": 0.3002, "step": 1527 }, { "epoch": 2.3803585346843334, "grad_norm": 0.23818184986570196, "learning_rate": 1.1416184971098267e-05, "loss": 0.2901, "step": 1528 }, { "epoch": 2.381917381137958, "grad_norm": 0.2177316843513184, "learning_rate": 1.138728323699422e-05, "loss": 0.2837, "step": 1529 }, { "epoch": 2.3834762275915824, "grad_norm": 0.2230616433485158, "learning_rate": 1.1358381502890174e-05, "loss": 0.2926, "step": 1530 }, { "epoch": 2.3850350740452066, "grad_norm": 0.2092816344234883, "learning_rate": 1.1329479768786127e-05, "loss": 0.2771, "step": 1531 }, { "epoch": 2.386593920498831, "grad_norm": 0.24426299651005548, "learning_rate": 1.1300578034682082e-05, "loss": 0.2893, "step": 1532 }, { "epoch": 2.388152766952455, "grad_norm": 0.2257044731930893, "learning_rate": 1.1271676300578036e-05, "loss": 0.3099, "step": 1533 }, { "epoch": 2.3897116134060794, "grad_norm": 0.22709943701947127, "learning_rate": 1.1242774566473989e-05, "loss": 0.2815, "step": 1534 }, { "epoch": 2.3912704598597037, "grad_norm": 0.23081380246295977, "learning_rate": 1.1213872832369943e-05, "loss": 0.2892, "step": 1535 }, { "epoch": 2.3928293063133284, "grad_norm": 0.2404518884149201, "learning_rate": 1.1184971098265896e-05, "loss": 0.2784, "step": 1536 }, { "epoch": 2.3943881527669526, "grad_norm": 0.21874834698362006, "learning_rate": 1.115606936416185e-05, "loss": 0.2838, "step": 1537 }, { "epoch": 2.395946999220577, "grad_norm": 0.2170285592301821, "learning_rate": 1.1127167630057805e-05, "loss": 0.2836, "step": 1538 }, { "epoch": 2.397505845674201, "grad_norm": 0.22151366255017815, "learning_rate": 1.1098265895953758e-05, "loss": 0.2945, "step": 1539 }, { "epoch": 2.3990646921278254, "grad_norm": 0.2272145173574314, "learning_rate": 1.1069364161849711e-05, "loss": 0.2836, "step": 1540 }, { "epoch": 2.4006235385814496, "grad_norm": 0.22611588795882234, "learning_rate": 1.1040462427745665e-05, "loss": 0.3038, "step": 1541 }, { "epoch": 2.402182385035074, "grad_norm": 0.21786269575794123, "learning_rate": 1.1011560693641618e-05, "loss": 0.2955, "step": 1542 }, { "epoch": 2.4037412314886986, "grad_norm": 0.2081735505823338, "learning_rate": 1.0982658959537573e-05, "loss": 0.2845, "step": 1543 }, { "epoch": 2.405300077942323, "grad_norm": 0.21791564332112892, "learning_rate": 1.0953757225433527e-05, "loss": 0.2888, "step": 1544 }, { "epoch": 2.406858924395947, "grad_norm": 0.21343814806147593, "learning_rate": 1.092485549132948e-05, "loss": 0.3027, "step": 1545 }, { "epoch": 2.4084177708495713, "grad_norm": 0.24305021560110593, "learning_rate": 1.0895953757225434e-05, "loss": 0.3096, "step": 1546 }, { "epoch": 2.4099766173031956, "grad_norm": 0.2428823537320873, "learning_rate": 1.0867052023121387e-05, "loss": 0.2787, "step": 1547 }, { "epoch": 2.41153546375682, "grad_norm": 0.22533176874540484, "learning_rate": 1.0838150289017342e-05, "loss": 0.3003, "step": 1548 }, { "epoch": 2.413094310210444, "grad_norm": 0.23052788059445348, "learning_rate": 1.0809248554913296e-05, "loss": 0.2937, "step": 1549 }, { "epoch": 2.4146531566640688, "grad_norm": 0.21873125390687612, "learning_rate": 1.078034682080925e-05, "loss": 0.2852, "step": 1550 }, { "epoch": 2.416212003117693, "grad_norm": 0.23262040078023866, "learning_rate": 1.0751445086705203e-05, "loss": 0.2867, "step": 1551 }, { "epoch": 2.4177708495713173, "grad_norm": 0.2061110656033279, "learning_rate": 1.0722543352601156e-05, "loss": 0.2834, "step": 1552 }, { "epoch": 2.4193296960249415, "grad_norm": 0.21742353821980845, "learning_rate": 1.0693641618497111e-05, "loss": 0.286, "step": 1553 }, { "epoch": 2.4208885424785658, "grad_norm": 0.21761958606143322, "learning_rate": 1.0664739884393065e-05, "loss": 0.2934, "step": 1554 }, { "epoch": 2.42244738893219, "grad_norm": 0.22413107731125181, "learning_rate": 1.0635838150289018e-05, "loss": 0.3017, "step": 1555 }, { "epoch": 2.4240062353858143, "grad_norm": 0.21834093708352184, "learning_rate": 1.0606936416184972e-05, "loss": 0.304, "step": 1556 }, { "epoch": 2.425565081839439, "grad_norm": 0.22975540033681002, "learning_rate": 1.0578034682080925e-05, "loss": 0.2926, "step": 1557 }, { "epoch": 2.4271239282930632, "grad_norm": 0.2274868161114814, "learning_rate": 1.0549132947976879e-05, "loss": 0.2999, "step": 1558 }, { "epoch": 2.4286827747466875, "grad_norm": 0.22082603952846616, "learning_rate": 1.0520231213872834e-05, "loss": 0.2782, "step": 1559 }, { "epoch": 2.4302416212003117, "grad_norm": 0.23461317288098227, "learning_rate": 1.0491329479768787e-05, "loss": 0.2842, "step": 1560 }, { "epoch": 2.431800467653936, "grad_norm": 0.20440782512445527, "learning_rate": 1.046242774566474e-05, "loss": 0.2853, "step": 1561 }, { "epoch": 2.4333593141075602, "grad_norm": 0.23611049627402994, "learning_rate": 1.0433526011560694e-05, "loss": 0.2841, "step": 1562 }, { "epoch": 2.4349181605611845, "grad_norm": 0.23437404672528528, "learning_rate": 1.0404624277456647e-05, "loss": 0.2963, "step": 1563 }, { "epoch": 2.436477007014809, "grad_norm": 0.2204494699300127, "learning_rate": 1.0375722543352603e-05, "loss": 0.2806, "step": 1564 }, { "epoch": 2.4380358534684334, "grad_norm": 0.20490594460203101, "learning_rate": 1.0346820809248556e-05, "loss": 0.2761, "step": 1565 }, { "epoch": 2.4395946999220577, "grad_norm": 0.6956010247527492, "learning_rate": 1.031791907514451e-05, "loss": 0.2901, "step": 1566 }, { "epoch": 2.441153546375682, "grad_norm": 0.26224821751440014, "learning_rate": 1.0289017341040463e-05, "loss": 0.3034, "step": 1567 }, { "epoch": 2.442712392829306, "grad_norm": 0.25126259259462486, "learning_rate": 1.0260115606936416e-05, "loss": 0.2899, "step": 1568 }, { "epoch": 2.4442712392829304, "grad_norm": 0.22870053657745457, "learning_rate": 1.0231213872832372e-05, "loss": 0.2822, "step": 1569 }, { "epoch": 2.445830085736555, "grad_norm": 0.23587258742639972, "learning_rate": 1.0202312138728325e-05, "loss": 0.2593, "step": 1570 }, { "epoch": 2.4473889321901794, "grad_norm": 0.22897162723371334, "learning_rate": 1.0173410404624278e-05, "loss": 0.2912, "step": 1571 }, { "epoch": 2.4489477786438036, "grad_norm": 0.24149402834986947, "learning_rate": 1.0144508670520232e-05, "loss": 0.2878, "step": 1572 }, { "epoch": 2.450506625097428, "grad_norm": 0.23350074250406747, "learning_rate": 1.0115606936416185e-05, "loss": 0.2869, "step": 1573 }, { "epoch": 2.452065471551052, "grad_norm": 0.22195416376987834, "learning_rate": 1.008670520231214e-05, "loss": 0.2956, "step": 1574 }, { "epoch": 2.4536243180046764, "grad_norm": 1.3487509546561793, "learning_rate": 1.0057803468208094e-05, "loss": 0.3012, "step": 1575 }, { "epoch": 2.455183164458301, "grad_norm": 0.2873786952181812, "learning_rate": 1.0028901734104047e-05, "loss": 0.284, "step": 1576 }, { "epoch": 2.4567420109119253, "grad_norm": 0.281096191572859, "learning_rate": 1e-05, "loss": 0.2839, "step": 1577 }, { "epoch": 2.4583008573655496, "grad_norm": 0.26582966306416833, "learning_rate": 9.971098265895954e-06, "loss": 0.2868, "step": 1578 }, { "epoch": 2.459859703819174, "grad_norm": 0.23786440274496637, "learning_rate": 9.94219653179191e-06, "loss": 0.2874, "step": 1579 }, { "epoch": 2.461418550272798, "grad_norm": 0.26380977501613667, "learning_rate": 9.913294797687863e-06, "loss": 0.298, "step": 1580 }, { "epoch": 2.4629773967264224, "grad_norm": 0.28090632623520245, "learning_rate": 9.884393063583816e-06, "loss": 0.2824, "step": 1581 }, { "epoch": 2.4645362431800466, "grad_norm": 0.24654785691378972, "learning_rate": 9.85549132947977e-06, "loss": 0.2842, "step": 1582 }, { "epoch": 2.4660950896336713, "grad_norm": 0.22185993368845872, "learning_rate": 9.826589595375723e-06, "loss": 0.2784, "step": 1583 }, { "epoch": 2.4676539360872956, "grad_norm": 0.25759773055873747, "learning_rate": 9.797687861271677e-06, "loss": 0.3065, "step": 1584 }, { "epoch": 2.46921278254092, "grad_norm": 0.2448676371644166, "learning_rate": 9.768786127167632e-06, "loss": 0.2951, "step": 1585 }, { "epoch": 2.470771628994544, "grad_norm": 0.22038630080496294, "learning_rate": 9.739884393063585e-06, "loss": 0.2766, "step": 1586 }, { "epoch": 2.4723304754481683, "grad_norm": 0.21067846046967328, "learning_rate": 9.710982658959537e-06, "loss": 0.2772, "step": 1587 }, { "epoch": 2.4738893219017926, "grad_norm": 0.2196571148859021, "learning_rate": 9.68208092485549e-06, "loss": 0.2944, "step": 1588 }, { "epoch": 2.475448168355417, "grad_norm": 0.20661464746790095, "learning_rate": 9.653179190751446e-06, "loss": 0.3023, "step": 1589 }, { "epoch": 2.4770070148090415, "grad_norm": 0.20946455345772869, "learning_rate": 9.624277456647399e-06, "loss": 0.2768, "step": 1590 }, { "epoch": 2.4785658612626658, "grad_norm": 0.21800665843244352, "learning_rate": 9.595375722543352e-06, "loss": 0.3014, "step": 1591 }, { "epoch": 2.48012470771629, "grad_norm": 0.21469256658865496, "learning_rate": 9.566473988439306e-06, "loss": 0.2865, "step": 1592 }, { "epoch": 2.4816835541699143, "grad_norm": 0.20483902813627355, "learning_rate": 9.53757225433526e-06, "loss": 0.2734, "step": 1593 }, { "epoch": 2.4832424006235385, "grad_norm": 0.22797586254976032, "learning_rate": 9.508670520231214e-06, "loss": 0.2839, "step": 1594 }, { "epoch": 2.4848012470771628, "grad_norm": 0.20269734125823377, "learning_rate": 9.479768786127168e-06, "loss": 0.2853, "step": 1595 }, { "epoch": 2.486360093530787, "grad_norm": 0.21209747614140953, "learning_rate": 9.450867052023121e-06, "loss": 0.2846, "step": 1596 }, { "epoch": 2.4879189399844117, "grad_norm": 0.20886949964376036, "learning_rate": 9.421965317919075e-06, "loss": 0.2768, "step": 1597 }, { "epoch": 2.489477786438036, "grad_norm": 0.21987185193696723, "learning_rate": 9.393063583815028e-06, "loss": 0.289, "step": 1598 }, { "epoch": 2.49103663289166, "grad_norm": 0.22992122520081743, "learning_rate": 9.364161849710983e-06, "loss": 0.317, "step": 1599 }, { "epoch": 2.4925954793452845, "grad_norm": 0.2216415692962526, "learning_rate": 9.335260115606937e-06, "loss": 0.2872, "step": 1600 }, { "epoch": 2.4941543257989087, "grad_norm": 0.22333309804066181, "learning_rate": 9.30635838150289e-06, "loss": 0.2762, "step": 1601 }, { "epoch": 2.495713172252533, "grad_norm": 0.2088510278237382, "learning_rate": 9.277456647398844e-06, "loss": 0.2849, "step": 1602 }, { "epoch": 2.4972720187061572, "grad_norm": 0.21834916127206458, "learning_rate": 9.248554913294797e-06, "loss": 0.2945, "step": 1603 }, { "epoch": 2.498830865159782, "grad_norm": 0.23784512272220631, "learning_rate": 9.21965317919075e-06, "loss": 0.29, "step": 1604 }, { "epoch": 2.500389711613406, "grad_norm": 0.22622134161606003, "learning_rate": 9.190751445086706e-06, "loss": 0.2855, "step": 1605 }, { "epoch": 2.5019485580670304, "grad_norm": 0.20987648910752202, "learning_rate": 9.16184971098266e-06, "loss": 0.2886, "step": 1606 }, { "epoch": 2.5035074045206547, "grad_norm": 0.2271896769680405, "learning_rate": 9.132947976878613e-06, "loss": 0.2849, "step": 1607 }, { "epoch": 2.505066250974279, "grad_norm": 0.20703729358511513, "learning_rate": 9.104046242774566e-06, "loss": 0.285, "step": 1608 }, { "epoch": 2.5066250974279036, "grad_norm": 0.2210043646064696, "learning_rate": 9.07514450867052e-06, "loss": 0.2793, "step": 1609 }, { "epoch": 2.5081839438815274, "grad_norm": 0.20832438994217642, "learning_rate": 9.046242774566475e-06, "loss": 0.2706, "step": 1610 }, { "epoch": 2.509742790335152, "grad_norm": 0.22663700329890946, "learning_rate": 9.017341040462428e-06, "loss": 0.2926, "step": 1611 }, { "epoch": 2.5113016367887764, "grad_norm": 0.2321310919459722, "learning_rate": 8.988439306358382e-06, "loss": 0.2929, "step": 1612 }, { "epoch": 2.5128604832424006, "grad_norm": 0.20367730242743393, "learning_rate": 8.959537572254335e-06, "loss": 0.273, "step": 1613 }, { "epoch": 2.514419329696025, "grad_norm": 0.2573291858309685, "learning_rate": 8.930635838150288e-06, "loss": 0.2937, "step": 1614 }, { "epoch": 2.515978176149649, "grad_norm": 0.21765613776271725, "learning_rate": 8.901734104046244e-06, "loss": 0.3041, "step": 1615 }, { "epoch": 2.517537022603274, "grad_norm": 0.22719439248378498, "learning_rate": 8.872832369942197e-06, "loss": 0.3025, "step": 1616 }, { "epoch": 2.5190958690568976, "grad_norm": 0.2233584630382556, "learning_rate": 8.84393063583815e-06, "loss": 0.2901, "step": 1617 }, { "epoch": 2.5206547155105223, "grad_norm": 0.2756546250082013, "learning_rate": 8.815028901734104e-06, "loss": 0.299, "step": 1618 }, { "epoch": 2.5222135619641466, "grad_norm": 0.2471624019129573, "learning_rate": 8.786127167630057e-06, "loss": 0.2856, "step": 1619 }, { "epoch": 2.523772408417771, "grad_norm": 0.22928484558281476, "learning_rate": 8.757225433526013e-06, "loss": 0.2885, "step": 1620 }, { "epoch": 2.525331254871395, "grad_norm": 0.22619496618870039, "learning_rate": 8.728323699421966e-06, "loss": 0.2858, "step": 1621 }, { "epoch": 2.5268901013250193, "grad_norm": 0.23523904557731568, "learning_rate": 8.69942196531792e-06, "loss": 0.2879, "step": 1622 }, { "epoch": 2.528448947778644, "grad_norm": 0.23102152485027722, "learning_rate": 8.670520231213873e-06, "loss": 0.2978, "step": 1623 }, { "epoch": 2.530007794232268, "grad_norm": 0.23596999054388898, "learning_rate": 8.641618497109826e-06, "loss": 0.2955, "step": 1624 }, { "epoch": 2.5315666406858925, "grad_norm": 0.23775289137632036, "learning_rate": 8.61271676300578e-06, "loss": 0.303, "step": 1625 }, { "epoch": 2.533125487139517, "grad_norm": 0.22544643461170122, "learning_rate": 8.583815028901735e-06, "loss": 0.2907, "step": 1626 }, { "epoch": 2.534684333593141, "grad_norm": 0.2598040215067987, "learning_rate": 8.554913294797688e-06, "loss": 0.2946, "step": 1627 }, { "epoch": 2.5362431800467653, "grad_norm": 0.21413673157343954, "learning_rate": 8.526011560693642e-06, "loss": 0.2836, "step": 1628 }, { "epoch": 2.5378020265003896, "grad_norm": 0.4266061422700911, "learning_rate": 8.497109826589595e-06, "loss": 0.2864, "step": 1629 }, { "epoch": 2.5393608729540142, "grad_norm": 0.21799108371646567, "learning_rate": 8.468208092485549e-06, "loss": 0.2813, "step": 1630 }, { "epoch": 2.5409197194076385, "grad_norm": 0.2322583381309059, "learning_rate": 8.439306358381504e-06, "loss": 0.3071, "step": 1631 }, { "epoch": 2.5424785658612628, "grad_norm": 0.2202382316089639, "learning_rate": 8.410404624277457e-06, "loss": 0.2926, "step": 1632 }, { "epoch": 2.544037412314887, "grad_norm": 0.22009637471013366, "learning_rate": 8.38150289017341e-06, "loss": 0.2963, "step": 1633 }, { "epoch": 2.5455962587685113, "grad_norm": 0.20540911119232894, "learning_rate": 8.352601156069364e-06, "loss": 0.275, "step": 1634 }, { "epoch": 2.5471551052221355, "grad_norm": 0.21777131689422363, "learning_rate": 8.323699421965318e-06, "loss": 0.2996, "step": 1635 }, { "epoch": 2.5487139516757598, "grad_norm": 0.20908427654761658, "learning_rate": 8.294797687861273e-06, "loss": 0.2742, "step": 1636 }, { "epoch": 2.5502727981293845, "grad_norm": 0.21275798202193402, "learning_rate": 8.265895953757226e-06, "loss": 0.2921, "step": 1637 }, { "epoch": 2.5518316445830087, "grad_norm": 0.22079763984850975, "learning_rate": 8.23699421965318e-06, "loss": 0.3023, "step": 1638 }, { "epoch": 2.553390491036633, "grad_norm": 0.20167698447825708, "learning_rate": 8.208092485549133e-06, "loss": 0.3018, "step": 1639 }, { "epoch": 2.554949337490257, "grad_norm": 0.204370882797588, "learning_rate": 8.179190751445087e-06, "loss": 0.2934, "step": 1640 }, { "epoch": 2.5565081839438815, "grad_norm": 0.20933271636814302, "learning_rate": 8.150289017341042e-06, "loss": 0.2884, "step": 1641 }, { "epoch": 2.5580670303975057, "grad_norm": 0.2364244855094284, "learning_rate": 8.121387283236995e-06, "loss": 0.291, "step": 1642 }, { "epoch": 2.55962587685113, "grad_norm": 0.21280205055543394, "learning_rate": 8.092485549132949e-06, "loss": 0.2953, "step": 1643 }, { "epoch": 2.5611847233047547, "grad_norm": 0.2225834531719899, "learning_rate": 8.063583815028902e-06, "loss": 0.2931, "step": 1644 }, { "epoch": 2.562743569758379, "grad_norm": 0.20003247002848737, "learning_rate": 8.034682080924856e-06, "loss": 0.2769, "step": 1645 }, { "epoch": 2.564302416212003, "grad_norm": 0.21372713472537674, "learning_rate": 8.00578034682081e-06, "loss": 0.2832, "step": 1646 }, { "epoch": 2.5658612626656274, "grad_norm": 0.21988186302060084, "learning_rate": 7.976878612716764e-06, "loss": 0.2816, "step": 1647 }, { "epoch": 2.5674201091192517, "grad_norm": 0.2114883685070932, "learning_rate": 7.947976878612718e-06, "loss": 0.2716, "step": 1648 }, { "epoch": 2.568978955572876, "grad_norm": 0.22479614996029462, "learning_rate": 7.919075144508671e-06, "loss": 0.2895, "step": 1649 }, { "epoch": 2.5705378020265, "grad_norm": 0.20995160083973538, "learning_rate": 7.890173410404624e-06, "loss": 0.2938, "step": 1650 }, { "epoch": 2.572096648480125, "grad_norm": 0.19074058893303877, "learning_rate": 7.861271676300578e-06, "loss": 0.2879, "step": 1651 }, { "epoch": 2.573655494933749, "grad_norm": 0.21433796239796818, "learning_rate": 7.832369942196533e-06, "loss": 0.2801, "step": 1652 }, { "epoch": 2.5752143413873734, "grad_norm": 0.23148461464412992, "learning_rate": 7.803468208092486e-06, "loss": 0.3049, "step": 1653 }, { "epoch": 2.5767731878409976, "grad_norm": 0.22404297663771336, "learning_rate": 7.77456647398844e-06, "loss": 0.2844, "step": 1654 }, { "epoch": 2.578332034294622, "grad_norm": 0.22573478963343901, "learning_rate": 7.745664739884393e-06, "loss": 0.2907, "step": 1655 }, { "epoch": 2.5798908807482466, "grad_norm": 0.19465572962012764, "learning_rate": 7.716763005780347e-06, "loss": 0.2786, "step": 1656 }, { "epoch": 2.5814497272018704, "grad_norm": 0.21095454192122828, "learning_rate": 7.687861271676302e-06, "loss": 0.2845, "step": 1657 }, { "epoch": 2.583008573655495, "grad_norm": 0.20373213583538066, "learning_rate": 7.658959537572255e-06, "loss": 0.2809, "step": 1658 }, { "epoch": 2.5845674201091193, "grad_norm": 0.20909977447044456, "learning_rate": 7.630057803468209e-06, "loss": 0.2918, "step": 1659 }, { "epoch": 2.5861262665627436, "grad_norm": 0.20642027240366984, "learning_rate": 7.601156069364162e-06, "loss": 0.2951, "step": 1660 }, { "epoch": 2.587685113016368, "grad_norm": 0.20983785591383902, "learning_rate": 7.5722543352601166e-06, "loss": 0.2958, "step": 1661 }, { "epoch": 2.589243959469992, "grad_norm": 0.23042841089901345, "learning_rate": 7.54335260115607e-06, "loss": 0.3101, "step": 1662 }, { "epoch": 2.590802805923617, "grad_norm": 0.24194824337916648, "learning_rate": 7.514450867052024e-06, "loss": 0.3112, "step": 1663 }, { "epoch": 2.5923616523772406, "grad_norm": 0.21401069536687056, "learning_rate": 7.485549132947978e-06, "loss": 0.2846, "step": 1664 }, { "epoch": 2.5939204988308653, "grad_norm": 0.20182835176068856, "learning_rate": 7.456647398843931e-06, "loss": 0.2884, "step": 1665 }, { "epoch": 2.5954793452844895, "grad_norm": 0.2438186993870951, "learning_rate": 7.4277456647398855e-06, "loss": 0.2914, "step": 1666 }, { "epoch": 2.597038191738114, "grad_norm": 0.24739122146108788, "learning_rate": 7.398843930635839e-06, "loss": 0.2981, "step": 1667 }, { "epoch": 2.598597038191738, "grad_norm": 0.25582873700332, "learning_rate": 7.369942196531792e-06, "loss": 0.2857, "step": 1668 }, { "epoch": 2.6001558846453623, "grad_norm": 0.23788093624422235, "learning_rate": 7.341040462427747e-06, "loss": 0.2992, "step": 1669 }, { "epoch": 2.601714731098987, "grad_norm": 0.2200261335858021, "learning_rate": 7.3121387283237e-06, "loss": 0.2869, "step": 1670 }, { "epoch": 2.603273577552611, "grad_norm": 0.22579760963949985, "learning_rate": 7.2832369942196544e-06, "loss": 0.2982, "step": 1671 }, { "epoch": 2.6048324240062355, "grad_norm": 0.23297289522530723, "learning_rate": 7.254335260115608e-06, "loss": 0.2825, "step": 1672 }, { "epoch": 2.6063912704598597, "grad_norm": 0.20686966606749702, "learning_rate": 7.225433526011561e-06, "loss": 0.2852, "step": 1673 }, { "epoch": 2.607950116913484, "grad_norm": 0.2266441955874942, "learning_rate": 7.196531791907516e-06, "loss": 0.3128, "step": 1674 }, { "epoch": 2.6095089633671082, "grad_norm": 0.23213362522888187, "learning_rate": 7.167630057803469e-06, "loss": 0.2865, "step": 1675 }, { "epoch": 2.6110678098207325, "grad_norm": 0.22468659409132077, "learning_rate": 7.1387283236994225e-06, "loss": 0.2654, "step": 1676 }, { "epoch": 2.612626656274357, "grad_norm": 0.22695879901341415, "learning_rate": 7.109826589595377e-06, "loss": 0.2846, "step": 1677 }, { "epoch": 2.6141855027279814, "grad_norm": 0.2055809379257179, "learning_rate": 7.08092485549133e-06, "loss": 0.2799, "step": 1678 }, { "epoch": 2.6157443491816057, "grad_norm": 0.21550005636993283, "learning_rate": 7.0520231213872845e-06, "loss": 0.276, "step": 1679 }, { "epoch": 2.61730319563523, "grad_norm": 0.2057285652964607, "learning_rate": 7.023121387283238e-06, "loss": 0.2871, "step": 1680 }, { "epoch": 2.618862042088854, "grad_norm": 0.2329408962798249, "learning_rate": 6.9942196531791914e-06, "loss": 0.2969, "step": 1681 }, { "epoch": 2.6204208885424785, "grad_norm": 0.21413261205015083, "learning_rate": 6.965317919075146e-06, "loss": 0.2887, "step": 1682 }, { "epoch": 2.6219797349961027, "grad_norm": 0.20032603391697887, "learning_rate": 6.936416184971098e-06, "loss": 0.268, "step": 1683 }, { "epoch": 2.6235385814497274, "grad_norm": 0.22915281553263556, "learning_rate": 6.907514450867052e-06, "loss": 0.2961, "step": 1684 }, { "epoch": 2.6250974279033517, "grad_norm": 0.22923520682538653, "learning_rate": 6.878612716763005e-06, "loss": 0.2789, "step": 1685 }, { "epoch": 2.626656274356976, "grad_norm": 0.20610407028884006, "learning_rate": 6.8497109826589595e-06, "loss": 0.2793, "step": 1686 }, { "epoch": 2.6282151208106, "grad_norm": 0.21052444570238973, "learning_rate": 6.820809248554913e-06, "loss": 0.2864, "step": 1687 }, { "epoch": 2.6297739672642244, "grad_norm": 0.2116958939607887, "learning_rate": 6.791907514450866e-06, "loss": 0.2893, "step": 1688 }, { "epoch": 2.6313328137178487, "grad_norm": 0.2277230086812389, "learning_rate": 6.763005780346821e-06, "loss": 0.2912, "step": 1689 }, { "epoch": 2.632891660171473, "grad_norm": 0.22320147334511584, "learning_rate": 6.734104046242774e-06, "loss": 0.3009, "step": 1690 }, { "epoch": 2.6344505066250976, "grad_norm": 0.22390700490954193, "learning_rate": 6.7052023121387284e-06, "loss": 0.2818, "step": 1691 }, { "epoch": 2.636009353078722, "grad_norm": 0.22080083950392762, "learning_rate": 6.676300578034682e-06, "loss": 0.3002, "step": 1692 }, { "epoch": 2.637568199532346, "grad_norm": 0.2045566499206754, "learning_rate": 6.647398843930635e-06, "loss": 0.2845, "step": 1693 }, { "epoch": 2.6391270459859704, "grad_norm": 0.2058593680536001, "learning_rate": 6.61849710982659e-06, "loss": 0.2963, "step": 1694 }, { "epoch": 2.6406858924395946, "grad_norm": 0.21761267846258828, "learning_rate": 6.589595375722543e-06, "loss": 0.2958, "step": 1695 }, { "epoch": 2.642244738893219, "grad_norm": 0.21267130911051824, "learning_rate": 6.5606936416184965e-06, "loss": 0.282, "step": 1696 }, { "epoch": 2.643803585346843, "grad_norm": 0.2725520542350063, "learning_rate": 6.531791907514451e-06, "loss": 0.3082, "step": 1697 }, { "epoch": 2.645362431800468, "grad_norm": 0.22632776119108647, "learning_rate": 6.502890173410404e-06, "loss": 0.2888, "step": 1698 }, { "epoch": 2.646921278254092, "grad_norm": 0.21287462722249717, "learning_rate": 6.4739884393063585e-06, "loss": 0.2873, "step": 1699 }, { "epoch": 2.6484801247077163, "grad_norm": 0.20006610540308692, "learning_rate": 6.445086705202312e-06, "loss": 0.2891, "step": 1700 }, { "epoch": 2.6500389711613406, "grad_norm": 0.23392897268756666, "learning_rate": 6.4161849710982654e-06, "loss": 0.299, "step": 1701 }, { "epoch": 2.651597817614965, "grad_norm": 0.2094557655435245, "learning_rate": 6.38728323699422e-06, "loss": 0.2893, "step": 1702 }, { "epoch": 2.6531566640685895, "grad_norm": 0.2162667697074946, "learning_rate": 6.358381502890173e-06, "loss": 0.3018, "step": 1703 }, { "epoch": 2.6547155105222133, "grad_norm": 0.20292202156574402, "learning_rate": 6.3294797687861275e-06, "loss": 0.289, "step": 1704 }, { "epoch": 2.656274356975838, "grad_norm": 0.19930348804964854, "learning_rate": 6.300578034682081e-06, "loss": 0.2911, "step": 1705 }, { "epoch": 2.6578332034294623, "grad_norm": 0.21358543639853841, "learning_rate": 6.271676300578034e-06, "loss": 0.2961, "step": 1706 }, { "epoch": 2.6593920498830865, "grad_norm": 0.20342382560761232, "learning_rate": 6.242774566473989e-06, "loss": 0.2964, "step": 1707 }, { "epoch": 2.660950896336711, "grad_norm": 0.21527984552285181, "learning_rate": 6.213872832369942e-06, "loss": 0.2757, "step": 1708 }, { "epoch": 2.662509742790335, "grad_norm": 0.21467701298854663, "learning_rate": 6.1849710982658956e-06, "loss": 0.2901, "step": 1709 }, { "epoch": 2.6640685892439597, "grad_norm": 0.19947891783905247, "learning_rate": 6.15606936416185e-06, "loss": 0.2821, "step": 1710 }, { "epoch": 2.6656274356975835, "grad_norm": 0.2082997257189345, "learning_rate": 6.127167630057803e-06, "loss": 0.2845, "step": 1711 }, { "epoch": 2.6671862821512082, "grad_norm": 0.21230363198815436, "learning_rate": 6.098265895953758e-06, "loss": 0.2933, "step": 1712 }, { "epoch": 2.6687451286048325, "grad_norm": 0.19933909904198954, "learning_rate": 6.069364161849711e-06, "loss": 0.2837, "step": 1713 }, { "epoch": 2.6703039750584567, "grad_norm": 0.22579536637132155, "learning_rate": 6.0404624277456645e-06, "loss": 0.2944, "step": 1714 }, { "epoch": 2.671862821512081, "grad_norm": 0.2083571147340846, "learning_rate": 6.011560693641619e-06, "loss": 0.2858, "step": 1715 }, { "epoch": 2.6734216679657052, "grad_norm": 0.21447279328154586, "learning_rate": 5.982658959537572e-06, "loss": 0.2796, "step": 1716 }, { "epoch": 2.67498051441933, "grad_norm": 0.2577104816893999, "learning_rate": 5.9537572254335265e-06, "loss": 0.2777, "step": 1717 }, { "epoch": 2.6765393608729537, "grad_norm": 0.2134999870561008, "learning_rate": 5.92485549132948e-06, "loss": 0.288, "step": 1718 }, { "epoch": 2.6780982073265784, "grad_norm": 0.2057246513397782, "learning_rate": 5.895953757225433e-06, "loss": 0.2758, "step": 1719 }, { "epoch": 2.6796570537802027, "grad_norm": 0.2291012037170924, "learning_rate": 5.867052023121388e-06, "loss": 0.2913, "step": 1720 }, { "epoch": 2.681215900233827, "grad_norm": 0.2100404936453596, "learning_rate": 5.838150289017341e-06, "loss": 0.2817, "step": 1721 }, { "epoch": 2.682774746687451, "grad_norm": 0.20420008955113567, "learning_rate": 5.809248554913295e-06, "loss": 0.2963, "step": 1722 }, { "epoch": 2.6843335931410754, "grad_norm": 0.21002723265147977, "learning_rate": 5.780346820809249e-06, "loss": 0.3049, "step": 1723 }, { "epoch": 2.6858924395947, "grad_norm": 0.20340290578451115, "learning_rate": 5.751445086705202e-06, "loss": 0.2847, "step": 1724 }, { "epoch": 2.6874512860483244, "grad_norm": 0.21104439985364004, "learning_rate": 5.722543352601157e-06, "loss": 0.2821, "step": 1725 }, { "epoch": 2.6890101325019486, "grad_norm": 0.21696962161072067, "learning_rate": 5.69364161849711e-06, "loss": 0.2721, "step": 1726 }, { "epoch": 2.690568978955573, "grad_norm": 0.23015064920584716, "learning_rate": 5.6647398843930635e-06, "loss": 0.2922, "step": 1727 }, { "epoch": 2.692127825409197, "grad_norm": 0.20813991146352637, "learning_rate": 5.635838150289018e-06, "loss": 0.2971, "step": 1728 }, { "epoch": 2.6936866718628214, "grad_norm": 0.20649336747594163, "learning_rate": 5.606936416184971e-06, "loss": 0.3036, "step": 1729 }, { "epoch": 2.6952455183164457, "grad_norm": 0.20331720787397545, "learning_rate": 5.578034682080925e-06, "loss": 0.2886, "step": 1730 }, { "epoch": 2.6968043647700704, "grad_norm": 0.21078699708857532, "learning_rate": 5.549132947976879e-06, "loss": 0.2849, "step": 1731 }, { "epoch": 2.6983632112236946, "grad_norm": 0.22036177053042216, "learning_rate": 5.5202312138728324e-06, "loss": 0.2861, "step": 1732 }, { "epoch": 2.699922057677319, "grad_norm": 0.20172166515293294, "learning_rate": 5.491329479768787e-06, "loss": 0.2866, "step": 1733 }, { "epoch": 2.701480904130943, "grad_norm": 0.19522835656935336, "learning_rate": 5.46242774566474e-06, "loss": 0.2681, "step": 1734 }, { "epoch": 2.7030397505845674, "grad_norm": 0.2212443755898533, "learning_rate": 5.433526011560694e-06, "loss": 0.2958, "step": 1735 }, { "epoch": 2.7045985970381916, "grad_norm": 0.19883348976389542, "learning_rate": 5.404624277456648e-06, "loss": 0.283, "step": 1736 }, { "epoch": 2.706157443491816, "grad_norm": 0.22774567151755043, "learning_rate": 5.375722543352601e-06, "loss": 0.2817, "step": 1737 }, { "epoch": 2.7077162899454406, "grad_norm": 0.19653848981003447, "learning_rate": 5.346820809248556e-06, "loss": 0.2807, "step": 1738 }, { "epoch": 2.709275136399065, "grad_norm": 0.2155411949245046, "learning_rate": 5.317919075144509e-06, "loss": 0.2874, "step": 1739 }, { "epoch": 2.710833982852689, "grad_norm": 0.21309269576853238, "learning_rate": 5.2890173410404626e-06, "loss": 0.2922, "step": 1740 }, { "epoch": 2.7123928293063133, "grad_norm": 0.20084602651111383, "learning_rate": 5.260115606936417e-06, "loss": 0.2861, "step": 1741 }, { "epoch": 2.7139516757599376, "grad_norm": 0.20949324216365292, "learning_rate": 5.23121387283237e-06, "loss": 0.2906, "step": 1742 }, { "epoch": 2.715510522213562, "grad_norm": 0.19787967361130746, "learning_rate": 5.202312138728324e-06, "loss": 0.2851, "step": 1743 }, { "epoch": 2.717069368667186, "grad_norm": 0.19777902230004707, "learning_rate": 5.173410404624278e-06, "loss": 0.2724, "step": 1744 }, { "epoch": 2.7186282151208108, "grad_norm": 0.21000085676745223, "learning_rate": 5.1445086705202315e-06, "loss": 0.2863, "step": 1745 }, { "epoch": 2.720187061574435, "grad_norm": 0.20127920458103757, "learning_rate": 5.115606936416186e-06, "loss": 0.2857, "step": 1746 }, { "epoch": 2.7217459080280593, "grad_norm": 0.2090573509510013, "learning_rate": 5.086705202312139e-06, "loss": 0.2866, "step": 1747 }, { "epoch": 2.7233047544816835, "grad_norm": 0.21739367096274437, "learning_rate": 5.057803468208093e-06, "loss": 0.2814, "step": 1748 }, { "epoch": 2.7248636009353078, "grad_norm": 0.20833581410322963, "learning_rate": 5.028901734104047e-06, "loss": 0.2888, "step": 1749 }, { "epoch": 2.726422447388932, "grad_norm": 0.19909078368387395, "learning_rate": 5e-06, "loss": 0.2911, "step": 1750 }, { "epoch": 2.7279812938425563, "grad_norm": 0.20109886859179885, "learning_rate": 4.971098265895955e-06, "loss": 0.2895, "step": 1751 }, { "epoch": 2.729540140296181, "grad_norm": 0.20247469803387927, "learning_rate": 4.942196531791908e-06, "loss": 0.276, "step": 1752 }, { "epoch": 2.7310989867498052, "grad_norm": 0.20421110219784483, "learning_rate": 4.913294797687862e-06, "loss": 0.2792, "step": 1753 }, { "epoch": 2.7326578332034295, "grad_norm": 0.19858031749032556, "learning_rate": 4.884393063583816e-06, "loss": 0.285, "step": 1754 }, { "epoch": 2.7342166796570537, "grad_norm": 0.22215186509011026, "learning_rate": 4.8554913294797685e-06, "loss": 0.3038, "step": 1755 }, { "epoch": 2.735775526110678, "grad_norm": 0.20256389152709856, "learning_rate": 4.826589595375723e-06, "loss": 0.2726, "step": 1756 }, { "epoch": 2.7373343725643027, "grad_norm": 0.21034046831696598, "learning_rate": 4.797687861271676e-06, "loss": 0.296, "step": 1757 }, { "epoch": 2.7388932190179265, "grad_norm": 0.20958728755008627, "learning_rate": 4.76878612716763e-06, "loss": 0.2868, "step": 1758 }, { "epoch": 2.740452065471551, "grad_norm": 0.20383033185004612, "learning_rate": 4.739884393063584e-06, "loss": 0.2854, "step": 1759 }, { "epoch": 2.7420109119251754, "grad_norm": 0.19907487917140468, "learning_rate": 4.710982658959537e-06, "loss": 0.285, "step": 1760 }, { "epoch": 2.7435697583787997, "grad_norm": 0.20044771305150624, "learning_rate": 4.682080924855492e-06, "loss": 0.2816, "step": 1761 }, { "epoch": 2.745128604832424, "grad_norm": 0.2164138120105804, "learning_rate": 4.653179190751445e-06, "loss": 0.2939, "step": 1762 }, { "epoch": 2.746687451286048, "grad_norm": 0.24028484525744273, "learning_rate": 4.624277456647399e-06, "loss": 0.3133, "step": 1763 }, { "epoch": 2.748246297739673, "grad_norm": 0.19984994036225665, "learning_rate": 4.595375722543353e-06, "loss": 0.2868, "step": 1764 }, { "epoch": 2.7498051441932967, "grad_norm": 0.21636457592977681, "learning_rate": 4.566473988439306e-06, "loss": 0.2871, "step": 1765 }, { "epoch": 2.7513639906469214, "grad_norm": 0.19990931527574785, "learning_rate": 4.53757225433526e-06, "loss": 0.2708, "step": 1766 }, { "epoch": 2.7529228371005456, "grad_norm": 0.217527953437702, "learning_rate": 4.508670520231214e-06, "loss": 0.2972, "step": 1767 }, { "epoch": 2.75448168355417, "grad_norm": 0.19223678719292475, "learning_rate": 4.4797687861271675e-06, "loss": 0.2898, "step": 1768 }, { "epoch": 2.756040530007794, "grad_norm": 0.21317778941850382, "learning_rate": 4.450867052023122e-06, "loss": 0.2857, "step": 1769 }, { "epoch": 2.7575993764614184, "grad_norm": 0.21322464171901906, "learning_rate": 4.421965317919075e-06, "loss": 0.2889, "step": 1770 }, { "epoch": 2.759158222915043, "grad_norm": 0.21331351232714513, "learning_rate": 4.393063583815029e-06, "loss": 0.2787, "step": 1771 }, { "epoch": 2.7607170693686673, "grad_norm": 0.20670483032146136, "learning_rate": 4.364161849710983e-06, "loss": 0.2961, "step": 1772 }, { "epoch": 2.7622759158222916, "grad_norm": 0.22435785586885243, "learning_rate": 4.3352601156069365e-06, "loss": 0.2943, "step": 1773 }, { "epoch": 2.763834762275916, "grad_norm": 0.19815210166141742, "learning_rate": 4.30635838150289e-06, "loss": 0.2935, "step": 1774 }, { "epoch": 2.76539360872954, "grad_norm": 0.20921721259317946, "learning_rate": 4.277456647398844e-06, "loss": 0.2711, "step": 1775 }, { "epoch": 2.7669524551831643, "grad_norm": 0.2239327598983314, "learning_rate": 4.248554913294798e-06, "loss": 0.3043, "step": 1776 }, { "epoch": 2.7685113016367886, "grad_norm": 0.2004589789574725, "learning_rate": 4.219653179190752e-06, "loss": 0.2964, "step": 1777 }, { "epoch": 2.7700701480904133, "grad_norm": 0.20625941199900238, "learning_rate": 4.190751445086705e-06, "loss": 0.2907, "step": 1778 }, { "epoch": 2.7716289945440375, "grad_norm": 0.20519545434025108, "learning_rate": 4.161849710982659e-06, "loss": 0.3068, "step": 1779 }, { "epoch": 2.773187840997662, "grad_norm": 0.21554367954746417, "learning_rate": 4.132947976878613e-06, "loss": 0.2952, "step": 1780 }, { "epoch": 2.774746687451286, "grad_norm": 0.21272563121891813, "learning_rate": 4.1040462427745666e-06, "loss": 0.2842, "step": 1781 }, { "epoch": 2.7763055339049103, "grad_norm": 0.21611401595167964, "learning_rate": 4.075144508670521e-06, "loss": 0.3046, "step": 1782 }, { "epoch": 2.7778643803585346, "grad_norm": 0.1997924444181988, "learning_rate": 4.046242774566474e-06, "loss": 0.2931, "step": 1783 }, { "epoch": 2.779423226812159, "grad_norm": 0.21506298028146192, "learning_rate": 4.017341040462428e-06, "loss": 0.2832, "step": 1784 }, { "epoch": 2.7809820732657835, "grad_norm": 0.20021184702963296, "learning_rate": 3.988439306358382e-06, "loss": 0.29, "step": 1785 }, { "epoch": 2.7825409197194078, "grad_norm": 0.20970397205026278, "learning_rate": 3.9595375722543355e-06, "loss": 0.3047, "step": 1786 }, { "epoch": 2.784099766173032, "grad_norm": 0.20062422216247, "learning_rate": 3.930635838150289e-06, "loss": 0.2844, "step": 1787 }, { "epoch": 2.7856586126266563, "grad_norm": 0.20179644813019854, "learning_rate": 3.901734104046243e-06, "loss": 0.2842, "step": 1788 }, { "epoch": 2.7872174590802805, "grad_norm": 0.20018746664410972, "learning_rate": 3.872832369942197e-06, "loss": 0.2892, "step": 1789 }, { "epoch": 2.7887763055339048, "grad_norm": 0.1946501903871324, "learning_rate": 3.843930635838151e-06, "loss": 0.29, "step": 1790 }, { "epoch": 2.790335151987529, "grad_norm": 0.19085065858233305, "learning_rate": 3.815028901734104e-06, "loss": 0.275, "step": 1791 }, { "epoch": 2.7918939984411537, "grad_norm": 0.20722259921075883, "learning_rate": 3.7861271676300583e-06, "loss": 0.2804, "step": 1792 }, { "epoch": 2.793452844894778, "grad_norm": 0.2103035874157165, "learning_rate": 3.757225433526012e-06, "loss": 0.2896, "step": 1793 }, { "epoch": 2.795011691348402, "grad_norm": 0.2102336939252641, "learning_rate": 3.7283236994219656e-06, "loss": 0.2779, "step": 1794 }, { "epoch": 2.7965705378020265, "grad_norm": 0.20336255858279942, "learning_rate": 3.6994219653179195e-06, "loss": 0.2933, "step": 1795 }, { "epoch": 2.7981293842556507, "grad_norm": 0.21206172168707907, "learning_rate": 3.6705202312138733e-06, "loss": 0.2786, "step": 1796 }, { "epoch": 2.799688230709275, "grad_norm": 0.19753691873632548, "learning_rate": 3.6416184971098272e-06, "loss": 0.2925, "step": 1797 }, { "epoch": 2.8012470771628992, "grad_norm": 0.21025169769964883, "learning_rate": 3.6127167630057807e-06, "loss": 0.2968, "step": 1798 }, { "epoch": 2.802805923616524, "grad_norm": 0.22618633232114826, "learning_rate": 3.5838150289017345e-06, "loss": 0.2959, "step": 1799 }, { "epoch": 2.804364770070148, "grad_norm": 0.21822500972127065, "learning_rate": 3.5549132947976884e-06, "loss": 0.3092, "step": 1800 }, { "epoch": 2.8059236165237724, "grad_norm": 0.220482613302332, "learning_rate": 3.5260115606936423e-06, "loss": 0.2955, "step": 1801 }, { "epoch": 2.8074824629773967, "grad_norm": 0.20780910132598235, "learning_rate": 3.4971098265895957e-06, "loss": 0.2761, "step": 1802 }, { "epoch": 2.809041309431021, "grad_norm": 0.2072002447217666, "learning_rate": 3.468208092485549e-06, "loss": 0.2915, "step": 1803 }, { "epoch": 2.8106001558846456, "grad_norm": 0.20725271527567482, "learning_rate": 3.4393063583815026e-06, "loss": 0.2876, "step": 1804 }, { "epoch": 2.8121590023382694, "grad_norm": 0.2101105745995178, "learning_rate": 3.4104046242774565e-06, "loss": 0.2856, "step": 1805 }, { "epoch": 2.813717848791894, "grad_norm": 0.20498941201400542, "learning_rate": 3.3815028901734103e-06, "loss": 0.2796, "step": 1806 }, { "epoch": 2.8152766952455184, "grad_norm": 0.20462911144204035, "learning_rate": 3.3526011560693642e-06, "loss": 0.3016, "step": 1807 }, { "epoch": 2.8168355416991426, "grad_norm": 0.2019060859936445, "learning_rate": 3.3236994219653177e-06, "loss": 0.2805, "step": 1808 }, { "epoch": 2.818394388152767, "grad_norm": 0.2124738066306522, "learning_rate": 3.2947976878612715e-06, "loss": 0.2893, "step": 1809 }, { "epoch": 2.819953234606391, "grad_norm": 0.2088568665930041, "learning_rate": 3.2658959537572254e-06, "loss": 0.2844, "step": 1810 }, { "epoch": 2.821512081060016, "grad_norm": 0.19645688715836024, "learning_rate": 3.2369942196531793e-06, "loss": 0.2749, "step": 1811 }, { "epoch": 2.8230709275136396, "grad_norm": 0.19409466738213488, "learning_rate": 3.2080924855491327e-06, "loss": 0.2812, "step": 1812 }, { "epoch": 2.8246297739672643, "grad_norm": 0.2120230001481611, "learning_rate": 3.1791907514450866e-06, "loss": 0.2954, "step": 1813 }, { "epoch": 2.8261886204208886, "grad_norm": 0.1984563439139553, "learning_rate": 3.1502890173410405e-06, "loss": 0.2984, "step": 1814 }, { "epoch": 2.827747466874513, "grad_norm": 0.20703019708211207, "learning_rate": 3.1213872832369943e-06, "loss": 0.2818, "step": 1815 }, { "epoch": 2.829306313328137, "grad_norm": 0.2073483200049451, "learning_rate": 3.0924855491329478e-06, "loss": 0.2945, "step": 1816 }, { "epoch": 2.8308651597817613, "grad_norm": 0.20322071664588748, "learning_rate": 3.0635838150289016e-06, "loss": 0.28, "step": 1817 }, { "epoch": 2.832424006235386, "grad_norm": 0.19200180838930353, "learning_rate": 3.0346820809248555e-06, "loss": 0.2899, "step": 1818 }, { "epoch": 2.8339828526890103, "grad_norm": 0.19563958688827382, "learning_rate": 3.0057803468208094e-06, "loss": 0.282, "step": 1819 }, { "epoch": 2.8355416991426345, "grad_norm": 0.1999256570053609, "learning_rate": 2.9768786127167633e-06, "loss": 0.2915, "step": 1820 }, { "epoch": 2.837100545596259, "grad_norm": 0.19601345181511165, "learning_rate": 2.9479768786127167e-06, "loss": 0.2846, "step": 1821 }, { "epoch": 2.838659392049883, "grad_norm": 0.2099789860276315, "learning_rate": 2.9190751445086706e-06, "loss": 0.2807, "step": 1822 }, { "epoch": 2.8402182385035073, "grad_norm": 0.20747261090512478, "learning_rate": 2.8901734104046244e-06, "loss": 0.279, "step": 1823 }, { "epoch": 2.8417770849571315, "grad_norm": 0.22257411158836346, "learning_rate": 2.8612716763005783e-06, "loss": 0.2997, "step": 1824 }, { "epoch": 2.8433359314107562, "grad_norm": 0.20554349028187938, "learning_rate": 2.8323699421965318e-06, "loss": 0.2831, "step": 1825 }, { "epoch": 2.8448947778643805, "grad_norm": 0.20735600722529712, "learning_rate": 2.8034682080924856e-06, "loss": 0.294, "step": 1826 }, { "epoch": 2.8464536243180047, "grad_norm": 0.21980476595040552, "learning_rate": 2.7745664739884395e-06, "loss": 0.3026, "step": 1827 }, { "epoch": 2.848012470771629, "grad_norm": 0.1925143943177367, "learning_rate": 2.7456647398843934e-06, "loss": 0.2876, "step": 1828 }, { "epoch": 2.8495713172252533, "grad_norm": 0.1957010171931748, "learning_rate": 2.716763005780347e-06, "loss": 0.2859, "step": 1829 }, { "epoch": 2.8511301636788775, "grad_norm": 0.20358594514165249, "learning_rate": 2.6878612716763007e-06, "loss": 0.282, "step": 1830 }, { "epoch": 2.8526890101325018, "grad_norm": 0.20541392956189472, "learning_rate": 2.6589595375722546e-06, "loss": 0.2795, "step": 1831 }, { "epoch": 2.8542478565861265, "grad_norm": 0.1956976623325346, "learning_rate": 2.6300578034682084e-06, "loss": 0.2753, "step": 1832 }, { "epoch": 2.8558067030397507, "grad_norm": 0.1944502240686237, "learning_rate": 2.601156069364162e-06, "loss": 0.2779, "step": 1833 }, { "epoch": 2.857365549493375, "grad_norm": 0.1842303161830082, "learning_rate": 2.5722543352601157e-06, "loss": 0.2747, "step": 1834 }, { "epoch": 2.858924395946999, "grad_norm": 0.19270326923177308, "learning_rate": 2.5433526011560696e-06, "loss": 0.2895, "step": 1835 }, { "epoch": 2.8604832424006235, "grad_norm": 0.20578389463380026, "learning_rate": 2.5144508670520235e-06, "loss": 0.2784, "step": 1836 }, { "epoch": 2.8620420888542477, "grad_norm": 0.18959748823440806, "learning_rate": 2.4855491329479774e-06, "loss": 0.2941, "step": 1837 }, { "epoch": 2.863600935307872, "grad_norm": 0.189041300234095, "learning_rate": 2.456647398843931e-06, "loss": 0.2863, "step": 1838 }, { "epoch": 2.8651597817614967, "grad_norm": 0.21031320630114367, "learning_rate": 2.4277456647398842e-06, "loss": 0.2996, "step": 1839 }, { "epoch": 2.866718628215121, "grad_norm": 0.20690718603806962, "learning_rate": 2.398843930635838e-06, "loss": 0.2812, "step": 1840 }, { "epoch": 2.868277474668745, "grad_norm": 0.2073404491146693, "learning_rate": 2.369942196531792e-06, "loss": 0.2947, "step": 1841 }, { "epoch": 2.8698363211223694, "grad_norm": 0.19029334104582032, "learning_rate": 2.341040462427746e-06, "loss": 0.2747, "step": 1842 }, { "epoch": 2.8713951675759937, "grad_norm": 0.19017238272079218, "learning_rate": 2.3121387283236993e-06, "loss": 0.2847, "step": 1843 }, { "epoch": 2.872954014029618, "grad_norm": 0.1979579277272625, "learning_rate": 2.283236994219653e-06, "loss": 0.2966, "step": 1844 }, { "epoch": 2.874512860483242, "grad_norm": 0.20597366624414118, "learning_rate": 2.254335260115607e-06, "loss": 0.2967, "step": 1845 }, { "epoch": 2.876071706936867, "grad_norm": 0.1914660504884028, "learning_rate": 2.225433526011561e-06, "loss": 0.2847, "step": 1846 }, { "epoch": 2.877630553390491, "grad_norm": 0.21004523817562415, "learning_rate": 2.1965317919075144e-06, "loss": 0.2844, "step": 1847 }, { "epoch": 2.8791893998441154, "grad_norm": 0.2050510627988167, "learning_rate": 2.1676300578034682e-06, "loss": 0.3026, "step": 1848 }, { "epoch": 2.8807482462977396, "grad_norm": 0.19875825849528217, "learning_rate": 2.138728323699422e-06, "loss": 0.29, "step": 1849 }, { "epoch": 2.882307092751364, "grad_norm": 0.19201938735986301, "learning_rate": 2.109826589595376e-06, "loss": 0.2814, "step": 1850 }, { "epoch": 2.8838659392049886, "grad_norm": 0.19276619456964975, "learning_rate": 2.0809248554913294e-06, "loss": 0.2794, "step": 1851 }, { "epoch": 2.8854247856586124, "grad_norm": 0.20837326668805686, "learning_rate": 2.0520231213872833e-06, "loss": 0.274, "step": 1852 }, { "epoch": 2.886983632112237, "grad_norm": 0.20834168208438708, "learning_rate": 2.023121387283237e-06, "loss": 0.2817, "step": 1853 }, { "epoch": 2.8885424785658613, "grad_norm": 0.20432457017704334, "learning_rate": 1.994219653179191e-06, "loss": 0.2949, "step": 1854 }, { "epoch": 2.8901013250194856, "grad_norm": 0.20163276098807092, "learning_rate": 1.9653179190751445e-06, "loss": 0.2706, "step": 1855 }, { "epoch": 2.89166017147311, "grad_norm": 0.20977626122832665, "learning_rate": 1.9364161849710983e-06, "loss": 0.3086, "step": 1856 }, { "epoch": 2.893219017926734, "grad_norm": 0.20060594756785594, "learning_rate": 1.907514450867052e-06, "loss": 0.2911, "step": 1857 }, { "epoch": 2.894777864380359, "grad_norm": 0.20293612568917765, "learning_rate": 1.878612716763006e-06, "loss": 0.2911, "step": 1858 }, { "epoch": 2.8963367108339826, "grad_norm": 0.2096162037257313, "learning_rate": 1.8497109826589597e-06, "loss": 0.2918, "step": 1859 }, { "epoch": 2.8978955572876073, "grad_norm": 0.21042720816672483, "learning_rate": 1.8208092485549136e-06, "loss": 0.3119, "step": 1860 }, { "epoch": 2.8994544037412315, "grad_norm": 0.18793473718435338, "learning_rate": 1.7919075144508673e-06, "loss": 0.2752, "step": 1861 }, { "epoch": 2.901013250194856, "grad_norm": 0.2079973480121488, "learning_rate": 1.7630057803468211e-06, "loss": 0.3029, "step": 1862 }, { "epoch": 2.90257209664848, "grad_norm": 0.1961639560813693, "learning_rate": 1.7341040462427746e-06, "loss": 0.2916, "step": 1863 }, { "epoch": 2.9041309431021043, "grad_norm": 0.1907492865888087, "learning_rate": 1.7052023121387282e-06, "loss": 0.2789, "step": 1864 }, { "epoch": 2.905689789555729, "grad_norm": 0.20235805559023728, "learning_rate": 1.6763005780346821e-06, "loss": 0.2849, "step": 1865 }, { "epoch": 2.9072486360093532, "grad_norm": 0.20288572024690002, "learning_rate": 1.6473988439306358e-06, "loss": 0.2654, "step": 1866 }, { "epoch": 2.9088074824629775, "grad_norm": 0.20289343651469097, "learning_rate": 1.6184971098265896e-06, "loss": 0.2951, "step": 1867 }, { "epoch": 2.9103663289166017, "grad_norm": 0.18922720141621407, "learning_rate": 1.5895953757225433e-06, "loss": 0.2762, "step": 1868 }, { "epoch": 2.911925175370226, "grad_norm": 0.19241261968232612, "learning_rate": 1.5606936416184972e-06, "loss": 0.2881, "step": 1869 }, { "epoch": 2.9134840218238502, "grad_norm": 0.19674852841020815, "learning_rate": 1.5317919075144508e-06, "loss": 0.2832, "step": 1870 }, { "epoch": 2.9150428682774745, "grad_norm": 0.18740622725667494, "learning_rate": 1.5028901734104047e-06, "loss": 0.283, "step": 1871 }, { "epoch": 2.916601714731099, "grad_norm": 0.20669295826778195, "learning_rate": 1.4739884393063584e-06, "loss": 0.2808, "step": 1872 }, { "epoch": 2.9181605611847234, "grad_norm": 0.18108268743966033, "learning_rate": 1.4450867052023122e-06, "loss": 0.2708, "step": 1873 }, { "epoch": 2.9197194076383477, "grad_norm": 0.19216377105239985, "learning_rate": 1.4161849710982659e-06, "loss": 0.2798, "step": 1874 }, { "epoch": 2.921278254091972, "grad_norm": 0.2036871353826851, "learning_rate": 1.3872832369942197e-06, "loss": 0.2871, "step": 1875 }, { "epoch": 2.922837100545596, "grad_norm": 0.19352818374916894, "learning_rate": 1.3583815028901734e-06, "loss": 0.2982, "step": 1876 }, { "epoch": 2.9243959469992205, "grad_norm": 0.19567432642222407, "learning_rate": 1.3294797687861273e-06, "loss": 0.2832, "step": 1877 }, { "epoch": 2.9259547934528447, "grad_norm": 0.19512011366905269, "learning_rate": 1.300578034682081e-06, "loss": 0.2839, "step": 1878 }, { "epoch": 2.9275136399064694, "grad_norm": 0.202990327713183, "learning_rate": 1.2716763005780348e-06, "loss": 0.2823, "step": 1879 }, { "epoch": 2.9290724863600937, "grad_norm": 0.20583389454978893, "learning_rate": 1.2427745664739887e-06, "loss": 0.2988, "step": 1880 }, { "epoch": 2.930631332813718, "grad_norm": 0.2014751871494088, "learning_rate": 1.2138728323699421e-06, "loss": 0.293, "step": 1881 }, { "epoch": 2.932190179267342, "grad_norm": 0.19892906821370257, "learning_rate": 1.184971098265896e-06, "loss": 0.2949, "step": 1882 }, { "epoch": 2.9337490257209664, "grad_norm": 0.18958700311963878, "learning_rate": 1.1560693641618497e-06, "loss": 0.2826, "step": 1883 }, { "epoch": 2.9353078721745907, "grad_norm": 0.19062254575039358, "learning_rate": 1.1271676300578035e-06, "loss": 0.2734, "step": 1884 }, { "epoch": 2.936866718628215, "grad_norm": 0.19081647607459015, "learning_rate": 1.0982658959537572e-06, "loss": 0.2719, "step": 1885 }, { "epoch": 2.9384255650818396, "grad_norm": 0.20749991571237006, "learning_rate": 1.069364161849711e-06, "loss": 0.2902, "step": 1886 }, { "epoch": 2.939984411535464, "grad_norm": 0.201338883635779, "learning_rate": 1.0404624277456647e-06, "loss": 0.2866, "step": 1887 }, { "epoch": 2.941543257989088, "grad_norm": 0.185374324630895, "learning_rate": 1.0115606936416186e-06, "loss": 0.291, "step": 1888 }, { "epoch": 2.9431021044427124, "grad_norm": 0.19169182164611995, "learning_rate": 9.826589595375722e-07, "loss": 0.2919, "step": 1889 }, { "epoch": 2.9446609508963366, "grad_norm": 0.19500681675014075, "learning_rate": 9.53757225433526e-07, "loss": 0.289, "step": 1890 }, { "epoch": 2.946219797349961, "grad_norm": 0.20336352790343418, "learning_rate": 9.248554913294799e-07, "loss": 0.2888, "step": 1891 }, { "epoch": 2.947778643803585, "grad_norm": 0.18996694302336917, "learning_rate": 8.959537572254336e-07, "loss": 0.2837, "step": 1892 }, { "epoch": 2.94933749025721, "grad_norm": 0.2156301542305652, "learning_rate": 8.670520231213873e-07, "loss": 0.2799, "step": 1893 }, { "epoch": 2.950896336710834, "grad_norm": 0.19727279525402547, "learning_rate": 8.381502890173411e-07, "loss": 0.2906, "step": 1894 }, { "epoch": 2.9524551831644583, "grad_norm": 0.19284045282449663, "learning_rate": 8.092485549132948e-07, "loss": 0.2795, "step": 1895 }, { "epoch": 2.9540140296180826, "grad_norm": 0.20089648135420463, "learning_rate": 7.803468208092486e-07, "loss": 0.2857, "step": 1896 }, { "epoch": 2.955572876071707, "grad_norm": 0.18636348684721002, "learning_rate": 7.514450867052023e-07, "loss": 0.28, "step": 1897 }, { "epoch": 2.9571317225253315, "grad_norm": 0.1842759397599328, "learning_rate": 7.225433526011561e-07, "loss": 0.2811, "step": 1898 }, { "epoch": 2.9586905689789553, "grad_norm": 0.198586721483869, "learning_rate": 6.936416184971099e-07, "loss": 0.2795, "step": 1899 }, { "epoch": 2.96024941543258, "grad_norm": 0.20175920964436003, "learning_rate": 6.647398843930636e-07, "loss": 0.2963, "step": 1900 }, { "epoch": 2.9618082618862043, "grad_norm": 0.1894255377343828, "learning_rate": 6.358381502890174e-07, "loss": 0.2786, "step": 1901 }, { "epoch": 2.9633671083398285, "grad_norm": 0.18846655912423485, "learning_rate": 6.069364161849711e-07, "loss": 0.2755, "step": 1902 }, { "epoch": 2.964925954793453, "grad_norm": 0.1928139829341696, "learning_rate": 5.780346820809248e-07, "loss": 0.2999, "step": 1903 }, { "epoch": 2.966484801247077, "grad_norm": 0.19323847526640792, "learning_rate": 5.491329479768786e-07, "loss": 0.2937, "step": 1904 }, { "epoch": 2.9680436477007017, "grad_norm": 0.20120482130965053, "learning_rate": 5.202312138728324e-07, "loss": 0.2947, "step": 1905 }, { "epoch": 2.9696024941543255, "grad_norm": 0.21061414202296763, "learning_rate": 4.913294797687861e-07, "loss": 0.2905, "step": 1906 }, { "epoch": 2.9711613406079502, "grad_norm": 0.21063326211832292, "learning_rate": 4.6242774566473993e-07, "loss": 0.2965, "step": 1907 }, { "epoch": 2.9727201870615745, "grad_norm": 0.20410862279580308, "learning_rate": 4.3352601156069365e-07, "loss": 0.2891, "step": 1908 }, { "epoch": 2.9742790335151987, "grad_norm": 0.19836165458820448, "learning_rate": 4.046242774566474e-07, "loss": 0.2962, "step": 1909 }, { "epoch": 2.975837879968823, "grad_norm": 0.19040393729360364, "learning_rate": 3.7572254335260117e-07, "loss": 0.2986, "step": 1910 }, { "epoch": 2.9773967264224472, "grad_norm": 0.18498159624839047, "learning_rate": 3.4682080924855494e-07, "loss": 0.2855, "step": 1911 }, { "epoch": 2.978955572876072, "grad_norm": 0.19089075468074318, "learning_rate": 3.179190751445087e-07, "loss": 0.2913, "step": 1912 }, { "epoch": 2.980514419329696, "grad_norm": 0.19143218775583679, "learning_rate": 2.890173410404624e-07, "loss": 0.2739, "step": 1913 }, { "epoch": 2.9820732657833204, "grad_norm": 0.19760714438972718, "learning_rate": 2.601156069364162e-07, "loss": 0.2955, "step": 1914 }, { "epoch": 2.9836321122369447, "grad_norm": 0.18341652706228365, "learning_rate": 2.3121387283236997e-07, "loss": 0.2713, "step": 1915 }, { "epoch": 2.985190958690569, "grad_norm": 0.19428086051572635, "learning_rate": 2.023121387283237e-07, "loss": 0.2867, "step": 1916 }, { "epoch": 2.986749805144193, "grad_norm": 0.19485137014886658, "learning_rate": 1.7341040462427747e-07, "loss": 0.2761, "step": 1917 }, { "epoch": 2.9883086515978174, "grad_norm": 0.19777851419652856, "learning_rate": 1.445086705202312e-07, "loss": 0.2823, "step": 1918 }, { "epoch": 2.989867498051442, "grad_norm": 0.19065668774947628, "learning_rate": 1.1560693641618498e-07, "loss": 0.2954, "step": 1919 }, { "epoch": 2.9914263445050664, "grad_norm": 0.1974490266526018, "learning_rate": 8.670520231213873e-08, "loss": 0.2813, "step": 1920 }, { "epoch": 2.9929851909586906, "grad_norm": 0.19818812256280618, "learning_rate": 5.780346820809249e-08, "loss": 0.2773, "step": 1921 }, { "epoch": 2.994544037412315, "grad_norm": 0.1943618741709864, "learning_rate": 2.8901734104046246e-08, "loss": 0.2787, "step": 1922 }, { "epoch": 2.996102883865939, "grad_norm": 0.1954427128274177, "learning_rate": 0.0, "loss": 0.278, "step": 1923 }, { "epoch": 2.996102883865939, "step": 1923, "total_flos": 1.6424334483479593e+18, "train_loss": 0.4532742842497208, "train_runtime": 111641.3311, "train_samples_per_second": 0.276, "train_steps_per_second": 0.017 } ], "logging_steps": 1, "max_steps": 1923, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.6424334483479593e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }