{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.2443223063362023, "eval_steps": 500, "global_step": 3000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00041480866950119257, "grad_norm": 5.84375, "learning_rate": 0.0, "loss": 5.3678, "step": 1 }, { "epoch": 0.0008296173390023851, "grad_norm": 5.15625, "learning_rate": 4e-05, "loss": 5.3815, "step": 2 }, { "epoch": 0.0012444260085035777, "grad_norm": 8.0625, "learning_rate": 8e-05, "loss": 5.578, "step": 3 }, { "epoch": 0.0016592346780047703, "grad_norm": 5.15625, "learning_rate": 0.00012, "loss": 5.1489, "step": 4 }, { "epoch": 0.002074043347505963, "grad_norm": 4.96875, "learning_rate": 0.00016, "loss": 5.1569, "step": 5 }, { "epoch": 0.0024888520170071554, "grad_norm": 4.3125, "learning_rate": 0.0002, "loss": 5.2508, "step": 6 }, { "epoch": 0.002903660686508348, "grad_norm": 3.828125, "learning_rate": 0.00019997998999499752, "loss": 5.2444, "step": 7 }, { "epoch": 0.0033184693560095405, "grad_norm": 3.453125, "learning_rate": 0.000199959979989995, "loss": 5.1866, "step": 8 }, { "epoch": 0.0037332780255107333, "grad_norm": 4.5625, "learning_rate": 0.00019993996998499253, "loss": 4.7983, "step": 9 }, { "epoch": 0.004148086695011926, "grad_norm": 4.0, "learning_rate": 0.00019991995997999, "loss": 5.2917, "step": 10 }, { "epoch": 0.0045628953645131184, "grad_norm": 3.90625, "learning_rate": 0.0001998999499749875, "loss": 5.1454, "step": 11 }, { "epoch": 0.004977704034014311, "grad_norm": 3.65625, "learning_rate": 0.000199879939969985, "loss": 5.2539, "step": 12 }, { "epoch": 0.005392512703515503, "grad_norm": 3.671875, "learning_rate": 0.0001998599299649825, "loss": 5.2084, "step": 13 }, { "epoch": 0.005807321373016696, "grad_norm": 3.65625, "learning_rate": 0.00019983991995998, "loss": 5.28, "step": 14 }, { "epoch": 0.006222130042517889, "grad_norm": 3.359375, "learning_rate": 0.0001998199099549775, "loss": 5.2941, "step": 15 }, { "epoch": 0.006636938712019081, "grad_norm": 3.328125, "learning_rate": 0.000199799899949975, "loss": 5.0754, "step": 16 }, { "epoch": 0.007051747381520273, "grad_norm": 3.953125, "learning_rate": 0.0001997798899449725, "loss": 5.0806, "step": 17 }, { "epoch": 0.007466556051021467, "grad_norm": 3.59375, "learning_rate": 0.00019975987993997, "loss": 5.303, "step": 18 }, { "epoch": 0.007881364720522659, "grad_norm": 3.234375, "learning_rate": 0.0001997398699349675, "loss": 5.0535, "step": 19 }, { "epoch": 0.008296173390023852, "grad_norm": 3.046875, "learning_rate": 0.000199719859929965, "loss": 5.0371, "step": 20 }, { "epoch": 0.008710982059525044, "grad_norm": 2.90625, "learning_rate": 0.00019969984992496248, "loss": 4.9977, "step": 21 }, { "epoch": 0.009125790729026237, "grad_norm": 3.0, "learning_rate": 0.00019967983991996, "loss": 5.3496, "step": 22 }, { "epoch": 0.009540599398527428, "grad_norm": 2.5, "learning_rate": 0.00019965982991495748, "loss": 4.8964, "step": 23 }, { "epoch": 0.009955408068028622, "grad_norm": 3.15625, "learning_rate": 0.000199639819909955, "loss": 5.0368, "step": 24 }, { "epoch": 0.010370216737529815, "grad_norm": 2.875, "learning_rate": 0.00019961980990495248, "loss": 5.5296, "step": 25 }, { "epoch": 0.010785025407031006, "grad_norm": 2.46875, "learning_rate": 0.00019959979989995, "loss": 5.1468, "step": 26 }, { "epoch": 0.0111998340765322, "grad_norm": 3.09375, "learning_rate": 0.00019957978989494749, "loss": 5.0455, "step": 27 }, { "epoch": 0.011614642746033393, "grad_norm": 3.09375, "learning_rate": 0.00019955977988994497, "loss": 5.1206, "step": 28 }, { "epoch": 0.012029451415534584, "grad_norm": 3.03125, "learning_rate": 0.00019953976988494246, "loss": 5.0164, "step": 29 }, { "epoch": 0.012444260085035777, "grad_norm": 2.765625, "learning_rate": 0.00019951975987993998, "loss": 5.0402, "step": 30 }, { "epoch": 0.01285906875453697, "grad_norm": 2.5625, "learning_rate": 0.00019949974987493746, "loss": 5.1637, "step": 31 }, { "epoch": 0.013273877424038162, "grad_norm": 3.078125, "learning_rate": 0.00019947973986993498, "loss": 5.1842, "step": 32 }, { "epoch": 0.013688686093539355, "grad_norm": 2.40625, "learning_rate": 0.00019945972986493246, "loss": 4.8599, "step": 33 }, { "epoch": 0.014103494763040547, "grad_norm": 2.546875, "learning_rate": 0.00019943971985992998, "loss": 5.1072, "step": 34 }, { "epoch": 0.01451830343254174, "grad_norm": 2.359375, "learning_rate": 0.0001994197098549275, "loss": 5.1639, "step": 35 }, { "epoch": 0.014933112102042933, "grad_norm": 2.375, "learning_rate": 0.00019939969984992498, "loss": 5.0431, "step": 36 }, { "epoch": 0.015347920771544125, "grad_norm": 3.140625, "learning_rate": 0.00019937968984492247, "loss": 5.0079, "step": 37 }, { "epoch": 0.015762729441045318, "grad_norm": 2.421875, "learning_rate": 0.00019935967983991995, "loss": 5.1109, "step": 38 }, { "epoch": 0.01617753811054651, "grad_norm": 2.390625, "learning_rate": 0.00019933966983491747, "loss": 5.0062, "step": 39 }, { "epoch": 0.016592346780047704, "grad_norm": 2.453125, "learning_rate": 0.00019931965982991496, "loss": 5.1522, "step": 40 }, { "epoch": 0.017007155449548894, "grad_norm": 2.578125, "learning_rate": 0.00019929964982491247, "loss": 4.9839, "step": 41 }, { "epoch": 0.017421964119050087, "grad_norm": 2.25, "learning_rate": 0.00019927963981990996, "loss": 5.0712, "step": 42 }, { "epoch": 0.01783677278855128, "grad_norm": 2.65625, "learning_rate": 0.00019925962981490747, "loss": 5.2342, "step": 43 }, { "epoch": 0.018251581458052474, "grad_norm": 2.46875, "learning_rate": 0.00019923961980990496, "loss": 5.0108, "step": 44 }, { "epoch": 0.018666390127553667, "grad_norm": 2.671875, "learning_rate": 0.00019921960980490247, "loss": 5.1698, "step": 45 }, { "epoch": 0.019081198797054857, "grad_norm": 3.296875, "learning_rate": 0.00019919959979989996, "loss": 5.1331, "step": 46 }, { "epoch": 0.01949600746655605, "grad_norm": 3.203125, "learning_rate": 0.00019917958979489745, "loss": 5.1377, "step": 47 }, { "epoch": 0.019910816136057243, "grad_norm": 2.59375, "learning_rate": 0.00019915957978989496, "loss": 4.8147, "step": 48 }, { "epoch": 0.020325624805558436, "grad_norm": 4.0, "learning_rate": 0.00019913956978489245, "loss": 4.7886, "step": 49 }, { "epoch": 0.02074043347505963, "grad_norm": 2.796875, "learning_rate": 0.00019911955977988996, "loss": 5.0484, "step": 50 }, { "epoch": 0.021155242144560823, "grad_norm": 3.078125, "learning_rate": 0.00019909954977488745, "loss": 5.22, "step": 51 }, { "epoch": 0.021570050814062013, "grad_norm": 2.359375, "learning_rate": 0.00019907953976988497, "loss": 5.0641, "step": 52 }, { "epoch": 0.021984859483563206, "grad_norm": 2.71875, "learning_rate": 0.00019905952976488245, "loss": 4.9759, "step": 53 }, { "epoch": 0.0223996681530644, "grad_norm": 2.640625, "learning_rate": 0.00019903951975987997, "loss": 5.0263, "step": 54 }, { "epoch": 0.022814476822565592, "grad_norm": 3.171875, "learning_rate": 0.00019901950975487746, "loss": 5.1059, "step": 55 }, { "epoch": 0.023229285492066785, "grad_norm": 2.875, "learning_rate": 0.00019899949974987494, "loss": 5.0911, "step": 56 }, { "epoch": 0.023644094161567975, "grad_norm": 3.390625, "learning_rate": 0.00019897948974487243, "loss": 4.9596, "step": 57 }, { "epoch": 0.02405890283106917, "grad_norm": 2.4375, "learning_rate": 0.00019895947973986994, "loss": 4.9327, "step": 58 }, { "epoch": 0.02447371150057036, "grad_norm": 2.421875, "learning_rate": 0.00019893946973486743, "loss": 5.2353, "step": 59 }, { "epoch": 0.024888520170071555, "grad_norm": 2.609375, "learning_rate": 0.00019891945972986495, "loss": 4.7571, "step": 60 }, { "epoch": 0.025303328839572748, "grad_norm": 2.75, "learning_rate": 0.00019889944972486243, "loss": 5.0165, "step": 61 }, { "epoch": 0.02571813750907394, "grad_norm": 2.421875, "learning_rate": 0.00019887943971985995, "loss": 5.0253, "step": 62 }, { "epoch": 0.02613294617857513, "grad_norm": 2.640625, "learning_rate": 0.00019885942971485744, "loss": 5.0432, "step": 63 }, { "epoch": 0.026547754848076324, "grad_norm": 2.640625, "learning_rate": 0.00019883941970985492, "loss": 5.2195, "step": 64 }, { "epoch": 0.026962563517577517, "grad_norm": 2.265625, "learning_rate": 0.00019881940970485244, "loss": 5.157, "step": 65 }, { "epoch": 0.02737737218707871, "grad_norm": 2.4375, "learning_rate": 0.00019879939969984992, "loss": 5.0014, "step": 66 }, { "epoch": 0.027792180856579904, "grad_norm": 2.421875, "learning_rate": 0.00019877938969484744, "loss": 5.05, "step": 67 }, { "epoch": 0.028206989526081094, "grad_norm": 2.71875, "learning_rate": 0.00019875937968984493, "loss": 4.9888, "step": 68 }, { "epoch": 0.028621798195582287, "grad_norm": 2.71875, "learning_rate": 0.00019873936968484244, "loss": 4.7699, "step": 69 }, { "epoch": 0.02903660686508348, "grad_norm": 2.328125, "learning_rate": 0.00019871935967983993, "loss": 4.6892, "step": 70 }, { "epoch": 0.029451415534584673, "grad_norm": 2.875, "learning_rate": 0.00019869934967483744, "loss": 4.8417, "step": 71 }, { "epoch": 0.029866224204085867, "grad_norm": 2.65625, "learning_rate": 0.00019867933966983493, "loss": 4.9361, "step": 72 }, { "epoch": 0.030281032873587056, "grad_norm": 3.21875, "learning_rate": 0.00019865932966483242, "loss": 4.86, "step": 73 }, { "epoch": 0.03069584154308825, "grad_norm": 2.671875, "learning_rate": 0.0001986393196598299, "loss": 5.232, "step": 74 }, { "epoch": 0.031110650212589443, "grad_norm": 2.578125, "learning_rate": 0.00019861930965482742, "loss": 5.0538, "step": 75 }, { "epoch": 0.031525458882090636, "grad_norm": 2.828125, "learning_rate": 0.00019859929964982493, "loss": 4.9119, "step": 76 }, { "epoch": 0.03194026755159183, "grad_norm": 2.5, "learning_rate": 0.00019857928964482242, "loss": 4.9387, "step": 77 }, { "epoch": 0.03235507622109302, "grad_norm": 2.78125, "learning_rate": 0.00019855927963981993, "loss": 5.0123, "step": 78 }, { "epoch": 0.032769884890594216, "grad_norm": 3.65625, "learning_rate": 0.00019853926963481742, "loss": 4.8198, "step": 79 }, { "epoch": 0.03318469356009541, "grad_norm": 2.84375, "learning_rate": 0.00019851925962981494, "loss": 5.1624, "step": 80 }, { "epoch": 0.0335995022295966, "grad_norm": 2.765625, "learning_rate": 0.00019849924962481242, "loss": 4.7054, "step": 81 }, { "epoch": 0.03401431089909779, "grad_norm": 3.1875, "learning_rate": 0.0001984792396198099, "loss": 4.8135, "step": 82 }, { "epoch": 0.03442911956859898, "grad_norm": 2.921875, "learning_rate": 0.0001984592296148074, "loss": 5.0179, "step": 83 }, { "epoch": 0.034843928238100175, "grad_norm": 2.859375, "learning_rate": 0.0001984392196098049, "loss": 4.8094, "step": 84 }, { "epoch": 0.03525873690760137, "grad_norm": 2.96875, "learning_rate": 0.0001984192096048024, "loss": 4.8648, "step": 85 }, { "epoch": 0.03567354557710256, "grad_norm": 2.484375, "learning_rate": 0.0001983991995997999, "loss": 4.8648, "step": 86 }, { "epoch": 0.036088354246603754, "grad_norm": 2.484375, "learning_rate": 0.0001983791895947974, "loss": 5.2012, "step": 87 }, { "epoch": 0.03650316291610495, "grad_norm": 2.625, "learning_rate": 0.00019835917958979492, "loss": 5.0743, "step": 88 }, { "epoch": 0.03691797158560614, "grad_norm": 2.4375, "learning_rate": 0.0001983391695847924, "loss": 4.9473, "step": 89 }, { "epoch": 0.037332780255107334, "grad_norm": 2.546875, "learning_rate": 0.00019831915957978992, "loss": 4.8153, "step": 90 }, { "epoch": 0.03774758892460853, "grad_norm": 2.5625, "learning_rate": 0.0001982991495747874, "loss": 4.9615, "step": 91 }, { "epoch": 0.038162397594109713, "grad_norm": 2.359375, "learning_rate": 0.0001982791395697849, "loss": 4.7808, "step": 92 }, { "epoch": 0.03857720626361091, "grad_norm": 3.578125, "learning_rate": 0.0001982591295647824, "loss": 4.7397, "step": 93 }, { "epoch": 0.0389920149331121, "grad_norm": 3.171875, "learning_rate": 0.0001982391195597799, "loss": 4.7318, "step": 94 }, { "epoch": 0.03940682360261329, "grad_norm": 2.71875, "learning_rate": 0.0001982191095547774, "loss": 4.7505, "step": 95 }, { "epoch": 0.039821632272114486, "grad_norm": 2.84375, "learning_rate": 0.0001981990995497749, "loss": 4.9952, "step": 96 }, { "epoch": 0.04023644094161568, "grad_norm": 3.390625, "learning_rate": 0.0001981790895447724, "loss": 4.7346, "step": 97 }, { "epoch": 0.04065124961111687, "grad_norm": 2.828125, "learning_rate": 0.0001981590795397699, "loss": 4.9629, "step": 98 }, { "epoch": 0.041066058280618066, "grad_norm": 2.5625, "learning_rate": 0.00019813906953476738, "loss": 4.9172, "step": 99 }, { "epoch": 0.04148086695011926, "grad_norm": 3.09375, "learning_rate": 0.00019811905952976487, "loss": 4.7065, "step": 100 }, { "epoch": 0.04189567561962045, "grad_norm": 3.25, "learning_rate": 0.00019809904952476239, "loss": 4.8555, "step": 101 }, { "epoch": 0.042310484289121646, "grad_norm": 3.171875, "learning_rate": 0.00019807903951975987, "loss": 5.1348, "step": 102 }, { "epoch": 0.04272529295862283, "grad_norm": 2.734375, "learning_rate": 0.0001980590295147574, "loss": 4.6186, "step": 103 }, { "epoch": 0.043140101628124025, "grad_norm": 2.75, "learning_rate": 0.0001980390195097549, "loss": 4.9732, "step": 104 }, { "epoch": 0.04355491029762522, "grad_norm": 3.1875, "learning_rate": 0.0001980190095047524, "loss": 4.6935, "step": 105 }, { "epoch": 0.04396971896712641, "grad_norm": 2.640625, "learning_rate": 0.0001979989994997499, "loss": 4.8805, "step": 106 }, { "epoch": 0.044384527636627605, "grad_norm": 2.828125, "learning_rate": 0.0001979789894947474, "loss": 4.7667, "step": 107 }, { "epoch": 0.0447993363061288, "grad_norm": 2.515625, "learning_rate": 0.00019795897948974488, "loss": 4.5853, "step": 108 }, { "epoch": 0.04521414497562999, "grad_norm": 2.640625, "learning_rate": 0.00019793896948474236, "loss": 4.8158, "step": 109 }, { "epoch": 0.045628953645131184, "grad_norm": 3.046875, "learning_rate": 0.00019791895947973988, "loss": 4.7036, "step": 110 }, { "epoch": 0.04604376231463238, "grad_norm": 2.5, "learning_rate": 0.00019789894947473737, "loss": 4.8263, "step": 111 }, { "epoch": 0.04645857098413357, "grad_norm": 2.890625, "learning_rate": 0.00019787893946973488, "loss": 4.6605, "step": 112 }, { "epoch": 0.046873379653634764, "grad_norm": 2.625, "learning_rate": 0.00019785892946473237, "loss": 4.941, "step": 113 }, { "epoch": 0.04728818832313595, "grad_norm": 2.296875, "learning_rate": 0.00019783891945972988, "loss": 4.8322, "step": 114 }, { "epoch": 0.047702996992637144, "grad_norm": 2.65625, "learning_rate": 0.00019781890945472737, "loss": 4.7647, "step": 115 }, { "epoch": 0.04811780566213834, "grad_norm": 2.640625, "learning_rate": 0.00019779889944972488, "loss": 4.9316, "step": 116 }, { "epoch": 0.04853261433163953, "grad_norm": 3.078125, "learning_rate": 0.00019777888944472237, "loss": 5.0164, "step": 117 }, { "epoch": 0.04894742300114072, "grad_norm": 2.84375, "learning_rate": 0.00019775887943971986, "loss": 5.037, "step": 118 }, { "epoch": 0.049362231670641916, "grad_norm": 2.578125, "learning_rate": 0.00019773886943471737, "loss": 4.8183, "step": 119 }, { "epoch": 0.04977704034014311, "grad_norm": 2.6875, "learning_rate": 0.00019771885942971486, "loss": 4.6524, "step": 120 }, { "epoch": 0.0501918490096443, "grad_norm": 2.578125, "learning_rate": 0.00019769884942471237, "loss": 4.8141, "step": 121 }, { "epoch": 0.050606657679145496, "grad_norm": 2.515625, "learning_rate": 0.00019767883941970986, "loss": 4.719, "step": 122 }, { "epoch": 0.05102146634864669, "grad_norm": 3.0, "learning_rate": 0.00019765882941470738, "loss": 4.6229, "step": 123 }, { "epoch": 0.05143627501814788, "grad_norm": 3.03125, "learning_rate": 0.00019763881940970486, "loss": 4.7896, "step": 124 }, { "epoch": 0.05185108368764907, "grad_norm": 2.78125, "learning_rate": 0.00019761880940470238, "loss": 4.6164, "step": 125 }, { "epoch": 0.05226589235715026, "grad_norm": 2.46875, "learning_rate": 0.00019759879939969987, "loss": 4.777, "step": 126 }, { "epoch": 0.052680701026651455, "grad_norm": 2.40625, "learning_rate": 0.00019757878939469735, "loss": 4.7161, "step": 127 }, { "epoch": 0.05309550969615265, "grad_norm": 2.484375, "learning_rate": 0.00019755877938969484, "loss": 4.5956, "step": 128 }, { "epoch": 0.05351031836565384, "grad_norm": 2.8125, "learning_rate": 0.00019753876938469235, "loss": 4.8752, "step": 129 }, { "epoch": 0.053925127035155035, "grad_norm": 2.09375, "learning_rate": 0.00019751875937968984, "loss": 4.8244, "step": 130 }, { "epoch": 0.05433993570465623, "grad_norm": 2.6875, "learning_rate": 0.00019749874937468736, "loss": 4.9454, "step": 131 }, { "epoch": 0.05475474437415742, "grad_norm": 2.390625, "learning_rate": 0.00019747873936968487, "loss": 4.8617, "step": 132 }, { "epoch": 0.055169553043658615, "grad_norm": 2.65625, "learning_rate": 0.00019745872936468236, "loss": 4.9938, "step": 133 }, { "epoch": 0.05558436171315981, "grad_norm": 2.390625, "learning_rate": 0.00019743871935967985, "loss": 4.6219, "step": 134 }, { "epoch": 0.055999170382661, "grad_norm": 3.296875, "learning_rate": 0.00019741870935467733, "loss": 4.699, "step": 135 }, { "epoch": 0.05641397905216219, "grad_norm": 2.78125, "learning_rate": 0.00019739869934967485, "loss": 4.8001, "step": 136 }, { "epoch": 0.05682878772166338, "grad_norm": 2.40625, "learning_rate": 0.00019737868934467233, "loss": 4.617, "step": 137 }, { "epoch": 0.057243596391164574, "grad_norm": 2.59375, "learning_rate": 0.00019735867933966985, "loss": 4.6295, "step": 138 }, { "epoch": 0.05765840506066577, "grad_norm": 2.828125, "learning_rate": 0.00019733866933466734, "loss": 4.716, "step": 139 }, { "epoch": 0.05807321373016696, "grad_norm": 2.265625, "learning_rate": 0.00019731865932966485, "loss": 4.7463, "step": 140 }, { "epoch": 0.05848802239966815, "grad_norm": 2.484375, "learning_rate": 0.00019729864932466234, "loss": 4.6761, "step": 141 }, { "epoch": 0.05890283106916935, "grad_norm": 2.703125, "learning_rate": 0.00019727863931965985, "loss": 4.7429, "step": 142 }, { "epoch": 0.05931763973867054, "grad_norm": 2.671875, "learning_rate": 0.00019725862931465734, "loss": 4.9624, "step": 143 }, { "epoch": 0.05973244840817173, "grad_norm": 3.171875, "learning_rate": 0.00019723861930965483, "loss": 5.0382, "step": 144 }, { "epoch": 0.060147257077672926, "grad_norm": 2.71875, "learning_rate": 0.00019721860930465234, "loss": 4.7497, "step": 145 }, { "epoch": 0.06056206574717411, "grad_norm": 2.3125, "learning_rate": 0.00019719859929964983, "loss": 5.0408, "step": 146 }, { "epoch": 0.060976874416675306, "grad_norm": 2.375, "learning_rate": 0.00019717858929464734, "loss": 4.8933, "step": 147 }, { "epoch": 0.0613916830861765, "grad_norm": 2.328125, "learning_rate": 0.00019715857928964483, "loss": 4.7968, "step": 148 }, { "epoch": 0.06180649175567769, "grad_norm": 2.46875, "learning_rate": 0.00019713856928464234, "loss": 4.8065, "step": 149 }, { "epoch": 0.062221300425178885, "grad_norm": 2.765625, "learning_rate": 0.00019711855927963983, "loss": 4.8984, "step": 150 }, { "epoch": 0.06263610909468008, "grad_norm": 2.3125, "learning_rate": 0.00019709854927463735, "loss": 4.8585, "step": 151 }, { "epoch": 0.06305091776418127, "grad_norm": 2.546875, "learning_rate": 0.00019707853926963483, "loss": 4.6844, "step": 152 }, { "epoch": 0.06346572643368247, "grad_norm": 2.890625, "learning_rate": 0.00019705852926463232, "loss": 4.7681, "step": 153 }, { "epoch": 0.06388053510318366, "grad_norm": 2.953125, "learning_rate": 0.0001970385192596298, "loss": 4.9683, "step": 154 }, { "epoch": 0.06429534377268485, "grad_norm": 2.609375, "learning_rate": 0.00019701850925462732, "loss": 4.6499, "step": 155 }, { "epoch": 0.06471015244218604, "grad_norm": 3.046875, "learning_rate": 0.0001969984992496248, "loss": 4.8674, "step": 156 }, { "epoch": 0.06512496111168724, "grad_norm": 2.53125, "learning_rate": 0.00019697848924462232, "loss": 4.8091, "step": 157 }, { "epoch": 0.06553976978118843, "grad_norm": 2.421875, "learning_rate": 0.0001969584792396198, "loss": 4.7716, "step": 158 }, { "epoch": 0.06595457845068962, "grad_norm": 3.046875, "learning_rate": 0.00019693846923461733, "loss": 4.9851, "step": 159 }, { "epoch": 0.06636938712019082, "grad_norm": 2.5625, "learning_rate": 0.00019691845922961484, "loss": 4.7429, "step": 160 }, { "epoch": 0.06678419578969201, "grad_norm": 2.703125, "learning_rate": 0.00019689844922461233, "loss": 4.7494, "step": 161 }, { "epoch": 0.0671990044591932, "grad_norm": 2.234375, "learning_rate": 0.00019687843921960981, "loss": 4.6275, "step": 162 }, { "epoch": 0.06761381312869438, "grad_norm": 2.140625, "learning_rate": 0.0001968584292146073, "loss": 4.6852, "step": 163 }, { "epoch": 0.06802862179819558, "grad_norm": 2.40625, "learning_rate": 0.00019683841920960482, "loss": 4.9879, "step": 164 }, { "epoch": 0.06844343046769677, "grad_norm": 2.15625, "learning_rate": 0.0001968184092046023, "loss": 4.7699, "step": 165 }, { "epoch": 0.06885823913719796, "grad_norm": 2.546875, "learning_rate": 0.00019679839919959982, "loss": 4.719, "step": 166 }, { "epoch": 0.06927304780669916, "grad_norm": 3.0, "learning_rate": 0.0001967783891945973, "loss": 4.7444, "step": 167 }, { "epoch": 0.06968785647620035, "grad_norm": 2.921875, "learning_rate": 0.00019675837918959482, "loss": 4.5795, "step": 168 }, { "epoch": 0.07010266514570154, "grad_norm": 2.578125, "learning_rate": 0.0001967383691845923, "loss": 4.5533, "step": 169 }, { "epoch": 0.07051747381520274, "grad_norm": 2.8125, "learning_rate": 0.0001967183591795898, "loss": 4.6815, "step": 170 }, { "epoch": 0.07093228248470393, "grad_norm": 2.921875, "learning_rate": 0.00019669834917458728, "loss": 4.5554, "step": 171 }, { "epoch": 0.07134709115420512, "grad_norm": 2.921875, "learning_rate": 0.0001966783391695848, "loss": 4.5954, "step": 172 }, { "epoch": 0.07176189982370632, "grad_norm": 2.4375, "learning_rate": 0.0001966583291645823, "loss": 4.4794, "step": 173 }, { "epoch": 0.07217670849320751, "grad_norm": 2.5625, "learning_rate": 0.0001966383191595798, "loss": 4.7955, "step": 174 }, { "epoch": 0.0725915171627087, "grad_norm": 3.71875, "learning_rate": 0.0001966183091545773, "loss": 4.7153, "step": 175 }, { "epoch": 0.0730063258322099, "grad_norm": 2.546875, "learning_rate": 0.0001965982991495748, "loss": 4.9269, "step": 176 }, { "epoch": 0.07342113450171109, "grad_norm": 2.6875, "learning_rate": 0.0001965782891445723, "loss": 4.5882, "step": 177 }, { "epoch": 0.07383594317121228, "grad_norm": 2.375, "learning_rate": 0.0001965582791395698, "loss": 4.6883, "step": 178 }, { "epoch": 0.07425075184071347, "grad_norm": 2.75, "learning_rate": 0.0001965382691345673, "loss": 4.6313, "step": 179 }, { "epoch": 0.07466556051021467, "grad_norm": 2.421875, "learning_rate": 0.00019651825912956477, "loss": 4.676, "step": 180 }, { "epoch": 0.07508036917971586, "grad_norm": 2.546875, "learning_rate": 0.0001964982491245623, "loss": 4.5967, "step": 181 }, { "epoch": 0.07549517784921705, "grad_norm": 2.53125, "learning_rate": 0.00019647823911955978, "loss": 4.863, "step": 182 }, { "epoch": 0.07590998651871825, "grad_norm": 2.765625, "learning_rate": 0.0001964582291145573, "loss": 4.4148, "step": 183 }, { "epoch": 0.07632479518821943, "grad_norm": 2.453125, "learning_rate": 0.00019643821910955478, "loss": 4.8396, "step": 184 }, { "epoch": 0.07673960385772062, "grad_norm": 2.625, "learning_rate": 0.0001964182091045523, "loss": 4.741, "step": 185 }, { "epoch": 0.07715441252722181, "grad_norm": 2.890625, "learning_rate": 0.00019639819909954978, "loss": 4.6253, "step": 186 }, { "epoch": 0.077569221196723, "grad_norm": 2.90625, "learning_rate": 0.0001963781890945473, "loss": 4.6309, "step": 187 }, { "epoch": 0.0779840298662242, "grad_norm": 2.640625, "learning_rate": 0.00019635817908954478, "loss": 4.8369, "step": 188 }, { "epoch": 0.0783988385357254, "grad_norm": 2.59375, "learning_rate": 0.00019633816908454227, "loss": 4.8035, "step": 189 }, { "epoch": 0.07881364720522659, "grad_norm": 2.75, "learning_rate": 0.00019631815907953978, "loss": 4.8848, "step": 190 }, { "epoch": 0.07922845587472778, "grad_norm": 2.421875, "learning_rate": 0.00019629814907453727, "loss": 4.8717, "step": 191 }, { "epoch": 0.07964326454422897, "grad_norm": 2.84375, "learning_rate": 0.00019627813906953478, "loss": 5.0118, "step": 192 }, { "epoch": 0.08005807321373017, "grad_norm": 2.6875, "learning_rate": 0.00019625812906453227, "loss": 4.7332, "step": 193 }, { "epoch": 0.08047288188323136, "grad_norm": 2.8125, "learning_rate": 0.0001962381190595298, "loss": 4.6317, "step": 194 }, { "epoch": 0.08088769055273255, "grad_norm": 2.109375, "learning_rate": 0.00019621810905452727, "loss": 4.6825, "step": 195 }, { "epoch": 0.08130249922223375, "grad_norm": 2.703125, "learning_rate": 0.0001961980990495248, "loss": 4.5926, "step": 196 }, { "epoch": 0.08171730789173494, "grad_norm": 2.546875, "learning_rate": 0.00019617808904452225, "loss": 4.6585, "step": 197 }, { "epoch": 0.08213211656123613, "grad_norm": 2.609375, "learning_rate": 0.00019615807903951976, "loss": 4.7669, "step": 198 }, { "epoch": 0.08254692523073733, "grad_norm": 2.703125, "learning_rate": 0.00019613806903451725, "loss": 4.7337, "step": 199 }, { "epoch": 0.08296173390023852, "grad_norm": 2.375, "learning_rate": 0.00019611805902951476, "loss": 4.9821, "step": 200 }, { "epoch": 0.08337654256973971, "grad_norm": 2.75, "learning_rate": 0.00019609804902451228, "loss": 4.6123, "step": 201 }, { "epoch": 0.0837913512392409, "grad_norm": 3.140625, "learning_rate": 0.00019607803901950977, "loss": 4.9772, "step": 202 }, { "epoch": 0.0842061599087421, "grad_norm": 3.546875, "learning_rate": 0.00019605802901450728, "loss": 4.5804, "step": 203 }, { "epoch": 0.08462096857824329, "grad_norm": 2.875, "learning_rate": 0.00019603801900950477, "loss": 4.7674, "step": 204 }, { "epoch": 0.08503577724774448, "grad_norm": 2.515625, "learning_rate": 0.00019601800900450226, "loss": 4.8421, "step": 205 }, { "epoch": 0.08545058591724566, "grad_norm": 2.484375, "learning_rate": 0.00019599799899949974, "loss": 4.6861, "step": 206 }, { "epoch": 0.08586539458674686, "grad_norm": 2.578125, "learning_rate": 0.00019597798899449726, "loss": 4.5915, "step": 207 }, { "epoch": 0.08628020325624805, "grad_norm": 2.9375, "learning_rate": 0.00019595797898949474, "loss": 4.7162, "step": 208 }, { "epoch": 0.08669501192574924, "grad_norm": 2.46875, "learning_rate": 0.00019593796898449226, "loss": 4.8773, "step": 209 }, { "epoch": 0.08710982059525044, "grad_norm": 2.40625, "learning_rate": 0.00019591795897948975, "loss": 4.6635, "step": 210 }, { "epoch": 0.08752462926475163, "grad_norm": 2.75, "learning_rate": 0.00019589794897448726, "loss": 4.8939, "step": 211 }, { "epoch": 0.08793943793425282, "grad_norm": 2.53125, "learning_rate": 0.00019587793896948475, "loss": 4.8248, "step": 212 }, { "epoch": 0.08835424660375402, "grad_norm": 2.609375, "learning_rate": 0.00019585792896448226, "loss": 4.5695, "step": 213 }, { "epoch": 0.08876905527325521, "grad_norm": 2.515625, "learning_rate": 0.00019583791895947975, "loss": 4.7652, "step": 214 }, { "epoch": 0.0891838639427564, "grad_norm": 2.359375, "learning_rate": 0.00019581790895447724, "loss": 4.6485, "step": 215 }, { "epoch": 0.0895986726122576, "grad_norm": 2.171875, "learning_rate": 0.00019579789894947475, "loss": 4.6893, "step": 216 }, { "epoch": 0.09001348128175879, "grad_norm": 2.515625, "learning_rate": 0.00019577788894447224, "loss": 4.7114, "step": 217 }, { "epoch": 0.09042828995125998, "grad_norm": 2.296875, "learning_rate": 0.00019575787893946975, "loss": 4.6796, "step": 218 }, { "epoch": 0.09084309862076118, "grad_norm": 2.359375, "learning_rate": 0.00019573786893446724, "loss": 4.7153, "step": 219 }, { "epoch": 0.09125790729026237, "grad_norm": 2.6875, "learning_rate": 0.00019571785892946475, "loss": 4.5696, "step": 220 }, { "epoch": 0.09167271595976356, "grad_norm": 2.453125, "learning_rate": 0.00019569784892446224, "loss": 5.0882, "step": 221 }, { "epoch": 0.09208752462926476, "grad_norm": 2.390625, "learning_rate": 0.00019567783891945976, "loss": 4.7033, "step": 222 }, { "epoch": 0.09250233329876595, "grad_norm": 2.609375, "learning_rate": 0.00019565782891445724, "loss": 4.813, "step": 223 }, { "epoch": 0.09291714196826714, "grad_norm": 2.234375, "learning_rate": 0.00019563781890945473, "loss": 4.4582, "step": 224 }, { "epoch": 0.09333195063776833, "grad_norm": 2.234375, "learning_rate": 0.00019561780890445222, "loss": 4.7579, "step": 225 }, { "epoch": 0.09374675930726953, "grad_norm": 2.40625, "learning_rate": 0.00019559779889944973, "loss": 4.6935, "step": 226 }, { "epoch": 0.09416156797677072, "grad_norm": 2.46875, "learning_rate": 0.00019557778889444722, "loss": 4.7105, "step": 227 }, { "epoch": 0.0945763766462719, "grad_norm": 2.546875, "learning_rate": 0.00019555777888944473, "loss": 4.5942, "step": 228 }, { "epoch": 0.0949911853157731, "grad_norm": 2.5, "learning_rate": 0.00019553776888444225, "loss": 4.6961, "step": 229 }, { "epoch": 0.09540599398527429, "grad_norm": 2.53125, "learning_rate": 0.00019551775887943974, "loss": 4.9044, "step": 230 }, { "epoch": 0.09582080265477548, "grad_norm": 2.3125, "learning_rate": 0.00019549774887443725, "loss": 4.7508, "step": 231 }, { "epoch": 0.09623561132427667, "grad_norm": 2.203125, "learning_rate": 0.0001954777388694347, "loss": 4.7008, "step": 232 }, { "epoch": 0.09665041999377787, "grad_norm": 2.3125, "learning_rate": 0.00019545772886443222, "loss": 4.9721, "step": 233 }, { "epoch": 0.09706522866327906, "grad_norm": 2.53125, "learning_rate": 0.0001954377188594297, "loss": 4.54, "step": 234 }, { "epoch": 0.09748003733278025, "grad_norm": 2.265625, "learning_rate": 0.00019541770885442723, "loss": 4.8986, "step": 235 }, { "epoch": 0.09789484600228145, "grad_norm": 2.421875, "learning_rate": 0.0001953976988494247, "loss": 4.3516, "step": 236 }, { "epoch": 0.09830965467178264, "grad_norm": 2.890625, "learning_rate": 0.00019537768884442223, "loss": 4.6487, "step": 237 }, { "epoch": 0.09872446334128383, "grad_norm": 2.234375, "learning_rate": 0.00019535767883941971, "loss": 4.7861, "step": 238 }, { "epoch": 0.09913927201078503, "grad_norm": 2.171875, "learning_rate": 0.00019533766883441723, "loss": 4.4747, "step": 239 }, { "epoch": 0.09955408068028622, "grad_norm": 2.640625, "learning_rate": 0.00019531765882941472, "loss": 4.6611, "step": 240 }, { "epoch": 0.09996888934978741, "grad_norm": 2.359375, "learning_rate": 0.0001952976488244122, "loss": 4.8161, "step": 241 }, { "epoch": 0.1003836980192886, "grad_norm": 2.34375, "learning_rate": 0.00019527763881940972, "loss": 4.6394, "step": 242 }, { "epoch": 0.1007985066887898, "grad_norm": 2.421875, "learning_rate": 0.0001952576288144072, "loss": 4.5768, "step": 243 }, { "epoch": 0.10121331535829099, "grad_norm": 2.484375, "learning_rate": 0.00019523761880940472, "loss": 4.5237, "step": 244 }, { "epoch": 0.10162812402779219, "grad_norm": 2.171875, "learning_rate": 0.0001952176088044022, "loss": 4.5366, "step": 245 }, { "epoch": 0.10204293269729338, "grad_norm": 2.859375, "learning_rate": 0.00019519759879939972, "loss": 4.647, "step": 246 }, { "epoch": 0.10245774136679457, "grad_norm": 2.078125, "learning_rate": 0.0001951775887943972, "loss": 4.5444, "step": 247 }, { "epoch": 0.10287255003629577, "grad_norm": 2.125, "learning_rate": 0.00019515757878939472, "loss": 4.5602, "step": 248 }, { "epoch": 0.10328735870579694, "grad_norm": 2.25, "learning_rate": 0.0001951375687843922, "loss": 4.5301, "step": 249 }, { "epoch": 0.10370216737529814, "grad_norm": 3.046875, "learning_rate": 0.0001951175587793897, "loss": 4.691, "step": 250 }, { "epoch": 0.10411697604479933, "grad_norm": 2.375, "learning_rate": 0.00019509754877438718, "loss": 4.6174, "step": 251 }, { "epoch": 0.10453178471430052, "grad_norm": 2.15625, "learning_rate": 0.0001950775387693847, "loss": 4.7134, "step": 252 }, { "epoch": 0.10494659338380172, "grad_norm": 2.453125, "learning_rate": 0.0001950575287643822, "loss": 4.8516, "step": 253 }, { "epoch": 0.10536140205330291, "grad_norm": 2.828125, "learning_rate": 0.0001950375187593797, "loss": 4.6675, "step": 254 }, { "epoch": 0.1057762107228041, "grad_norm": 2.53125, "learning_rate": 0.0001950175087543772, "loss": 4.7508, "step": 255 }, { "epoch": 0.1061910193923053, "grad_norm": 2.65625, "learning_rate": 0.0001949974987493747, "loss": 4.6769, "step": 256 }, { "epoch": 0.10660582806180649, "grad_norm": 2.3125, "learning_rate": 0.00019497748874437222, "loss": 4.3861, "step": 257 }, { "epoch": 0.10702063673130768, "grad_norm": 2.34375, "learning_rate": 0.0001949574787393697, "loss": 4.6051, "step": 258 }, { "epoch": 0.10743544540080888, "grad_norm": 2.265625, "learning_rate": 0.0001949374687343672, "loss": 4.5368, "step": 259 }, { "epoch": 0.10785025407031007, "grad_norm": 2.546875, "learning_rate": 0.00019491745872936468, "loss": 4.7008, "step": 260 }, { "epoch": 0.10826506273981126, "grad_norm": 2.21875, "learning_rate": 0.0001948974487243622, "loss": 4.6742, "step": 261 }, { "epoch": 0.10867987140931246, "grad_norm": 2.453125, "learning_rate": 0.00019487743871935968, "loss": 4.3499, "step": 262 }, { "epoch": 0.10909468007881365, "grad_norm": 2.390625, "learning_rate": 0.0001948574287143572, "loss": 4.6754, "step": 263 }, { "epoch": 0.10950948874831484, "grad_norm": 2.296875, "learning_rate": 0.00019483741870935468, "loss": 4.6755, "step": 264 }, { "epoch": 0.10992429741781604, "grad_norm": 2.46875, "learning_rate": 0.0001948174087043522, "loss": 4.6689, "step": 265 }, { "epoch": 0.11033910608731723, "grad_norm": 2.109375, "learning_rate": 0.00019479739869934968, "loss": 4.6618, "step": 266 }, { "epoch": 0.11075391475681842, "grad_norm": 2.53125, "learning_rate": 0.0001947773886943472, "loss": 4.7117, "step": 267 }, { "epoch": 0.11116872342631962, "grad_norm": 2.21875, "learning_rate": 0.00019475737868934466, "loss": 4.6274, "step": 268 }, { "epoch": 0.11158353209582081, "grad_norm": 2.46875, "learning_rate": 0.00019473736868434217, "loss": 4.6007, "step": 269 }, { "epoch": 0.111998340765322, "grad_norm": 2.53125, "learning_rate": 0.0001947173586793397, "loss": 4.8718, "step": 270 }, { "epoch": 0.11241314943482318, "grad_norm": 2.734375, "learning_rate": 0.00019469734867433717, "loss": 4.7442, "step": 271 }, { "epoch": 0.11282795810432437, "grad_norm": 2.390625, "learning_rate": 0.0001946773386693347, "loss": 4.6158, "step": 272 }, { "epoch": 0.11324276677382557, "grad_norm": 2.453125, "learning_rate": 0.00019465732866433218, "loss": 4.7251, "step": 273 }, { "epoch": 0.11365757544332676, "grad_norm": 2.453125, "learning_rate": 0.0001946373186593297, "loss": 4.9831, "step": 274 }, { "epoch": 0.11407238411282795, "grad_norm": 2.421875, "learning_rate": 0.00019461730865432718, "loss": 4.6253, "step": 275 }, { "epoch": 0.11448719278232915, "grad_norm": 2.125, "learning_rate": 0.00019459729864932467, "loss": 4.5307, "step": 276 }, { "epoch": 0.11490200145183034, "grad_norm": 2.546875, "learning_rate": 0.00019457728864432215, "loss": 4.7805, "step": 277 }, { "epoch": 0.11531681012133153, "grad_norm": 2.265625, "learning_rate": 0.00019455727863931967, "loss": 4.5133, "step": 278 }, { "epoch": 0.11573161879083273, "grad_norm": 2.5, "learning_rate": 0.00019453726863431715, "loss": 4.4652, "step": 279 }, { "epoch": 0.11614642746033392, "grad_norm": 2.5625, "learning_rate": 0.00019451725862931467, "loss": 4.7797, "step": 280 }, { "epoch": 0.11656123612983511, "grad_norm": 2.546875, "learning_rate": 0.00019449724862431216, "loss": 4.6143, "step": 281 }, { "epoch": 0.1169760447993363, "grad_norm": 2.53125, "learning_rate": 0.00019447723861930967, "loss": 4.7538, "step": 282 }, { "epoch": 0.1173908534688375, "grad_norm": 2.53125, "learning_rate": 0.00019445722861430716, "loss": 4.6332, "step": 283 }, { "epoch": 0.1178056621383387, "grad_norm": 2.484375, "learning_rate": 0.00019443721860930467, "loss": 4.5163, "step": 284 }, { "epoch": 0.11822047080783989, "grad_norm": 2.34375, "learning_rate": 0.00019441720860430216, "loss": 4.4752, "step": 285 }, { "epoch": 0.11863527947734108, "grad_norm": 2.3125, "learning_rate": 0.00019439719859929965, "loss": 4.5883, "step": 286 }, { "epoch": 0.11905008814684227, "grad_norm": 2.40625, "learning_rate": 0.00019437718859429716, "loss": 4.723, "step": 287 }, { "epoch": 0.11946489681634347, "grad_norm": 2.8125, "learning_rate": 0.00019435717858929465, "loss": 4.8239, "step": 288 }, { "epoch": 0.11987970548584466, "grad_norm": 2.765625, "learning_rate": 0.00019433716858429216, "loss": 4.7219, "step": 289 }, { "epoch": 0.12029451415534585, "grad_norm": 2.515625, "learning_rate": 0.00019431715857928965, "loss": 4.9378, "step": 290 }, { "epoch": 0.12070932282484705, "grad_norm": 2.359375, "learning_rate": 0.00019429714857428716, "loss": 4.4392, "step": 291 }, { "epoch": 0.12112413149434823, "grad_norm": 2.390625, "learning_rate": 0.00019427713856928465, "loss": 5.0761, "step": 292 }, { "epoch": 0.12153894016384942, "grad_norm": 2.484375, "learning_rate": 0.00019425712856428217, "loss": 4.5624, "step": 293 }, { "epoch": 0.12195374883335061, "grad_norm": 2.453125, "learning_rate": 0.00019423711855927965, "loss": 4.889, "step": 294 }, { "epoch": 0.1223685575028518, "grad_norm": 2.28125, "learning_rate": 0.00019421710855427714, "loss": 4.7606, "step": 295 }, { "epoch": 0.122783366172353, "grad_norm": 2.359375, "learning_rate": 0.00019419709854927463, "loss": 4.4913, "step": 296 }, { "epoch": 0.12319817484185419, "grad_norm": 2.40625, "learning_rate": 0.00019417708854427214, "loss": 4.6842, "step": 297 }, { "epoch": 0.12361298351135538, "grad_norm": 2.4375, "learning_rate": 0.00019415707853926966, "loss": 4.5627, "step": 298 }, { "epoch": 0.12402779218085658, "grad_norm": 2.28125, "learning_rate": 0.00019413706853426714, "loss": 4.68, "step": 299 }, { "epoch": 0.12444260085035777, "grad_norm": 2.703125, "learning_rate": 0.00019411705852926466, "loss": 4.5748, "step": 300 }, { "epoch": 0.12485740951985896, "grad_norm": 2.953125, "learning_rate": 0.00019409704852426215, "loss": 4.5055, "step": 301 }, { "epoch": 0.12527221818936016, "grad_norm": 2.515625, "learning_rate": 0.00019407703851925966, "loss": 4.6107, "step": 302 }, { "epoch": 0.12568702685886135, "grad_norm": 2.984375, "learning_rate": 0.00019405702851425712, "loss": 4.531, "step": 303 }, { "epoch": 0.12610183552836254, "grad_norm": 2.28125, "learning_rate": 0.00019403701850925463, "loss": 4.6775, "step": 304 }, { "epoch": 0.12651664419786374, "grad_norm": 2.859375, "learning_rate": 0.00019401700850425212, "loss": 4.6577, "step": 305 }, { "epoch": 0.12693145286736493, "grad_norm": 2.484375, "learning_rate": 0.00019399699849924964, "loss": 4.5784, "step": 306 }, { "epoch": 0.12734626153686612, "grad_norm": 3.28125, "learning_rate": 0.00019397698849424712, "loss": 4.6397, "step": 307 }, { "epoch": 0.12776107020636732, "grad_norm": 2.359375, "learning_rate": 0.00019395697848924464, "loss": 4.7788, "step": 308 }, { "epoch": 0.1281758788758685, "grad_norm": 2.546875, "learning_rate": 0.00019393696848424212, "loss": 5.0678, "step": 309 }, { "epoch": 0.1285906875453697, "grad_norm": 2.328125, "learning_rate": 0.00019391695847923964, "loss": 4.4743, "step": 310 }, { "epoch": 0.1290054962148709, "grad_norm": 2.84375, "learning_rate": 0.00019389694847423713, "loss": 4.6317, "step": 311 }, { "epoch": 0.1294203048843721, "grad_norm": 2.3125, "learning_rate": 0.00019387693846923461, "loss": 4.6397, "step": 312 }, { "epoch": 0.12983511355387328, "grad_norm": 2.40625, "learning_rate": 0.00019385692846423213, "loss": 4.5141, "step": 313 }, { "epoch": 0.13024992222337448, "grad_norm": 2.390625, "learning_rate": 0.00019383691845922962, "loss": 4.6465, "step": 314 }, { "epoch": 0.13066473089287567, "grad_norm": 2.390625, "learning_rate": 0.00019381690845422713, "loss": 4.8338, "step": 315 }, { "epoch": 0.13107953956237686, "grad_norm": 2.421875, "learning_rate": 0.00019379689844922462, "loss": 4.6736, "step": 316 }, { "epoch": 0.13149434823187806, "grad_norm": 2.859375, "learning_rate": 0.00019377688844422213, "loss": 4.7102, "step": 317 }, { "epoch": 0.13190915690137925, "grad_norm": 2.46875, "learning_rate": 0.00019375687843921962, "loss": 4.6496, "step": 318 }, { "epoch": 0.13232396557088044, "grad_norm": 2.140625, "learning_rate": 0.00019373686843421713, "loss": 4.8149, "step": 319 }, { "epoch": 0.13273877424038164, "grad_norm": 2.21875, "learning_rate": 0.00019371685842921462, "loss": 4.5535, "step": 320 }, { "epoch": 0.13315358290988283, "grad_norm": 2.78125, "learning_rate": 0.0001936968484242121, "loss": 4.5018, "step": 321 }, { "epoch": 0.13356839157938402, "grad_norm": 2.484375, "learning_rate": 0.0001936768384192096, "loss": 4.6553, "step": 322 }, { "epoch": 0.13398320024888521, "grad_norm": 2.390625, "learning_rate": 0.0001936568284142071, "loss": 4.8589, "step": 323 }, { "epoch": 0.1343980089183864, "grad_norm": 2.515625, "learning_rate": 0.0001936368184092046, "loss": 4.8949, "step": 324 }, { "epoch": 0.13481281758788757, "grad_norm": 2.3125, "learning_rate": 0.0001936168084042021, "loss": 4.5917, "step": 325 }, { "epoch": 0.13522762625738877, "grad_norm": 2.765625, "learning_rate": 0.00019359679839919963, "loss": 4.5207, "step": 326 }, { "epoch": 0.13564243492688996, "grad_norm": 3.015625, "learning_rate": 0.0001935767883941971, "loss": 4.6245, "step": 327 }, { "epoch": 0.13605724359639115, "grad_norm": 2.265625, "learning_rate": 0.00019355677838919463, "loss": 4.5721, "step": 328 }, { "epoch": 0.13647205226589235, "grad_norm": 2.171875, "learning_rate": 0.00019353676838419211, "loss": 4.5226, "step": 329 }, { "epoch": 0.13688686093539354, "grad_norm": 2.203125, "learning_rate": 0.0001935167583791896, "loss": 4.4595, "step": 330 }, { "epoch": 0.13730166960489473, "grad_norm": 2.328125, "learning_rate": 0.0001934967483741871, "loss": 4.941, "step": 331 }, { "epoch": 0.13771647827439593, "grad_norm": 2.296875, "learning_rate": 0.0001934767383691846, "loss": 4.523, "step": 332 }, { "epoch": 0.13813128694389712, "grad_norm": 2.234375, "learning_rate": 0.0001934567283641821, "loss": 4.5625, "step": 333 }, { "epoch": 0.1385460956133983, "grad_norm": 3.140625, "learning_rate": 0.0001934367183591796, "loss": 4.7129, "step": 334 }, { "epoch": 0.1389609042828995, "grad_norm": 2.609375, "learning_rate": 0.0001934167083541771, "loss": 4.5963, "step": 335 }, { "epoch": 0.1393757129524007, "grad_norm": 2.5, "learning_rate": 0.0001933966983491746, "loss": 4.4981, "step": 336 }, { "epoch": 0.1397905216219019, "grad_norm": 2.453125, "learning_rate": 0.0001933766883441721, "loss": 4.4765, "step": 337 }, { "epoch": 0.14020533029140309, "grad_norm": 2.28125, "learning_rate": 0.00019335667833916958, "loss": 4.7399, "step": 338 }, { "epoch": 0.14062013896090428, "grad_norm": 2.59375, "learning_rate": 0.0001933366683341671, "loss": 4.5903, "step": 339 }, { "epoch": 0.14103494763040547, "grad_norm": 3.40625, "learning_rate": 0.00019331665832916458, "loss": 4.7435, "step": 340 }, { "epoch": 0.14144975629990666, "grad_norm": 2.734375, "learning_rate": 0.0001932966483241621, "loss": 4.7956, "step": 341 }, { "epoch": 0.14186456496940786, "grad_norm": 2.390625, "learning_rate": 0.00019327663831915958, "loss": 4.7103, "step": 342 }, { "epoch": 0.14227937363890905, "grad_norm": 2.4375, "learning_rate": 0.0001932566283141571, "loss": 4.6113, "step": 343 }, { "epoch": 0.14269418230841024, "grad_norm": 2.46875, "learning_rate": 0.00019323661830915459, "loss": 4.7041, "step": 344 }, { "epoch": 0.14310899097791144, "grad_norm": 2.390625, "learning_rate": 0.0001932166083041521, "loss": 4.6125, "step": 345 }, { "epoch": 0.14352379964741263, "grad_norm": 2.3125, "learning_rate": 0.0001931965982991496, "loss": 4.4779, "step": 346 }, { "epoch": 0.14393860831691382, "grad_norm": 2.75, "learning_rate": 0.00019317658829414708, "loss": 4.4845, "step": 347 }, { "epoch": 0.14435341698641502, "grad_norm": 2.703125, "learning_rate": 0.00019315657828914456, "loss": 4.5883, "step": 348 }, { "epoch": 0.1447682256559162, "grad_norm": 2.1875, "learning_rate": 0.00019313656828414208, "loss": 4.6283, "step": 349 }, { "epoch": 0.1451830343254174, "grad_norm": 2.28125, "learning_rate": 0.00019311655827913956, "loss": 4.6821, "step": 350 }, { "epoch": 0.1455978429949186, "grad_norm": 2.5, "learning_rate": 0.00019309654827413708, "loss": 4.2464, "step": 351 }, { "epoch": 0.1460126516644198, "grad_norm": 2.484375, "learning_rate": 0.00019307653826913457, "loss": 4.4201, "step": 352 }, { "epoch": 0.14642746033392098, "grad_norm": 2.984375, "learning_rate": 0.00019305652826413208, "loss": 4.8415, "step": 353 }, { "epoch": 0.14684226900342218, "grad_norm": 2.53125, "learning_rate": 0.0001930365182591296, "loss": 4.612, "step": 354 }, { "epoch": 0.14725707767292337, "grad_norm": 2.25, "learning_rate": 0.00019301650825412708, "loss": 4.7386, "step": 355 }, { "epoch": 0.14767188634242456, "grad_norm": 2.171875, "learning_rate": 0.00019299649824912457, "loss": 4.6206, "step": 356 }, { "epoch": 0.14808669501192576, "grad_norm": 2.28125, "learning_rate": 0.00019297648824412206, "loss": 4.7842, "step": 357 }, { "epoch": 0.14850150368142695, "grad_norm": 2.3125, "learning_rate": 0.00019295647823911957, "loss": 4.4395, "step": 358 }, { "epoch": 0.14891631235092814, "grad_norm": 2.09375, "learning_rate": 0.00019293646823411706, "loss": 4.8042, "step": 359 }, { "epoch": 0.14933112102042934, "grad_norm": 2.09375, "learning_rate": 0.00019291645822911457, "loss": 4.5574, "step": 360 }, { "epoch": 0.14974592968993053, "grad_norm": 2.234375, "learning_rate": 0.00019289644822411206, "loss": 4.5097, "step": 361 }, { "epoch": 0.15016073835943172, "grad_norm": 2.4375, "learning_rate": 0.00019287643821910957, "loss": 4.6248, "step": 362 }, { "epoch": 0.15057554702893292, "grad_norm": 2.34375, "learning_rate": 0.00019285642821410706, "loss": 4.5815, "step": 363 }, { "epoch": 0.1509903556984341, "grad_norm": 2.3125, "learning_rate": 0.00019283641820910458, "loss": 4.6993, "step": 364 }, { "epoch": 0.1514051643679353, "grad_norm": 2.6875, "learning_rate": 0.00019281640820410206, "loss": 4.9004, "step": 365 }, { "epoch": 0.1518199730374365, "grad_norm": 2.40625, "learning_rate": 0.00019279639819909955, "loss": 4.6554, "step": 366 }, { "epoch": 0.1522347817069377, "grad_norm": 2.15625, "learning_rate": 0.00019277638819409706, "loss": 4.3681, "step": 367 }, { "epoch": 0.15264959037643885, "grad_norm": 2.328125, "learning_rate": 0.00019275637818909455, "loss": 4.3656, "step": 368 }, { "epoch": 0.15306439904594005, "grad_norm": 2.40625, "learning_rate": 0.00019273636818409207, "loss": 4.5683, "step": 369 }, { "epoch": 0.15347920771544124, "grad_norm": 2.265625, "learning_rate": 0.00019271635817908955, "loss": 4.7063, "step": 370 }, { "epoch": 0.15389401638494243, "grad_norm": 2.515625, "learning_rate": 0.00019269634817408707, "loss": 4.7527, "step": 371 }, { "epoch": 0.15430882505444363, "grad_norm": 2.328125, "learning_rate": 0.00019267633816908456, "loss": 4.6669, "step": 372 }, { "epoch": 0.15472363372394482, "grad_norm": 2.1875, "learning_rate": 0.00019265632816408204, "loss": 4.5687, "step": 373 }, { "epoch": 0.155138442393446, "grad_norm": 2.28125, "learning_rate": 0.00019263631815907953, "loss": 4.8767, "step": 374 }, { "epoch": 0.1555532510629472, "grad_norm": 2.21875, "learning_rate": 0.00019261630815407704, "loss": 4.5858, "step": 375 }, { "epoch": 0.1559680597324484, "grad_norm": 2.09375, "learning_rate": 0.00019259629814907453, "loss": 4.4694, "step": 376 }, { "epoch": 0.1563828684019496, "grad_norm": 2.5, "learning_rate": 0.00019257628814407205, "loss": 4.4371, "step": 377 }, { "epoch": 0.1567976770714508, "grad_norm": 2.171875, "learning_rate": 0.00019255627813906953, "loss": 4.4846, "step": 378 }, { "epoch": 0.15721248574095198, "grad_norm": 2.328125, "learning_rate": 0.00019253626813406705, "loss": 4.7829, "step": 379 }, { "epoch": 0.15762729441045317, "grad_norm": 2.234375, "learning_rate": 0.00019251625812906453, "loss": 4.7909, "step": 380 }, { "epoch": 0.15804210307995437, "grad_norm": 2.296875, "learning_rate": 0.00019249624812406205, "loss": 4.349, "step": 381 }, { "epoch": 0.15845691174945556, "grad_norm": 2.375, "learning_rate": 0.00019247623811905954, "loss": 4.518, "step": 382 }, { "epoch": 0.15887172041895675, "grad_norm": 2.296875, "learning_rate": 0.00019245622811405702, "loss": 4.7616, "step": 383 }, { "epoch": 0.15928652908845795, "grad_norm": 2.796875, "learning_rate": 0.00019243621810905454, "loss": 4.5117, "step": 384 }, { "epoch": 0.15970133775795914, "grad_norm": 2.28125, "learning_rate": 0.00019241620810405203, "loss": 4.5693, "step": 385 }, { "epoch": 0.16011614642746033, "grad_norm": 2.234375, "learning_rate": 0.00019239619809904954, "loss": 4.748, "step": 386 }, { "epoch": 0.16053095509696153, "grad_norm": 2.3125, "learning_rate": 0.00019237618809404703, "loss": 4.6349, "step": 387 }, { "epoch": 0.16094576376646272, "grad_norm": 2.359375, "learning_rate": 0.00019235617808904454, "loss": 4.6104, "step": 388 }, { "epoch": 0.1613605724359639, "grad_norm": 2.71875, "learning_rate": 0.00019233616808404203, "loss": 4.6278, "step": 389 }, { "epoch": 0.1617753811054651, "grad_norm": 2.21875, "learning_rate": 0.00019231615807903954, "loss": 4.6366, "step": 390 }, { "epoch": 0.1621901897749663, "grad_norm": 2.390625, "learning_rate": 0.00019229614807403703, "loss": 4.6784, "step": 391 }, { "epoch": 0.1626049984444675, "grad_norm": 2.3125, "learning_rate": 0.00019227613806903452, "loss": 4.5775, "step": 392 }, { "epoch": 0.16301980711396868, "grad_norm": 2.5, "learning_rate": 0.000192256128064032, "loss": 4.7694, "step": 393 }, { "epoch": 0.16343461578346988, "grad_norm": 2.4375, "learning_rate": 0.00019223611805902952, "loss": 4.3015, "step": 394 }, { "epoch": 0.16384942445297107, "grad_norm": 2.25, "learning_rate": 0.00019221610805402703, "loss": 4.5376, "step": 395 }, { "epoch": 0.16426423312247226, "grad_norm": 2.578125, "learning_rate": 0.00019219609804902452, "loss": 4.5119, "step": 396 }, { "epoch": 0.16467904179197346, "grad_norm": 2.5625, "learning_rate": 0.00019217608804402204, "loss": 4.6437, "step": 397 }, { "epoch": 0.16509385046147465, "grad_norm": 2.765625, "learning_rate": 0.00019215607803901952, "loss": 4.7193, "step": 398 }, { "epoch": 0.16550865913097584, "grad_norm": 2.0625, "learning_rate": 0.00019213606803401704, "loss": 4.5158, "step": 399 }, { "epoch": 0.16592346780047704, "grad_norm": 2.34375, "learning_rate": 0.00019211605802901452, "loss": 4.9949, "step": 400 }, { "epoch": 0.16633827646997823, "grad_norm": 2.15625, "learning_rate": 0.000192096048024012, "loss": 4.7789, "step": 401 }, { "epoch": 0.16675308513947942, "grad_norm": 2.515625, "learning_rate": 0.0001920760380190095, "loss": 4.6586, "step": 402 }, { "epoch": 0.16716789380898062, "grad_norm": 2.5, "learning_rate": 0.000192056028014007, "loss": 4.5197, "step": 403 }, { "epoch": 0.1675827024784818, "grad_norm": 2.484375, "learning_rate": 0.0001920360180090045, "loss": 4.6449, "step": 404 }, { "epoch": 0.167997511147983, "grad_norm": 2.125, "learning_rate": 0.00019201600800400202, "loss": 4.6291, "step": 405 }, { "epoch": 0.1684123198174842, "grad_norm": 2.171875, "learning_rate": 0.0001919959979989995, "loss": 4.7104, "step": 406 }, { "epoch": 0.1688271284869854, "grad_norm": 2.0625, "learning_rate": 0.00019197598799399702, "loss": 4.7399, "step": 407 }, { "epoch": 0.16924193715648658, "grad_norm": 2.28125, "learning_rate": 0.0001919559779889945, "loss": 4.3293, "step": 408 }, { "epoch": 0.16965674582598778, "grad_norm": 2.1875, "learning_rate": 0.000191935967983992, "loss": 4.4642, "step": 409 }, { "epoch": 0.17007155449548897, "grad_norm": 2.453125, "learning_rate": 0.0001919159579789895, "loss": 4.5354, "step": 410 }, { "epoch": 0.17048636316499013, "grad_norm": 2.15625, "learning_rate": 0.000191895947973987, "loss": 4.4258, "step": 411 }, { "epoch": 0.17090117183449133, "grad_norm": 2.328125, "learning_rate": 0.0001918759379689845, "loss": 4.7604, "step": 412 }, { "epoch": 0.17131598050399252, "grad_norm": 2.6875, "learning_rate": 0.000191855927963982, "loss": 4.9124, "step": 413 }, { "epoch": 0.17173078917349371, "grad_norm": 2.0625, "learning_rate": 0.0001918359179589795, "loss": 4.8301, "step": 414 }, { "epoch": 0.1721455978429949, "grad_norm": 2.296875, "learning_rate": 0.000191815907953977, "loss": 4.7501, "step": 415 }, { "epoch": 0.1725604065124961, "grad_norm": 2.171875, "learning_rate": 0.0001917958979489745, "loss": 4.5007, "step": 416 }, { "epoch": 0.1729752151819973, "grad_norm": 2.75, "learning_rate": 0.000191775887943972, "loss": 4.7086, "step": 417 }, { "epoch": 0.1733900238514985, "grad_norm": 2.359375, "learning_rate": 0.00019175587793896949, "loss": 4.5778, "step": 418 }, { "epoch": 0.17380483252099968, "grad_norm": 2.296875, "learning_rate": 0.00019173586793396697, "loss": 4.529, "step": 419 }, { "epoch": 0.17421964119050087, "grad_norm": 2.53125, "learning_rate": 0.0001917158579289645, "loss": 4.4016, "step": 420 }, { "epoch": 0.17463444986000207, "grad_norm": 2.1875, "learning_rate": 0.00019169584792396197, "loss": 4.4806, "step": 421 }, { "epoch": 0.17504925852950326, "grad_norm": 2.453125, "learning_rate": 0.0001916758379189595, "loss": 4.6308, "step": 422 }, { "epoch": 0.17546406719900445, "grad_norm": 2.28125, "learning_rate": 0.000191655827913957, "loss": 4.5926, "step": 423 }, { "epoch": 0.17587887586850565, "grad_norm": 2.328125, "learning_rate": 0.0001916358179089545, "loss": 4.3179, "step": 424 }, { "epoch": 0.17629368453800684, "grad_norm": 2.625, "learning_rate": 0.000191615807903952, "loss": 4.5057, "step": 425 }, { "epoch": 0.17670849320750803, "grad_norm": 2.484375, "learning_rate": 0.0001915957978989495, "loss": 4.6446, "step": 426 }, { "epoch": 0.17712330187700923, "grad_norm": 2.21875, "learning_rate": 0.00019157578789394698, "loss": 4.7132, "step": 427 }, { "epoch": 0.17753811054651042, "grad_norm": 2.296875, "learning_rate": 0.00019155577788894447, "loss": 4.4011, "step": 428 }, { "epoch": 0.1779529192160116, "grad_norm": 2.3125, "learning_rate": 0.00019153576788394198, "loss": 4.7595, "step": 429 }, { "epoch": 0.1783677278855128, "grad_norm": 2.171875, "learning_rate": 0.00019151575787893947, "loss": 4.5743, "step": 430 }, { "epoch": 0.178782536555014, "grad_norm": 2.265625, "learning_rate": 0.00019149574787393698, "loss": 4.6784, "step": 431 }, { "epoch": 0.1791973452245152, "grad_norm": 2.4375, "learning_rate": 0.00019147573786893447, "loss": 4.4957, "step": 432 }, { "epoch": 0.17961215389401639, "grad_norm": 2.59375, "learning_rate": 0.00019145572786393198, "loss": 4.4072, "step": 433 }, { "epoch": 0.18002696256351758, "grad_norm": 2.4375, "learning_rate": 0.00019143571785892947, "loss": 4.6275, "step": 434 }, { "epoch": 0.18044177123301877, "grad_norm": 2.59375, "learning_rate": 0.00019141570785392699, "loss": 4.6934, "step": 435 }, { "epoch": 0.18085657990251996, "grad_norm": 2.234375, "learning_rate": 0.00019139569784892447, "loss": 4.6939, "step": 436 }, { "epoch": 0.18127138857202116, "grad_norm": 2.5, "learning_rate": 0.00019137568784392196, "loss": 4.6987, "step": 437 }, { "epoch": 0.18168619724152235, "grad_norm": 2.421875, "learning_rate": 0.00019135567783891947, "loss": 4.5695, "step": 438 }, { "epoch": 0.18210100591102354, "grad_norm": 2.171875, "learning_rate": 0.00019133566783391696, "loss": 4.7376, "step": 439 }, { "epoch": 0.18251581458052474, "grad_norm": 2.640625, "learning_rate": 0.00019131565782891448, "loss": 4.5068, "step": 440 }, { "epoch": 0.18293062325002593, "grad_norm": 2.734375, "learning_rate": 0.00019129564782391196, "loss": 4.4134, "step": 441 }, { "epoch": 0.18334543191952712, "grad_norm": 2.453125, "learning_rate": 0.00019127563781890948, "loss": 4.5897, "step": 442 }, { "epoch": 0.18376024058902832, "grad_norm": 2.15625, "learning_rate": 0.00019125562781390697, "loss": 4.5844, "step": 443 }, { "epoch": 0.1841750492585295, "grad_norm": 2.328125, "learning_rate": 0.00019123561780890445, "loss": 4.6493, "step": 444 }, { "epoch": 0.1845898579280307, "grad_norm": 2.515625, "learning_rate": 0.00019121560780390194, "loss": 4.4406, "step": 445 }, { "epoch": 0.1850046665975319, "grad_norm": 2.34375, "learning_rate": 0.00019119559779889945, "loss": 4.5867, "step": 446 }, { "epoch": 0.1854194752670331, "grad_norm": 2.453125, "learning_rate": 0.00019117558779389694, "loss": 4.7002, "step": 447 }, { "epoch": 0.18583428393653428, "grad_norm": 2.703125, "learning_rate": 0.00019115557778889446, "loss": 4.7154, "step": 448 }, { "epoch": 0.18624909260603548, "grad_norm": 2.25, "learning_rate": 0.00019113556778389194, "loss": 4.5751, "step": 449 }, { "epoch": 0.18666390127553667, "grad_norm": 2.5, "learning_rate": 0.00019111555777888946, "loss": 4.448, "step": 450 }, { "epoch": 0.18707870994503786, "grad_norm": 2.234375, "learning_rate": 0.00019109554777388697, "loss": 4.4763, "step": 451 }, { "epoch": 0.18749351861453906, "grad_norm": 2.234375, "learning_rate": 0.00019107553776888446, "loss": 4.795, "step": 452 }, { "epoch": 0.18790832728404025, "grad_norm": 2.203125, "learning_rate": 0.00019105552776388195, "loss": 4.6979, "step": 453 }, { "epoch": 0.18832313595354144, "grad_norm": 2.421875, "learning_rate": 0.00019103551775887943, "loss": 4.3613, "step": 454 }, { "epoch": 0.1887379446230426, "grad_norm": 2.203125, "learning_rate": 0.00019101550775387695, "loss": 4.4992, "step": 455 }, { "epoch": 0.1891527532925438, "grad_norm": 2.703125, "learning_rate": 0.00019099549774887444, "loss": 4.6909, "step": 456 }, { "epoch": 0.189567561962045, "grad_norm": 2.390625, "learning_rate": 0.00019097548774387195, "loss": 4.6363, "step": 457 }, { "epoch": 0.1899823706315462, "grad_norm": 2.28125, "learning_rate": 0.00019095547773886944, "loss": 4.551, "step": 458 }, { "epoch": 0.19039717930104738, "grad_norm": 2.40625, "learning_rate": 0.00019093546773386695, "loss": 4.5158, "step": 459 }, { "epoch": 0.19081198797054857, "grad_norm": 2.578125, "learning_rate": 0.00019091545772886444, "loss": 4.6661, "step": 460 }, { "epoch": 0.19122679664004977, "grad_norm": 2.3125, "learning_rate": 0.00019089544772386195, "loss": 4.2847, "step": 461 }, { "epoch": 0.19164160530955096, "grad_norm": 2.328125, "learning_rate": 0.00019087543771885944, "loss": 4.4697, "step": 462 }, { "epoch": 0.19205641397905215, "grad_norm": 2.078125, "learning_rate": 0.00019085542771385693, "loss": 4.4851, "step": 463 }, { "epoch": 0.19247122264855335, "grad_norm": 2.015625, "learning_rate": 0.00019083541770885444, "loss": 4.4844, "step": 464 }, { "epoch": 0.19288603131805454, "grad_norm": 2.203125, "learning_rate": 0.00019081540770385193, "loss": 4.7539, "step": 465 }, { "epoch": 0.19330083998755573, "grad_norm": 1.9765625, "learning_rate": 0.00019079539769884944, "loss": 4.3977, "step": 466 }, { "epoch": 0.19371564865705693, "grad_norm": 2.171875, "learning_rate": 0.00019077538769384693, "loss": 4.3886, "step": 467 }, { "epoch": 0.19413045732655812, "grad_norm": 2.40625, "learning_rate": 0.00019075537768884445, "loss": 4.7598, "step": 468 }, { "epoch": 0.1945452659960593, "grad_norm": 2.171875, "learning_rate": 0.00019073536768384193, "loss": 4.6188, "step": 469 }, { "epoch": 0.1949600746655605, "grad_norm": 2.25, "learning_rate": 0.00019071535767883945, "loss": 4.4196, "step": 470 }, { "epoch": 0.1953748833350617, "grad_norm": 2.046875, "learning_rate": 0.00019069534767383693, "loss": 4.5675, "step": 471 }, { "epoch": 0.1957896920045629, "grad_norm": 2.265625, "learning_rate": 0.00019067533766883442, "loss": 4.5032, "step": 472 }, { "epoch": 0.1962045006740641, "grad_norm": 2.46875, "learning_rate": 0.0001906553276638319, "loss": 4.5395, "step": 473 }, { "epoch": 0.19661930934356528, "grad_norm": 2.46875, "learning_rate": 0.00019063531765882942, "loss": 4.5965, "step": 474 }, { "epoch": 0.19703411801306647, "grad_norm": 2.40625, "learning_rate": 0.0001906153076538269, "loss": 4.8952, "step": 475 }, { "epoch": 0.19744892668256767, "grad_norm": 2.234375, "learning_rate": 0.00019059529764882443, "loss": 4.5312, "step": 476 }, { "epoch": 0.19786373535206886, "grad_norm": 2.046875, "learning_rate": 0.0001905752876438219, "loss": 4.576, "step": 477 }, { "epoch": 0.19827854402157005, "grad_norm": 2.421875, "learning_rate": 0.00019055527763881943, "loss": 4.3351, "step": 478 }, { "epoch": 0.19869335269107125, "grad_norm": 2.078125, "learning_rate": 0.00019053526763381691, "loss": 4.5906, "step": 479 }, { "epoch": 0.19910816136057244, "grad_norm": 2.296875, "learning_rate": 0.0001905152576288144, "loss": 4.4678, "step": 480 }, { "epoch": 0.19952297003007363, "grad_norm": 2.21875, "learning_rate": 0.00019049524762381192, "loss": 4.3641, "step": 481 }, { "epoch": 0.19993777869957483, "grad_norm": 2.328125, "learning_rate": 0.0001904752376188094, "loss": 4.4432, "step": 482 }, { "epoch": 0.20035258736907602, "grad_norm": 2.390625, "learning_rate": 0.00019045522761380692, "loss": 4.4244, "step": 483 }, { "epoch": 0.2007673960385772, "grad_norm": 2.21875, "learning_rate": 0.0001904352176088044, "loss": 4.5019, "step": 484 }, { "epoch": 0.2011822047080784, "grad_norm": 2.34375, "learning_rate": 0.00019041520760380192, "loss": 4.5012, "step": 485 }, { "epoch": 0.2015970133775796, "grad_norm": 2.203125, "learning_rate": 0.0001903951975987994, "loss": 4.6173, "step": 486 }, { "epoch": 0.2020118220470808, "grad_norm": 2.59375, "learning_rate": 0.00019037518759379692, "loss": 4.6062, "step": 487 }, { "epoch": 0.20242663071658198, "grad_norm": 2.109375, "learning_rate": 0.0001903551775887944, "loss": 4.4623, "step": 488 }, { "epoch": 0.20284143938608318, "grad_norm": 2.15625, "learning_rate": 0.0001903351675837919, "loss": 4.5057, "step": 489 }, { "epoch": 0.20325624805558437, "grad_norm": 1.984375, "learning_rate": 0.00019031515757878938, "loss": 4.3361, "step": 490 }, { "epoch": 0.20367105672508556, "grad_norm": 2.21875, "learning_rate": 0.0001902951475737869, "loss": 4.3838, "step": 491 }, { "epoch": 0.20408586539458676, "grad_norm": 2.3125, "learning_rate": 0.0001902751375687844, "loss": 4.5128, "step": 492 }, { "epoch": 0.20450067406408795, "grad_norm": 2.515625, "learning_rate": 0.0001902551275637819, "loss": 4.4631, "step": 493 }, { "epoch": 0.20491548273358914, "grad_norm": 2.328125, "learning_rate": 0.0001902351175587794, "loss": 4.7555, "step": 494 }, { "epoch": 0.20533029140309034, "grad_norm": 3.078125, "learning_rate": 0.0001902151075537769, "loss": 4.8614, "step": 495 }, { "epoch": 0.20574510007259153, "grad_norm": 2.1875, "learning_rate": 0.00019019509754877441, "loss": 4.5543, "step": 496 }, { "epoch": 0.20615990874209272, "grad_norm": 2.4375, "learning_rate": 0.0001901750875437719, "loss": 4.5363, "step": 497 }, { "epoch": 0.2065747174115939, "grad_norm": 2.28125, "learning_rate": 0.0001901550775387694, "loss": 4.6619, "step": 498 }, { "epoch": 0.20698952608109508, "grad_norm": 2.265625, "learning_rate": 0.00019013506753376688, "loss": 4.3184, "step": 499 }, { "epoch": 0.20740433475059628, "grad_norm": 2.28125, "learning_rate": 0.0001901150575287644, "loss": 4.3842, "step": 500 }, { "epoch": 0.20781914342009747, "grad_norm": 2.375, "learning_rate": 0.00019009504752376188, "loss": 4.5367, "step": 501 }, { "epoch": 0.20823395208959866, "grad_norm": 2.359375, "learning_rate": 0.0001900750375187594, "loss": 4.9936, "step": 502 }, { "epoch": 0.20864876075909985, "grad_norm": 2.1875, "learning_rate": 0.00019005502751375688, "loss": 4.4494, "step": 503 }, { "epoch": 0.20906356942860105, "grad_norm": 2.203125, "learning_rate": 0.0001900350175087544, "loss": 4.5118, "step": 504 }, { "epoch": 0.20947837809810224, "grad_norm": 2.15625, "learning_rate": 0.00019001500750375188, "loss": 4.6193, "step": 505 }, { "epoch": 0.20989318676760343, "grad_norm": 2.359375, "learning_rate": 0.0001899949974987494, "loss": 4.364, "step": 506 }, { "epoch": 0.21030799543710463, "grad_norm": 2.234375, "learning_rate": 0.00018997498749374688, "loss": 4.5113, "step": 507 }, { "epoch": 0.21072280410660582, "grad_norm": 2.125, "learning_rate": 0.00018995497748874437, "loss": 4.4587, "step": 508 }, { "epoch": 0.21113761277610701, "grad_norm": 2.109375, "learning_rate": 0.00018993496748374188, "loss": 4.5137, "step": 509 }, { "epoch": 0.2115524214456082, "grad_norm": 2.296875, "learning_rate": 0.00018991495747873937, "loss": 4.4468, "step": 510 }, { "epoch": 0.2119672301151094, "grad_norm": 2.140625, "learning_rate": 0.00018989494747373689, "loss": 4.395, "step": 511 }, { "epoch": 0.2123820387846106, "grad_norm": 2.140625, "learning_rate": 0.00018987493746873437, "loss": 4.3873, "step": 512 }, { "epoch": 0.2127968474541118, "grad_norm": 2.28125, "learning_rate": 0.0001898549274637319, "loss": 4.3558, "step": 513 }, { "epoch": 0.21321165612361298, "grad_norm": 2.421875, "learning_rate": 0.00018983491745872938, "loss": 4.548, "step": 514 }, { "epoch": 0.21362646479311417, "grad_norm": 2.359375, "learning_rate": 0.00018981490745372686, "loss": 4.5822, "step": 515 }, { "epoch": 0.21404127346261537, "grad_norm": 2.359375, "learning_rate": 0.00018979489744872435, "loss": 4.5558, "step": 516 }, { "epoch": 0.21445608213211656, "grad_norm": 2.15625, "learning_rate": 0.00018977488744372186, "loss": 4.5747, "step": 517 }, { "epoch": 0.21487089080161775, "grad_norm": 2.140625, "learning_rate": 0.00018975487743871935, "loss": 4.3788, "step": 518 }, { "epoch": 0.21528569947111895, "grad_norm": 2.140625, "learning_rate": 0.00018973486743371687, "loss": 4.5583, "step": 519 }, { "epoch": 0.21570050814062014, "grad_norm": 2.265625, "learning_rate": 0.00018971485742871438, "loss": 4.583, "step": 520 }, { "epoch": 0.21611531681012133, "grad_norm": 2.359375, "learning_rate": 0.00018969484742371187, "loss": 4.5199, "step": 521 }, { "epoch": 0.21653012547962253, "grad_norm": 2.03125, "learning_rate": 0.00018967483741870938, "loss": 4.5162, "step": 522 }, { "epoch": 0.21694493414912372, "grad_norm": 2.140625, "learning_rate": 0.00018965482741370687, "loss": 4.5838, "step": 523 }, { "epoch": 0.2173597428186249, "grad_norm": 2.265625, "learning_rate": 0.00018963481740870436, "loss": 4.5335, "step": 524 }, { "epoch": 0.2177745514881261, "grad_norm": 2.484375, "learning_rate": 0.00018961480740370184, "loss": 4.7659, "step": 525 }, { "epoch": 0.2181893601576273, "grad_norm": 2.328125, "learning_rate": 0.00018959479739869936, "loss": 4.2365, "step": 526 }, { "epoch": 0.2186041688271285, "grad_norm": 2.15625, "learning_rate": 0.00018957478739369685, "loss": 4.7451, "step": 527 }, { "epoch": 0.21901897749662969, "grad_norm": 2.34375, "learning_rate": 0.00018955477738869436, "loss": 4.5823, "step": 528 }, { "epoch": 0.21943378616613088, "grad_norm": 2.1875, "learning_rate": 0.00018953476738369185, "loss": 4.4656, "step": 529 }, { "epoch": 0.21984859483563207, "grad_norm": 2.078125, "learning_rate": 0.00018951475737868936, "loss": 4.9049, "step": 530 }, { "epoch": 0.22026340350513327, "grad_norm": 2.1875, "learning_rate": 0.00018949474737368685, "loss": 4.6767, "step": 531 }, { "epoch": 0.22067821217463446, "grad_norm": 2.046875, "learning_rate": 0.00018947473736868436, "loss": 4.6346, "step": 532 }, { "epoch": 0.22109302084413565, "grad_norm": 2.453125, "learning_rate": 0.00018945472736368185, "loss": 4.4386, "step": 533 }, { "epoch": 0.22150782951363684, "grad_norm": 2.359375, "learning_rate": 0.00018943471735867934, "loss": 4.6902, "step": 534 }, { "epoch": 0.22192263818313804, "grad_norm": 2.046875, "learning_rate": 0.00018941470735367685, "loss": 4.4004, "step": 535 }, { "epoch": 0.22233744685263923, "grad_norm": 2.984375, "learning_rate": 0.00018939469734867434, "loss": 4.3807, "step": 536 }, { "epoch": 0.22275225552214042, "grad_norm": 2.375, "learning_rate": 0.00018937468734367185, "loss": 4.5815, "step": 537 }, { "epoch": 0.22316706419164162, "grad_norm": 2.265625, "learning_rate": 0.00018935467733866934, "loss": 4.7156, "step": 538 }, { "epoch": 0.2235818728611428, "grad_norm": 2.171875, "learning_rate": 0.00018933466733366686, "loss": 4.5454, "step": 539 }, { "epoch": 0.223996681530644, "grad_norm": 2.328125, "learning_rate": 0.00018931465732866434, "loss": 4.5494, "step": 540 }, { "epoch": 0.22441149020014517, "grad_norm": 2.3125, "learning_rate": 0.00018929464732366186, "loss": 4.6564, "step": 541 }, { "epoch": 0.22482629886964636, "grad_norm": 2.453125, "learning_rate": 0.00018927463731865934, "loss": 4.7024, "step": 542 }, { "epoch": 0.22524110753914756, "grad_norm": 2.28125, "learning_rate": 0.00018925462731365683, "loss": 4.6705, "step": 543 }, { "epoch": 0.22565591620864875, "grad_norm": 2.40625, "learning_rate": 0.00018923461730865432, "loss": 4.5733, "step": 544 }, { "epoch": 0.22607072487814994, "grad_norm": 1.984375, "learning_rate": 0.00018921460730365183, "loss": 4.5865, "step": 545 }, { "epoch": 0.22648553354765114, "grad_norm": 2.140625, "learning_rate": 0.00018919459729864932, "loss": 4.6588, "step": 546 }, { "epoch": 0.22690034221715233, "grad_norm": 2.421875, "learning_rate": 0.00018917458729364684, "loss": 4.362, "step": 547 }, { "epoch": 0.22731515088665352, "grad_norm": 2.078125, "learning_rate": 0.00018915457728864435, "loss": 4.434, "step": 548 }, { "epoch": 0.22772995955615472, "grad_norm": 2.359375, "learning_rate": 0.00018913456728364184, "loss": 4.5605, "step": 549 }, { "epoch": 0.2281447682256559, "grad_norm": 2.5625, "learning_rate": 0.00018911455727863932, "loss": 4.3094, "step": 550 }, { "epoch": 0.2285595768951571, "grad_norm": 2.734375, "learning_rate": 0.0001890945472736368, "loss": 4.3491, "step": 551 }, { "epoch": 0.2289743855646583, "grad_norm": 2.25, "learning_rate": 0.00018907453726863433, "loss": 4.5983, "step": 552 }, { "epoch": 0.2293891942341595, "grad_norm": 1.9140625, "learning_rate": 0.0001890545272636318, "loss": 4.6936, "step": 553 }, { "epoch": 0.22980400290366068, "grad_norm": 2.40625, "learning_rate": 0.00018903451725862933, "loss": 4.4533, "step": 554 }, { "epoch": 0.23021881157316187, "grad_norm": 2.203125, "learning_rate": 0.00018901450725362681, "loss": 4.3935, "step": 555 }, { "epoch": 0.23063362024266307, "grad_norm": 2.125, "learning_rate": 0.00018899449724862433, "loss": 4.5972, "step": 556 }, { "epoch": 0.23104842891216426, "grad_norm": 2.4375, "learning_rate": 0.00018897448724362182, "loss": 4.4789, "step": 557 }, { "epoch": 0.23146323758166545, "grad_norm": 2.203125, "learning_rate": 0.00018895447723861933, "loss": 4.5092, "step": 558 }, { "epoch": 0.23187804625116665, "grad_norm": 2.390625, "learning_rate": 0.00018893446723361682, "loss": 4.5099, "step": 559 }, { "epoch": 0.23229285492066784, "grad_norm": 2.453125, "learning_rate": 0.0001889144572286143, "loss": 4.5015, "step": 560 }, { "epoch": 0.23270766359016903, "grad_norm": 2.25, "learning_rate": 0.00018889444722361182, "loss": 4.2881, "step": 561 }, { "epoch": 0.23312247225967023, "grad_norm": 2.078125, "learning_rate": 0.0001888744372186093, "loss": 4.309, "step": 562 }, { "epoch": 0.23353728092917142, "grad_norm": 2.296875, "learning_rate": 0.00018885442721360682, "loss": 4.5298, "step": 563 }, { "epoch": 0.2339520895986726, "grad_norm": 2.125, "learning_rate": 0.0001888344172086043, "loss": 4.4788, "step": 564 }, { "epoch": 0.2343668982681738, "grad_norm": 2.03125, "learning_rate": 0.00018881440720360182, "loss": 4.3574, "step": 565 }, { "epoch": 0.234781706937675, "grad_norm": 2.265625, "learning_rate": 0.0001887943971985993, "loss": 4.5106, "step": 566 }, { "epoch": 0.2351965156071762, "grad_norm": 2.21875, "learning_rate": 0.00018877438719359682, "loss": 4.6805, "step": 567 }, { "epoch": 0.2356113242766774, "grad_norm": 2.171875, "learning_rate": 0.0001887543771885943, "loss": 4.4447, "step": 568 }, { "epoch": 0.23602613294617858, "grad_norm": 2.15625, "learning_rate": 0.0001887343671835918, "loss": 4.4886, "step": 569 }, { "epoch": 0.23644094161567977, "grad_norm": 2.109375, "learning_rate": 0.0001887143571785893, "loss": 4.5327, "step": 570 }, { "epoch": 0.23685575028518097, "grad_norm": 2.0625, "learning_rate": 0.0001886943471735868, "loss": 4.4291, "step": 571 }, { "epoch": 0.23727055895468216, "grad_norm": 2.328125, "learning_rate": 0.0001886743371685843, "loss": 4.4504, "step": 572 }, { "epoch": 0.23768536762418335, "grad_norm": 2.71875, "learning_rate": 0.0001886543271635818, "loss": 4.5148, "step": 573 }, { "epoch": 0.23810017629368455, "grad_norm": 2.078125, "learning_rate": 0.0001886343171585793, "loss": 4.6872, "step": 574 }, { "epoch": 0.23851498496318574, "grad_norm": 2.28125, "learning_rate": 0.0001886143071535768, "loss": 4.3635, "step": 575 }, { "epoch": 0.23892979363268693, "grad_norm": 2.390625, "learning_rate": 0.00018859429714857432, "loss": 4.5316, "step": 576 }, { "epoch": 0.23934460230218813, "grad_norm": 2.765625, "learning_rate": 0.0001885742871435718, "loss": 4.5817, "step": 577 }, { "epoch": 0.23975941097168932, "grad_norm": 2.1875, "learning_rate": 0.0001885542771385693, "loss": 4.6379, "step": 578 }, { "epoch": 0.2401742196411905, "grad_norm": 2.421875, "learning_rate": 0.00018853426713356678, "loss": 4.5374, "step": 579 }, { "epoch": 0.2405890283106917, "grad_norm": 2.171875, "learning_rate": 0.0001885142571285643, "loss": 4.2486, "step": 580 }, { "epoch": 0.2410038369801929, "grad_norm": 2.234375, "learning_rate": 0.00018849424712356178, "loss": 4.395, "step": 581 }, { "epoch": 0.2414186456496941, "grad_norm": 2.234375, "learning_rate": 0.0001884742371185593, "loss": 4.7067, "step": 582 }, { "epoch": 0.24183345431919528, "grad_norm": 2.328125, "learning_rate": 0.00018845422711355678, "loss": 4.3778, "step": 583 }, { "epoch": 0.24224826298869645, "grad_norm": 2.34375, "learning_rate": 0.0001884342171085543, "loss": 4.6454, "step": 584 }, { "epoch": 0.24266307165819764, "grad_norm": 2.6875, "learning_rate": 0.00018841420710355179, "loss": 4.4081, "step": 585 }, { "epoch": 0.24307788032769884, "grad_norm": 2.203125, "learning_rate": 0.00018839419709854927, "loss": 4.4475, "step": 586 }, { "epoch": 0.24349268899720003, "grad_norm": 2.34375, "learning_rate": 0.00018837418709354676, "loss": 4.6707, "step": 587 }, { "epoch": 0.24390749766670122, "grad_norm": 2.359375, "learning_rate": 0.00018835417708854427, "loss": 4.6686, "step": 588 }, { "epoch": 0.24432230633620242, "grad_norm": 2.578125, "learning_rate": 0.0001883341670835418, "loss": 4.4247, "step": 589 }, { "epoch": 0.2447371150057036, "grad_norm": 2.09375, "learning_rate": 0.00018831415707853928, "loss": 4.5128, "step": 590 }, { "epoch": 0.2451519236752048, "grad_norm": 2.171875, "learning_rate": 0.0001882941470735368, "loss": 4.5771, "step": 591 }, { "epoch": 0.245566732344706, "grad_norm": 2.25, "learning_rate": 0.00018827413706853428, "loss": 4.4879, "step": 592 }, { "epoch": 0.2459815410142072, "grad_norm": 2.203125, "learning_rate": 0.0001882541270635318, "loss": 4.6341, "step": 593 }, { "epoch": 0.24639634968370838, "grad_norm": 2.03125, "learning_rate": 0.00018823411705852928, "loss": 4.5878, "step": 594 }, { "epoch": 0.24681115835320958, "grad_norm": 2.296875, "learning_rate": 0.00018821410705352677, "loss": 4.4684, "step": 595 }, { "epoch": 0.24722596702271077, "grad_norm": 2.4375, "learning_rate": 0.00018819409704852425, "loss": 4.6463, "step": 596 }, { "epoch": 0.24764077569221196, "grad_norm": 2.375, "learning_rate": 0.00018817408704352177, "loss": 4.2628, "step": 597 }, { "epoch": 0.24805558436171316, "grad_norm": 2.03125, "learning_rate": 0.00018815407703851926, "loss": 4.3581, "step": 598 }, { "epoch": 0.24847039303121435, "grad_norm": 2.140625, "learning_rate": 0.00018813406703351677, "loss": 4.5657, "step": 599 }, { "epoch": 0.24888520170071554, "grad_norm": 2.328125, "learning_rate": 0.00018811405702851426, "loss": 4.4318, "step": 600 }, { "epoch": 0.24930001037021673, "grad_norm": 2.140625, "learning_rate": 0.00018809404702351177, "loss": 4.4151, "step": 601 }, { "epoch": 0.24971481903971793, "grad_norm": 2.171875, "learning_rate": 0.00018807403701850926, "loss": 4.5038, "step": 602 }, { "epoch": 0.25012962770921915, "grad_norm": 2.3125, "learning_rate": 0.00018805402701350677, "loss": 4.62, "step": 603 }, { "epoch": 0.2505444363787203, "grad_norm": 2.5, "learning_rate": 0.00018803401700850426, "loss": 4.5212, "step": 604 }, { "epoch": 0.25095924504822154, "grad_norm": 2.09375, "learning_rate": 0.00018801400700350175, "loss": 4.4718, "step": 605 }, { "epoch": 0.2513740537177227, "grad_norm": 1.9765625, "learning_rate": 0.00018799399699849926, "loss": 4.453, "step": 606 }, { "epoch": 0.25178886238722387, "grad_norm": 2.234375, "learning_rate": 0.00018797398699349675, "loss": 4.3431, "step": 607 }, { "epoch": 0.2522036710567251, "grad_norm": 2.421875, "learning_rate": 0.00018795397698849426, "loss": 4.5271, "step": 608 }, { "epoch": 0.25261847972622625, "grad_norm": 2.03125, "learning_rate": 0.00018793396698349175, "loss": 4.4361, "step": 609 }, { "epoch": 0.2530332883957275, "grad_norm": 2.515625, "learning_rate": 0.00018791395697848927, "loss": 4.4498, "step": 610 }, { "epoch": 0.25344809706522864, "grad_norm": 2.078125, "learning_rate": 0.00018789394697348675, "loss": 4.6997, "step": 611 }, { "epoch": 0.25386290573472986, "grad_norm": 2.25, "learning_rate": 0.00018787393696848427, "loss": 4.5507, "step": 612 }, { "epoch": 0.254277714404231, "grad_norm": 2.890625, "learning_rate": 0.00018785392696348175, "loss": 4.6352, "step": 613 }, { "epoch": 0.25469252307373225, "grad_norm": 2.359375, "learning_rate": 0.00018783391695847924, "loss": 4.5464, "step": 614 }, { "epoch": 0.2551073317432334, "grad_norm": 2.234375, "learning_rate": 0.00018781390695347673, "loss": 4.6704, "step": 615 }, { "epoch": 0.25552214041273463, "grad_norm": 2.390625, "learning_rate": 0.00018779389694847424, "loss": 4.578, "step": 616 }, { "epoch": 0.2559369490822358, "grad_norm": 2.953125, "learning_rate": 0.00018777388694347176, "loss": 4.6426, "step": 617 }, { "epoch": 0.256351757751737, "grad_norm": 2.265625, "learning_rate": 0.00018775387693846925, "loss": 4.4142, "step": 618 }, { "epoch": 0.2567665664212382, "grad_norm": 2.359375, "learning_rate": 0.00018773386693346676, "loss": 4.8122, "step": 619 }, { "epoch": 0.2571813750907394, "grad_norm": 2.25, "learning_rate": 0.00018771385692846425, "loss": 4.6555, "step": 620 }, { "epoch": 0.25759618376024057, "grad_norm": 2.65625, "learning_rate": 0.00018769384692346173, "loss": 4.5801, "step": 621 }, { "epoch": 0.2580109924297418, "grad_norm": 2.265625, "learning_rate": 0.00018767383691845922, "loss": 4.6506, "step": 622 }, { "epoch": 0.25842580109924296, "grad_norm": 2.09375, "learning_rate": 0.00018765382691345674, "loss": 4.6701, "step": 623 }, { "epoch": 0.2588406097687442, "grad_norm": 2.578125, "learning_rate": 0.00018763381690845422, "loss": 4.3571, "step": 624 }, { "epoch": 0.25925541843824534, "grad_norm": 2.109375, "learning_rate": 0.00018761380690345174, "loss": 4.4263, "step": 625 }, { "epoch": 0.25967022710774657, "grad_norm": 2.0625, "learning_rate": 0.00018759379689844922, "loss": 4.6005, "step": 626 }, { "epoch": 0.26008503577724773, "grad_norm": 1.9296875, "learning_rate": 0.00018757378689344674, "loss": 4.4602, "step": 627 }, { "epoch": 0.26049984444674895, "grad_norm": 2.265625, "learning_rate": 0.00018755377688844423, "loss": 4.4964, "step": 628 }, { "epoch": 0.2609146531162501, "grad_norm": 1.9296875, "learning_rate": 0.00018753376688344174, "loss": 4.2681, "step": 629 }, { "epoch": 0.26132946178575134, "grad_norm": 2.0625, "learning_rate": 0.00018751375687843923, "loss": 4.4756, "step": 630 }, { "epoch": 0.2617442704552525, "grad_norm": 2.046875, "learning_rate": 0.00018749374687343672, "loss": 4.4144, "step": 631 }, { "epoch": 0.2621590791247537, "grad_norm": 2.203125, "learning_rate": 0.00018747373686843423, "loss": 4.322, "step": 632 }, { "epoch": 0.2625738877942549, "grad_norm": 2.15625, "learning_rate": 0.00018745372686343172, "loss": 4.5444, "step": 633 }, { "epoch": 0.2629886964637561, "grad_norm": 2.421875, "learning_rate": 0.00018743371685842923, "loss": 4.5651, "step": 634 }, { "epoch": 0.2634035051332573, "grad_norm": 2.3125, "learning_rate": 0.00018741370685342672, "loss": 4.6755, "step": 635 }, { "epoch": 0.2638183138027585, "grad_norm": 2.453125, "learning_rate": 0.00018739369684842423, "loss": 4.5744, "step": 636 }, { "epoch": 0.26423312247225966, "grad_norm": 2.421875, "learning_rate": 0.00018737368684342172, "loss": 4.5153, "step": 637 }, { "epoch": 0.2646479311417609, "grad_norm": 2.296875, "learning_rate": 0.00018735367683841923, "loss": 4.6289, "step": 638 }, { "epoch": 0.26506273981126205, "grad_norm": 2.265625, "learning_rate": 0.00018733366683341672, "loss": 4.4528, "step": 639 }, { "epoch": 0.26547754848076327, "grad_norm": 2.265625, "learning_rate": 0.0001873136568284142, "loss": 4.7298, "step": 640 }, { "epoch": 0.26589235715026444, "grad_norm": 2.078125, "learning_rate": 0.0001872936468234117, "loss": 4.3405, "step": 641 }, { "epoch": 0.26630716581976566, "grad_norm": 2.234375, "learning_rate": 0.0001872736368184092, "loss": 4.6529, "step": 642 }, { "epoch": 0.2667219744892668, "grad_norm": 2.40625, "learning_rate": 0.0001872536268134067, "loss": 4.3613, "step": 643 }, { "epoch": 0.26713678315876804, "grad_norm": 2.46875, "learning_rate": 0.0001872336168084042, "loss": 4.8304, "step": 644 }, { "epoch": 0.2675515918282692, "grad_norm": 2.484375, "learning_rate": 0.00018721360680340173, "loss": 4.5604, "step": 645 }, { "epoch": 0.26796640049777043, "grad_norm": 2.03125, "learning_rate": 0.00018719359679839921, "loss": 4.5692, "step": 646 }, { "epoch": 0.2683812091672716, "grad_norm": 2.015625, "learning_rate": 0.00018717358679339673, "loss": 4.2713, "step": 647 }, { "epoch": 0.2687960178367728, "grad_norm": 2.21875, "learning_rate": 0.00018715357678839422, "loss": 4.2112, "step": 648 }, { "epoch": 0.269210826506274, "grad_norm": 2.140625, "learning_rate": 0.0001871335667833917, "loss": 4.3731, "step": 649 }, { "epoch": 0.26962563517577515, "grad_norm": 2.296875, "learning_rate": 0.0001871135567783892, "loss": 4.5213, "step": 650 }, { "epoch": 0.27004044384527637, "grad_norm": 2.203125, "learning_rate": 0.0001870935467733867, "loss": 4.323, "step": 651 }, { "epoch": 0.27045525251477753, "grad_norm": 2.375, "learning_rate": 0.0001870735367683842, "loss": 4.7545, "step": 652 }, { "epoch": 0.27087006118427875, "grad_norm": 2.0625, "learning_rate": 0.0001870535267633817, "loss": 4.5885, "step": 653 }, { "epoch": 0.2712848698537799, "grad_norm": 2.03125, "learning_rate": 0.0001870335167583792, "loss": 4.2861, "step": 654 }, { "epoch": 0.27169967852328114, "grad_norm": 2.1875, "learning_rate": 0.0001870135067533767, "loss": 4.3931, "step": 655 }, { "epoch": 0.2721144871927823, "grad_norm": 2.1875, "learning_rate": 0.0001869934967483742, "loss": 4.6776, "step": 656 }, { "epoch": 0.2725292958622835, "grad_norm": 2.4375, "learning_rate": 0.00018697348674337168, "loss": 4.529, "step": 657 }, { "epoch": 0.2729441045317847, "grad_norm": 2.375, "learning_rate": 0.0001869534767383692, "loss": 4.48, "step": 658 }, { "epoch": 0.2733589132012859, "grad_norm": 2.484375, "learning_rate": 0.00018693346673336668, "loss": 4.6314, "step": 659 }, { "epoch": 0.2737737218707871, "grad_norm": 2.3125, "learning_rate": 0.0001869134567283642, "loss": 4.5495, "step": 660 }, { "epoch": 0.2741885305402883, "grad_norm": 2.296875, "learning_rate": 0.00018689344672336169, "loss": 4.5729, "step": 661 }, { "epoch": 0.27460333920978947, "grad_norm": 2.078125, "learning_rate": 0.0001868734367183592, "loss": 4.3969, "step": 662 }, { "epoch": 0.2750181478792907, "grad_norm": 1.8984375, "learning_rate": 0.0001868534267133567, "loss": 4.5123, "step": 663 }, { "epoch": 0.27543295654879185, "grad_norm": 2.234375, "learning_rate": 0.0001868334167083542, "loss": 4.2258, "step": 664 }, { "epoch": 0.2758477652182931, "grad_norm": 1.9609375, "learning_rate": 0.0001868134067033517, "loss": 4.2612, "step": 665 }, { "epoch": 0.27626257388779424, "grad_norm": 1.9921875, "learning_rate": 0.00018679339669834918, "loss": 4.6057, "step": 666 }, { "epoch": 0.27667738255729546, "grad_norm": 2.296875, "learning_rate": 0.00018677338669334666, "loss": 4.5225, "step": 667 }, { "epoch": 0.2770921912267966, "grad_norm": 2.203125, "learning_rate": 0.00018675337668834418, "loss": 4.8157, "step": 668 }, { "epoch": 0.27750699989629785, "grad_norm": 2.171875, "learning_rate": 0.00018673336668334167, "loss": 4.6119, "step": 669 }, { "epoch": 0.277921808565799, "grad_norm": 2.109375, "learning_rate": 0.00018671335667833918, "loss": 4.663, "step": 670 }, { "epoch": 0.27833661723530023, "grad_norm": 1.921875, "learning_rate": 0.00018669334667333667, "loss": 4.3608, "step": 671 }, { "epoch": 0.2787514259048014, "grad_norm": 2.015625, "learning_rate": 0.00018667333666833418, "loss": 4.4261, "step": 672 }, { "epoch": 0.2791662345743026, "grad_norm": 2.234375, "learning_rate": 0.0001866533266633317, "loss": 4.2453, "step": 673 }, { "epoch": 0.2795810432438038, "grad_norm": 2.140625, "learning_rate": 0.00018663331665832918, "loss": 4.6417, "step": 674 }, { "epoch": 0.279995851913305, "grad_norm": 2.09375, "learning_rate": 0.00018661330665332667, "loss": 4.5089, "step": 675 }, { "epoch": 0.28041066058280617, "grad_norm": 2.328125, "learning_rate": 0.00018659329664832416, "loss": 4.6126, "step": 676 }, { "epoch": 0.2808254692523074, "grad_norm": 2.546875, "learning_rate": 0.00018657328664332167, "loss": 4.4829, "step": 677 }, { "epoch": 0.28124027792180856, "grad_norm": 2.203125, "learning_rate": 0.00018655327663831916, "loss": 4.5772, "step": 678 }, { "epoch": 0.2816550865913098, "grad_norm": 2.09375, "learning_rate": 0.00018653326663331667, "loss": 4.6509, "step": 679 }, { "epoch": 0.28206989526081094, "grad_norm": 1.9453125, "learning_rate": 0.00018651325662831416, "loss": 4.586, "step": 680 }, { "epoch": 0.28248470393031216, "grad_norm": 2.046875, "learning_rate": 0.00018649324662331168, "loss": 4.3017, "step": 681 }, { "epoch": 0.28289951259981333, "grad_norm": 2.234375, "learning_rate": 0.00018647323661830916, "loss": 4.372, "step": 682 }, { "epoch": 0.28331432126931455, "grad_norm": 2.328125, "learning_rate": 0.00018645322661330668, "loss": 4.6157, "step": 683 }, { "epoch": 0.2837291299388157, "grad_norm": 2.53125, "learning_rate": 0.00018643321660830414, "loss": 4.4688, "step": 684 }, { "epoch": 0.28414393860831694, "grad_norm": 2.078125, "learning_rate": 0.00018641320660330165, "loss": 4.4668, "step": 685 }, { "epoch": 0.2845587472778181, "grad_norm": 2.234375, "learning_rate": 0.00018639319659829917, "loss": 4.6026, "step": 686 }, { "epoch": 0.2849735559473193, "grad_norm": 2.4375, "learning_rate": 0.00018637318659329665, "loss": 4.506, "step": 687 }, { "epoch": 0.2853883646168205, "grad_norm": 2.25, "learning_rate": 0.00018635317658829417, "loss": 4.4546, "step": 688 }, { "epoch": 0.2858031732863217, "grad_norm": 2.1875, "learning_rate": 0.00018633316658329166, "loss": 4.6656, "step": 689 }, { "epoch": 0.2862179819558229, "grad_norm": 2.09375, "learning_rate": 0.00018631315657828917, "loss": 4.3545, "step": 690 }, { "epoch": 0.2866327906253241, "grad_norm": 2.171875, "learning_rate": 0.00018629314657328666, "loss": 4.4224, "step": 691 }, { "epoch": 0.28704759929482526, "grad_norm": 2.421875, "learning_rate": 0.00018627313656828414, "loss": 4.4423, "step": 692 }, { "epoch": 0.2874624079643264, "grad_norm": 2.28125, "learning_rate": 0.00018625312656328163, "loss": 4.5748, "step": 693 }, { "epoch": 0.28787721663382765, "grad_norm": 2.25, "learning_rate": 0.00018623311655827915, "loss": 4.4786, "step": 694 }, { "epoch": 0.2882920253033288, "grad_norm": 2.265625, "learning_rate": 0.00018621310655327663, "loss": 4.29, "step": 695 }, { "epoch": 0.28870683397283003, "grad_norm": 2.0625, "learning_rate": 0.00018619309654827415, "loss": 4.5001, "step": 696 }, { "epoch": 0.2891216426423312, "grad_norm": 2.046875, "learning_rate": 0.00018617308654327163, "loss": 4.2712, "step": 697 }, { "epoch": 0.2895364513118324, "grad_norm": 2.140625, "learning_rate": 0.00018615307653826915, "loss": 4.4952, "step": 698 }, { "epoch": 0.2899512599813336, "grad_norm": 2.125, "learning_rate": 0.00018613306653326664, "loss": 4.4281, "step": 699 }, { "epoch": 0.2903660686508348, "grad_norm": 2.265625, "learning_rate": 0.00018611305652826415, "loss": 4.431, "step": 700 }, { "epoch": 0.290780877320336, "grad_norm": 2.171875, "learning_rate": 0.00018609304652326164, "loss": 4.5756, "step": 701 }, { "epoch": 0.2911956859898372, "grad_norm": 2.375, "learning_rate": 0.00018607303651825913, "loss": 4.4734, "step": 702 }, { "epoch": 0.29161049465933836, "grad_norm": 2.0, "learning_rate": 0.00018605302651325664, "loss": 4.4126, "step": 703 }, { "epoch": 0.2920253033288396, "grad_norm": 2.75, "learning_rate": 0.00018603301650825413, "loss": 4.412, "step": 704 }, { "epoch": 0.29244011199834075, "grad_norm": 2.015625, "learning_rate": 0.00018601300650325164, "loss": 4.3727, "step": 705 }, { "epoch": 0.29285492066784197, "grad_norm": 2.484375, "learning_rate": 0.00018599299649824913, "loss": 4.5848, "step": 706 }, { "epoch": 0.29326972933734313, "grad_norm": 2.25, "learning_rate": 0.00018597298649324664, "loss": 4.4878, "step": 707 }, { "epoch": 0.29368453800684435, "grad_norm": 2.296875, "learning_rate": 0.00018595297648824413, "loss": 4.5346, "step": 708 }, { "epoch": 0.2940993466763455, "grad_norm": 2.703125, "learning_rate": 0.00018593296648324164, "loss": 4.6369, "step": 709 }, { "epoch": 0.29451415534584674, "grad_norm": 2.546875, "learning_rate": 0.00018591295647823913, "loss": 4.4056, "step": 710 }, { "epoch": 0.2949289640153479, "grad_norm": 2.078125, "learning_rate": 0.00018589294647323662, "loss": 4.4915, "step": 711 }, { "epoch": 0.2953437726848491, "grad_norm": 2.234375, "learning_rate": 0.0001858729364682341, "loss": 4.4464, "step": 712 }, { "epoch": 0.2957585813543503, "grad_norm": 2.125, "learning_rate": 0.00018585292646323162, "loss": 4.241, "step": 713 }, { "epoch": 0.2961733900238515, "grad_norm": 2.265625, "learning_rate": 0.00018583291645822914, "loss": 4.7127, "step": 714 }, { "epoch": 0.2965881986933527, "grad_norm": 2.171875, "learning_rate": 0.00018581290645322662, "loss": 4.0667, "step": 715 }, { "epoch": 0.2970030073628539, "grad_norm": 2.078125, "learning_rate": 0.00018579289644822414, "loss": 4.4533, "step": 716 }, { "epoch": 0.29741781603235506, "grad_norm": 2.265625, "learning_rate": 0.00018577288644322162, "loss": 4.551, "step": 717 }, { "epoch": 0.2978326247018563, "grad_norm": 2.046875, "learning_rate": 0.00018575287643821914, "loss": 4.4095, "step": 718 }, { "epoch": 0.29824743337135745, "grad_norm": 2.234375, "learning_rate": 0.00018573286643321663, "loss": 4.4535, "step": 719 }, { "epoch": 0.29866224204085867, "grad_norm": 2.265625, "learning_rate": 0.0001857128564282141, "loss": 4.5547, "step": 720 }, { "epoch": 0.29907705071035984, "grad_norm": 2.46875, "learning_rate": 0.0001856928464232116, "loss": 4.4773, "step": 721 }, { "epoch": 0.29949185937986106, "grad_norm": 2.09375, "learning_rate": 0.00018567283641820911, "loss": 4.681, "step": 722 }, { "epoch": 0.2999066680493622, "grad_norm": 2.09375, "learning_rate": 0.0001856528264132066, "loss": 4.7683, "step": 723 }, { "epoch": 0.30032147671886344, "grad_norm": 2.421875, "learning_rate": 0.00018563281640820412, "loss": 4.4252, "step": 724 }, { "epoch": 0.3007362853883646, "grad_norm": 2.3125, "learning_rate": 0.0001856128064032016, "loss": 4.5449, "step": 725 }, { "epoch": 0.30115109405786583, "grad_norm": 2.03125, "learning_rate": 0.00018559279639819912, "loss": 4.7446, "step": 726 }, { "epoch": 0.301565902727367, "grad_norm": 2.0625, "learning_rate": 0.0001855727863931966, "loss": 4.5691, "step": 727 }, { "epoch": 0.3019807113968682, "grad_norm": 2.109375, "learning_rate": 0.0001855527763881941, "loss": 4.4674, "step": 728 }, { "epoch": 0.3023955200663694, "grad_norm": 2.25, "learning_rate": 0.0001855327663831916, "loss": 4.4785, "step": 729 }, { "epoch": 0.3028103287358706, "grad_norm": 2.296875, "learning_rate": 0.0001855127563781891, "loss": 4.5722, "step": 730 }, { "epoch": 0.30322513740537177, "grad_norm": 2.328125, "learning_rate": 0.0001854927463731866, "loss": 4.3957, "step": 731 }, { "epoch": 0.303639946074873, "grad_norm": 2.703125, "learning_rate": 0.0001854727363681841, "loss": 4.3544, "step": 732 }, { "epoch": 0.30405475474437416, "grad_norm": 2.09375, "learning_rate": 0.0001854527263631816, "loss": 4.4401, "step": 733 }, { "epoch": 0.3044695634138754, "grad_norm": 2.234375, "learning_rate": 0.0001854327163581791, "loss": 4.622, "step": 734 }, { "epoch": 0.30488437208337654, "grad_norm": 1.90625, "learning_rate": 0.0001854127063531766, "loss": 4.6069, "step": 735 }, { "epoch": 0.3052991807528777, "grad_norm": 2.171875, "learning_rate": 0.0001853926963481741, "loss": 4.2986, "step": 736 }, { "epoch": 0.30571398942237893, "grad_norm": 2.28125, "learning_rate": 0.0001853726863431716, "loss": 4.498, "step": 737 }, { "epoch": 0.3061287980918801, "grad_norm": 2.359375, "learning_rate": 0.00018535267633816907, "loss": 4.2426, "step": 738 }, { "epoch": 0.3065436067613813, "grad_norm": 2.1875, "learning_rate": 0.0001853326663331666, "loss": 4.8175, "step": 739 }, { "epoch": 0.3069584154308825, "grad_norm": 2.15625, "learning_rate": 0.00018531265632816408, "loss": 4.4716, "step": 740 }, { "epoch": 0.3073732241003837, "grad_norm": 1.828125, "learning_rate": 0.0001852926463231616, "loss": 4.2615, "step": 741 }, { "epoch": 0.30778803276988487, "grad_norm": 2.484375, "learning_rate": 0.0001852726363181591, "loss": 4.5818, "step": 742 }, { "epoch": 0.3082028414393861, "grad_norm": 2.078125, "learning_rate": 0.0001852526263131566, "loss": 4.5623, "step": 743 }, { "epoch": 0.30861765010888725, "grad_norm": 2.125, "learning_rate": 0.0001852326163081541, "loss": 4.2151, "step": 744 }, { "epoch": 0.3090324587783885, "grad_norm": 2.3125, "learning_rate": 0.0001852126063031516, "loss": 4.4879, "step": 745 }, { "epoch": 0.30944726744788964, "grad_norm": 2.265625, "learning_rate": 0.00018519259629814908, "loss": 4.4464, "step": 746 }, { "epoch": 0.30986207611739086, "grad_norm": 2.234375, "learning_rate": 0.00018517258629314657, "loss": 4.4975, "step": 747 }, { "epoch": 0.310276884786892, "grad_norm": 1.921875, "learning_rate": 0.00018515257628814408, "loss": 4.2955, "step": 748 }, { "epoch": 0.31069169345639325, "grad_norm": 2.40625, "learning_rate": 0.00018513256628314157, "loss": 4.4464, "step": 749 }, { "epoch": 0.3111065021258944, "grad_norm": 2.0625, "learning_rate": 0.00018511255627813908, "loss": 4.5451, "step": 750 }, { "epoch": 0.31152131079539563, "grad_norm": 2.125, "learning_rate": 0.00018509254627313657, "loss": 4.4628, "step": 751 }, { "epoch": 0.3119361194648968, "grad_norm": 2.125, "learning_rate": 0.00018507253626813409, "loss": 4.624, "step": 752 }, { "epoch": 0.312350928134398, "grad_norm": 2.15625, "learning_rate": 0.00018505252626313157, "loss": 4.494, "step": 753 }, { "epoch": 0.3127657368038992, "grad_norm": 2.046875, "learning_rate": 0.0001850325162581291, "loss": 4.4368, "step": 754 }, { "epoch": 0.3131805454734004, "grad_norm": 2.15625, "learning_rate": 0.00018501250625312657, "loss": 4.3604, "step": 755 }, { "epoch": 0.3135953541429016, "grad_norm": 2.046875, "learning_rate": 0.00018499249624812406, "loss": 4.4223, "step": 756 }, { "epoch": 0.3140101628124028, "grad_norm": 2.265625, "learning_rate": 0.00018497248624312158, "loss": 4.4868, "step": 757 }, { "epoch": 0.31442497148190396, "grad_norm": 2.078125, "learning_rate": 0.00018495247623811906, "loss": 4.5598, "step": 758 }, { "epoch": 0.3148397801514052, "grad_norm": 2.578125, "learning_rate": 0.00018493246623311658, "loss": 4.4284, "step": 759 }, { "epoch": 0.31525458882090635, "grad_norm": 2.140625, "learning_rate": 0.00018491245622811407, "loss": 4.2679, "step": 760 }, { "epoch": 0.31566939749040757, "grad_norm": 2.140625, "learning_rate": 0.00018489244622311158, "loss": 4.6374, "step": 761 }, { "epoch": 0.31608420615990873, "grad_norm": 2.0, "learning_rate": 0.00018487243621810907, "loss": 4.4247, "step": 762 }, { "epoch": 0.31649901482940995, "grad_norm": 2.140625, "learning_rate": 0.00018485242621310655, "loss": 4.3802, "step": 763 }, { "epoch": 0.3169138234989111, "grad_norm": 2.46875, "learning_rate": 0.00018483241620810404, "loss": 4.7218, "step": 764 }, { "epoch": 0.31732863216841234, "grad_norm": 2.03125, "learning_rate": 0.00018481240620310156, "loss": 4.5808, "step": 765 }, { "epoch": 0.3177434408379135, "grad_norm": 2.1875, "learning_rate": 0.00018479239619809904, "loss": 4.3449, "step": 766 }, { "epoch": 0.3181582495074147, "grad_norm": 2.21875, "learning_rate": 0.00018477238619309656, "loss": 4.2355, "step": 767 }, { "epoch": 0.3185730581769159, "grad_norm": 2.34375, "learning_rate": 0.00018475237618809404, "loss": 4.408, "step": 768 }, { "epoch": 0.3189878668464171, "grad_norm": 3.0, "learning_rate": 0.00018473236618309156, "loss": 4.4664, "step": 769 }, { "epoch": 0.3194026755159183, "grad_norm": 2.25, "learning_rate": 0.00018471235617808907, "loss": 4.3633, "step": 770 }, { "epoch": 0.3198174841854195, "grad_norm": 1.984375, "learning_rate": 0.00018469234617308656, "loss": 4.4451, "step": 771 }, { "epoch": 0.32023229285492066, "grad_norm": 1.8515625, "learning_rate": 0.00018467233616808405, "loss": 4.6828, "step": 772 }, { "epoch": 0.3206471015244219, "grad_norm": 2.28125, "learning_rate": 0.00018465232616308154, "loss": 4.3011, "step": 773 }, { "epoch": 0.32106191019392305, "grad_norm": 2.171875, "learning_rate": 0.00018463231615807905, "loss": 4.3755, "step": 774 }, { "epoch": 0.32147671886342427, "grad_norm": 2.859375, "learning_rate": 0.00018461230615307654, "loss": 4.5426, "step": 775 }, { "epoch": 0.32189152753292544, "grad_norm": 2.34375, "learning_rate": 0.00018459229614807405, "loss": 4.4585, "step": 776 }, { "epoch": 0.32230633620242666, "grad_norm": 2.3125, "learning_rate": 0.00018457228614307154, "loss": 4.3362, "step": 777 }, { "epoch": 0.3227211448719278, "grad_norm": 2.328125, "learning_rate": 0.00018455227613806905, "loss": 4.2984, "step": 778 }, { "epoch": 0.323135953541429, "grad_norm": 2.125, "learning_rate": 0.00018453226613306654, "loss": 4.3185, "step": 779 }, { "epoch": 0.3235507622109302, "grad_norm": 2.0625, "learning_rate": 0.00018451225612806405, "loss": 4.2773, "step": 780 }, { "epoch": 0.3239655708804314, "grad_norm": 2.015625, "learning_rate": 0.00018449224612306154, "loss": 4.3669, "step": 781 }, { "epoch": 0.3243803795499326, "grad_norm": 2.140625, "learning_rate": 0.00018447223611805903, "loss": 4.2738, "step": 782 }, { "epoch": 0.32479518821943376, "grad_norm": 2.046875, "learning_rate": 0.00018445222611305654, "loss": 4.3231, "step": 783 }, { "epoch": 0.325209996888935, "grad_norm": 2.234375, "learning_rate": 0.00018443221610805403, "loss": 4.6122, "step": 784 }, { "epoch": 0.32562480555843615, "grad_norm": 2.25, "learning_rate": 0.00018441220610305155, "loss": 4.3661, "step": 785 }, { "epoch": 0.32603961422793737, "grad_norm": 2.140625, "learning_rate": 0.00018439219609804903, "loss": 4.4894, "step": 786 }, { "epoch": 0.32645442289743853, "grad_norm": 2.3125, "learning_rate": 0.00018437218609304655, "loss": 4.7167, "step": 787 }, { "epoch": 0.32686923156693976, "grad_norm": 1.96875, "learning_rate": 0.00018435217608804403, "loss": 4.4352, "step": 788 }, { "epoch": 0.3272840402364409, "grad_norm": 2.078125, "learning_rate": 0.00018433216608304155, "loss": 4.2495, "step": 789 }, { "epoch": 0.32769884890594214, "grad_norm": 1.984375, "learning_rate": 0.000184312156078039, "loss": 4.46, "step": 790 }, { "epoch": 0.3281136575754433, "grad_norm": 2.203125, "learning_rate": 0.00018429214607303652, "loss": 4.4045, "step": 791 }, { "epoch": 0.32852846624494453, "grad_norm": 2.09375, "learning_rate": 0.000184272136068034, "loss": 4.4309, "step": 792 }, { "epoch": 0.3289432749144457, "grad_norm": 2.015625, "learning_rate": 0.00018425212606303152, "loss": 4.2624, "step": 793 }, { "epoch": 0.3293580835839469, "grad_norm": 2.09375, "learning_rate": 0.000184232116058029, "loss": 4.6596, "step": 794 }, { "epoch": 0.3297728922534481, "grad_norm": 2.203125, "learning_rate": 0.00018421210605302653, "loss": 4.6687, "step": 795 }, { "epoch": 0.3301877009229493, "grad_norm": 2.109375, "learning_rate": 0.00018419209604802401, "loss": 4.4726, "step": 796 }, { "epoch": 0.33060250959245047, "grad_norm": 2.265625, "learning_rate": 0.00018417208604302153, "loss": 4.7115, "step": 797 }, { "epoch": 0.3310173182619517, "grad_norm": 2.140625, "learning_rate": 0.00018415207603801902, "loss": 4.5506, "step": 798 }, { "epoch": 0.33143212693145285, "grad_norm": 2.140625, "learning_rate": 0.0001841320660330165, "loss": 4.2231, "step": 799 }, { "epoch": 0.3318469356009541, "grad_norm": 2.296875, "learning_rate": 0.00018411205602801402, "loss": 4.351, "step": 800 }, { "epoch": 0.33226174427045524, "grad_norm": 2.125, "learning_rate": 0.0001840920460230115, "loss": 4.4115, "step": 801 }, { "epoch": 0.33267655293995646, "grad_norm": 2.328125, "learning_rate": 0.00018407203601800902, "loss": 4.3118, "step": 802 }, { "epoch": 0.3330913616094576, "grad_norm": 2.234375, "learning_rate": 0.0001840520260130065, "loss": 4.2783, "step": 803 }, { "epoch": 0.33350617027895885, "grad_norm": 2.078125, "learning_rate": 0.00018403201600800402, "loss": 4.3474, "step": 804 }, { "epoch": 0.33392097894846, "grad_norm": 2.015625, "learning_rate": 0.0001840120060030015, "loss": 4.6202, "step": 805 }, { "epoch": 0.33433578761796123, "grad_norm": 2.0, "learning_rate": 0.00018399199599799902, "loss": 4.5117, "step": 806 }, { "epoch": 0.3347505962874624, "grad_norm": 2.203125, "learning_rate": 0.0001839719859929965, "loss": 4.5406, "step": 807 }, { "epoch": 0.3351654049569636, "grad_norm": 2.65625, "learning_rate": 0.000183951975987994, "loss": 4.6213, "step": 808 }, { "epoch": 0.3355802136264648, "grad_norm": 2.28125, "learning_rate": 0.00018393196598299148, "loss": 4.4354, "step": 809 }, { "epoch": 0.335995022295966, "grad_norm": 1.8984375, "learning_rate": 0.000183911955977989, "loss": 4.5277, "step": 810 }, { "epoch": 0.33640983096546717, "grad_norm": 2.125, "learning_rate": 0.0001838919459729865, "loss": 4.4133, "step": 811 }, { "epoch": 0.3368246396349684, "grad_norm": 1.8359375, "learning_rate": 0.000183871935967984, "loss": 4.5978, "step": 812 }, { "epoch": 0.33723944830446956, "grad_norm": 1.96875, "learning_rate": 0.00018385192596298151, "loss": 4.292, "step": 813 }, { "epoch": 0.3376542569739708, "grad_norm": 2.09375, "learning_rate": 0.000183831915957979, "loss": 4.4707, "step": 814 }, { "epoch": 0.33806906564347194, "grad_norm": 2.1875, "learning_rate": 0.00018381190595297652, "loss": 4.3632, "step": 815 }, { "epoch": 0.33848387431297317, "grad_norm": 2.046875, "learning_rate": 0.000183791895947974, "loss": 4.5217, "step": 816 }, { "epoch": 0.33889868298247433, "grad_norm": 2.125, "learning_rate": 0.0001837718859429715, "loss": 4.3052, "step": 817 }, { "epoch": 0.33931349165197555, "grad_norm": 2.28125, "learning_rate": 0.00018375187593796898, "loss": 4.3678, "step": 818 }, { "epoch": 0.3397283003214767, "grad_norm": 2.25, "learning_rate": 0.0001837318659329665, "loss": 4.3474, "step": 819 }, { "epoch": 0.34014310899097794, "grad_norm": 2.28125, "learning_rate": 0.00018371185592796398, "loss": 4.1707, "step": 820 }, { "epoch": 0.3405579176604791, "grad_norm": 2.203125, "learning_rate": 0.0001836918459229615, "loss": 4.3324, "step": 821 }, { "epoch": 0.34097272632998027, "grad_norm": 2.125, "learning_rate": 0.00018367183591795898, "loss": 4.3408, "step": 822 }, { "epoch": 0.3413875349994815, "grad_norm": 2.40625, "learning_rate": 0.0001836518259129565, "loss": 4.4247, "step": 823 }, { "epoch": 0.34180234366898266, "grad_norm": 2.125, "learning_rate": 0.000183631815907954, "loss": 4.4371, "step": 824 }, { "epoch": 0.3422171523384839, "grad_norm": 2.5625, "learning_rate": 0.00018361180590295147, "loss": 4.4755, "step": 825 }, { "epoch": 0.34263196100798504, "grad_norm": 2.0, "learning_rate": 0.00018359179589794898, "loss": 4.4183, "step": 826 }, { "epoch": 0.34304676967748626, "grad_norm": 2.140625, "learning_rate": 0.00018357178589294647, "loss": 4.4067, "step": 827 }, { "epoch": 0.34346157834698743, "grad_norm": 1.9375, "learning_rate": 0.00018355177588794399, "loss": 4.6639, "step": 828 }, { "epoch": 0.34387638701648865, "grad_norm": 2.28125, "learning_rate": 0.00018353176588294147, "loss": 4.3251, "step": 829 }, { "epoch": 0.3442911956859898, "grad_norm": 2.03125, "learning_rate": 0.000183511755877939, "loss": 4.6155, "step": 830 }, { "epoch": 0.34470600435549104, "grad_norm": 2.0625, "learning_rate": 0.00018349174587293648, "loss": 4.5124, "step": 831 }, { "epoch": 0.3451208130249922, "grad_norm": 1.96875, "learning_rate": 0.000183471735867934, "loss": 4.3172, "step": 832 }, { "epoch": 0.3455356216944934, "grad_norm": 2.203125, "learning_rate": 0.00018345172586293148, "loss": 4.2967, "step": 833 }, { "epoch": 0.3459504303639946, "grad_norm": 1.9453125, "learning_rate": 0.00018343171585792896, "loss": 4.3185, "step": 834 }, { "epoch": 0.3463652390334958, "grad_norm": 2.390625, "learning_rate": 0.00018341170585292645, "loss": 4.4256, "step": 835 }, { "epoch": 0.346780047702997, "grad_norm": 2.203125, "learning_rate": 0.00018339169584792397, "loss": 4.3809, "step": 836 }, { "epoch": 0.3471948563724982, "grad_norm": 2.265625, "learning_rate": 0.00018337168584292145, "loss": 4.2845, "step": 837 }, { "epoch": 0.34760966504199936, "grad_norm": 2.09375, "learning_rate": 0.00018335167583791897, "loss": 4.5889, "step": 838 }, { "epoch": 0.3480244737115006, "grad_norm": 2.21875, "learning_rate": 0.00018333166583291648, "loss": 4.5076, "step": 839 }, { "epoch": 0.34843928238100175, "grad_norm": 2.234375, "learning_rate": 0.00018331165582791397, "loss": 4.4128, "step": 840 }, { "epoch": 0.34885409105050297, "grad_norm": 2.3125, "learning_rate": 0.00018329164582291148, "loss": 4.6423, "step": 841 }, { "epoch": 0.34926889972000413, "grad_norm": 2.0625, "learning_rate": 0.00018327163581790897, "loss": 4.2914, "step": 842 }, { "epoch": 0.34968370838950535, "grad_norm": 2.234375, "learning_rate": 0.00018325162581290646, "loss": 4.4937, "step": 843 }, { "epoch": 0.3500985170590065, "grad_norm": 2.296875, "learning_rate": 0.00018323161580790395, "loss": 4.3421, "step": 844 }, { "epoch": 0.35051332572850774, "grad_norm": 2.15625, "learning_rate": 0.00018321160580290146, "loss": 4.8162, "step": 845 }, { "epoch": 0.3509281343980089, "grad_norm": 2.15625, "learning_rate": 0.00018319159579789895, "loss": 4.3132, "step": 846 }, { "epoch": 0.3513429430675101, "grad_norm": 1.9140625, "learning_rate": 0.00018317158579289646, "loss": 4.4433, "step": 847 }, { "epoch": 0.3517577517370113, "grad_norm": 2.09375, "learning_rate": 0.00018315157578789395, "loss": 4.508, "step": 848 }, { "epoch": 0.3521725604065125, "grad_norm": 2.0, "learning_rate": 0.00018313156578289146, "loss": 4.544, "step": 849 }, { "epoch": 0.3525873690760137, "grad_norm": 2.3125, "learning_rate": 0.00018311155577788895, "loss": 4.5758, "step": 850 }, { "epoch": 0.3530021777455149, "grad_norm": 2.109375, "learning_rate": 0.00018309154577288646, "loss": 4.3671, "step": 851 }, { "epoch": 0.35341698641501607, "grad_norm": 2.1875, "learning_rate": 0.00018307153576788395, "loss": 4.5324, "step": 852 }, { "epoch": 0.3538317950845173, "grad_norm": 1.9375, "learning_rate": 0.00018305152576288144, "loss": 4.6805, "step": 853 }, { "epoch": 0.35424660375401845, "grad_norm": 1.9609375, "learning_rate": 0.00018303151575787895, "loss": 4.4482, "step": 854 }, { "epoch": 0.3546614124235197, "grad_norm": 2.203125, "learning_rate": 0.00018301150575287644, "loss": 4.4146, "step": 855 }, { "epoch": 0.35507622109302084, "grad_norm": 2.421875, "learning_rate": 0.00018299149574787396, "loss": 4.4826, "step": 856 }, { "epoch": 0.35549102976252206, "grad_norm": 2.171875, "learning_rate": 0.00018297148574287144, "loss": 4.3543, "step": 857 }, { "epoch": 0.3559058384320232, "grad_norm": 1.953125, "learning_rate": 0.00018295147573786896, "loss": 4.4718, "step": 858 }, { "epoch": 0.35632064710152445, "grad_norm": 2.078125, "learning_rate": 0.00018293146573286644, "loss": 4.5553, "step": 859 }, { "epoch": 0.3567354557710256, "grad_norm": 1.921875, "learning_rate": 0.00018291145572786393, "loss": 4.522, "step": 860 }, { "epoch": 0.35715026444052683, "grad_norm": 2.03125, "learning_rate": 0.00018289144572286142, "loss": 4.2493, "step": 861 }, { "epoch": 0.357565073110028, "grad_norm": 2.09375, "learning_rate": 0.00018287143571785893, "loss": 4.4382, "step": 862 }, { "epoch": 0.3579798817795292, "grad_norm": 2.28125, "learning_rate": 0.00018285142571285642, "loss": 4.7082, "step": 863 }, { "epoch": 0.3583946904490304, "grad_norm": 2.0625, "learning_rate": 0.00018283141570785393, "loss": 4.584, "step": 864 }, { "epoch": 0.35880949911853155, "grad_norm": 2.625, "learning_rate": 0.00018281140570285142, "loss": 4.4695, "step": 865 }, { "epoch": 0.35922430778803277, "grad_norm": 2.171875, "learning_rate": 0.00018279139569784894, "loss": 4.3282, "step": 866 }, { "epoch": 0.35963911645753394, "grad_norm": 1.921875, "learning_rate": 0.00018277138569284645, "loss": 4.6258, "step": 867 }, { "epoch": 0.36005392512703516, "grad_norm": 2.21875, "learning_rate": 0.00018275137568784394, "loss": 4.6066, "step": 868 }, { "epoch": 0.3604687337965363, "grad_norm": 1.96875, "learning_rate": 0.00018273136568284143, "loss": 4.382, "step": 869 }, { "epoch": 0.36088354246603754, "grad_norm": 2.0, "learning_rate": 0.0001827113556778389, "loss": 4.769, "step": 870 }, { "epoch": 0.3612983511355387, "grad_norm": 2.0, "learning_rate": 0.00018269134567283643, "loss": 4.2924, "step": 871 }, { "epoch": 0.36171315980503993, "grad_norm": 1.984375, "learning_rate": 0.00018267133566783391, "loss": 4.5543, "step": 872 }, { "epoch": 0.3621279684745411, "grad_norm": 1.890625, "learning_rate": 0.00018265132566283143, "loss": 4.3232, "step": 873 }, { "epoch": 0.3625427771440423, "grad_norm": 2.09375, "learning_rate": 0.00018263131565782892, "loss": 4.2604, "step": 874 }, { "epoch": 0.3629575858135435, "grad_norm": 2.203125, "learning_rate": 0.00018261130565282643, "loss": 4.3332, "step": 875 }, { "epoch": 0.3633723944830447, "grad_norm": 2.171875, "learning_rate": 0.00018259129564782392, "loss": 4.3428, "step": 876 }, { "epoch": 0.36378720315254587, "grad_norm": 2.015625, "learning_rate": 0.00018257128564282143, "loss": 4.2177, "step": 877 }, { "epoch": 0.3642020118220471, "grad_norm": 2.09375, "learning_rate": 0.00018255127563781892, "loss": 4.4712, "step": 878 }, { "epoch": 0.36461682049154825, "grad_norm": 2.234375, "learning_rate": 0.0001825312656328164, "loss": 4.5336, "step": 879 }, { "epoch": 0.3650316291610495, "grad_norm": 2.125, "learning_rate": 0.00018251125562781392, "loss": 4.6374, "step": 880 }, { "epoch": 0.36544643783055064, "grad_norm": 1.875, "learning_rate": 0.0001824912456228114, "loss": 4.4747, "step": 881 }, { "epoch": 0.36586124650005186, "grad_norm": 2.28125, "learning_rate": 0.00018247123561780892, "loss": 4.2962, "step": 882 }, { "epoch": 0.366276055169553, "grad_norm": 1.8515625, "learning_rate": 0.0001824512256128064, "loss": 4.2172, "step": 883 }, { "epoch": 0.36669086383905425, "grad_norm": 2.0625, "learning_rate": 0.00018243121560780392, "loss": 4.3837, "step": 884 }, { "epoch": 0.3671056725085554, "grad_norm": 1.9765625, "learning_rate": 0.0001824112056028014, "loss": 4.3049, "step": 885 }, { "epoch": 0.36752048117805663, "grad_norm": 2.0625, "learning_rate": 0.00018239119559779893, "loss": 4.4653, "step": 886 }, { "epoch": 0.3679352898475578, "grad_norm": 1.96875, "learning_rate": 0.0001823711855927964, "loss": 4.517, "step": 887 }, { "epoch": 0.368350098517059, "grad_norm": 1.90625, "learning_rate": 0.0001823511755877939, "loss": 4.3587, "step": 888 }, { "epoch": 0.3687649071865602, "grad_norm": 2.015625, "learning_rate": 0.0001823311655827914, "loss": 4.5023, "step": 889 }, { "epoch": 0.3691797158560614, "grad_norm": 2.109375, "learning_rate": 0.0001823111555777889, "loss": 4.4634, "step": 890 }, { "epoch": 0.3695945245255626, "grad_norm": 1.875, "learning_rate": 0.0001822911455727864, "loss": 4.2693, "step": 891 }, { "epoch": 0.3700093331950638, "grad_norm": 2.28125, "learning_rate": 0.0001822711355677839, "loss": 4.4683, "step": 892 }, { "epoch": 0.37042414186456496, "grad_norm": 2.015625, "learning_rate": 0.0001822511255627814, "loss": 4.2299, "step": 893 }, { "epoch": 0.3708389505340662, "grad_norm": 2.484375, "learning_rate": 0.0001822311155577789, "loss": 4.4003, "step": 894 }, { "epoch": 0.37125375920356735, "grad_norm": 1.9453125, "learning_rate": 0.00018221110555277642, "loss": 4.2237, "step": 895 }, { "epoch": 0.37166856787306857, "grad_norm": 2.09375, "learning_rate": 0.00018219109554777388, "loss": 4.3543, "step": 896 }, { "epoch": 0.37208337654256973, "grad_norm": 2.203125, "learning_rate": 0.0001821710855427714, "loss": 4.2995, "step": 897 }, { "epoch": 0.37249818521207095, "grad_norm": 1.8515625, "learning_rate": 0.00018215107553776888, "loss": 4.2978, "step": 898 }, { "epoch": 0.3729129938815721, "grad_norm": 2.3125, "learning_rate": 0.0001821310655327664, "loss": 4.5876, "step": 899 }, { "epoch": 0.37332780255107334, "grad_norm": 2.15625, "learning_rate": 0.00018211105552776388, "loss": 4.3065, "step": 900 }, { "epoch": 0.3737426112205745, "grad_norm": 2.109375, "learning_rate": 0.0001820910455227614, "loss": 4.2268, "step": 901 }, { "epoch": 0.3741574198900757, "grad_norm": 2.140625, "learning_rate": 0.00018207103551775889, "loss": 4.4453, "step": 902 }, { "epoch": 0.3745722285595769, "grad_norm": 2.234375, "learning_rate": 0.0001820510255127564, "loss": 4.337, "step": 903 }, { "epoch": 0.3749870372290781, "grad_norm": 2.03125, "learning_rate": 0.0001820310155077539, "loss": 4.4857, "step": 904 }, { "epoch": 0.3754018458985793, "grad_norm": 2.265625, "learning_rate": 0.00018201100550275137, "loss": 4.3252, "step": 905 }, { "epoch": 0.3758166545680805, "grad_norm": 2.015625, "learning_rate": 0.00018199099549774886, "loss": 4.5325, "step": 906 }, { "epoch": 0.37623146323758166, "grad_norm": 2.5, "learning_rate": 0.00018197098549274638, "loss": 4.3182, "step": 907 }, { "epoch": 0.3766462719070829, "grad_norm": 2.125, "learning_rate": 0.0001819509754877439, "loss": 4.5372, "step": 908 }, { "epoch": 0.37706108057658405, "grad_norm": 2.125, "learning_rate": 0.00018193096548274138, "loss": 4.3708, "step": 909 }, { "epoch": 0.3774758892460852, "grad_norm": 2.171875, "learning_rate": 0.0001819109554777389, "loss": 4.2993, "step": 910 }, { "epoch": 0.37789069791558644, "grad_norm": 2.265625, "learning_rate": 0.00018189094547273638, "loss": 4.6262, "step": 911 }, { "epoch": 0.3783055065850876, "grad_norm": 2.125, "learning_rate": 0.0001818709354677339, "loss": 4.2815, "step": 912 }, { "epoch": 0.3787203152545888, "grad_norm": 2.046875, "learning_rate": 0.00018185092546273138, "loss": 4.2242, "step": 913 }, { "epoch": 0.37913512392409, "grad_norm": 2.1875, "learning_rate": 0.00018183091545772887, "loss": 4.6269, "step": 914 }, { "epoch": 0.3795499325935912, "grad_norm": 2.15625, "learning_rate": 0.00018181090545272636, "loss": 4.5047, "step": 915 }, { "epoch": 0.3799647412630924, "grad_norm": 2.203125, "learning_rate": 0.00018179089544772387, "loss": 4.348, "step": 916 }, { "epoch": 0.3803795499325936, "grad_norm": 2.171875, "learning_rate": 0.00018177088544272136, "loss": 4.3026, "step": 917 }, { "epoch": 0.38079435860209476, "grad_norm": 2.21875, "learning_rate": 0.00018175087543771887, "loss": 4.3912, "step": 918 }, { "epoch": 0.381209167271596, "grad_norm": 2.46875, "learning_rate": 0.00018173086543271636, "loss": 4.3818, "step": 919 }, { "epoch": 0.38162397594109715, "grad_norm": 1.9296875, "learning_rate": 0.00018171085542771387, "loss": 4.3962, "step": 920 }, { "epoch": 0.38203878461059837, "grad_norm": 1.9765625, "learning_rate": 0.0001816908454227114, "loss": 4.2943, "step": 921 }, { "epoch": 0.38245359328009954, "grad_norm": 2.125, "learning_rate": 0.00018167083541770887, "loss": 4.6705, "step": 922 }, { "epoch": 0.38286840194960076, "grad_norm": 2.28125, "learning_rate": 0.00018165082541270636, "loss": 4.3915, "step": 923 }, { "epoch": 0.3832832106191019, "grad_norm": 1.8828125, "learning_rate": 0.00018163081540770385, "loss": 4.2542, "step": 924 }, { "epoch": 0.38369801928860314, "grad_norm": 1.953125, "learning_rate": 0.00018161080540270136, "loss": 4.513, "step": 925 }, { "epoch": 0.3841128279581043, "grad_norm": 1.9765625, "learning_rate": 0.00018159079539769885, "loss": 4.5005, "step": 926 }, { "epoch": 0.38452763662760553, "grad_norm": 2.015625, "learning_rate": 0.00018157078539269637, "loss": 4.5435, "step": 927 }, { "epoch": 0.3849424452971067, "grad_norm": 1.96875, "learning_rate": 0.00018155077538769385, "loss": 4.5259, "step": 928 }, { "epoch": 0.3853572539666079, "grad_norm": 2.140625, "learning_rate": 0.00018153076538269137, "loss": 4.2063, "step": 929 }, { "epoch": 0.3857720626361091, "grad_norm": 1.9375, "learning_rate": 0.00018151075537768885, "loss": 4.2209, "step": 930 }, { "epoch": 0.3861868713056103, "grad_norm": 2.03125, "learning_rate": 0.00018149074537268634, "loss": 4.3294, "step": 931 }, { "epoch": 0.38660167997511147, "grad_norm": 2.046875, "learning_rate": 0.00018147073536768383, "loss": 4.2802, "step": 932 }, { "epoch": 0.3870164886446127, "grad_norm": 2.234375, "learning_rate": 0.00018145072536268134, "loss": 4.5184, "step": 933 }, { "epoch": 0.38743129731411385, "grad_norm": 2.140625, "learning_rate": 0.00018143071535767883, "loss": 4.2314, "step": 934 }, { "epoch": 0.3878461059836151, "grad_norm": 2.015625, "learning_rate": 0.00018141070535267634, "loss": 4.4718, "step": 935 }, { "epoch": 0.38826091465311624, "grad_norm": 2.015625, "learning_rate": 0.00018139069534767386, "loss": 4.5467, "step": 936 }, { "epoch": 0.38867572332261746, "grad_norm": 2.421875, "learning_rate": 0.00018137068534267135, "loss": 4.2893, "step": 937 }, { "epoch": 0.3890905319921186, "grad_norm": 2.28125, "learning_rate": 0.00018135067533766886, "loss": 4.3072, "step": 938 }, { "epoch": 0.38950534066161985, "grad_norm": 1.96875, "learning_rate": 0.00018133066533266635, "loss": 4.5215, "step": 939 }, { "epoch": 0.389920149331121, "grad_norm": 1.9296875, "learning_rate": 0.00018131065532766384, "loss": 4.2413, "step": 940 }, { "epoch": 0.39033495800062223, "grad_norm": 2.140625, "learning_rate": 0.00018129064532266132, "loss": 4.638, "step": 941 }, { "epoch": 0.3907497666701234, "grad_norm": 2.15625, "learning_rate": 0.00018127063531765884, "loss": 4.3446, "step": 942 }, { "epoch": 0.3911645753396246, "grad_norm": 1.984375, "learning_rate": 0.00018125062531265632, "loss": 4.4265, "step": 943 }, { "epoch": 0.3915793840091258, "grad_norm": 2.0625, "learning_rate": 0.00018123061530765384, "loss": 4.4044, "step": 944 }, { "epoch": 0.391994192678627, "grad_norm": 2.203125, "learning_rate": 0.00018121060530265133, "loss": 4.3234, "step": 945 }, { "epoch": 0.3924090013481282, "grad_norm": 2.09375, "learning_rate": 0.00018119059529764884, "loss": 4.3193, "step": 946 }, { "epoch": 0.3928238100176294, "grad_norm": 2.234375, "learning_rate": 0.00018117058529264633, "loss": 4.3301, "step": 947 }, { "epoch": 0.39323861868713056, "grad_norm": 2.265625, "learning_rate": 0.00018115057528764384, "loss": 4.4719, "step": 948 }, { "epoch": 0.3936534273566318, "grad_norm": 2.140625, "learning_rate": 0.00018113056528264133, "loss": 4.4303, "step": 949 }, { "epoch": 0.39406823602613295, "grad_norm": 1.9375, "learning_rate": 0.00018111055527763882, "loss": 4.3204, "step": 950 }, { "epoch": 0.39448304469563417, "grad_norm": 2.0625, "learning_rate": 0.00018109054527263633, "loss": 4.7624, "step": 951 }, { "epoch": 0.39489785336513533, "grad_norm": 2.046875, "learning_rate": 0.00018107053526763382, "loss": 4.3146, "step": 952 }, { "epoch": 0.3953126620346365, "grad_norm": 2.140625, "learning_rate": 0.00018105052526263133, "loss": 4.4074, "step": 953 }, { "epoch": 0.3957274707041377, "grad_norm": 2.125, "learning_rate": 0.00018103051525762882, "loss": 4.6061, "step": 954 }, { "epoch": 0.3961422793736389, "grad_norm": 2.1875, "learning_rate": 0.00018101050525262633, "loss": 4.3081, "step": 955 }, { "epoch": 0.3965570880431401, "grad_norm": 2.28125, "learning_rate": 0.00018099049524762382, "loss": 4.6302, "step": 956 }, { "epoch": 0.39697189671264127, "grad_norm": 1.953125, "learning_rate": 0.00018097048524262134, "loss": 4.4911, "step": 957 }, { "epoch": 0.3973867053821425, "grad_norm": 2.25, "learning_rate": 0.00018095047523761882, "loss": 4.2517, "step": 958 }, { "epoch": 0.39780151405164366, "grad_norm": 1.8359375, "learning_rate": 0.0001809304652326163, "loss": 4.2793, "step": 959 }, { "epoch": 0.3982163227211449, "grad_norm": 1.859375, "learning_rate": 0.0001809104552276138, "loss": 4.4896, "step": 960 }, { "epoch": 0.39863113139064604, "grad_norm": 2.0625, "learning_rate": 0.0001808904452226113, "loss": 4.3398, "step": 961 }, { "epoch": 0.39904594006014726, "grad_norm": 2.234375, "learning_rate": 0.0001808704352176088, "loss": 4.3402, "step": 962 }, { "epoch": 0.39946074872964843, "grad_norm": 2.046875, "learning_rate": 0.00018085042521260631, "loss": 4.3537, "step": 963 }, { "epoch": 0.39987555739914965, "grad_norm": 1.984375, "learning_rate": 0.00018083041520760383, "loss": 4.4073, "step": 964 }, { "epoch": 0.4002903660686508, "grad_norm": 2.203125, "learning_rate": 0.00018081040520260132, "loss": 4.3917, "step": 965 }, { "epoch": 0.40070517473815204, "grad_norm": 2.046875, "learning_rate": 0.0001807903951975988, "loss": 4.2131, "step": 966 }, { "epoch": 0.4011199834076532, "grad_norm": 2.140625, "learning_rate": 0.0001807703851925963, "loss": 4.4373, "step": 967 }, { "epoch": 0.4015347920771544, "grad_norm": 1.9296875, "learning_rate": 0.0001807503751875938, "loss": 4.756, "step": 968 }, { "epoch": 0.4019496007466556, "grad_norm": 2.140625, "learning_rate": 0.0001807303651825913, "loss": 4.3509, "step": 969 }, { "epoch": 0.4023644094161568, "grad_norm": 1.921875, "learning_rate": 0.0001807103551775888, "loss": 4.5064, "step": 970 }, { "epoch": 0.402779218085658, "grad_norm": 2.296875, "learning_rate": 0.0001806903451725863, "loss": 4.3962, "step": 971 }, { "epoch": 0.4031940267551592, "grad_norm": 1.9609375, "learning_rate": 0.0001806703351675838, "loss": 4.449, "step": 972 }, { "epoch": 0.40360883542466036, "grad_norm": 2.203125, "learning_rate": 0.0001806503251625813, "loss": 4.284, "step": 973 }, { "epoch": 0.4040236440941616, "grad_norm": 2.09375, "learning_rate": 0.0001806303151575788, "loss": 4.4059, "step": 974 }, { "epoch": 0.40443845276366275, "grad_norm": 2.09375, "learning_rate": 0.0001806103051525763, "loss": 4.4427, "step": 975 }, { "epoch": 0.40485326143316397, "grad_norm": 1.953125, "learning_rate": 0.00018059029514757378, "loss": 4.305, "step": 976 }, { "epoch": 0.40526807010266513, "grad_norm": 1.9765625, "learning_rate": 0.0001805702851425713, "loss": 4.3912, "step": 977 }, { "epoch": 0.40568287877216636, "grad_norm": 1.875, "learning_rate": 0.00018055027513756879, "loss": 4.2529, "step": 978 }, { "epoch": 0.4060976874416675, "grad_norm": 2.03125, "learning_rate": 0.0001805302651325663, "loss": 4.262, "step": 979 }, { "epoch": 0.40651249611116874, "grad_norm": 1.8671875, "learning_rate": 0.0001805102551275638, "loss": 4.45, "step": 980 }, { "epoch": 0.4069273047806699, "grad_norm": 2.09375, "learning_rate": 0.0001804902451225613, "loss": 4.4505, "step": 981 }, { "epoch": 0.40734211345017113, "grad_norm": 2.046875, "learning_rate": 0.0001804702351175588, "loss": 4.3363, "step": 982 }, { "epoch": 0.4077569221196723, "grad_norm": 2.203125, "learning_rate": 0.0001804502251125563, "loss": 4.4308, "step": 983 }, { "epoch": 0.4081717307891735, "grad_norm": 1.9609375, "learning_rate": 0.0001804302151075538, "loss": 4.7916, "step": 984 }, { "epoch": 0.4085865394586747, "grad_norm": 1.9375, "learning_rate": 0.00018041020510255128, "loss": 4.2049, "step": 985 }, { "epoch": 0.4090013481281759, "grad_norm": 2.078125, "learning_rate": 0.00018039019509754877, "loss": 4.3206, "step": 986 }, { "epoch": 0.40941615679767707, "grad_norm": 2.03125, "learning_rate": 0.00018037018509254628, "loss": 4.4308, "step": 987 }, { "epoch": 0.4098309654671783, "grad_norm": 2.0, "learning_rate": 0.00018035017508754377, "loss": 4.2344, "step": 988 }, { "epoch": 0.41024577413667945, "grad_norm": 1.8984375, "learning_rate": 0.00018033016508254128, "loss": 4.368, "step": 989 }, { "epoch": 0.4106605828061807, "grad_norm": 2.125, "learning_rate": 0.00018031015507753877, "loss": 4.6057, "step": 990 }, { "epoch": 0.41107539147568184, "grad_norm": 2.140625, "learning_rate": 0.00018029014507253628, "loss": 4.1587, "step": 991 }, { "epoch": 0.41149020014518306, "grad_norm": 2.234375, "learning_rate": 0.0001802701350675338, "loss": 4.3412, "step": 992 }, { "epoch": 0.4119050088146842, "grad_norm": 2.109375, "learning_rate": 0.00018025012506253128, "loss": 4.5148, "step": 993 }, { "epoch": 0.41231981748418545, "grad_norm": 1.90625, "learning_rate": 0.00018023011505752877, "loss": 4.2939, "step": 994 }, { "epoch": 0.4127346261536866, "grad_norm": 1.8984375, "learning_rate": 0.00018021010505252626, "loss": 4.3407, "step": 995 }, { "epoch": 0.4131494348231878, "grad_norm": 1.9609375, "learning_rate": 0.00018019009504752377, "loss": 4.3568, "step": 996 }, { "epoch": 0.413564243492689, "grad_norm": 1.8828125, "learning_rate": 0.00018017008504252126, "loss": 4.1772, "step": 997 }, { "epoch": 0.41397905216219016, "grad_norm": 1.8984375, "learning_rate": 0.00018015007503751878, "loss": 4.1354, "step": 998 }, { "epoch": 0.4143938608316914, "grad_norm": 2.09375, "learning_rate": 0.00018013006503251626, "loss": 4.259, "step": 999 }, { "epoch": 0.41480866950119255, "grad_norm": 2.046875, "learning_rate": 0.00018011005502751378, "loss": 4.3353, "step": 1000 }, { "epoch": 0.41522347817069377, "grad_norm": 1.859375, "learning_rate": 0.00018009004502251126, "loss": 4.305, "step": 1001 }, { "epoch": 0.41563828684019494, "grad_norm": 1.9921875, "learning_rate": 0.00018007003501750875, "loss": 4.4013, "step": 1002 }, { "epoch": 0.41605309550969616, "grad_norm": 2.1875, "learning_rate": 0.00018005002501250624, "loss": 4.339, "step": 1003 }, { "epoch": 0.4164679041791973, "grad_norm": 2.0625, "learning_rate": 0.00018003001500750375, "loss": 4.3102, "step": 1004 }, { "epoch": 0.41688271284869854, "grad_norm": 2.0, "learning_rate": 0.00018001000500250127, "loss": 4.2317, "step": 1005 }, { "epoch": 0.4172975215181997, "grad_norm": 1.953125, "learning_rate": 0.00017998999499749875, "loss": 4.2712, "step": 1006 }, { "epoch": 0.41771233018770093, "grad_norm": 2.140625, "learning_rate": 0.00017996998499249627, "loss": 4.3149, "step": 1007 }, { "epoch": 0.4181271388572021, "grad_norm": 2.203125, "learning_rate": 0.00017994997498749376, "loss": 4.2377, "step": 1008 }, { "epoch": 0.4185419475267033, "grad_norm": 2.015625, "learning_rate": 0.00017992996498249127, "loss": 4.7052, "step": 1009 }, { "epoch": 0.4189567561962045, "grad_norm": 1.9609375, "learning_rate": 0.00017990995497748876, "loss": 4.1802, "step": 1010 }, { "epoch": 0.4193715648657057, "grad_norm": 1.8984375, "learning_rate": 0.00017988994497248625, "loss": 4.6006, "step": 1011 }, { "epoch": 0.41978637353520687, "grad_norm": 2.03125, "learning_rate": 0.00017986993496748373, "loss": 4.5496, "step": 1012 }, { "epoch": 0.4202011822047081, "grad_norm": 2.046875, "learning_rate": 0.00017984992496248125, "loss": 4.3438, "step": 1013 }, { "epoch": 0.42061599087420926, "grad_norm": 2.171875, "learning_rate": 0.00017982991495747873, "loss": 4.3297, "step": 1014 }, { "epoch": 0.4210307995437105, "grad_norm": 2.015625, "learning_rate": 0.00017980990495247625, "loss": 4.4914, "step": 1015 }, { "epoch": 0.42144560821321164, "grad_norm": 1.9296875, "learning_rate": 0.00017978989494747374, "loss": 4.2198, "step": 1016 }, { "epoch": 0.42186041688271286, "grad_norm": 2.125, "learning_rate": 0.00017976988494247125, "loss": 4.7153, "step": 1017 }, { "epoch": 0.42227522555221403, "grad_norm": 2.046875, "learning_rate": 0.00017974987493746876, "loss": 4.4418, "step": 1018 }, { "epoch": 0.42269003422171525, "grad_norm": 1.9765625, "learning_rate": 0.00017972986493246625, "loss": 4.1992, "step": 1019 }, { "epoch": 0.4231048428912164, "grad_norm": 2.109375, "learning_rate": 0.00017970985492746374, "loss": 4.2828, "step": 1020 }, { "epoch": 0.42351965156071764, "grad_norm": 2.203125, "learning_rate": 0.00017968984492246123, "loss": 4.5623, "step": 1021 }, { "epoch": 0.4239344602302188, "grad_norm": 2.171875, "learning_rate": 0.00017966983491745874, "loss": 4.2724, "step": 1022 }, { "epoch": 0.42434926889972, "grad_norm": 2.0625, "learning_rate": 0.00017964982491245623, "loss": 4.6182, "step": 1023 }, { "epoch": 0.4247640775692212, "grad_norm": 2.09375, "learning_rate": 0.00017962981490745374, "loss": 4.2161, "step": 1024 }, { "epoch": 0.4251788862387224, "grad_norm": 2.03125, "learning_rate": 0.00017960980490245123, "loss": 4.3507, "step": 1025 }, { "epoch": 0.4255936949082236, "grad_norm": 2.15625, "learning_rate": 0.00017958979489744874, "loss": 4.3925, "step": 1026 }, { "epoch": 0.4260085035777248, "grad_norm": 1.90625, "learning_rate": 0.00017956978489244623, "loss": 4.4245, "step": 1027 }, { "epoch": 0.42642331224722596, "grad_norm": 1.9921875, "learning_rate": 0.00017954977488744375, "loss": 4.5282, "step": 1028 }, { "epoch": 0.4268381209167272, "grad_norm": 1.9765625, "learning_rate": 0.00017952976488244123, "loss": 4.3807, "step": 1029 }, { "epoch": 0.42725292958622835, "grad_norm": 2.0625, "learning_rate": 0.00017950975487743872, "loss": 4.4721, "step": 1030 }, { "epoch": 0.42766773825572957, "grad_norm": 1.9921875, "learning_rate": 0.0001794897448724362, "loss": 4.3272, "step": 1031 }, { "epoch": 0.42808254692523073, "grad_norm": 2.078125, "learning_rate": 0.00017946973486743372, "loss": 4.4598, "step": 1032 }, { "epoch": 0.42849735559473195, "grad_norm": 1.984375, "learning_rate": 0.00017944972486243124, "loss": 4.299, "step": 1033 }, { "epoch": 0.4289121642642331, "grad_norm": 2.28125, "learning_rate": 0.00017942971485742872, "loss": 4.4141, "step": 1034 }, { "epoch": 0.42932697293373434, "grad_norm": 2.03125, "learning_rate": 0.00017940970485242624, "loss": 4.2519, "step": 1035 }, { "epoch": 0.4297417816032355, "grad_norm": 2.078125, "learning_rate": 0.00017938969484742373, "loss": 4.3656, "step": 1036 }, { "epoch": 0.4301565902727367, "grad_norm": 2.421875, "learning_rate": 0.0001793696848424212, "loss": 4.6236, "step": 1037 }, { "epoch": 0.4305713989422379, "grad_norm": 2.375, "learning_rate": 0.0001793496748374187, "loss": 4.275, "step": 1038 }, { "epoch": 0.43098620761173906, "grad_norm": 1.9140625, "learning_rate": 0.00017932966483241621, "loss": 4.1828, "step": 1039 }, { "epoch": 0.4314010162812403, "grad_norm": 1.9921875, "learning_rate": 0.0001793096548274137, "loss": 4.4596, "step": 1040 }, { "epoch": 0.43181582495074144, "grad_norm": 2.15625, "learning_rate": 0.00017928964482241122, "loss": 4.4108, "step": 1041 }, { "epoch": 0.43223063362024267, "grad_norm": 2.09375, "learning_rate": 0.0001792696348174087, "loss": 4.5056, "step": 1042 }, { "epoch": 0.43264544228974383, "grad_norm": 2.078125, "learning_rate": 0.00017924962481240622, "loss": 4.3782, "step": 1043 }, { "epoch": 0.43306025095924505, "grad_norm": 2.1875, "learning_rate": 0.0001792296148074037, "loss": 4.1952, "step": 1044 }, { "epoch": 0.4334750596287462, "grad_norm": 1.9453125, "learning_rate": 0.00017920960480240122, "loss": 4.2819, "step": 1045 }, { "epoch": 0.43388986829824744, "grad_norm": 2.09375, "learning_rate": 0.0001791895947973987, "loss": 4.26, "step": 1046 }, { "epoch": 0.4343046769677486, "grad_norm": 2.046875, "learning_rate": 0.0001791695847923962, "loss": 4.4574, "step": 1047 }, { "epoch": 0.4347194856372498, "grad_norm": 2.046875, "learning_rate": 0.0001791495747873937, "loss": 4.2924, "step": 1048 }, { "epoch": 0.435134294306751, "grad_norm": 2.0, "learning_rate": 0.0001791295647823912, "loss": 4.1717, "step": 1049 }, { "epoch": 0.4355491029762522, "grad_norm": 2.21875, "learning_rate": 0.0001791095547773887, "loss": 4.3903, "step": 1050 }, { "epoch": 0.4359639116457534, "grad_norm": 2.09375, "learning_rate": 0.0001790895447723862, "loss": 4.2457, "step": 1051 }, { "epoch": 0.4363787203152546, "grad_norm": 2.25, "learning_rate": 0.0001790695347673837, "loss": 4.597, "step": 1052 }, { "epoch": 0.43679352898475576, "grad_norm": 2.140625, "learning_rate": 0.0001790495247623812, "loss": 4.5183, "step": 1053 }, { "epoch": 0.437208337654257, "grad_norm": 2.1875, "learning_rate": 0.00017902951475737871, "loss": 4.2368, "step": 1054 }, { "epoch": 0.43762314632375815, "grad_norm": 1.9453125, "learning_rate": 0.0001790095047523762, "loss": 4.271, "step": 1055 }, { "epoch": 0.43803795499325937, "grad_norm": 2.203125, "learning_rate": 0.0001789894947473737, "loss": 4.3001, "step": 1056 }, { "epoch": 0.43845276366276054, "grad_norm": 2.046875, "learning_rate": 0.00017896948474237118, "loss": 4.5858, "step": 1057 }, { "epoch": 0.43886757233226176, "grad_norm": 2.1875, "learning_rate": 0.0001789494747373687, "loss": 4.5801, "step": 1058 }, { "epoch": 0.4392823810017629, "grad_norm": 2.140625, "learning_rate": 0.00017892946473236618, "loss": 4.5272, "step": 1059 }, { "epoch": 0.43969718967126414, "grad_norm": 2.09375, "learning_rate": 0.0001789094547273637, "loss": 4.7148, "step": 1060 }, { "epoch": 0.4401119983407653, "grad_norm": 1.9921875, "learning_rate": 0.0001788894447223612, "loss": 4.4599, "step": 1061 }, { "epoch": 0.44052680701026653, "grad_norm": 1.96875, "learning_rate": 0.0001788694347173587, "loss": 4.2887, "step": 1062 }, { "epoch": 0.4409416156797677, "grad_norm": 2.03125, "learning_rate": 0.0001788494247123562, "loss": 4.5133, "step": 1063 }, { "epoch": 0.4413564243492689, "grad_norm": 2.03125, "learning_rate": 0.0001788294147073537, "loss": 4.1968, "step": 1064 }, { "epoch": 0.4417712330187701, "grad_norm": 2.5625, "learning_rate": 0.00017880940470235118, "loss": 4.5081, "step": 1065 }, { "epoch": 0.4421860416882713, "grad_norm": 2.28125, "learning_rate": 0.00017878939469734867, "loss": 4.2804, "step": 1066 }, { "epoch": 0.44260085035777247, "grad_norm": 1.984375, "learning_rate": 0.00017876938469234618, "loss": 4.2879, "step": 1067 }, { "epoch": 0.4430156590272737, "grad_norm": 2.15625, "learning_rate": 0.00017874937468734367, "loss": 4.3384, "step": 1068 }, { "epoch": 0.44343046769677485, "grad_norm": 2.421875, "learning_rate": 0.00017872936468234119, "loss": 4.4903, "step": 1069 }, { "epoch": 0.4438452763662761, "grad_norm": 2.140625, "learning_rate": 0.00017870935467733867, "loss": 4.4126, "step": 1070 }, { "epoch": 0.44426008503577724, "grad_norm": 2.125, "learning_rate": 0.0001786893446723362, "loss": 4.5079, "step": 1071 }, { "epoch": 0.44467489370527846, "grad_norm": 2.0625, "learning_rate": 0.00017866933466733367, "loss": 4.3872, "step": 1072 }, { "epoch": 0.4450897023747796, "grad_norm": 1.984375, "learning_rate": 0.00017864932466233116, "loss": 4.7223, "step": 1073 }, { "epoch": 0.44550451104428085, "grad_norm": 2.171875, "learning_rate": 0.00017862931465732868, "loss": 4.2778, "step": 1074 }, { "epoch": 0.445919319713782, "grad_norm": 1.90625, "learning_rate": 0.00017860930465232616, "loss": 4.4252, "step": 1075 }, { "epoch": 0.44633412838328324, "grad_norm": 2.125, "learning_rate": 0.00017858929464732368, "loss": 4.5733, "step": 1076 }, { "epoch": 0.4467489370527844, "grad_norm": 2.296875, "learning_rate": 0.00017856928464232117, "loss": 4.4882, "step": 1077 }, { "epoch": 0.4471637457222856, "grad_norm": 1.875, "learning_rate": 0.00017854927463731868, "loss": 4.337, "step": 1078 }, { "epoch": 0.4475785543917868, "grad_norm": 1.8828125, "learning_rate": 0.00017852926463231617, "loss": 4.1409, "step": 1079 }, { "epoch": 0.447993363061288, "grad_norm": 2.109375, "learning_rate": 0.00017850925462731368, "loss": 4.4273, "step": 1080 }, { "epoch": 0.4484081717307892, "grad_norm": 2.28125, "learning_rate": 0.00017848924462231117, "loss": 4.3459, "step": 1081 }, { "epoch": 0.44882298040029034, "grad_norm": 2.078125, "learning_rate": 0.00017846923461730866, "loss": 4.4212, "step": 1082 }, { "epoch": 0.44923778906979156, "grad_norm": 2.09375, "learning_rate": 0.00017844922461230614, "loss": 4.6168, "step": 1083 }, { "epoch": 0.4496525977392927, "grad_norm": 1.828125, "learning_rate": 0.00017842921460730366, "loss": 4.4838, "step": 1084 }, { "epoch": 0.45006740640879395, "grad_norm": 1.8671875, "learning_rate": 0.00017840920460230114, "loss": 4.4301, "step": 1085 }, { "epoch": 0.4504822150782951, "grad_norm": 1.8984375, "learning_rate": 0.00017838919459729866, "loss": 4.4215, "step": 1086 }, { "epoch": 0.45089702374779633, "grad_norm": 2.03125, "learning_rate": 0.00017836918459229615, "loss": 4.3412, "step": 1087 }, { "epoch": 0.4513118324172975, "grad_norm": 1.8046875, "learning_rate": 0.00017834917458729366, "loss": 4.3783, "step": 1088 }, { "epoch": 0.4517266410867987, "grad_norm": 2.328125, "learning_rate": 0.00017832916458229117, "loss": 4.6112, "step": 1089 }, { "epoch": 0.4521414497562999, "grad_norm": 2.140625, "learning_rate": 0.00017830915457728866, "loss": 4.3509, "step": 1090 }, { "epoch": 0.4525562584258011, "grad_norm": 2.078125, "learning_rate": 0.00017828914457228615, "loss": 4.383, "step": 1091 }, { "epoch": 0.45297106709530227, "grad_norm": 1.9140625, "learning_rate": 0.00017826913456728364, "loss": 4.2993, "step": 1092 }, { "epoch": 0.4533858757648035, "grad_norm": 1.953125, "learning_rate": 0.00017824912456228115, "loss": 4.2683, "step": 1093 }, { "epoch": 0.45380068443430466, "grad_norm": 1.8984375, "learning_rate": 0.00017822911455727864, "loss": 4.5543, "step": 1094 }, { "epoch": 0.4542154931038059, "grad_norm": 1.9296875, "learning_rate": 0.00017820910455227615, "loss": 4.344, "step": 1095 }, { "epoch": 0.45463030177330704, "grad_norm": 1.984375, "learning_rate": 0.00017818909454727364, "loss": 4.7335, "step": 1096 }, { "epoch": 0.45504511044280826, "grad_norm": 2.171875, "learning_rate": 0.00017816908454227115, "loss": 4.3833, "step": 1097 }, { "epoch": 0.45545991911230943, "grad_norm": 1.8203125, "learning_rate": 0.00017814907453726864, "loss": 4.5725, "step": 1098 }, { "epoch": 0.45587472778181065, "grad_norm": 2.015625, "learning_rate": 0.00017812906453226616, "loss": 4.5028, "step": 1099 }, { "epoch": 0.4562895364513118, "grad_norm": 1.9765625, "learning_rate": 0.00017810905452726364, "loss": 4.2995, "step": 1100 }, { "epoch": 0.45670434512081304, "grad_norm": 1.9765625, "learning_rate": 0.00017808904452226113, "loss": 4.2497, "step": 1101 }, { "epoch": 0.4571191537903142, "grad_norm": 1.9453125, "learning_rate": 0.00017806903451725865, "loss": 4.4419, "step": 1102 }, { "epoch": 0.4575339624598154, "grad_norm": 1.859375, "learning_rate": 0.00017804902451225613, "loss": 4.4091, "step": 1103 }, { "epoch": 0.4579487711293166, "grad_norm": 1.875, "learning_rate": 0.00017802901450725365, "loss": 4.2748, "step": 1104 }, { "epoch": 0.4583635797988178, "grad_norm": 1.921875, "learning_rate": 0.00017800900450225113, "loss": 4.5101, "step": 1105 }, { "epoch": 0.458778388468319, "grad_norm": 2.21875, "learning_rate": 0.00017798899449724865, "loss": 4.3646, "step": 1106 }, { "epoch": 0.4591931971378202, "grad_norm": 2.078125, "learning_rate": 0.00017796898449224614, "loss": 4.4547, "step": 1107 }, { "epoch": 0.45960800580732136, "grad_norm": 2.046875, "learning_rate": 0.00017794897448724362, "loss": 4.3562, "step": 1108 }, { "epoch": 0.4600228144768226, "grad_norm": 2.109375, "learning_rate": 0.0001779289644822411, "loss": 4.1776, "step": 1109 }, { "epoch": 0.46043762314632375, "grad_norm": 1.78125, "learning_rate": 0.00017790895447723862, "loss": 4.2128, "step": 1110 }, { "epoch": 0.46085243181582497, "grad_norm": 2.046875, "learning_rate": 0.0001778889444722361, "loss": 4.2559, "step": 1111 }, { "epoch": 0.46126724048532614, "grad_norm": 1.859375, "learning_rate": 0.00017786893446723363, "loss": 4.4851, "step": 1112 }, { "epoch": 0.46168204915482736, "grad_norm": 2.375, "learning_rate": 0.00017784892446223111, "loss": 4.4802, "step": 1113 }, { "epoch": 0.4620968578243285, "grad_norm": 1.9921875, "learning_rate": 0.00017782891445722863, "loss": 4.2383, "step": 1114 }, { "epoch": 0.46251166649382974, "grad_norm": 2.046875, "learning_rate": 0.00017780890445222612, "loss": 4.5444, "step": 1115 }, { "epoch": 0.4629264751633309, "grad_norm": 2.0, "learning_rate": 0.00017778889444722363, "loss": 4.4736, "step": 1116 }, { "epoch": 0.46334128383283213, "grad_norm": 2.234375, "learning_rate": 0.00017776888444222112, "loss": 4.37, "step": 1117 }, { "epoch": 0.4637560925023333, "grad_norm": 1.921875, "learning_rate": 0.0001777488744372186, "loss": 4.3285, "step": 1118 }, { "epoch": 0.4641709011718345, "grad_norm": 2.0625, "learning_rate": 0.00017772886443221612, "loss": 4.3134, "step": 1119 }, { "epoch": 0.4645857098413357, "grad_norm": 2.0, "learning_rate": 0.0001777088544272136, "loss": 4.3374, "step": 1120 }, { "epoch": 0.4650005185108369, "grad_norm": 1.7890625, "learning_rate": 0.00017768884442221112, "loss": 4.2235, "step": 1121 }, { "epoch": 0.46541532718033807, "grad_norm": 2.59375, "learning_rate": 0.0001776688344172086, "loss": 4.0565, "step": 1122 }, { "epoch": 0.4658301358498393, "grad_norm": 1.859375, "learning_rate": 0.00017764882441220612, "loss": 4.6, "step": 1123 }, { "epoch": 0.46624494451934045, "grad_norm": 1.8828125, "learning_rate": 0.0001776288144072036, "loss": 4.2149, "step": 1124 }, { "epoch": 0.4666597531888416, "grad_norm": 2.140625, "learning_rate": 0.00017760880440220112, "loss": 4.4413, "step": 1125 }, { "epoch": 0.46707456185834284, "grad_norm": 1.7734375, "learning_rate": 0.0001775887943971986, "loss": 4.3745, "step": 1126 }, { "epoch": 0.467489370527844, "grad_norm": 2.09375, "learning_rate": 0.0001775687843921961, "loss": 4.0623, "step": 1127 }, { "epoch": 0.4679041791973452, "grad_norm": 1.8984375, "learning_rate": 0.00017754877438719359, "loss": 4.3665, "step": 1128 }, { "epoch": 0.4683189878668464, "grad_norm": 2.0, "learning_rate": 0.0001775287643821911, "loss": 4.5509, "step": 1129 }, { "epoch": 0.4687337965363476, "grad_norm": 2.03125, "learning_rate": 0.00017750875437718861, "loss": 4.2777, "step": 1130 }, { "epoch": 0.4691486052058488, "grad_norm": 1.8828125, "learning_rate": 0.0001774887443721861, "loss": 4.3339, "step": 1131 }, { "epoch": 0.46956341387535, "grad_norm": 1.984375, "learning_rate": 0.00017746873436718362, "loss": 4.7568, "step": 1132 }, { "epoch": 0.46997822254485117, "grad_norm": 2.1875, "learning_rate": 0.0001774487243621811, "loss": 4.5942, "step": 1133 }, { "epoch": 0.4703930312143524, "grad_norm": 2.4375, "learning_rate": 0.00017742871435717862, "loss": 4.4972, "step": 1134 }, { "epoch": 0.47080783988385355, "grad_norm": 1.96875, "learning_rate": 0.0001774087043521761, "loss": 4.5464, "step": 1135 }, { "epoch": 0.4712226485533548, "grad_norm": 2.375, "learning_rate": 0.0001773886943471736, "loss": 4.3029, "step": 1136 }, { "epoch": 0.47163745722285594, "grad_norm": 2.09375, "learning_rate": 0.00017736868434217108, "loss": 4.1835, "step": 1137 }, { "epoch": 0.47205226589235716, "grad_norm": 2.34375, "learning_rate": 0.0001773486743371686, "loss": 4.3698, "step": 1138 }, { "epoch": 0.4724670745618583, "grad_norm": 2.125, "learning_rate": 0.00017732866433216608, "loss": 4.4014, "step": 1139 }, { "epoch": 0.47288188323135955, "grad_norm": 1.9921875, "learning_rate": 0.0001773086543271636, "loss": 4.3377, "step": 1140 }, { "epoch": 0.4732966919008607, "grad_norm": 1.9140625, "learning_rate": 0.00017728864432216108, "loss": 4.4091, "step": 1141 }, { "epoch": 0.47371150057036193, "grad_norm": 1.96875, "learning_rate": 0.0001772686343171586, "loss": 4.3561, "step": 1142 }, { "epoch": 0.4741263092398631, "grad_norm": 1.9609375, "learning_rate": 0.00017724862431215608, "loss": 4.488, "step": 1143 }, { "epoch": 0.4745411179093643, "grad_norm": 1.9140625, "learning_rate": 0.00017722861430715357, "loss": 4.5677, "step": 1144 }, { "epoch": 0.4749559265788655, "grad_norm": 2.125, "learning_rate": 0.00017720860430215109, "loss": 4.376, "step": 1145 }, { "epoch": 0.4753707352483667, "grad_norm": 2.0625, "learning_rate": 0.00017718859429714857, "loss": 4.4235, "step": 1146 }, { "epoch": 0.47578554391786787, "grad_norm": 2.125, "learning_rate": 0.0001771685842921461, "loss": 4.4235, "step": 1147 }, { "epoch": 0.4762003525873691, "grad_norm": 2.015625, "learning_rate": 0.00017714857428714358, "loss": 4.5709, "step": 1148 }, { "epoch": 0.47661516125687026, "grad_norm": 2.0625, "learning_rate": 0.0001771285642821411, "loss": 4.4402, "step": 1149 }, { "epoch": 0.4770299699263715, "grad_norm": 2.125, "learning_rate": 0.00017710855427713858, "loss": 4.1787, "step": 1150 }, { "epoch": 0.47744477859587264, "grad_norm": 1.9375, "learning_rate": 0.0001770885442721361, "loss": 4.3755, "step": 1151 }, { "epoch": 0.47785958726537386, "grad_norm": 2.46875, "learning_rate": 0.00017706853426713358, "loss": 4.7307, "step": 1152 }, { "epoch": 0.47827439593487503, "grad_norm": 1.9296875, "learning_rate": 0.00017704852426213107, "loss": 4.4037, "step": 1153 }, { "epoch": 0.47868920460437625, "grad_norm": 2.09375, "learning_rate": 0.00017702851425712855, "loss": 4.5645, "step": 1154 }, { "epoch": 0.4791040132738774, "grad_norm": 1.9375, "learning_rate": 0.00017700850425212607, "loss": 4.4413, "step": 1155 }, { "epoch": 0.47951882194337864, "grad_norm": 2.21875, "learning_rate": 0.00017698849424712355, "loss": 4.189, "step": 1156 }, { "epoch": 0.4799336306128798, "grad_norm": 2.078125, "learning_rate": 0.00017696848424212107, "loss": 4.4323, "step": 1157 }, { "epoch": 0.480348439282381, "grad_norm": 1.8515625, "learning_rate": 0.00017694847423711858, "loss": 4.2657, "step": 1158 }, { "epoch": 0.4807632479518822, "grad_norm": 2.09375, "learning_rate": 0.00017692846423211607, "loss": 4.6062, "step": 1159 }, { "epoch": 0.4811780566213834, "grad_norm": 2.375, "learning_rate": 0.00017690845422711359, "loss": 4.5159, "step": 1160 }, { "epoch": 0.4815928652908846, "grad_norm": 1.9921875, "learning_rate": 0.00017688844422211107, "loss": 4.4225, "step": 1161 }, { "epoch": 0.4820076739603858, "grad_norm": 1.796875, "learning_rate": 0.00017686843421710856, "loss": 4.6641, "step": 1162 }, { "epoch": 0.48242248262988696, "grad_norm": 1.890625, "learning_rate": 0.00017684842421210605, "loss": 4.3169, "step": 1163 }, { "epoch": 0.4828372912993882, "grad_norm": 2.015625, "learning_rate": 0.00017682841420710356, "loss": 4.2844, "step": 1164 }, { "epoch": 0.48325209996888935, "grad_norm": 2.015625, "learning_rate": 0.00017680840420210105, "loss": 4.3329, "step": 1165 }, { "epoch": 0.48366690863839057, "grad_norm": 2.0625, "learning_rate": 0.00017678839419709856, "loss": 4.4835, "step": 1166 }, { "epoch": 0.48408171730789173, "grad_norm": 2.0625, "learning_rate": 0.00017676838419209605, "loss": 4.3825, "step": 1167 }, { "epoch": 0.4844965259773929, "grad_norm": 2.015625, "learning_rate": 0.00017674837418709356, "loss": 4.5153, "step": 1168 }, { "epoch": 0.4849113346468941, "grad_norm": 2.140625, "learning_rate": 0.00017672836418209105, "loss": 4.3609, "step": 1169 }, { "epoch": 0.4853261433163953, "grad_norm": 2.4375, "learning_rate": 0.00017670835417708857, "loss": 4.3737, "step": 1170 }, { "epoch": 0.4857409519858965, "grad_norm": 2.1875, "learning_rate": 0.00017668834417208605, "loss": 4.3602, "step": 1171 }, { "epoch": 0.4861557606553977, "grad_norm": 2.234375, "learning_rate": 0.00017666833416708354, "loss": 4.3427, "step": 1172 }, { "epoch": 0.4865705693248989, "grad_norm": 2.234375, "learning_rate": 0.00017664832416208106, "loss": 4.4172, "step": 1173 }, { "epoch": 0.48698537799440006, "grad_norm": 2.078125, "learning_rate": 0.00017662831415707854, "loss": 4.3782, "step": 1174 }, { "epoch": 0.4874001866639013, "grad_norm": 2.375, "learning_rate": 0.00017660830415207606, "loss": 4.8118, "step": 1175 }, { "epoch": 0.48781499533340245, "grad_norm": 2.109375, "learning_rate": 0.00017658829414707354, "loss": 4.6522, "step": 1176 }, { "epoch": 0.48822980400290367, "grad_norm": 1.7421875, "learning_rate": 0.00017656828414207106, "loss": 4.2258, "step": 1177 }, { "epoch": 0.48864461267240483, "grad_norm": 1.8359375, "learning_rate": 0.00017654827413706855, "loss": 4.2992, "step": 1178 }, { "epoch": 0.48905942134190605, "grad_norm": 2.078125, "learning_rate": 0.00017652826413206603, "loss": 4.1931, "step": 1179 }, { "epoch": 0.4894742300114072, "grad_norm": 2.03125, "learning_rate": 0.00017650825412706352, "loss": 4.4329, "step": 1180 }, { "epoch": 0.48988903868090844, "grad_norm": 2.3125, "learning_rate": 0.00017648824412206103, "loss": 4.3167, "step": 1181 }, { "epoch": 0.4903038473504096, "grad_norm": 2.15625, "learning_rate": 0.00017646823411705852, "loss": 4.4636, "step": 1182 }, { "epoch": 0.4907186560199108, "grad_norm": 2.359375, "learning_rate": 0.00017644822411205604, "loss": 4.1785, "step": 1183 }, { "epoch": 0.491133464689412, "grad_norm": 2.5, "learning_rate": 0.00017642821410705352, "loss": 4.3716, "step": 1184 }, { "epoch": 0.4915482733589132, "grad_norm": 2.046875, "learning_rate": 0.00017640820410205104, "loss": 4.6145, "step": 1185 }, { "epoch": 0.4919630820284144, "grad_norm": 2.109375, "learning_rate": 0.00017638819409704855, "loss": 4.2539, "step": 1186 }, { "epoch": 0.4923778906979156, "grad_norm": 1.9375, "learning_rate": 0.00017636818409204604, "loss": 4.4768, "step": 1187 }, { "epoch": 0.49279269936741676, "grad_norm": 2.125, "learning_rate": 0.00017634817408704353, "loss": 4.3121, "step": 1188 }, { "epoch": 0.493207508036918, "grad_norm": 2.0625, "learning_rate": 0.00017632816408204101, "loss": 4.2803, "step": 1189 }, { "epoch": 0.49362231670641915, "grad_norm": 2.21875, "learning_rate": 0.00017630815407703853, "loss": 4.81, "step": 1190 }, { "epoch": 0.49403712537592037, "grad_norm": 1.8203125, "learning_rate": 0.00017628814407203602, "loss": 4.5307, "step": 1191 }, { "epoch": 0.49445193404542154, "grad_norm": 1.90625, "learning_rate": 0.00017626813406703353, "loss": 4.4134, "step": 1192 }, { "epoch": 0.49486674271492276, "grad_norm": 2.015625, "learning_rate": 0.00017624812406203102, "loss": 4.3406, "step": 1193 }, { "epoch": 0.4952815513844239, "grad_norm": 1.9453125, "learning_rate": 0.00017622811405702853, "loss": 4.2751, "step": 1194 }, { "epoch": 0.49569636005392514, "grad_norm": 2.125, "learning_rate": 0.00017620810405202602, "loss": 4.3858, "step": 1195 }, { "epoch": 0.4961111687234263, "grad_norm": 1.96875, "learning_rate": 0.00017618809404702353, "loss": 4.4253, "step": 1196 }, { "epoch": 0.49652597739292753, "grad_norm": 1.8515625, "learning_rate": 0.00017616808404202102, "loss": 4.4802, "step": 1197 }, { "epoch": 0.4969407860624287, "grad_norm": 1.875, "learning_rate": 0.0001761480740370185, "loss": 4.2476, "step": 1198 }, { "epoch": 0.4973555947319299, "grad_norm": 1.8984375, "learning_rate": 0.00017612806403201602, "loss": 4.5804, "step": 1199 }, { "epoch": 0.4977704034014311, "grad_norm": 2.28125, "learning_rate": 0.0001761080540270135, "loss": 4.3889, "step": 1200 }, { "epoch": 0.4981852120709323, "grad_norm": 1.7890625, "learning_rate": 0.00017608804402201102, "loss": 4.2759, "step": 1201 }, { "epoch": 0.49860002074043347, "grad_norm": 1.8046875, "learning_rate": 0.0001760680340170085, "loss": 4.4606, "step": 1202 }, { "epoch": 0.4990148294099347, "grad_norm": 1.984375, "learning_rate": 0.00017604802401200603, "loss": 4.3045, "step": 1203 }, { "epoch": 0.49942963807943586, "grad_norm": 1.921875, "learning_rate": 0.0001760280140070035, "loss": 4.4005, "step": 1204 }, { "epoch": 0.4998444467489371, "grad_norm": 1.890625, "learning_rate": 0.00017600800400200103, "loss": 4.7614, "step": 1205 }, { "epoch": 0.5002592554184383, "grad_norm": 2.234375, "learning_rate": 0.00017598799399699851, "loss": 4.4581, "step": 1206 }, { "epoch": 0.5006740640879395, "grad_norm": 1.96875, "learning_rate": 0.000175967983991996, "loss": 4.2721, "step": 1207 }, { "epoch": 0.5010888727574406, "grad_norm": 1.9765625, "learning_rate": 0.0001759479739869935, "loss": 4.5858, "step": 1208 }, { "epoch": 0.5015036814269418, "grad_norm": 1.875, "learning_rate": 0.000175927963981991, "loss": 4.2532, "step": 1209 }, { "epoch": 0.5019184900964431, "grad_norm": 1.984375, "learning_rate": 0.0001759079539769885, "loss": 4.5726, "step": 1210 }, { "epoch": 0.5023332987659442, "grad_norm": 2.28125, "learning_rate": 0.000175887943971986, "loss": 4.3911, "step": 1211 }, { "epoch": 0.5027481074354454, "grad_norm": 1.96875, "learning_rate": 0.0001758679339669835, "loss": 4.3685, "step": 1212 }, { "epoch": 0.5031629161049466, "grad_norm": 1.9140625, "learning_rate": 0.000175847923961981, "loss": 4.2716, "step": 1213 }, { "epoch": 0.5035777247744477, "grad_norm": 2.09375, "learning_rate": 0.0001758279139569785, "loss": 4.3746, "step": 1214 }, { "epoch": 0.503992533443949, "grad_norm": 1.8828125, "learning_rate": 0.00017580790395197598, "loss": 4.1584, "step": 1215 }, { "epoch": 0.5044073421134502, "grad_norm": 2.09375, "learning_rate": 0.0001757878939469735, "loss": 4.4824, "step": 1216 }, { "epoch": 0.5048221507829513, "grad_norm": 1.9609375, "learning_rate": 0.00017576788394197098, "loss": 4.2045, "step": 1217 }, { "epoch": 0.5052369594524525, "grad_norm": 2.0625, "learning_rate": 0.0001757478739369685, "loss": 4.4475, "step": 1218 }, { "epoch": 0.5056517681219538, "grad_norm": 1.9453125, "learning_rate": 0.00017572786393196599, "loss": 4.0906, "step": 1219 }, { "epoch": 0.506066576791455, "grad_norm": 2.0, "learning_rate": 0.0001757078539269635, "loss": 4.3062, "step": 1220 }, { "epoch": 0.5064813854609561, "grad_norm": 2.03125, "learning_rate": 0.000175687843921961, "loss": 4.2466, "step": 1221 }, { "epoch": 0.5068961941304573, "grad_norm": 1.8515625, "learning_rate": 0.0001756678339169585, "loss": 4.395, "step": 1222 }, { "epoch": 0.5073110027999586, "grad_norm": 1.859375, "learning_rate": 0.000175647823911956, "loss": 4.2161, "step": 1223 }, { "epoch": 0.5077258114694597, "grad_norm": 2.21875, "learning_rate": 0.00017562781390695348, "loss": 4.4077, "step": 1224 }, { "epoch": 0.5081406201389609, "grad_norm": 1.8046875, "learning_rate": 0.00017560780390195096, "loss": 4.2748, "step": 1225 }, { "epoch": 0.508555428808462, "grad_norm": 2.015625, "learning_rate": 0.00017558779389694848, "loss": 4.2839, "step": 1226 }, { "epoch": 0.5089702374779633, "grad_norm": 1.7578125, "learning_rate": 0.000175567783891946, "loss": 4.2471, "step": 1227 }, { "epoch": 0.5093850461474645, "grad_norm": 1.9921875, "learning_rate": 0.00017554777388694348, "loss": 4.2755, "step": 1228 }, { "epoch": 0.5097998548169657, "grad_norm": 1.7890625, "learning_rate": 0.000175527763881941, "loss": 4.3981, "step": 1229 }, { "epoch": 0.5102146634864668, "grad_norm": 2.140625, "learning_rate": 0.00017550775387693848, "loss": 4.489, "step": 1230 }, { "epoch": 0.5106294721559681, "grad_norm": 1.9375, "learning_rate": 0.000175487743871936, "loss": 4.2009, "step": 1231 }, { "epoch": 0.5110442808254693, "grad_norm": 2.09375, "learning_rate": 0.00017546773386693348, "loss": 4.5686, "step": 1232 }, { "epoch": 0.5114590894949704, "grad_norm": 2.734375, "learning_rate": 0.00017544772386193097, "loss": 4.3067, "step": 1233 }, { "epoch": 0.5118738981644716, "grad_norm": 2.140625, "learning_rate": 0.00017542771385692846, "loss": 4.3154, "step": 1234 }, { "epoch": 0.5122887068339729, "grad_norm": 2.09375, "learning_rate": 0.00017540770385192597, "loss": 4.75, "step": 1235 }, { "epoch": 0.512703515503474, "grad_norm": 2.015625, "learning_rate": 0.00017538769384692346, "loss": 4.2725, "step": 1236 }, { "epoch": 0.5131183241729752, "grad_norm": 2.125, "learning_rate": 0.00017536768384192097, "loss": 4.3889, "step": 1237 }, { "epoch": 0.5135331328424764, "grad_norm": 2.078125, "learning_rate": 0.00017534767383691846, "loss": 4.2123, "step": 1238 }, { "epoch": 0.5139479415119776, "grad_norm": 2.0625, "learning_rate": 0.00017532766383191597, "loss": 4.2747, "step": 1239 }, { "epoch": 0.5143627501814788, "grad_norm": 1.9765625, "learning_rate": 0.0001753076538269135, "loss": 4.4114, "step": 1240 }, { "epoch": 0.51477755885098, "grad_norm": 1.96875, "learning_rate": 0.00017528764382191098, "loss": 4.3279, "step": 1241 }, { "epoch": 0.5151923675204811, "grad_norm": 2.140625, "learning_rate": 0.00017526763381690846, "loss": 4.5109, "step": 1242 }, { "epoch": 0.5156071761899824, "grad_norm": 1.8984375, "learning_rate": 0.00017524762381190595, "loss": 4.2621, "step": 1243 }, { "epoch": 0.5160219848594836, "grad_norm": 2.21875, "learning_rate": 0.00017522761380690347, "loss": 4.3983, "step": 1244 }, { "epoch": 0.5164367935289847, "grad_norm": 1.890625, "learning_rate": 0.00017520760380190095, "loss": 4.6373, "step": 1245 }, { "epoch": 0.5168516021984859, "grad_norm": 2.0, "learning_rate": 0.00017518759379689847, "loss": 4.4741, "step": 1246 }, { "epoch": 0.5172664108679872, "grad_norm": 2.03125, "learning_rate": 0.00017516758379189595, "loss": 4.3331, "step": 1247 }, { "epoch": 0.5176812195374884, "grad_norm": 2.125, "learning_rate": 0.00017514757378689347, "loss": 4.4647, "step": 1248 }, { "epoch": 0.5180960282069895, "grad_norm": 1.7578125, "learning_rate": 0.00017512756378189096, "loss": 4.4374, "step": 1249 }, { "epoch": 0.5185108368764907, "grad_norm": 1.8046875, "learning_rate": 0.00017510755377688844, "loss": 4.2743, "step": 1250 }, { "epoch": 0.518925645545992, "grad_norm": 2.078125, "learning_rate": 0.00017508754377188593, "loss": 4.0994, "step": 1251 }, { "epoch": 0.5193404542154931, "grad_norm": 2.03125, "learning_rate": 0.00017506753376688344, "loss": 4.6526, "step": 1252 }, { "epoch": 0.5197552628849943, "grad_norm": 1.8671875, "learning_rate": 0.00017504752376188093, "loss": 4.3946, "step": 1253 }, { "epoch": 0.5201700715544955, "grad_norm": 1.8828125, "learning_rate": 0.00017502751375687845, "loss": 4.1943, "step": 1254 }, { "epoch": 0.5205848802239966, "grad_norm": 2.09375, "learning_rate": 0.00017500750375187596, "loss": 4.2717, "step": 1255 }, { "epoch": 0.5209996888934979, "grad_norm": 1.96875, "learning_rate": 0.00017498749374687345, "loss": 4.5654, "step": 1256 }, { "epoch": 0.5214144975629991, "grad_norm": 2.34375, "learning_rate": 0.00017496748374187096, "loss": 4.5405, "step": 1257 }, { "epoch": 0.5218293062325002, "grad_norm": 2.171875, "learning_rate": 0.00017494747373686845, "loss": 4.2844, "step": 1258 }, { "epoch": 0.5222441149020014, "grad_norm": 2.171875, "learning_rate": 0.00017492746373186594, "loss": 4.4466, "step": 1259 }, { "epoch": 0.5226589235715027, "grad_norm": 1.9375, "learning_rate": 0.00017490745372686342, "loss": 4.3552, "step": 1260 }, { "epoch": 0.5230737322410038, "grad_norm": 2.046875, "learning_rate": 0.00017488744372186094, "loss": 4.3936, "step": 1261 }, { "epoch": 0.523488540910505, "grad_norm": 2.15625, "learning_rate": 0.00017486743371685843, "loss": 4.3574, "step": 1262 }, { "epoch": 0.5239033495800062, "grad_norm": 2.0625, "learning_rate": 0.00017484742371185594, "loss": 4.4254, "step": 1263 }, { "epoch": 0.5243181582495074, "grad_norm": 1.8359375, "learning_rate": 0.00017482741370685343, "loss": 4.2937, "step": 1264 }, { "epoch": 0.5247329669190086, "grad_norm": 1.984375, "learning_rate": 0.00017480740370185094, "loss": 4.2344, "step": 1265 }, { "epoch": 0.5251477755885098, "grad_norm": 2.234375, "learning_rate": 0.00017478739369684843, "loss": 4.1864, "step": 1266 }, { "epoch": 0.525562584258011, "grad_norm": 1.984375, "learning_rate": 0.00017476738369184594, "loss": 4.1684, "step": 1267 }, { "epoch": 0.5259773929275122, "grad_norm": 1.765625, "learning_rate": 0.00017474737368684343, "loss": 4.4383, "step": 1268 }, { "epoch": 0.5263922015970134, "grad_norm": 2.0, "learning_rate": 0.00017472736368184092, "loss": 4.1107, "step": 1269 }, { "epoch": 0.5268070102665146, "grad_norm": 2.328125, "learning_rate": 0.00017470735367683843, "loss": 4.4619, "step": 1270 }, { "epoch": 0.5272218189360157, "grad_norm": 1.8203125, "learning_rate": 0.00017468734367183592, "loss": 4.1893, "step": 1271 }, { "epoch": 0.527636627605517, "grad_norm": 1.71875, "learning_rate": 0.00017466733366683343, "loss": 4.5942, "step": 1272 }, { "epoch": 0.5280514362750182, "grad_norm": 2.046875, "learning_rate": 0.00017464732366183092, "loss": 4.305, "step": 1273 }, { "epoch": 0.5284662449445193, "grad_norm": 1.8359375, "learning_rate": 0.00017462731365682844, "loss": 4.1635, "step": 1274 }, { "epoch": 0.5288810536140205, "grad_norm": 1.9453125, "learning_rate": 0.00017460730365182592, "loss": 4.4209, "step": 1275 }, { "epoch": 0.5292958622835218, "grad_norm": 1.9609375, "learning_rate": 0.00017458729364682344, "loss": 4.2263, "step": 1276 }, { "epoch": 0.5297106709530229, "grad_norm": 1.9609375, "learning_rate": 0.0001745672836418209, "loss": 4.355, "step": 1277 }, { "epoch": 0.5301254796225241, "grad_norm": 2.046875, "learning_rate": 0.0001745472736368184, "loss": 4.2577, "step": 1278 }, { "epoch": 0.5305402882920253, "grad_norm": 1.9375, "learning_rate": 0.0001745272636318159, "loss": 4.3233, "step": 1279 }, { "epoch": 0.5309550969615265, "grad_norm": 1.8359375, "learning_rate": 0.00017450725362681341, "loss": 4.3949, "step": 1280 }, { "epoch": 0.5313699056310277, "grad_norm": 1.9375, "learning_rate": 0.0001744872436218109, "loss": 4.0058, "step": 1281 }, { "epoch": 0.5317847143005289, "grad_norm": 2.03125, "learning_rate": 0.00017446723361680842, "loss": 4.4116, "step": 1282 }, { "epoch": 0.53219952297003, "grad_norm": 1.84375, "learning_rate": 0.00017444722361180593, "loss": 4.1986, "step": 1283 }, { "epoch": 0.5326143316395313, "grad_norm": 2.0625, "learning_rate": 0.00017442721360680342, "loss": 4.248, "step": 1284 }, { "epoch": 0.5330291403090325, "grad_norm": 2.171875, "learning_rate": 0.0001744072036018009, "loss": 4.5027, "step": 1285 }, { "epoch": 0.5334439489785336, "grad_norm": 2.046875, "learning_rate": 0.0001743871935967984, "loss": 4.3186, "step": 1286 }, { "epoch": 0.5338587576480348, "grad_norm": 1.9609375, "learning_rate": 0.0001743671835917959, "loss": 4.3903, "step": 1287 }, { "epoch": 0.5342735663175361, "grad_norm": 2.359375, "learning_rate": 0.0001743471735867934, "loss": 4.583, "step": 1288 }, { "epoch": 0.5346883749870373, "grad_norm": 1.90625, "learning_rate": 0.0001743271635817909, "loss": 4.2498, "step": 1289 }, { "epoch": 0.5351031836565384, "grad_norm": 1.9453125, "learning_rate": 0.0001743071535767884, "loss": 4.5248, "step": 1290 }, { "epoch": 0.5355179923260396, "grad_norm": 2.03125, "learning_rate": 0.0001742871435717859, "loss": 4.3364, "step": 1291 }, { "epoch": 0.5359328009955409, "grad_norm": 2.125, "learning_rate": 0.0001742671335667834, "loss": 4.1994, "step": 1292 }, { "epoch": 0.536347609665042, "grad_norm": 1.984375, "learning_rate": 0.0001742471235617809, "loss": 4.375, "step": 1293 }, { "epoch": 0.5367624183345432, "grad_norm": 1.9375, "learning_rate": 0.0001742271135567784, "loss": 4.301, "step": 1294 }, { "epoch": 0.5371772270040444, "grad_norm": 2.265625, "learning_rate": 0.00017420710355177589, "loss": 4.4581, "step": 1295 }, { "epoch": 0.5375920356735456, "grad_norm": 1.78125, "learning_rate": 0.0001741870935467734, "loss": 4.2014, "step": 1296 }, { "epoch": 0.5380068443430468, "grad_norm": 1.875, "learning_rate": 0.0001741670835417709, "loss": 4.4829, "step": 1297 }, { "epoch": 0.538421653012548, "grad_norm": 2.015625, "learning_rate": 0.0001741470735367684, "loss": 4.0274, "step": 1298 }, { "epoch": 0.5388364616820491, "grad_norm": 1.7578125, "learning_rate": 0.0001741270635317659, "loss": 4.3097, "step": 1299 }, { "epoch": 0.5392512703515503, "grad_norm": 1.921875, "learning_rate": 0.0001741070535267634, "loss": 4.3272, "step": 1300 }, { "epoch": 0.5396660790210516, "grad_norm": 2.359375, "learning_rate": 0.0001740870435217609, "loss": 4.6465, "step": 1301 }, { "epoch": 0.5400808876905527, "grad_norm": 1.9453125, "learning_rate": 0.0001740670335167584, "loss": 3.9895, "step": 1302 }, { "epoch": 0.5404956963600539, "grad_norm": 1.875, "learning_rate": 0.0001740470235117559, "loss": 4.4907, "step": 1303 }, { "epoch": 0.5409105050295551, "grad_norm": 1.8203125, "learning_rate": 0.00017402701350675338, "loss": 4.2979, "step": 1304 }, { "epoch": 0.5413253136990563, "grad_norm": 1.9375, "learning_rate": 0.00017400700350175087, "loss": 4.3366, "step": 1305 }, { "epoch": 0.5417401223685575, "grad_norm": 2.0625, "learning_rate": 0.00017398699349674838, "loss": 4.311, "step": 1306 }, { "epoch": 0.5421549310380587, "grad_norm": 2.203125, "learning_rate": 0.00017396698349174587, "loss": 4.4545, "step": 1307 }, { "epoch": 0.5425697397075598, "grad_norm": 1.953125, "learning_rate": 0.00017394697348674338, "loss": 4.3074, "step": 1308 }, { "epoch": 0.5429845483770611, "grad_norm": 2.03125, "learning_rate": 0.00017392696348174087, "loss": 4.2206, "step": 1309 }, { "epoch": 0.5433993570465623, "grad_norm": 1.8984375, "learning_rate": 0.00017390695347673838, "loss": 4.2822, "step": 1310 }, { "epoch": 0.5438141657160634, "grad_norm": 2.1875, "learning_rate": 0.0001738869434717359, "loss": 4.2488, "step": 1311 }, { "epoch": 0.5442289743855646, "grad_norm": 2.203125, "learning_rate": 0.00017386693346673336, "loss": 4.2637, "step": 1312 }, { "epoch": 0.5446437830550659, "grad_norm": 1.8359375, "learning_rate": 0.00017384692346173087, "loss": 4.1538, "step": 1313 }, { "epoch": 0.545058591724567, "grad_norm": 1.96875, "learning_rate": 0.00017382691345672836, "loss": 4.28, "step": 1314 }, { "epoch": 0.5454734003940682, "grad_norm": 2.03125, "learning_rate": 0.00017380690345172588, "loss": 4.0767, "step": 1315 }, { "epoch": 0.5458882090635694, "grad_norm": 2.140625, "learning_rate": 0.00017378689344672336, "loss": 4.4075, "step": 1316 }, { "epoch": 0.5463030177330707, "grad_norm": 2.1875, "learning_rate": 0.00017376688344172088, "loss": 4.3724, "step": 1317 }, { "epoch": 0.5467178264025718, "grad_norm": 2.140625, "learning_rate": 0.00017374687343671836, "loss": 4.4544, "step": 1318 }, { "epoch": 0.547132635072073, "grad_norm": 1.953125, "learning_rate": 0.00017372686343171588, "loss": 4.268, "step": 1319 }, { "epoch": 0.5475474437415742, "grad_norm": 1.984375, "learning_rate": 0.00017370685342671337, "loss": 4.1814, "step": 1320 }, { "epoch": 0.5479622524110754, "grad_norm": 1.875, "learning_rate": 0.00017368684342171085, "loss": 4.368, "step": 1321 }, { "epoch": 0.5483770610805766, "grad_norm": 2.03125, "learning_rate": 0.00017366683341670834, "loss": 4.2994, "step": 1322 }, { "epoch": 0.5487918697500778, "grad_norm": 1.9296875, "learning_rate": 0.00017364682341170585, "loss": 4.2659, "step": 1323 }, { "epoch": 0.5492066784195789, "grad_norm": 2.328125, "learning_rate": 0.00017362681340670337, "loss": 4.425, "step": 1324 }, { "epoch": 0.5496214870890802, "grad_norm": 2.171875, "learning_rate": 0.00017360680340170086, "loss": 4.5258, "step": 1325 }, { "epoch": 0.5500362957585814, "grad_norm": 2.1875, "learning_rate": 0.00017358679339669837, "loss": 4.526, "step": 1326 }, { "epoch": 0.5504511044280825, "grad_norm": 1.8359375, "learning_rate": 0.00017356678339169586, "loss": 4.4056, "step": 1327 }, { "epoch": 0.5508659130975837, "grad_norm": 1.9296875, "learning_rate": 0.00017354677338669337, "loss": 4.234, "step": 1328 }, { "epoch": 0.551280721767085, "grad_norm": 1.8359375, "learning_rate": 0.00017352676338169086, "loss": 4.3617, "step": 1329 }, { "epoch": 0.5516955304365861, "grad_norm": 2.015625, "learning_rate": 0.00017350675337668835, "loss": 4.4761, "step": 1330 }, { "epoch": 0.5521103391060873, "grad_norm": 1.96875, "learning_rate": 0.00017348674337168583, "loss": 4.5065, "step": 1331 }, { "epoch": 0.5525251477755885, "grad_norm": 1.9296875, "learning_rate": 0.00017346673336668335, "loss": 4.5396, "step": 1332 }, { "epoch": 0.5529399564450898, "grad_norm": 1.890625, "learning_rate": 0.00017344672336168084, "loss": 4.8258, "step": 1333 }, { "epoch": 0.5533547651145909, "grad_norm": 1.9453125, "learning_rate": 0.00017342671335667835, "loss": 4.424, "step": 1334 }, { "epoch": 0.5537695737840921, "grad_norm": 2.046875, "learning_rate": 0.00017340670335167584, "loss": 4.4027, "step": 1335 }, { "epoch": 0.5541843824535932, "grad_norm": 1.9375, "learning_rate": 0.00017338669334667335, "loss": 4.4958, "step": 1336 }, { "epoch": 0.5545991911230945, "grad_norm": 1.9140625, "learning_rate": 0.00017336668334167087, "loss": 4.3067, "step": 1337 }, { "epoch": 0.5550139997925957, "grad_norm": 1.9296875, "learning_rate": 0.00017334667333666835, "loss": 4.2119, "step": 1338 }, { "epoch": 0.5554288084620969, "grad_norm": 1.8046875, "learning_rate": 0.00017332666333166584, "loss": 4.4246, "step": 1339 }, { "epoch": 0.555843617131598, "grad_norm": 1.890625, "learning_rate": 0.00017330665332666333, "loss": 4.3445, "step": 1340 }, { "epoch": 0.5562584258010992, "grad_norm": 2.171875, "learning_rate": 0.00017328664332166084, "loss": 4.487, "step": 1341 }, { "epoch": 0.5566732344706005, "grad_norm": 1.796875, "learning_rate": 0.00017326663331665833, "loss": 4.316, "step": 1342 }, { "epoch": 0.5570880431401016, "grad_norm": 1.9765625, "learning_rate": 0.00017324662331165584, "loss": 4.3887, "step": 1343 }, { "epoch": 0.5575028518096028, "grad_norm": 1.8125, "learning_rate": 0.00017322661330665333, "loss": 4.2804, "step": 1344 }, { "epoch": 0.557917660479104, "grad_norm": 2.0, "learning_rate": 0.00017320660330165085, "loss": 4.6823, "step": 1345 }, { "epoch": 0.5583324691486052, "grad_norm": 2.0, "learning_rate": 0.00017318659329664833, "loss": 4.3116, "step": 1346 }, { "epoch": 0.5587472778181064, "grad_norm": 1.8828125, "learning_rate": 0.00017316658329164585, "loss": 4.3569, "step": 1347 }, { "epoch": 0.5591620864876076, "grad_norm": 2.0625, "learning_rate": 0.0001731465732866433, "loss": 4.3595, "step": 1348 }, { "epoch": 0.5595768951571087, "grad_norm": 2.015625, "learning_rate": 0.00017312656328164082, "loss": 4.0869, "step": 1349 }, { "epoch": 0.55999170382661, "grad_norm": 1.8828125, "learning_rate": 0.0001731065532766383, "loss": 4.3401, "step": 1350 }, { "epoch": 0.5604065124961112, "grad_norm": 1.90625, "learning_rate": 0.00017308654327163582, "loss": 4.2857, "step": 1351 }, { "epoch": 0.5608213211656123, "grad_norm": 2.09375, "learning_rate": 0.00017306653326663334, "loss": 4.1177, "step": 1352 }, { "epoch": 0.5612361298351135, "grad_norm": 1.859375, "learning_rate": 0.00017304652326163083, "loss": 4.2515, "step": 1353 }, { "epoch": 0.5616509385046148, "grad_norm": 2.046875, "learning_rate": 0.00017302651325662834, "loss": 4.122, "step": 1354 }, { "epoch": 0.562065747174116, "grad_norm": 2.078125, "learning_rate": 0.00017300650325162583, "loss": 4.3245, "step": 1355 }, { "epoch": 0.5624805558436171, "grad_norm": 1.8046875, "learning_rate": 0.00017298649324662331, "loss": 4.1449, "step": 1356 }, { "epoch": 0.5628953645131183, "grad_norm": 2.109375, "learning_rate": 0.0001729664832416208, "loss": 4.3922, "step": 1357 }, { "epoch": 0.5633101731826196, "grad_norm": 2.0, "learning_rate": 0.00017294647323661832, "loss": 4.2348, "step": 1358 }, { "epoch": 0.5637249818521207, "grad_norm": 2.015625, "learning_rate": 0.0001729264632316158, "loss": 4.4701, "step": 1359 }, { "epoch": 0.5641397905216219, "grad_norm": 1.8359375, "learning_rate": 0.00017290645322661332, "loss": 4.1999, "step": 1360 }, { "epoch": 0.564554599191123, "grad_norm": 2.0625, "learning_rate": 0.0001728864432216108, "loss": 4.3161, "step": 1361 }, { "epoch": 0.5649694078606243, "grad_norm": 2.0625, "learning_rate": 0.00017286643321660832, "loss": 4.1485, "step": 1362 }, { "epoch": 0.5653842165301255, "grad_norm": 1.96875, "learning_rate": 0.0001728464232116058, "loss": 4.4301, "step": 1363 }, { "epoch": 0.5657990251996267, "grad_norm": 1.9609375, "learning_rate": 0.00017282641320660332, "loss": 4.3702, "step": 1364 }, { "epoch": 0.5662138338691278, "grad_norm": 1.8828125, "learning_rate": 0.0001728064032016008, "loss": 4.2414, "step": 1365 }, { "epoch": 0.5666286425386291, "grad_norm": 1.9453125, "learning_rate": 0.0001727863931965983, "loss": 4.4038, "step": 1366 }, { "epoch": 0.5670434512081303, "grad_norm": 1.8359375, "learning_rate": 0.0001727663831915958, "loss": 4.3009, "step": 1367 }, { "epoch": 0.5674582598776314, "grad_norm": 1.921875, "learning_rate": 0.0001727463731865933, "loss": 4.3038, "step": 1368 }, { "epoch": 0.5678730685471326, "grad_norm": 2.3125, "learning_rate": 0.0001727263631815908, "loss": 4.3313, "step": 1369 }, { "epoch": 0.5682878772166339, "grad_norm": 1.9765625, "learning_rate": 0.0001727063531765883, "loss": 4.1563, "step": 1370 }, { "epoch": 0.568702685886135, "grad_norm": 1.9765625, "learning_rate": 0.0001726863431715858, "loss": 4.2363, "step": 1371 }, { "epoch": 0.5691174945556362, "grad_norm": 1.890625, "learning_rate": 0.0001726663331665833, "loss": 4.3705, "step": 1372 }, { "epoch": 0.5695323032251374, "grad_norm": 2.0, "learning_rate": 0.00017264632316158082, "loss": 4.3854, "step": 1373 }, { "epoch": 0.5699471118946386, "grad_norm": 2.09375, "learning_rate": 0.0001726263131565783, "loss": 4.2774, "step": 1374 }, { "epoch": 0.5703619205641398, "grad_norm": 1.9609375, "learning_rate": 0.0001726063031515758, "loss": 4.2757, "step": 1375 }, { "epoch": 0.570776729233641, "grad_norm": 1.890625, "learning_rate": 0.00017258629314657328, "loss": 4.673, "step": 1376 }, { "epoch": 0.5711915379031421, "grad_norm": 2.03125, "learning_rate": 0.0001725662831415708, "loss": 4.1743, "step": 1377 }, { "epoch": 0.5716063465726434, "grad_norm": 1.96875, "learning_rate": 0.00017254627313656828, "loss": 4.1254, "step": 1378 }, { "epoch": 0.5720211552421446, "grad_norm": 2.078125, "learning_rate": 0.0001725262631315658, "loss": 4.196, "step": 1379 }, { "epoch": 0.5724359639116458, "grad_norm": 1.84375, "learning_rate": 0.0001725062531265633, "loss": 4.2856, "step": 1380 }, { "epoch": 0.5728507725811469, "grad_norm": 2.515625, "learning_rate": 0.0001724862431215608, "loss": 4.2915, "step": 1381 }, { "epoch": 0.5732655812506482, "grad_norm": 1.8515625, "learning_rate": 0.0001724662331165583, "loss": 4.4261, "step": 1382 }, { "epoch": 0.5736803899201494, "grad_norm": 2.21875, "learning_rate": 0.00017244622311155577, "loss": 4.2462, "step": 1383 }, { "epoch": 0.5740951985896505, "grad_norm": 1.9453125, "learning_rate": 0.00017242621310655328, "loss": 4.4918, "step": 1384 }, { "epoch": 0.5745100072591517, "grad_norm": 2.203125, "learning_rate": 0.00017240620310155077, "loss": 4.1641, "step": 1385 }, { "epoch": 0.5749248159286529, "grad_norm": 1.8125, "learning_rate": 0.00017238619309654829, "loss": 4.4753, "step": 1386 }, { "epoch": 0.5753396245981541, "grad_norm": 1.9140625, "learning_rate": 0.00017236618309154577, "loss": 4.144, "step": 1387 }, { "epoch": 0.5757544332676553, "grad_norm": 1.890625, "learning_rate": 0.0001723461730865433, "loss": 4.3885, "step": 1388 }, { "epoch": 0.5761692419371565, "grad_norm": 2.03125, "learning_rate": 0.00017232616308154077, "loss": 4.1634, "step": 1389 }, { "epoch": 0.5765840506066576, "grad_norm": 2.0, "learning_rate": 0.0001723061530765383, "loss": 4.3891, "step": 1390 }, { "epoch": 0.5769988592761589, "grad_norm": 1.9140625, "learning_rate": 0.00017228614307153578, "loss": 4.2899, "step": 1391 }, { "epoch": 0.5774136679456601, "grad_norm": 2.015625, "learning_rate": 0.00017226613306653326, "loss": 4.2481, "step": 1392 }, { "epoch": 0.5778284766151612, "grad_norm": 1.8828125, "learning_rate": 0.00017224612306153078, "loss": 4.3697, "step": 1393 }, { "epoch": 0.5782432852846624, "grad_norm": 1.875, "learning_rate": 0.00017222611305652826, "loss": 4.5009, "step": 1394 }, { "epoch": 0.5786580939541637, "grad_norm": 1.984375, "learning_rate": 0.00017220610305152578, "loss": 4.2218, "step": 1395 }, { "epoch": 0.5790729026236648, "grad_norm": 2.109375, "learning_rate": 0.00017218609304652327, "loss": 4.5284, "step": 1396 }, { "epoch": 0.579487711293166, "grad_norm": 1.8515625, "learning_rate": 0.00017216608304152078, "loss": 4.4089, "step": 1397 }, { "epoch": 0.5799025199626672, "grad_norm": 1.71875, "learning_rate": 0.00017214607303651827, "loss": 4.3238, "step": 1398 }, { "epoch": 0.5803173286321684, "grad_norm": 1.96875, "learning_rate": 0.00017212606303151578, "loss": 4.1448, "step": 1399 }, { "epoch": 0.5807321373016696, "grad_norm": 1.9609375, "learning_rate": 0.00017210605302651327, "loss": 4.1475, "step": 1400 }, { "epoch": 0.5811469459711708, "grad_norm": 1.765625, "learning_rate": 0.00017208604302151076, "loss": 4.3953, "step": 1401 }, { "epoch": 0.581561754640672, "grad_norm": 2.015625, "learning_rate": 0.00017206603301650824, "loss": 4.4854, "step": 1402 }, { "epoch": 0.5819765633101732, "grad_norm": 2.046875, "learning_rate": 0.00017204602301150576, "loss": 4.461, "step": 1403 }, { "epoch": 0.5823913719796744, "grad_norm": 1.875, "learning_rate": 0.00017202601300650325, "loss": 4.3844, "step": 1404 }, { "epoch": 0.5828061806491756, "grad_norm": 1.8671875, "learning_rate": 0.00017200600300150076, "loss": 4.6786, "step": 1405 }, { "epoch": 0.5832209893186767, "grad_norm": 1.8828125, "learning_rate": 0.00017198599299649825, "loss": 4.3629, "step": 1406 }, { "epoch": 0.583635797988178, "grad_norm": 2.078125, "learning_rate": 0.00017196598299149576, "loss": 4.1442, "step": 1407 }, { "epoch": 0.5840506066576792, "grad_norm": 1.8515625, "learning_rate": 0.00017194597298649328, "loss": 4.3643, "step": 1408 }, { "epoch": 0.5844654153271803, "grad_norm": 2.15625, "learning_rate": 0.00017192596298149076, "loss": 4.1961, "step": 1409 }, { "epoch": 0.5848802239966815, "grad_norm": 1.9765625, "learning_rate": 0.00017190595297648825, "loss": 4.3112, "step": 1410 }, { "epoch": 0.5852950326661828, "grad_norm": 2.109375, "learning_rate": 0.00017188594297148574, "loss": 4.1584, "step": 1411 }, { "epoch": 0.5857098413356839, "grad_norm": 1.8828125, "learning_rate": 0.00017186593296648325, "loss": 4.4323, "step": 1412 }, { "epoch": 0.5861246500051851, "grad_norm": 2.15625, "learning_rate": 0.00017184592296148074, "loss": 4.3019, "step": 1413 }, { "epoch": 0.5865394586746863, "grad_norm": 2.125, "learning_rate": 0.00017182591295647825, "loss": 4.46, "step": 1414 }, { "epoch": 0.5869542673441875, "grad_norm": 1.9453125, "learning_rate": 0.00017180590295147574, "loss": 4.3739, "step": 1415 }, { "epoch": 0.5873690760136887, "grad_norm": 1.8359375, "learning_rate": 0.00017178589294647326, "loss": 4.2291, "step": 1416 }, { "epoch": 0.5877838846831899, "grad_norm": 1.890625, "learning_rate": 0.00017176588294147074, "loss": 4.1138, "step": 1417 }, { "epoch": 0.588198693352691, "grad_norm": 1.6953125, "learning_rate": 0.00017174587293646823, "loss": 4.2786, "step": 1418 }, { "epoch": 0.5886135020221923, "grad_norm": 2.109375, "learning_rate": 0.00017172586293146572, "loss": 4.4354, "step": 1419 }, { "epoch": 0.5890283106916935, "grad_norm": 1.8671875, "learning_rate": 0.00017170585292646323, "loss": 4.1951, "step": 1420 }, { "epoch": 0.5894431193611946, "grad_norm": 1.8359375, "learning_rate": 0.00017168584292146075, "loss": 4.3176, "step": 1421 }, { "epoch": 0.5898579280306958, "grad_norm": 1.9609375, "learning_rate": 0.00017166583291645823, "loss": 4.5763, "step": 1422 }, { "epoch": 0.5902727367001971, "grad_norm": 1.9375, "learning_rate": 0.00017164582291145575, "loss": 4.2788, "step": 1423 }, { "epoch": 0.5906875453696983, "grad_norm": 1.8828125, "learning_rate": 0.00017162581290645324, "loss": 4.3428, "step": 1424 }, { "epoch": 0.5911023540391994, "grad_norm": 2.109375, "learning_rate": 0.00017160580290145075, "loss": 4.4421, "step": 1425 }, { "epoch": 0.5915171627087006, "grad_norm": 1.8515625, "learning_rate": 0.00017158579289644824, "loss": 4.2542, "step": 1426 }, { "epoch": 0.5919319713782017, "grad_norm": 1.7265625, "learning_rate": 0.00017156578289144572, "loss": 4.261, "step": 1427 }, { "epoch": 0.592346780047703, "grad_norm": 2.265625, "learning_rate": 0.0001715457728864432, "loss": 4.2739, "step": 1428 }, { "epoch": 0.5927615887172042, "grad_norm": 2.046875, "learning_rate": 0.00017152576288144073, "loss": 4.1583, "step": 1429 }, { "epoch": 0.5931763973867054, "grad_norm": 1.8046875, "learning_rate": 0.0001715057528764382, "loss": 4.2409, "step": 1430 }, { "epoch": 0.5935912060562065, "grad_norm": 2.234375, "learning_rate": 0.00017148574287143573, "loss": 4.3583, "step": 1431 }, { "epoch": 0.5940060147257078, "grad_norm": 1.7265625, "learning_rate": 0.00017146573286643322, "loss": 4.0787, "step": 1432 }, { "epoch": 0.594420823395209, "grad_norm": 1.9296875, "learning_rate": 0.00017144572286143073, "loss": 4.3969, "step": 1433 }, { "epoch": 0.5948356320647101, "grad_norm": 2.28125, "learning_rate": 0.00017142571285642824, "loss": 4.4596, "step": 1434 }, { "epoch": 0.5952504407342113, "grad_norm": 1.8984375, "learning_rate": 0.00017140570285142573, "loss": 4.4686, "step": 1435 }, { "epoch": 0.5956652494037126, "grad_norm": 1.859375, "learning_rate": 0.00017138569284642322, "loss": 4.2836, "step": 1436 }, { "epoch": 0.5960800580732137, "grad_norm": 2.53125, "learning_rate": 0.0001713656828414207, "loss": 4.5226, "step": 1437 }, { "epoch": 0.5964948667427149, "grad_norm": 1.9609375, "learning_rate": 0.00017134567283641822, "loss": 4.6246, "step": 1438 }, { "epoch": 0.5969096754122161, "grad_norm": 1.8046875, "learning_rate": 0.0001713256628314157, "loss": 4.3064, "step": 1439 }, { "epoch": 0.5973244840817173, "grad_norm": 2.0625, "learning_rate": 0.00017130565282641322, "loss": 4.0792, "step": 1440 }, { "epoch": 0.5977392927512185, "grad_norm": 1.90625, "learning_rate": 0.0001712856428214107, "loss": 4.2498, "step": 1441 }, { "epoch": 0.5981541014207197, "grad_norm": 2.046875, "learning_rate": 0.00017126563281640822, "loss": 4.2767, "step": 1442 }, { "epoch": 0.5985689100902208, "grad_norm": 1.921875, "learning_rate": 0.0001712456228114057, "loss": 4.2968, "step": 1443 }, { "epoch": 0.5989837187597221, "grad_norm": 1.890625, "learning_rate": 0.00017122561280640323, "loss": 4.0411, "step": 1444 }, { "epoch": 0.5993985274292233, "grad_norm": 1.84375, "learning_rate": 0.0001712056028014007, "loss": 4.0592, "step": 1445 }, { "epoch": 0.5998133360987244, "grad_norm": 1.8203125, "learning_rate": 0.0001711855927963982, "loss": 4.2207, "step": 1446 }, { "epoch": 0.6002281447682256, "grad_norm": 1.9453125, "learning_rate": 0.0001711655827913957, "loss": 4.2818, "step": 1447 }, { "epoch": 0.6006429534377269, "grad_norm": 2.234375, "learning_rate": 0.0001711455727863932, "loss": 4.4416, "step": 1448 }, { "epoch": 0.601057762107228, "grad_norm": 2.015625, "learning_rate": 0.00017112556278139072, "loss": 4.2443, "step": 1449 }, { "epoch": 0.6014725707767292, "grad_norm": 2.078125, "learning_rate": 0.0001711055527763882, "loss": 4.3454, "step": 1450 }, { "epoch": 0.6018873794462304, "grad_norm": 2.078125, "learning_rate": 0.00017108554277138572, "loss": 4.1842, "step": 1451 }, { "epoch": 0.6023021881157317, "grad_norm": 2.234375, "learning_rate": 0.0001710655327663832, "loss": 4.4242, "step": 1452 }, { "epoch": 0.6027169967852328, "grad_norm": 2.046875, "learning_rate": 0.0001710455227613807, "loss": 4.5832, "step": 1453 }, { "epoch": 0.603131805454734, "grad_norm": 2.21875, "learning_rate": 0.00017102551275637818, "loss": 4.2929, "step": 1454 }, { "epoch": 0.6035466141242352, "grad_norm": 1.90625, "learning_rate": 0.0001710055027513757, "loss": 4.1691, "step": 1455 }, { "epoch": 0.6039614227937364, "grad_norm": 1.9921875, "learning_rate": 0.00017098549274637318, "loss": 4.5907, "step": 1456 }, { "epoch": 0.6043762314632376, "grad_norm": 2.4375, "learning_rate": 0.0001709654827413707, "loss": 4.3613, "step": 1457 }, { "epoch": 0.6047910401327388, "grad_norm": 1.9375, "learning_rate": 0.00017094547273636818, "loss": 4.2897, "step": 1458 }, { "epoch": 0.6052058488022399, "grad_norm": 2.03125, "learning_rate": 0.0001709254627313657, "loss": 4.2352, "step": 1459 }, { "epoch": 0.6056206574717412, "grad_norm": 1.7734375, "learning_rate": 0.00017090545272636318, "loss": 4.3734, "step": 1460 }, { "epoch": 0.6060354661412424, "grad_norm": 2.1875, "learning_rate": 0.0001708854427213607, "loss": 4.2973, "step": 1461 }, { "epoch": 0.6064502748107435, "grad_norm": 1.8359375, "learning_rate": 0.00017086543271635819, "loss": 4.355, "step": 1462 }, { "epoch": 0.6068650834802447, "grad_norm": 1.9453125, "learning_rate": 0.00017084542271135567, "loss": 4.4822, "step": 1463 }, { "epoch": 0.607279892149746, "grad_norm": 1.890625, "learning_rate": 0.0001708254127063532, "loss": 4.1659, "step": 1464 }, { "epoch": 0.6076947008192471, "grad_norm": 1.921875, "learning_rate": 0.00017080540270135067, "loss": 4.313, "step": 1465 }, { "epoch": 0.6081095094887483, "grad_norm": 1.984375, "learning_rate": 0.0001707853926963482, "loss": 4.4229, "step": 1466 }, { "epoch": 0.6085243181582495, "grad_norm": 1.921875, "learning_rate": 0.00017076538269134568, "loss": 4.1758, "step": 1467 }, { "epoch": 0.6089391268277508, "grad_norm": 1.96875, "learning_rate": 0.0001707453726863432, "loss": 4.2771, "step": 1468 }, { "epoch": 0.6093539354972519, "grad_norm": 2.15625, "learning_rate": 0.00017072536268134068, "loss": 4.3834, "step": 1469 }, { "epoch": 0.6097687441667531, "grad_norm": 1.796875, "learning_rate": 0.0001707053526763382, "loss": 4.1566, "step": 1470 }, { "epoch": 0.6101835528362543, "grad_norm": 1.859375, "learning_rate": 0.00017068534267133568, "loss": 4.3964, "step": 1471 }, { "epoch": 0.6105983615057554, "grad_norm": 1.9453125, "learning_rate": 0.00017066533266633317, "loss": 4.3178, "step": 1472 }, { "epoch": 0.6110131701752567, "grad_norm": 1.96875, "learning_rate": 0.00017064532266133065, "loss": 4.335, "step": 1473 }, { "epoch": 0.6114279788447579, "grad_norm": 1.9921875, "learning_rate": 0.00017062531265632817, "loss": 4.32, "step": 1474 }, { "epoch": 0.611842787514259, "grad_norm": 1.875, "learning_rate": 0.00017060530265132566, "loss": 4.4379, "step": 1475 }, { "epoch": 0.6122575961837602, "grad_norm": 2.078125, "learning_rate": 0.00017058529264632317, "loss": 4.3591, "step": 1476 }, { "epoch": 0.6126724048532615, "grad_norm": 2.109375, "learning_rate": 0.00017056528264132068, "loss": 4.2203, "step": 1477 }, { "epoch": 0.6130872135227626, "grad_norm": 2.015625, "learning_rate": 0.00017054527263631817, "loss": 4.5472, "step": 1478 }, { "epoch": 0.6135020221922638, "grad_norm": 2.109375, "learning_rate": 0.0001705252626313157, "loss": 4.2897, "step": 1479 }, { "epoch": 0.613916830861765, "grad_norm": 2.03125, "learning_rate": 0.00017050525262631317, "loss": 4.3161, "step": 1480 }, { "epoch": 0.6143316395312662, "grad_norm": 2.1875, "learning_rate": 0.00017048524262131066, "loss": 4.2874, "step": 1481 }, { "epoch": 0.6147464482007674, "grad_norm": 1.9296875, "learning_rate": 0.00017046523261630815, "loss": 4.4308, "step": 1482 }, { "epoch": 0.6151612568702686, "grad_norm": 1.953125, "learning_rate": 0.00017044522261130566, "loss": 4.3472, "step": 1483 }, { "epoch": 0.6155760655397697, "grad_norm": 1.828125, "learning_rate": 0.00017042521260630315, "loss": 4.1987, "step": 1484 }, { "epoch": 0.615990874209271, "grad_norm": 2.078125, "learning_rate": 0.00017040520260130066, "loss": 4.4538, "step": 1485 }, { "epoch": 0.6164056828787722, "grad_norm": 1.90625, "learning_rate": 0.00017038519259629815, "loss": 4.1197, "step": 1486 }, { "epoch": 0.6168204915482733, "grad_norm": 2.171875, "learning_rate": 0.00017036518259129567, "loss": 4.5562, "step": 1487 }, { "epoch": 0.6172353002177745, "grad_norm": 1.875, "learning_rate": 0.00017034517258629315, "loss": 4.0696, "step": 1488 }, { "epoch": 0.6176501088872758, "grad_norm": 1.7421875, "learning_rate": 0.00017032516258129064, "loss": 4.1875, "step": 1489 }, { "epoch": 0.618064917556777, "grad_norm": 1.8828125, "learning_rate": 0.00017030515257628816, "loss": 4.1146, "step": 1490 }, { "epoch": 0.6184797262262781, "grad_norm": 1.8828125, "learning_rate": 0.00017028514257128564, "loss": 4.2917, "step": 1491 }, { "epoch": 0.6188945348957793, "grad_norm": 1.8203125, "learning_rate": 0.00017026513256628316, "loss": 4.3181, "step": 1492 }, { "epoch": 0.6193093435652806, "grad_norm": 1.921875, "learning_rate": 0.00017024512256128064, "loss": 4.4676, "step": 1493 }, { "epoch": 0.6197241522347817, "grad_norm": 1.953125, "learning_rate": 0.00017022511255627816, "loss": 4.1105, "step": 1494 }, { "epoch": 0.6201389609042829, "grad_norm": 1.9609375, "learning_rate": 0.00017020510255127565, "loss": 4.2254, "step": 1495 }, { "epoch": 0.620553769573784, "grad_norm": 1.7578125, "learning_rate": 0.00017018509254627316, "loss": 4.242, "step": 1496 }, { "epoch": 0.6209685782432853, "grad_norm": 1.84375, "learning_rate": 0.00017016508254127065, "loss": 4.3203, "step": 1497 }, { "epoch": 0.6213833869127865, "grad_norm": 1.890625, "learning_rate": 0.00017014507253626813, "loss": 4.2968, "step": 1498 }, { "epoch": 0.6217981955822877, "grad_norm": 1.9375, "learning_rate": 0.00017012506253126562, "loss": 4.5162, "step": 1499 }, { "epoch": 0.6222130042517888, "grad_norm": 2.03125, "learning_rate": 0.00017010505252626314, "loss": 4.5584, "step": 1500 }, { "epoch": 0.6226278129212901, "grad_norm": 1.9765625, "learning_rate": 0.00017008504252126062, "loss": 4.3569, "step": 1501 }, { "epoch": 0.6230426215907913, "grad_norm": 1.90625, "learning_rate": 0.00017006503251625814, "loss": 4.3569, "step": 1502 }, { "epoch": 0.6234574302602924, "grad_norm": 1.8359375, "learning_rate": 0.00017004502251125563, "loss": 4.2581, "step": 1503 }, { "epoch": 0.6238722389297936, "grad_norm": 1.796875, "learning_rate": 0.00017002501250625314, "loss": 4.3999, "step": 1504 }, { "epoch": 0.6242870475992949, "grad_norm": 2.015625, "learning_rate": 0.00017000500250125065, "loss": 4.166, "step": 1505 }, { "epoch": 0.624701856268796, "grad_norm": 2.015625, "learning_rate": 0.00016998499249624814, "loss": 4.167, "step": 1506 }, { "epoch": 0.6251166649382972, "grad_norm": 1.859375, "learning_rate": 0.00016996498249124563, "loss": 4.2139, "step": 1507 }, { "epoch": 0.6255314736077984, "grad_norm": 1.96875, "learning_rate": 0.00016994497248624312, "loss": 4.13, "step": 1508 }, { "epoch": 0.6259462822772996, "grad_norm": 1.9140625, "learning_rate": 0.00016992496248124063, "loss": 4.3987, "step": 1509 }, { "epoch": 0.6263610909468008, "grad_norm": 1.9140625, "learning_rate": 0.00016990495247623812, "loss": 4.2547, "step": 1510 }, { "epoch": 0.626775899616302, "grad_norm": 1.9453125, "learning_rate": 0.00016988494247123563, "loss": 4.5593, "step": 1511 }, { "epoch": 0.6271907082858031, "grad_norm": 2.03125, "learning_rate": 0.00016986493246623312, "loss": 4.4413, "step": 1512 }, { "epoch": 0.6276055169553044, "grad_norm": 1.953125, "learning_rate": 0.00016984492246123063, "loss": 4.4413, "step": 1513 }, { "epoch": 0.6280203256248056, "grad_norm": 1.875, "learning_rate": 0.00016982491245622812, "loss": 4.249, "step": 1514 }, { "epoch": 0.6284351342943068, "grad_norm": 2.171875, "learning_rate": 0.00016980490245122564, "loss": 4.4125, "step": 1515 }, { "epoch": 0.6288499429638079, "grad_norm": 1.9453125, "learning_rate": 0.00016978489244622312, "loss": 4.2449, "step": 1516 }, { "epoch": 0.6292647516333091, "grad_norm": 1.9765625, "learning_rate": 0.0001697648824412206, "loss": 4.2113, "step": 1517 }, { "epoch": 0.6296795603028104, "grad_norm": 1.9765625, "learning_rate": 0.00016974487243621812, "loss": 4.1162, "step": 1518 }, { "epoch": 0.6300943689723115, "grad_norm": 1.828125, "learning_rate": 0.0001697248624312156, "loss": 4.2895, "step": 1519 }, { "epoch": 0.6305091776418127, "grad_norm": 1.9453125, "learning_rate": 0.00016970485242621313, "loss": 4.3082, "step": 1520 }, { "epoch": 0.6309239863113139, "grad_norm": 1.9296875, "learning_rate": 0.0001696848424212106, "loss": 4.4505, "step": 1521 }, { "epoch": 0.6313387949808151, "grad_norm": 2.046875, "learning_rate": 0.00016966483241620813, "loss": 4.1492, "step": 1522 }, { "epoch": 0.6317536036503163, "grad_norm": 1.7890625, "learning_rate": 0.00016964482241120561, "loss": 4.2697, "step": 1523 }, { "epoch": 0.6321684123198175, "grad_norm": 1.7109375, "learning_rate": 0.0001696248124062031, "loss": 4.4773, "step": 1524 }, { "epoch": 0.6325832209893186, "grad_norm": 2.109375, "learning_rate": 0.0001696048024012006, "loss": 4.1244, "step": 1525 }, { "epoch": 0.6329980296588199, "grad_norm": 1.7890625, "learning_rate": 0.0001695847923961981, "loss": 4.4032, "step": 1526 }, { "epoch": 0.6334128383283211, "grad_norm": 1.9609375, "learning_rate": 0.0001695647823911956, "loss": 4.14, "step": 1527 }, { "epoch": 0.6338276469978222, "grad_norm": 2.03125, "learning_rate": 0.0001695447723861931, "loss": 4.1219, "step": 1528 }, { "epoch": 0.6342424556673234, "grad_norm": 1.9921875, "learning_rate": 0.0001695247623811906, "loss": 4.8954, "step": 1529 }, { "epoch": 0.6346572643368247, "grad_norm": 1.796875, "learning_rate": 0.0001695047523761881, "loss": 4.2527, "step": 1530 }, { "epoch": 0.6350720730063258, "grad_norm": 1.8359375, "learning_rate": 0.00016948474237118562, "loss": 4.4145, "step": 1531 }, { "epoch": 0.635486881675827, "grad_norm": 1.75, "learning_rate": 0.0001694647323661831, "loss": 4.1724, "step": 1532 }, { "epoch": 0.6359016903453282, "grad_norm": 2.140625, "learning_rate": 0.0001694447223611806, "loss": 4.2807, "step": 1533 }, { "epoch": 0.6363164990148295, "grad_norm": 1.75, "learning_rate": 0.00016942471235617808, "loss": 4.5104, "step": 1534 }, { "epoch": 0.6367313076843306, "grad_norm": 1.9140625, "learning_rate": 0.0001694047023511756, "loss": 4.2648, "step": 1535 }, { "epoch": 0.6371461163538318, "grad_norm": 1.9921875, "learning_rate": 0.00016938469234617308, "loss": 4.1736, "step": 1536 }, { "epoch": 0.637560925023333, "grad_norm": 2.296875, "learning_rate": 0.0001693646823411706, "loss": 4.1461, "step": 1537 }, { "epoch": 0.6379757336928342, "grad_norm": 1.8125, "learning_rate": 0.0001693446723361681, "loss": 4.1611, "step": 1538 }, { "epoch": 0.6383905423623354, "grad_norm": 1.984375, "learning_rate": 0.0001693246623311656, "loss": 4.436, "step": 1539 }, { "epoch": 0.6388053510318366, "grad_norm": 1.859375, "learning_rate": 0.0001693046523261631, "loss": 4.4321, "step": 1540 }, { "epoch": 0.6392201597013377, "grad_norm": 2.078125, "learning_rate": 0.0001692846423211606, "loss": 4.3143, "step": 1541 }, { "epoch": 0.639634968370839, "grad_norm": 2.078125, "learning_rate": 0.0001692646323161581, "loss": 4.2163, "step": 1542 }, { "epoch": 0.6400497770403402, "grad_norm": 1.9296875, "learning_rate": 0.00016924462231115558, "loss": 4.4981, "step": 1543 }, { "epoch": 0.6404645857098413, "grad_norm": 1.921875, "learning_rate": 0.00016922461230615306, "loss": 4.2936, "step": 1544 }, { "epoch": 0.6408793943793425, "grad_norm": 1.7734375, "learning_rate": 0.00016920460230115058, "loss": 4.6549, "step": 1545 }, { "epoch": 0.6412942030488438, "grad_norm": 2.0, "learning_rate": 0.0001691845922961481, "loss": 4.2875, "step": 1546 }, { "epoch": 0.6417090117183449, "grad_norm": 1.9453125, "learning_rate": 0.00016916458229114558, "loss": 4.3488, "step": 1547 }, { "epoch": 0.6421238203878461, "grad_norm": 1.84375, "learning_rate": 0.0001691445722861431, "loss": 4.5145, "step": 1548 }, { "epoch": 0.6425386290573473, "grad_norm": 1.9921875, "learning_rate": 0.00016912456228114058, "loss": 3.982, "step": 1549 }, { "epoch": 0.6429534377268485, "grad_norm": 1.8359375, "learning_rate": 0.0001691045522761381, "loss": 4.6159, "step": 1550 }, { "epoch": 0.6433682463963497, "grad_norm": 1.8359375, "learning_rate": 0.00016908454227113558, "loss": 4.2575, "step": 1551 }, { "epoch": 0.6437830550658509, "grad_norm": 1.921875, "learning_rate": 0.00016906453226613307, "loss": 4.3645, "step": 1552 }, { "epoch": 0.644197863735352, "grad_norm": 2.046875, "learning_rate": 0.00016904452226113056, "loss": 4.129, "step": 1553 }, { "epoch": 0.6446126724048533, "grad_norm": 1.9140625, "learning_rate": 0.00016902451225612807, "loss": 4.345, "step": 1554 }, { "epoch": 0.6450274810743545, "grad_norm": 2.046875, "learning_rate": 0.00016900450225112556, "loss": 4.4727, "step": 1555 }, { "epoch": 0.6454422897438556, "grad_norm": 1.8984375, "learning_rate": 0.00016898449224612307, "loss": 4.2183, "step": 1556 }, { "epoch": 0.6458570984133568, "grad_norm": 2.015625, "learning_rate": 0.00016896448224112056, "loss": 4.1696, "step": 1557 }, { "epoch": 0.646271907082858, "grad_norm": 1.8984375, "learning_rate": 0.00016894447223611808, "loss": 4.1474, "step": 1558 }, { "epoch": 0.6466867157523593, "grad_norm": 1.9140625, "learning_rate": 0.00016892446223111556, "loss": 4.1353, "step": 1559 }, { "epoch": 0.6471015244218604, "grad_norm": 1.859375, "learning_rate": 0.00016890445222611305, "loss": 4.4505, "step": 1560 }, { "epoch": 0.6475163330913616, "grad_norm": 1.9140625, "learning_rate": 0.00016888444222111057, "loss": 4.1747, "step": 1561 }, { "epoch": 0.6479311417608627, "grad_norm": 2.15625, "learning_rate": 0.00016886443221610805, "loss": 4.5492, "step": 1562 }, { "epoch": 0.648345950430364, "grad_norm": 1.84375, "learning_rate": 0.00016884442221110557, "loss": 4.0619, "step": 1563 }, { "epoch": 0.6487607590998652, "grad_norm": 2.15625, "learning_rate": 0.00016882441220610305, "loss": 4.1753, "step": 1564 }, { "epoch": 0.6491755677693664, "grad_norm": 1.8984375, "learning_rate": 0.00016880440220110057, "loss": 4.2605, "step": 1565 }, { "epoch": 0.6495903764388675, "grad_norm": 2.109375, "learning_rate": 0.00016878439219609806, "loss": 4.3825, "step": 1566 }, { "epoch": 0.6500051851083688, "grad_norm": 1.8203125, "learning_rate": 0.00016876438219109557, "loss": 4.3008, "step": 1567 }, { "epoch": 0.65041999377787, "grad_norm": 1.7265625, "learning_rate": 0.00016874437218609306, "loss": 4.4515, "step": 1568 }, { "epoch": 0.6508348024473711, "grad_norm": 1.7890625, "learning_rate": 0.00016872436218109054, "loss": 4.2375, "step": 1569 }, { "epoch": 0.6512496111168723, "grad_norm": 2.046875, "learning_rate": 0.00016870435217608803, "loss": 4.3505, "step": 1570 }, { "epoch": 0.6516644197863736, "grad_norm": 2.015625, "learning_rate": 0.00016868434217108555, "loss": 4.2456, "step": 1571 }, { "epoch": 0.6520792284558747, "grad_norm": 2.03125, "learning_rate": 0.00016866433216608303, "loss": 4.3927, "step": 1572 }, { "epoch": 0.6524940371253759, "grad_norm": 1.921875, "learning_rate": 0.00016864432216108055, "loss": 4.4573, "step": 1573 }, { "epoch": 0.6529088457948771, "grad_norm": 1.8671875, "learning_rate": 0.00016862431215607806, "loss": 4.2828, "step": 1574 }, { "epoch": 0.6533236544643783, "grad_norm": 2.078125, "learning_rate": 0.00016860430215107555, "loss": 4.2621, "step": 1575 }, { "epoch": 0.6537384631338795, "grad_norm": 1.75, "learning_rate": 0.00016858429214607306, "loss": 4.3591, "step": 1576 }, { "epoch": 0.6541532718033807, "grad_norm": 1.875, "learning_rate": 0.00016856428214107055, "loss": 4.3693, "step": 1577 }, { "epoch": 0.6545680804728818, "grad_norm": 1.96875, "learning_rate": 0.00016854427213606804, "loss": 4.2123, "step": 1578 }, { "epoch": 0.6549828891423831, "grad_norm": 2.046875, "learning_rate": 0.00016852426213106553, "loss": 4.3163, "step": 1579 }, { "epoch": 0.6553976978118843, "grad_norm": 1.96875, "learning_rate": 0.00016850425212606304, "loss": 4.2154, "step": 1580 }, { "epoch": 0.6558125064813854, "grad_norm": 1.84375, "learning_rate": 0.00016848424212106053, "loss": 4.4759, "step": 1581 }, { "epoch": 0.6562273151508866, "grad_norm": 1.984375, "learning_rate": 0.00016846423211605804, "loss": 4.3124, "step": 1582 }, { "epoch": 0.6566421238203879, "grad_norm": 1.859375, "learning_rate": 0.00016844422211105553, "loss": 4.1963, "step": 1583 }, { "epoch": 0.6570569324898891, "grad_norm": 2.296875, "learning_rate": 0.00016842421210605304, "loss": 4.4128, "step": 1584 }, { "epoch": 0.6574717411593902, "grad_norm": 1.890625, "learning_rate": 0.00016840420210105053, "loss": 4.1885, "step": 1585 }, { "epoch": 0.6578865498288914, "grad_norm": 1.7734375, "learning_rate": 0.00016838419209604805, "loss": 4.4294, "step": 1586 }, { "epoch": 0.6583013584983927, "grad_norm": 2.03125, "learning_rate": 0.00016836418209104553, "loss": 4.3992, "step": 1587 }, { "epoch": 0.6587161671678938, "grad_norm": 1.9140625, "learning_rate": 0.00016834417208604302, "loss": 4.1201, "step": 1588 }, { "epoch": 0.659130975837395, "grad_norm": 2.046875, "learning_rate": 0.00016832416208104053, "loss": 4.4606, "step": 1589 }, { "epoch": 0.6595457845068962, "grad_norm": 1.7734375, "learning_rate": 0.00016830415207603802, "loss": 4.2413, "step": 1590 }, { "epoch": 0.6599605931763974, "grad_norm": 1.828125, "learning_rate": 0.00016828414207103554, "loss": 4.3569, "step": 1591 }, { "epoch": 0.6603754018458986, "grad_norm": 2.125, "learning_rate": 0.00016826413206603302, "loss": 4.3401, "step": 1592 }, { "epoch": 0.6607902105153998, "grad_norm": 1.859375, "learning_rate": 0.00016824412206103054, "loss": 4.1639, "step": 1593 }, { "epoch": 0.6612050191849009, "grad_norm": 1.890625, "learning_rate": 0.00016822411205602802, "loss": 4.4522, "step": 1594 }, { "epoch": 0.6616198278544022, "grad_norm": 1.8984375, "learning_rate": 0.0001682041020510255, "loss": 4.2886, "step": 1595 }, { "epoch": 0.6620346365239034, "grad_norm": 2.265625, "learning_rate": 0.000168184092046023, "loss": 4.6065, "step": 1596 }, { "epoch": 0.6624494451934045, "grad_norm": 1.8671875, "learning_rate": 0.00016816408204102051, "loss": 4.1547, "step": 1597 }, { "epoch": 0.6628642538629057, "grad_norm": 1.8125, "learning_rate": 0.000168144072036018, "loss": 4.2301, "step": 1598 }, { "epoch": 0.663279062532407, "grad_norm": 1.921875, "learning_rate": 0.00016812406203101552, "loss": 4.4017, "step": 1599 }, { "epoch": 0.6636938712019081, "grad_norm": 2.125, "learning_rate": 0.000168104052026013, "loss": 4.2473, "step": 1600 }, { "epoch": 0.6641086798714093, "grad_norm": 1.96875, "learning_rate": 0.00016808404202101052, "loss": 4.2751, "step": 1601 }, { "epoch": 0.6645234885409105, "grad_norm": 1.9296875, "learning_rate": 0.00016806403201600803, "loss": 4.2893, "step": 1602 }, { "epoch": 0.6649382972104116, "grad_norm": 1.8046875, "learning_rate": 0.00016804402201100552, "loss": 4.2818, "step": 1603 }, { "epoch": 0.6653531058799129, "grad_norm": 1.8125, "learning_rate": 0.000168024012006003, "loss": 4.4164, "step": 1604 }, { "epoch": 0.6657679145494141, "grad_norm": 1.9765625, "learning_rate": 0.0001680040020010005, "loss": 4.0619, "step": 1605 }, { "epoch": 0.6661827232189153, "grad_norm": 2.15625, "learning_rate": 0.000167983991995998, "loss": 4.1438, "step": 1606 }, { "epoch": 0.6665975318884164, "grad_norm": 1.8046875, "learning_rate": 0.0001679639819909955, "loss": 4.1344, "step": 1607 }, { "epoch": 0.6670123405579177, "grad_norm": 1.7421875, "learning_rate": 0.000167943971985993, "loss": 4.3448, "step": 1608 }, { "epoch": 0.6674271492274189, "grad_norm": 1.9140625, "learning_rate": 0.0001679239619809905, "loss": 4.1337, "step": 1609 }, { "epoch": 0.66784195789692, "grad_norm": 1.8203125, "learning_rate": 0.000167903951975988, "loss": 4.48, "step": 1610 }, { "epoch": 0.6682567665664212, "grad_norm": 1.9765625, "learning_rate": 0.0001678839419709855, "loss": 4.3283, "step": 1611 }, { "epoch": 0.6686715752359225, "grad_norm": 1.796875, "learning_rate": 0.000167863931965983, "loss": 4.297, "step": 1612 }, { "epoch": 0.6690863839054236, "grad_norm": 1.8828125, "learning_rate": 0.0001678439219609805, "loss": 4.2076, "step": 1613 }, { "epoch": 0.6695011925749248, "grad_norm": 2.3125, "learning_rate": 0.000167823911955978, "loss": 4.1852, "step": 1614 }, { "epoch": 0.669916001244426, "grad_norm": 1.8046875, "learning_rate": 0.0001678039019509755, "loss": 4.2659, "step": 1615 }, { "epoch": 0.6703308099139272, "grad_norm": 1.8828125, "learning_rate": 0.000167783891945973, "loss": 4.4305, "step": 1616 }, { "epoch": 0.6707456185834284, "grad_norm": 2.0625, "learning_rate": 0.0001677638819409705, "loss": 4.1105, "step": 1617 }, { "epoch": 0.6711604272529296, "grad_norm": 1.9296875, "learning_rate": 0.000167743871935968, "loss": 4.2092, "step": 1618 }, { "epoch": 0.6715752359224307, "grad_norm": 2.15625, "learning_rate": 0.0001677238619309655, "loss": 4.0741, "step": 1619 }, { "epoch": 0.671990044591932, "grad_norm": 1.8984375, "learning_rate": 0.000167703851925963, "loss": 4.1418, "step": 1620 }, { "epoch": 0.6724048532614332, "grad_norm": 1.9921875, "learning_rate": 0.0001676838419209605, "loss": 4.154, "step": 1621 }, { "epoch": 0.6728196619309343, "grad_norm": 2.078125, "learning_rate": 0.000167663831915958, "loss": 4.4036, "step": 1622 }, { "epoch": 0.6732344706004355, "grad_norm": 1.7578125, "learning_rate": 0.00016764382191095548, "loss": 4.218, "step": 1623 }, { "epoch": 0.6736492792699368, "grad_norm": 1.8359375, "learning_rate": 0.00016762381190595297, "loss": 4.2124, "step": 1624 }, { "epoch": 0.674064087939438, "grad_norm": 1.765625, "learning_rate": 0.00016760380190095048, "loss": 4.3872, "step": 1625 }, { "epoch": 0.6744788966089391, "grad_norm": 2.21875, "learning_rate": 0.00016758379189594797, "loss": 4.1842, "step": 1626 }, { "epoch": 0.6748937052784403, "grad_norm": 1.7421875, "learning_rate": 0.00016756378189094548, "loss": 4.2315, "step": 1627 }, { "epoch": 0.6753085139479416, "grad_norm": 1.8359375, "learning_rate": 0.000167543771885943, "loss": 3.9879, "step": 1628 }, { "epoch": 0.6757233226174427, "grad_norm": 2.0, "learning_rate": 0.00016752376188094049, "loss": 4.1754, "step": 1629 }, { "epoch": 0.6761381312869439, "grad_norm": 2.03125, "learning_rate": 0.00016750375187593797, "loss": 4.3931, "step": 1630 }, { "epoch": 0.676552939956445, "grad_norm": 1.7734375, "learning_rate": 0.00016748374187093546, "loss": 4.4069, "step": 1631 }, { "epoch": 0.6769677486259463, "grad_norm": 1.859375, "learning_rate": 0.00016746373186593298, "loss": 4.3356, "step": 1632 }, { "epoch": 0.6773825572954475, "grad_norm": 1.9296875, "learning_rate": 0.00016744372186093046, "loss": 4.3862, "step": 1633 }, { "epoch": 0.6777973659649487, "grad_norm": 2.453125, "learning_rate": 0.00016742371185592798, "loss": 4.2633, "step": 1634 }, { "epoch": 0.6782121746344498, "grad_norm": 2.25, "learning_rate": 0.00016740370185092546, "loss": 4.1539, "step": 1635 }, { "epoch": 0.6786269833039511, "grad_norm": 2.21875, "learning_rate": 0.00016738369184592298, "loss": 4.3728, "step": 1636 }, { "epoch": 0.6790417919734523, "grad_norm": 1.875, "learning_rate": 0.00016736368184092047, "loss": 4.5238, "step": 1637 }, { "epoch": 0.6794566006429534, "grad_norm": 1.8828125, "learning_rate": 0.00016734367183591798, "loss": 4.2689, "step": 1638 }, { "epoch": 0.6798714093124546, "grad_norm": 1.7890625, "learning_rate": 0.00016732366183091547, "loss": 4.2544, "step": 1639 }, { "epoch": 0.6802862179819559, "grad_norm": 2.0625, "learning_rate": 0.00016730365182591295, "loss": 4.1013, "step": 1640 }, { "epoch": 0.680701026651457, "grad_norm": 1.953125, "learning_rate": 0.00016728364182091044, "loss": 4.3, "step": 1641 }, { "epoch": 0.6811158353209582, "grad_norm": 1.7421875, "learning_rate": 0.00016726363181590796, "loss": 4.3177, "step": 1642 }, { "epoch": 0.6815306439904594, "grad_norm": 2.109375, "learning_rate": 0.00016724362181090547, "loss": 4.2262, "step": 1643 }, { "epoch": 0.6819454526599605, "grad_norm": 2.0, "learning_rate": 0.00016722361180590296, "loss": 4.2507, "step": 1644 }, { "epoch": 0.6823602613294618, "grad_norm": 2.015625, "learning_rate": 0.00016720360180090047, "loss": 4.3631, "step": 1645 }, { "epoch": 0.682775069998963, "grad_norm": 2.1875, "learning_rate": 0.00016718359179589796, "loss": 4.4338, "step": 1646 }, { "epoch": 0.6831898786684641, "grad_norm": 2.0625, "learning_rate": 0.00016716358179089547, "loss": 4.3531, "step": 1647 }, { "epoch": 0.6836046873379653, "grad_norm": 1.8046875, "learning_rate": 0.00016714357178589296, "loss": 4.3155, "step": 1648 }, { "epoch": 0.6840194960074666, "grad_norm": 2.421875, "learning_rate": 0.00016712356178089045, "loss": 4.2694, "step": 1649 }, { "epoch": 0.6844343046769678, "grad_norm": 2.0, "learning_rate": 0.00016710355177588794, "loss": 4.4195, "step": 1650 }, { "epoch": 0.6848491133464689, "grad_norm": 1.8125, "learning_rate": 0.00016708354177088545, "loss": 4.2231, "step": 1651 }, { "epoch": 0.6852639220159701, "grad_norm": 2.046875, "learning_rate": 0.00016706353176588294, "loss": 4.3835, "step": 1652 }, { "epoch": 0.6856787306854714, "grad_norm": 2.1875, "learning_rate": 0.00016704352176088045, "loss": 4.2942, "step": 1653 }, { "epoch": 0.6860935393549725, "grad_norm": 1.9453125, "learning_rate": 0.00016702351175587794, "loss": 4.385, "step": 1654 }, { "epoch": 0.6865083480244737, "grad_norm": 2.203125, "learning_rate": 0.00016700350175087545, "loss": 4.2112, "step": 1655 }, { "epoch": 0.6869231566939749, "grad_norm": 1.9375, "learning_rate": 0.00016698349174587297, "loss": 4.5042, "step": 1656 }, { "epoch": 0.6873379653634761, "grad_norm": 1.921875, "learning_rate": 0.00016696348174087046, "loss": 4.2647, "step": 1657 }, { "epoch": 0.6877527740329773, "grad_norm": 1.984375, "learning_rate": 0.00016694347173586794, "loss": 4.3357, "step": 1658 }, { "epoch": 0.6881675827024785, "grad_norm": 2.203125, "learning_rate": 0.00016692346173086543, "loss": 4.2829, "step": 1659 }, { "epoch": 0.6885823913719796, "grad_norm": 1.921875, "learning_rate": 0.00016690345172586294, "loss": 4.2207, "step": 1660 }, { "epoch": 0.6889972000414809, "grad_norm": 1.8828125, "learning_rate": 0.00016688344172086043, "loss": 4.3693, "step": 1661 }, { "epoch": 0.6894120087109821, "grad_norm": 1.9765625, "learning_rate": 0.00016686343171585795, "loss": 4.1872, "step": 1662 }, { "epoch": 0.6898268173804832, "grad_norm": 1.8984375, "learning_rate": 0.00016684342171085543, "loss": 4.3783, "step": 1663 }, { "epoch": 0.6902416260499844, "grad_norm": 1.8203125, "learning_rate": 0.00016682341170585295, "loss": 4.3232, "step": 1664 }, { "epoch": 0.6906564347194857, "grad_norm": 2.0, "learning_rate": 0.00016680340170085043, "loss": 4.3469, "step": 1665 }, { "epoch": 0.6910712433889868, "grad_norm": 2.078125, "learning_rate": 0.00016678339169584792, "loss": 4.3845, "step": 1666 }, { "epoch": 0.691486052058488, "grad_norm": 2.0, "learning_rate": 0.0001667633816908454, "loss": 4.3378, "step": 1667 }, { "epoch": 0.6919008607279892, "grad_norm": 2.21875, "learning_rate": 0.00016674337168584292, "loss": 4.2599, "step": 1668 }, { "epoch": 0.6923156693974905, "grad_norm": 2.140625, "learning_rate": 0.0001667233616808404, "loss": 4.2731, "step": 1669 }, { "epoch": 0.6927304780669916, "grad_norm": 1.8828125, "learning_rate": 0.00016670335167583793, "loss": 4.1236, "step": 1670 }, { "epoch": 0.6931452867364928, "grad_norm": 2.046875, "learning_rate": 0.00016668334167083544, "loss": 4.1105, "step": 1671 }, { "epoch": 0.693560095405994, "grad_norm": 1.8203125, "learning_rate": 0.00016666333166583293, "loss": 4.1263, "step": 1672 }, { "epoch": 0.6939749040754952, "grad_norm": 2.328125, "learning_rate": 0.00016664332166083044, "loss": 4.0781, "step": 1673 }, { "epoch": 0.6943897127449964, "grad_norm": 2.015625, "learning_rate": 0.00016662331165582793, "loss": 4.2547, "step": 1674 }, { "epoch": 0.6948045214144976, "grad_norm": 1.75, "learning_rate": 0.00016660330165082542, "loss": 4.149, "step": 1675 }, { "epoch": 0.6952193300839987, "grad_norm": 1.96875, "learning_rate": 0.0001665832916458229, "loss": 4.1475, "step": 1676 }, { "epoch": 0.6956341387535, "grad_norm": 2.09375, "learning_rate": 0.00016656328164082042, "loss": 4.1184, "step": 1677 }, { "epoch": 0.6960489474230012, "grad_norm": 2.109375, "learning_rate": 0.0001665432716358179, "loss": 4.2829, "step": 1678 }, { "epoch": 0.6964637560925023, "grad_norm": 1.9609375, "learning_rate": 0.00016652326163081542, "loss": 4.2481, "step": 1679 }, { "epoch": 0.6968785647620035, "grad_norm": 2.015625, "learning_rate": 0.0001665032516258129, "loss": 4.5068, "step": 1680 }, { "epoch": 0.6972933734315048, "grad_norm": 2.25, "learning_rate": 0.00016648324162081042, "loss": 4.2291, "step": 1681 }, { "epoch": 0.6977081821010059, "grad_norm": 2.640625, "learning_rate": 0.0001664632316158079, "loss": 4.378, "step": 1682 }, { "epoch": 0.6981229907705071, "grad_norm": 2.03125, "learning_rate": 0.00016644322161080542, "loss": 4.3239, "step": 1683 }, { "epoch": 0.6985377994400083, "grad_norm": 1.7734375, "learning_rate": 0.0001664232116058029, "loss": 4.2392, "step": 1684 }, { "epoch": 0.6989526081095095, "grad_norm": 1.8671875, "learning_rate": 0.0001664032016008004, "loss": 4.1904, "step": 1685 }, { "epoch": 0.6993674167790107, "grad_norm": 2.234375, "learning_rate": 0.0001663831915957979, "loss": 4.1115, "step": 1686 }, { "epoch": 0.6997822254485119, "grad_norm": 1.828125, "learning_rate": 0.0001663631815907954, "loss": 4.076, "step": 1687 }, { "epoch": 0.700197034118013, "grad_norm": 2.125, "learning_rate": 0.0001663431715857929, "loss": 4.5088, "step": 1688 }, { "epoch": 0.7006118427875142, "grad_norm": 2.015625, "learning_rate": 0.0001663231615807904, "loss": 4.3093, "step": 1689 }, { "epoch": 0.7010266514570155, "grad_norm": 1.734375, "learning_rate": 0.00016630315157578791, "loss": 4.1884, "step": 1690 }, { "epoch": 0.7014414601265166, "grad_norm": 1.9453125, "learning_rate": 0.0001662831415707854, "loss": 4.0619, "step": 1691 }, { "epoch": 0.7018562687960178, "grad_norm": 1.671875, "learning_rate": 0.00016626313156578292, "loss": 4.2123, "step": 1692 }, { "epoch": 0.702271077465519, "grad_norm": 2.078125, "learning_rate": 0.0001662431215607804, "loss": 4.0962, "step": 1693 }, { "epoch": 0.7026858861350203, "grad_norm": 2.109375, "learning_rate": 0.0001662231115557779, "loss": 4.3124, "step": 1694 }, { "epoch": 0.7031006948045214, "grad_norm": 1.7578125, "learning_rate": 0.00016620310155077538, "loss": 4.1273, "step": 1695 }, { "epoch": 0.7035155034740226, "grad_norm": 2.046875, "learning_rate": 0.0001661830915457729, "loss": 4.1936, "step": 1696 }, { "epoch": 0.7039303121435238, "grad_norm": 2.359375, "learning_rate": 0.00016616308154077038, "loss": 4.2628, "step": 1697 }, { "epoch": 0.704345120813025, "grad_norm": 1.9765625, "learning_rate": 0.0001661430715357679, "loss": 4.6984, "step": 1698 }, { "epoch": 0.7047599294825262, "grad_norm": 1.90625, "learning_rate": 0.0001661230615307654, "loss": 4.2766, "step": 1699 }, { "epoch": 0.7051747381520274, "grad_norm": 1.5859375, "learning_rate": 0.0001661030515257629, "loss": 4.2213, "step": 1700 }, { "epoch": 0.7055895468215285, "grad_norm": 1.890625, "learning_rate": 0.00016608304152076038, "loss": 4.4998, "step": 1701 }, { "epoch": 0.7060043554910298, "grad_norm": 1.984375, "learning_rate": 0.00016606303151575787, "loss": 4.3587, "step": 1702 }, { "epoch": 0.706419164160531, "grad_norm": 1.8984375, "learning_rate": 0.00016604302151075539, "loss": 4.2724, "step": 1703 }, { "epoch": 0.7068339728300321, "grad_norm": 2.03125, "learning_rate": 0.00016602301150575287, "loss": 4.3734, "step": 1704 }, { "epoch": 0.7072487814995333, "grad_norm": 2.140625, "learning_rate": 0.0001660030015007504, "loss": 4.1035, "step": 1705 }, { "epoch": 0.7076635901690346, "grad_norm": 1.984375, "learning_rate": 0.00016598299149574787, "loss": 4.4057, "step": 1706 }, { "epoch": 0.7080783988385357, "grad_norm": 1.8671875, "learning_rate": 0.0001659629814907454, "loss": 4.2721, "step": 1707 }, { "epoch": 0.7084932075080369, "grad_norm": 2.140625, "learning_rate": 0.00016594297148574288, "loss": 4.168, "step": 1708 }, { "epoch": 0.7089080161775381, "grad_norm": 1.734375, "learning_rate": 0.0001659229614807404, "loss": 4.2883, "step": 1709 }, { "epoch": 0.7093228248470393, "grad_norm": 1.890625, "learning_rate": 0.00016590295147573788, "loss": 4.3137, "step": 1710 }, { "epoch": 0.7097376335165405, "grad_norm": 2.5625, "learning_rate": 0.00016588294147073536, "loss": 4.4237, "step": 1711 }, { "epoch": 0.7101524421860417, "grad_norm": 1.796875, "learning_rate": 0.00016586293146573288, "loss": 4.4452, "step": 1712 }, { "epoch": 0.7105672508555428, "grad_norm": 2.21875, "learning_rate": 0.00016584292146073037, "loss": 4.4238, "step": 1713 }, { "epoch": 0.7109820595250441, "grad_norm": 2.046875, "learning_rate": 0.00016582291145572788, "loss": 4.1477, "step": 1714 }, { "epoch": 0.7113968681945453, "grad_norm": 2.0, "learning_rate": 0.00016580290145072537, "loss": 4.2112, "step": 1715 }, { "epoch": 0.7118116768640464, "grad_norm": 1.796875, "learning_rate": 0.00016578289144572288, "loss": 4.0174, "step": 1716 }, { "epoch": 0.7122264855335476, "grad_norm": 1.9375, "learning_rate": 0.00016576288144072037, "loss": 4.5133, "step": 1717 }, { "epoch": 0.7126412942030489, "grad_norm": 2.015625, "learning_rate": 0.00016574287143571788, "loss": 4.5635, "step": 1718 }, { "epoch": 0.7130561028725501, "grad_norm": 1.859375, "learning_rate": 0.00016572286143071537, "loss": 4.0972, "step": 1719 }, { "epoch": 0.7134709115420512, "grad_norm": 1.9375, "learning_rate": 0.00016570285142571286, "loss": 4.4887, "step": 1720 }, { "epoch": 0.7138857202115524, "grad_norm": 1.953125, "learning_rate": 0.00016568284142071035, "loss": 4.4066, "step": 1721 }, { "epoch": 0.7143005288810537, "grad_norm": 1.9921875, "learning_rate": 0.00016566283141570786, "loss": 3.9391, "step": 1722 }, { "epoch": 0.7147153375505548, "grad_norm": 1.9375, "learning_rate": 0.00016564282141070535, "loss": 4.3862, "step": 1723 }, { "epoch": 0.715130146220056, "grad_norm": 2.34375, "learning_rate": 0.00016562281140570286, "loss": 4.0075, "step": 1724 }, { "epoch": 0.7155449548895572, "grad_norm": 1.8515625, "learning_rate": 0.00016560280140070038, "loss": 4.2231, "step": 1725 }, { "epoch": 0.7159597635590584, "grad_norm": 1.9296875, "learning_rate": 0.00016558279139569786, "loss": 4.2065, "step": 1726 }, { "epoch": 0.7163745722285596, "grad_norm": 1.9140625, "learning_rate": 0.00016556278139069538, "loss": 4.2309, "step": 1727 }, { "epoch": 0.7167893808980608, "grad_norm": 1.7734375, "learning_rate": 0.00016554277138569287, "loss": 4.0488, "step": 1728 }, { "epoch": 0.7172041895675619, "grad_norm": 1.765625, "learning_rate": 0.00016552276138069035, "loss": 4.2463, "step": 1729 }, { "epoch": 0.7176189982370631, "grad_norm": 1.78125, "learning_rate": 0.00016550275137568784, "loss": 4.528, "step": 1730 }, { "epoch": 0.7180338069065644, "grad_norm": 1.8359375, "learning_rate": 0.00016548274137068535, "loss": 4.1754, "step": 1731 }, { "epoch": 0.7184486155760655, "grad_norm": 1.9375, "learning_rate": 0.00016546273136568284, "loss": 4.1617, "step": 1732 }, { "epoch": 0.7188634242455667, "grad_norm": 1.78125, "learning_rate": 0.00016544272136068036, "loss": 4.315, "step": 1733 }, { "epoch": 0.7192782329150679, "grad_norm": 1.9296875, "learning_rate": 0.00016542271135567784, "loss": 4.27, "step": 1734 }, { "epoch": 0.7196930415845691, "grad_norm": 1.90625, "learning_rate": 0.00016540270135067536, "loss": 4.273, "step": 1735 }, { "epoch": 0.7201078502540703, "grad_norm": 1.90625, "learning_rate": 0.00016538269134567284, "loss": 4.2505, "step": 1736 }, { "epoch": 0.7205226589235715, "grad_norm": 2.046875, "learning_rate": 0.00016536268134067033, "loss": 4.4599, "step": 1737 }, { "epoch": 0.7209374675930726, "grad_norm": 2.046875, "learning_rate": 0.00016534267133566782, "loss": 4.193, "step": 1738 }, { "epoch": 0.7213522762625739, "grad_norm": 2.078125, "learning_rate": 0.00016532266133066533, "loss": 4.1589, "step": 1739 }, { "epoch": 0.7217670849320751, "grad_norm": 1.8203125, "learning_rate": 0.00016530265132566285, "loss": 4.2311, "step": 1740 }, { "epoch": 0.7221818936015763, "grad_norm": 2.046875, "learning_rate": 0.00016528264132066034, "loss": 4.0857, "step": 1741 }, { "epoch": 0.7225967022710774, "grad_norm": 2.09375, "learning_rate": 0.00016526263131565785, "loss": 4.213, "step": 1742 }, { "epoch": 0.7230115109405787, "grad_norm": 1.875, "learning_rate": 0.00016524262131065534, "loss": 4.2577, "step": 1743 }, { "epoch": 0.7234263196100799, "grad_norm": 1.890625, "learning_rate": 0.00016522261130565285, "loss": 4.5884, "step": 1744 }, { "epoch": 0.723841128279581, "grad_norm": 2.09375, "learning_rate": 0.00016520260130065034, "loss": 4.2585, "step": 1745 }, { "epoch": 0.7242559369490822, "grad_norm": 1.9453125, "learning_rate": 0.00016518259129564783, "loss": 4.4407, "step": 1746 }, { "epoch": 0.7246707456185835, "grad_norm": 1.96875, "learning_rate": 0.0001651625812906453, "loss": 4.621, "step": 1747 }, { "epoch": 0.7250855542880846, "grad_norm": 1.890625, "learning_rate": 0.00016514257128564283, "loss": 4.5925, "step": 1748 }, { "epoch": 0.7255003629575858, "grad_norm": 1.8984375, "learning_rate": 0.00016512256128064031, "loss": 4.1396, "step": 1749 }, { "epoch": 0.725915171627087, "grad_norm": 1.953125, "learning_rate": 0.00016510255127563783, "loss": 4.409, "step": 1750 }, { "epoch": 0.7263299802965882, "grad_norm": 1.859375, "learning_rate": 0.00016508254127063532, "loss": 4.4109, "step": 1751 }, { "epoch": 0.7267447889660894, "grad_norm": 2.0625, "learning_rate": 0.00016506253126563283, "loss": 4.3669, "step": 1752 }, { "epoch": 0.7271595976355906, "grad_norm": 1.9453125, "learning_rate": 0.00016504252126063035, "loss": 4.3425, "step": 1753 }, { "epoch": 0.7275744063050917, "grad_norm": 1.8984375, "learning_rate": 0.00016502251125562783, "loss": 4.2442, "step": 1754 }, { "epoch": 0.727989214974593, "grad_norm": 1.8203125, "learning_rate": 0.00016500250125062532, "loss": 4.1008, "step": 1755 }, { "epoch": 0.7284040236440942, "grad_norm": 2.140625, "learning_rate": 0.0001649824912456228, "loss": 4.5, "step": 1756 }, { "epoch": 0.7288188323135953, "grad_norm": 2.140625, "learning_rate": 0.00016496248124062032, "loss": 4.0528, "step": 1757 }, { "epoch": 0.7292336409830965, "grad_norm": 2.234375, "learning_rate": 0.0001649424712356178, "loss": 4.3436, "step": 1758 }, { "epoch": 0.7296484496525978, "grad_norm": 2.265625, "learning_rate": 0.00016492246123061532, "loss": 4.214, "step": 1759 }, { "epoch": 0.730063258322099, "grad_norm": 1.6640625, "learning_rate": 0.0001649024512256128, "loss": 4.262, "step": 1760 }, { "epoch": 0.7304780669916001, "grad_norm": 1.8359375, "learning_rate": 0.00016488244122061032, "loss": 4.417, "step": 1761 }, { "epoch": 0.7308928756611013, "grad_norm": 1.859375, "learning_rate": 0.0001648624312156078, "loss": 4.0478, "step": 1762 }, { "epoch": 0.7313076843306026, "grad_norm": 1.8984375, "learning_rate": 0.00016484242121060533, "loss": 4.3204, "step": 1763 }, { "epoch": 0.7317224930001037, "grad_norm": 1.9765625, "learning_rate": 0.0001648224112056028, "loss": 3.9769, "step": 1764 }, { "epoch": 0.7321373016696049, "grad_norm": 1.7109375, "learning_rate": 0.0001648024012006003, "loss": 4.3183, "step": 1765 }, { "epoch": 0.732552110339106, "grad_norm": 2.015625, "learning_rate": 0.0001647823911955978, "loss": 4.2596, "step": 1766 }, { "epoch": 0.7329669190086073, "grad_norm": 1.734375, "learning_rate": 0.0001647623811905953, "loss": 4.1137, "step": 1767 }, { "epoch": 0.7333817276781085, "grad_norm": 2.140625, "learning_rate": 0.00016474237118559282, "loss": 4.4027, "step": 1768 }, { "epoch": 0.7337965363476097, "grad_norm": 2.0625, "learning_rate": 0.0001647223611805903, "loss": 4.3033, "step": 1769 }, { "epoch": 0.7342113450171108, "grad_norm": 1.765625, "learning_rate": 0.00016470235117558782, "loss": 4.4344, "step": 1770 }, { "epoch": 0.7346261536866121, "grad_norm": 1.8203125, "learning_rate": 0.0001646823411705853, "loss": 4.4358, "step": 1771 }, { "epoch": 0.7350409623561133, "grad_norm": 1.78125, "learning_rate": 0.0001646623311655828, "loss": 4.2787, "step": 1772 }, { "epoch": 0.7354557710256144, "grad_norm": 1.796875, "learning_rate": 0.00016464232116058028, "loss": 4.303, "step": 1773 }, { "epoch": 0.7358705796951156, "grad_norm": 1.9140625, "learning_rate": 0.0001646223111555778, "loss": 4.3666, "step": 1774 }, { "epoch": 0.7362853883646168, "grad_norm": 1.9921875, "learning_rate": 0.00016460230115057528, "loss": 4.6831, "step": 1775 }, { "epoch": 0.736700197034118, "grad_norm": 1.859375, "learning_rate": 0.0001645822911455728, "loss": 4.2366, "step": 1776 }, { "epoch": 0.7371150057036192, "grad_norm": 1.7890625, "learning_rate": 0.00016456228114057028, "loss": 4.2226, "step": 1777 }, { "epoch": 0.7375298143731204, "grad_norm": 1.796875, "learning_rate": 0.0001645422711355678, "loss": 4.1475, "step": 1778 }, { "epoch": 0.7379446230426215, "grad_norm": 2.0, "learning_rate": 0.00016452226113056529, "loss": 4.1387, "step": 1779 }, { "epoch": 0.7383594317121228, "grad_norm": 1.8046875, "learning_rate": 0.0001645022511255628, "loss": 4.2031, "step": 1780 }, { "epoch": 0.738774240381624, "grad_norm": 1.9453125, "learning_rate": 0.0001644822411205603, "loss": 4.5157, "step": 1781 }, { "epoch": 0.7391890490511251, "grad_norm": 2.171875, "learning_rate": 0.00016446223111555777, "loss": 4.2089, "step": 1782 }, { "epoch": 0.7396038577206263, "grad_norm": 1.8359375, "learning_rate": 0.0001644422211105553, "loss": 4.1667, "step": 1783 }, { "epoch": 0.7400186663901276, "grad_norm": 2.1875, "learning_rate": 0.00016442221110555278, "loss": 4.168, "step": 1784 }, { "epoch": 0.7404334750596288, "grad_norm": 2.59375, "learning_rate": 0.0001644022011005503, "loss": 4.2752, "step": 1785 }, { "epoch": 0.7408482837291299, "grad_norm": 1.984375, "learning_rate": 0.00016438219109554778, "loss": 4.6423, "step": 1786 }, { "epoch": 0.7412630923986311, "grad_norm": 1.765625, "learning_rate": 0.0001643621810905453, "loss": 4.1835, "step": 1787 }, { "epoch": 0.7416779010681324, "grad_norm": 1.828125, "learning_rate": 0.00016434217108554278, "loss": 4.3036, "step": 1788 }, { "epoch": 0.7420927097376335, "grad_norm": 1.8359375, "learning_rate": 0.0001643221610805403, "loss": 4.3894, "step": 1789 }, { "epoch": 0.7425075184071347, "grad_norm": 1.9609375, "learning_rate": 0.00016430215107553778, "loss": 4.2872, "step": 1790 }, { "epoch": 0.7429223270766359, "grad_norm": 1.8671875, "learning_rate": 0.00016428214107053527, "loss": 4.3255, "step": 1791 }, { "epoch": 0.7433371357461371, "grad_norm": 2.0625, "learning_rate": 0.00016426213106553276, "loss": 4.183, "step": 1792 }, { "epoch": 0.7437519444156383, "grad_norm": 1.9609375, "learning_rate": 0.00016424212106053027, "loss": 4.2447, "step": 1793 }, { "epoch": 0.7441667530851395, "grad_norm": 1.75, "learning_rate": 0.00016422211105552776, "loss": 4.0575, "step": 1794 }, { "epoch": 0.7445815617546406, "grad_norm": 1.8828125, "learning_rate": 0.00016420210105052527, "loss": 4.0873, "step": 1795 }, { "epoch": 0.7449963704241419, "grad_norm": 2.03125, "learning_rate": 0.00016418209104552279, "loss": 4.3977, "step": 1796 }, { "epoch": 0.7454111790936431, "grad_norm": 1.8203125, "learning_rate": 0.00016416208104052027, "loss": 4.3011, "step": 1797 }, { "epoch": 0.7458259877631442, "grad_norm": 1.8515625, "learning_rate": 0.0001641420710355178, "loss": 4.0408, "step": 1798 }, { "epoch": 0.7462407964326454, "grad_norm": 1.9609375, "learning_rate": 0.00016412206103051525, "loss": 4.4989, "step": 1799 }, { "epoch": 0.7466556051021467, "grad_norm": 1.9921875, "learning_rate": 0.00016410205102551276, "loss": 4.4159, "step": 1800 }, { "epoch": 0.7470704137716478, "grad_norm": 1.7890625, "learning_rate": 0.00016408204102051025, "loss": 4.1536, "step": 1801 }, { "epoch": 0.747485222441149, "grad_norm": 1.953125, "learning_rate": 0.00016406203101550776, "loss": 4.2909, "step": 1802 }, { "epoch": 0.7479000311106502, "grad_norm": 1.953125, "learning_rate": 0.00016404202101050525, "loss": 4.523, "step": 1803 }, { "epoch": 0.7483148397801515, "grad_norm": 1.8359375, "learning_rate": 0.00016402201100550277, "loss": 4.1955, "step": 1804 }, { "epoch": 0.7487296484496526, "grad_norm": 1.9609375, "learning_rate": 0.00016400200100050025, "loss": 4.2691, "step": 1805 }, { "epoch": 0.7491444571191538, "grad_norm": 2.015625, "learning_rate": 0.00016398199099549777, "loss": 4.2078, "step": 1806 }, { "epoch": 0.749559265788655, "grad_norm": 1.7890625, "learning_rate": 0.00016396198099049525, "loss": 4.3648, "step": 1807 }, { "epoch": 0.7499740744581562, "grad_norm": 2.0625, "learning_rate": 0.00016394197098549274, "loss": 4.1677, "step": 1808 }, { "epoch": 0.7503888831276574, "grad_norm": 2.078125, "learning_rate": 0.00016392196098049026, "loss": 4.3633, "step": 1809 }, { "epoch": 0.7508036917971586, "grad_norm": 1.8046875, "learning_rate": 0.00016390195097548774, "loss": 4.2876, "step": 1810 }, { "epoch": 0.7512185004666597, "grad_norm": 1.9453125, "learning_rate": 0.00016388194097048526, "loss": 4.217, "step": 1811 }, { "epoch": 0.751633309136161, "grad_norm": 1.953125, "learning_rate": 0.00016386193096548275, "loss": 4.2404, "step": 1812 }, { "epoch": 0.7520481178056622, "grad_norm": 1.9140625, "learning_rate": 0.00016384192096048026, "loss": 4.5665, "step": 1813 }, { "epoch": 0.7524629264751633, "grad_norm": 2.015625, "learning_rate": 0.00016382191095547775, "loss": 4.4474, "step": 1814 }, { "epoch": 0.7528777351446645, "grad_norm": 2.1875, "learning_rate": 0.00016380190095047526, "loss": 4.1574, "step": 1815 }, { "epoch": 0.7532925438141658, "grad_norm": 1.796875, "learning_rate": 0.00016378189094547275, "loss": 4.1432, "step": 1816 }, { "epoch": 0.7537073524836669, "grad_norm": 1.890625, "learning_rate": 0.00016376188094047024, "loss": 4.2979, "step": 1817 }, { "epoch": 0.7541221611531681, "grad_norm": 2.0625, "learning_rate": 0.00016374187093546772, "loss": 4.1143, "step": 1818 }, { "epoch": 0.7545369698226693, "grad_norm": 1.765625, "learning_rate": 0.00016372186093046524, "loss": 4.2105, "step": 1819 }, { "epoch": 0.7549517784921704, "grad_norm": 2.125, "learning_rate": 0.00016370185092546272, "loss": 4.3099, "step": 1820 }, { "epoch": 0.7553665871616717, "grad_norm": 1.7734375, "learning_rate": 0.00016368184092046024, "loss": 4.0391, "step": 1821 }, { "epoch": 0.7557813958311729, "grad_norm": 2.015625, "learning_rate": 0.00016366183091545773, "loss": 4.3743, "step": 1822 }, { "epoch": 0.756196204500674, "grad_norm": 1.9453125, "learning_rate": 0.00016364182091045524, "loss": 4.5217, "step": 1823 }, { "epoch": 0.7566110131701752, "grad_norm": 2.203125, "learning_rate": 0.00016362181090545276, "loss": 4.1872, "step": 1824 }, { "epoch": 0.7570258218396765, "grad_norm": 1.8359375, "learning_rate": 0.00016360180090045024, "loss": 4.1789, "step": 1825 }, { "epoch": 0.7574406305091776, "grad_norm": 1.6640625, "learning_rate": 0.00016358179089544773, "loss": 4.352, "step": 1826 }, { "epoch": 0.7578554391786788, "grad_norm": 1.875, "learning_rate": 0.00016356178089044522, "loss": 4.2776, "step": 1827 }, { "epoch": 0.75827024784818, "grad_norm": 1.9375, "learning_rate": 0.00016354177088544273, "loss": 4.4706, "step": 1828 }, { "epoch": 0.7586850565176813, "grad_norm": 1.9609375, "learning_rate": 0.00016352176088044022, "loss": 4.3542, "step": 1829 }, { "epoch": 0.7590998651871824, "grad_norm": 1.9296875, "learning_rate": 0.00016350175087543773, "loss": 4.5175, "step": 1830 }, { "epoch": 0.7595146738566836, "grad_norm": 1.7734375, "learning_rate": 0.00016348174087043522, "loss": 4.4048, "step": 1831 }, { "epoch": 0.7599294825261848, "grad_norm": 1.8203125, "learning_rate": 0.00016346173086543273, "loss": 4.3196, "step": 1832 }, { "epoch": 0.760344291195686, "grad_norm": 1.921875, "learning_rate": 0.00016344172086043022, "loss": 4.2025, "step": 1833 }, { "epoch": 0.7607590998651872, "grad_norm": 1.75, "learning_rate": 0.00016342171085542774, "loss": 3.9915, "step": 1834 }, { "epoch": 0.7611739085346884, "grad_norm": 1.875, "learning_rate": 0.0001634017008504252, "loss": 4.2192, "step": 1835 }, { "epoch": 0.7615887172041895, "grad_norm": 1.9375, "learning_rate": 0.0001633816908454227, "loss": 4.0859, "step": 1836 }, { "epoch": 0.7620035258736908, "grad_norm": 1.9140625, "learning_rate": 0.00016336168084042023, "loss": 4.0278, "step": 1837 }, { "epoch": 0.762418334543192, "grad_norm": 2.0, "learning_rate": 0.0001633416708354177, "loss": 4.1441, "step": 1838 }, { "epoch": 0.7628331432126931, "grad_norm": 2.0, "learning_rate": 0.00016332166083041523, "loss": 4.198, "step": 1839 }, { "epoch": 0.7632479518821943, "grad_norm": 1.7734375, "learning_rate": 0.00016330165082541271, "loss": 4.103, "step": 1840 }, { "epoch": 0.7636627605516956, "grad_norm": 2.078125, "learning_rate": 0.00016328164082041023, "loss": 4.271, "step": 1841 }, { "epoch": 0.7640775692211967, "grad_norm": 1.9765625, "learning_rate": 0.00016326163081540772, "loss": 4.3573, "step": 1842 }, { "epoch": 0.7644923778906979, "grad_norm": 1.9140625, "learning_rate": 0.0001632416208104052, "loss": 4.4857, "step": 1843 }, { "epoch": 0.7649071865601991, "grad_norm": 1.8125, "learning_rate": 0.0001632216108054027, "loss": 4.3643, "step": 1844 }, { "epoch": 0.7653219952297003, "grad_norm": 2.140625, "learning_rate": 0.0001632016008004002, "loss": 4.2624, "step": 1845 }, { "epoch": 0.7657368038992015, "grad_norm": 1.953125, "learning_rate": 0.0001631815907953977, "loss": 4.2398, "step": 1846 }, { "epoch": 0.7661516125687027, "grad_norm": 1.96875, "learning_rate": 0.0001631615807903952, "loss": 4.2691, "step": 1847 }, { "epoch": 0.7665664212382038, "grad_norm": 2.046875, "learning_rate": 0.0001631415707853927, "loss": 4.1937, "step": 1848 }, { "epoch": 0.7669812299077051, "grad_norm": 1.84375, "learning_rate": 0.0001631215607803902, "loss": 4.2516, "step": 1849 }, { "epoch": 0.7673960385772063, "grad_norm": 1.8984375, "learning_rate": 0.00016310155077538772, "loss": 4.224, "step": 1850 }, { "epoch": 0.7678108472467075, "grad_norm": 1.875, "learning_rate": 0.0001630815407703852, "loss": 4.0366, "step": 1851 }, { "epoch": 0.7682256559162086, "grad_norm": 1.953125, "learning_rate": 0.0001630615307653827, "loss": 4.0647, "step": 1852 }, { "epoch": 0.7686404645857099, "grad_norm": 1.734375, "learning_rate": 0.00016304152076038018, "loss": 4.4412, "step": 1853 }, { "epoch": 0.7690552732552111, "grad_norm": 1.75, "learning_rate": 0.0001630215107553777, "loss": 4.2287, "step": 1854 }, { "epoch": 0.7694700819247122, "grad_norm": 1.78125, "learning_rate": 0.00016300150075037519, "loss": 4.1925, "step": 1855 }, { "epoch": 0.7698848905942134, "grad_norm": 1.8359375, "learning_rate": 0.0001629814907453727, "loss": 4.4409, "step": 1856 }, { "epoch": 0.7702996992637147, "grad_norm": 2.5, "learning_rate": 0.0001629614807403702, "loss": 4.2517, "step": 1857 }, { "epoch": 0.7707145079332158, "grad_norm": 1.890625, "learning_rate": 0.0001629414707353677, "loss": 4.4236, "step": 1858 }, { "epoch": 0.771129316602717, "grad_norm": 1.7421875, "learning_rate": 0.0001629214607303652, "loss": 4.2584, "step": 1859 }, { "epoch": 0.7715441252722182, "grad_norm": 2.09375, "learning_rate": 0.0001629014507253627, "loss": 4.2856, "step": 1860 }, { "epoch": 0.7719589339417193, "grad_norm": 1.890625, "learning_rate": 0.0001628814407203602, "loss": 4.3121, "step": 1861 }, { "epoch": 0.7723737426112206, "grad_norm": 1.9296875, "learning_rate": 0.00016286143071535768, "loss": 4.5599, "step": 1862 }, { "epoch": 0.7727885512807218, "grad_norm": 2.03125, "learning_rate": 0.00016284142071035517, "loss": 4.1843, "step": 1863 }, { "epoch": 0.7732033599502229, "grad_norm": 1.6640625, "learning_rate": 0.00016282141070535268, "loss": 4.2699, "step": 1864 }, { "epoch": 0.7736181686197241, "grad_norm": 2.046875, "learning_rate": 0.0001628014007003502, "loss": 4.5623, "step": 1865 }, { "epoch": 0.7740329772892254, "grad_norm": 1.90625, "learning_rate": 0.00016278139069534768, "loss": 4.1239, "step": 1866 }, { "epoch": 0.7744477859587265, "grad_norm": 1.7890625, "learning_rate": 0.0001627613806903452, "loss": 4.0881, "step": 1867 }, { "epoch": 0.7748625946282277, "grad_norm": 1.7890625, "learning_rate": 0.00016274137068534268, "loss": 4.3054, "step": 1868 }, { "epoch": 0.7752774032977289, "grad_norm": 1.84375, "learning_rate": 0.0001627213606803402, "loss": 4.3178, "step": 1869 }, { "epoch": 0.7756922119672301, "grad_norm": 1.6875, "learning_rate": 0.00016270135067533766, "loss": 4.1595, "step": 1870 }, { "epoch": 0.7761070206367313, "grad_norm": 1.8828125, "learning_rate": 0.00016268134067033517, "loss": 4.4073, "step": 1871 }, { "epoch": 0.7765218293062325, "grad_norm": 1.7734375, "learning_rate": 0.00016266133066533266, "loss": 4.418, "step": 1872 }, { "epoch": 0.7769366379757336, "grad_norm": 2.078125, "learning_rate": 0.00016264132066033017, "loss": 4.334, "step": 1873 }, { "epoch": 0.7773514466452349, "grad_norm": 1.7109375, "learning_rate": 0.00016262131065532766, "loss": 4.2096, "step": 1874 }, { "epoch": 0.7777662553147361, "grad_norm": 1.90625, "learning_rate": 0.00016260130065032518, "loss": 4.145, "step": 1875 }, { "epoch": 0.7781810639842373, "grad_norm": 1.8515625, "learning_rate": 0.00016258129064532266, "loss": 4.3219, "step": 1876 }, { "epoch": 0.7785958726537384, "grad_norm": 1.734375, "learning_rate": 0.00016256128064032018, "loss": 4.4551, "step": 1877 }, { "epoch": 0.7790106813232397, "grad_norm": 1.9296875, "learning_rate": 0.00016254127063531766, "loss": 4.2221, "step": 1878 }, { "epoch": 0.7794254899927409, "grad_norm": 1.75, "learning_rate": 0.00016252126063031515, "loss": 4.0949, "step": 1879 }, { "epoch": 0.779840298662242, "grad_norm": 2.0625, "learning_rate": 0.00016250125062531267, "loss": 4.455, "step": 1880 }, { "epoch": 0.7802551073317432, "grad_norm": 1.890625, "learning_rate": 0.00016248124062031015, "loss": 4.3139, "step": 1881 }, { "epoch": 0.7806699160012445, "grad_norm": 1.828125, "learning_rate": 0.00016246123061530767, "loss": 4.1647, "step": 1882 }, { "epoch": 0.7810847246707456, "grad_norm": 1.859375, "learning_rate": 0.00016244122061030516, "loss": 4.0756, "step": 1883 }, { "epoch": 0.7814995333402468, "grad_norm": 1.6796875, "learning_rate": 0.00016242121060530267, "loss": 4.4128, "step": 1884 }, { "epoch": 0.781914342009748, "grad_norm": 1.8046875, "learning_rate": 0.00016240120060030016, "loss": 4.12, "step": 1885 }, { "epoch": 0.7823291506792492, "grad_norm": 1.9921875, "learning_rate": 0.00016238119059529767, "loss": 4.333, "step": 1886 }, { "epoch": 0.7827439593487504, "grad_norm": 1.8515625, "learning_rate": 0.00016236118059029516, "loss": 4.2122, "step": 1887 }, { "epoch": 0.7831587680182516, "grad_norm": 1.75, "learning_rate": 0.00016234117058529265, "loss": 4.3084, "step": 1888 }, { "epoch": 0.7835735766877527, "grad_norm": 1.9453125, "learning_rate": 0.00016232116058029013, "loss": 4.5061, "step": 1889 }, { "epoch": 0.783988385357254, "grad_norm": 1.9296875, "learning_rate": 0.00016230115057528765, "loss": 4.1753, "step": 1890 }, { "epoch": 0.7844031940267552, "grad_norm": 2.125, "learning_rate": 0.00016228114057028514, "loss": 4.1175, "step": 1891 }, { "epoch": 0.7848180026962563, "grad_norm": 1.7734375, "learning_rate": 0.00016226113056528265, "loss": 4.4553, "step": 1892 }, { "epoch": 0.7852328113657575, "grad_norm": 1.984375, "learning_rate": 0.00016224112056028016, "loss": 4.3577, "step": 1893 }, { "epoch": 0.7856476200352588, "grad_norm": 1.96875, "learning_rate": 0.00016222111055527765, "loss": 4.3326, "step": 1894 }, { "epoch": 0.78606242870476, "grad_norm": 2.0, "learning_rate": 0.00016220110055027517, "loss": 4.2376, "step": 1895 }, { "epoch": 0.7864772373742611, "grad_norm": 1.7578125, "learning_rate": 0.00016218109054527265, "loss": 4.1652, "step": 1896 }, { "epoch": 0.7868920460437623, "grad_norm": 1.9453125, "learning_rate": 0.00016216108054027014, "loss": 4.1028, "step": 1897 }, { "epoch": 0.7873068547132636, "grad_norm": 1.96875, "learning_rate": 0.00016214107053526763, "loss": 4.166, "step": 1898 }, { "epoch": 0.7877216633827647, "grad_norm": 1.96875, "learning_rate": 0.00016212106053026514, "loss": 4.2311, "step": 1899 }, { "epoch": 0.7881364720522659, "grad_norm": 1.9296875, "learning_rate": 0.00016210105052526263, "loss": 4.3764, "step": 1900 }, { "epoch": 0.7885512807217671, "grad_norm": 1.953125, "learning_rate": 0.00016208104052026014, "loss": 4.4922, "step": 1901 }, { "epoch": 0.7889660893912683, "grad_norm": 1.703125, "learning_rate": 0.00016206103051525763, "loss": 4.5501, "step": 1902 }, { "epoch": 0.7893808980607695, "grad_norm": 1.8359375, "learning_rate": 0.00016204102051025515, "loss": 4.098, "step": 1903 }, { "epoch": 0.7897957067302707, "grad_norm": 1.7109375, "learning_rate": 0.00016202101050525263, "loss": 4.1942, "step": 1904 }, { "epoch": 0.7902105153997718, "grad_norm": 2.015625, "learning_rate": 0.00016200100050025012, "loss": 4.1957, "step": 1905 }, { "epoch": 0.790625324069273, "grad_norm": 1.8203125, "learning_rate": 0.00016198099049524763, "loss": 4.1782, "step": 1906 }, { "epoch": 0.7910401327387743, "grad_norm": 1.8515625, "learning_rate": 0.00016196098049024512, "loss": 4.083, "step": 1907 }, { "epoch": 0.7914549414082754, "grad_norm": 1.84375, "learning_rate": 0.00016194097048524264, "loss": 4.2352, "step": 1908 }, { "epoch": 0.7918697500777766, "grad_norm": 1.84375, "learning_rate": 0.00016192096048024012, "loss": 4.2171, "step": 1909 }, { "epoch": 0.7922845587472778, "grad_norm": 1.8828125, "learning_rate": 0.00016190095047523764, "loss": 4.4288, "step": 1910 }, { "epoch": 0.792699367416779, "grad_norm": 2.015625, "learning_rate": 0.00016188094047023512, "loss": 4.3989, "step": 1911 }, { "epoch": 0.7931141760862802, "grad_norm": 1.828125, "learning_rate": 0.00016186093046523264, "loss": 4.2165, "step": 1912 }, { "epoch": 0.7935289847557814, "grad_norm": 1.9296875, "learning_rate": 0.00016184092046023013, "loss": 4.4195, "step": 1913 }, { "epoch": 0.7939437934252825, "grad_norm": 1.859375, "learning_rate": 0.0001618209104552276, "loss": 4.4894, "step": 1914 }, { "epoch": 0.7943586020947838, "grad_norm": 1.953125, "learning_rate": 0.0001618009004502251, "loss": 4.1961, "step": 1915 }, { "epoch": 0.794773410764285, "grad_norm": 2.0625, "learning_rate": 0.00016178089044522262, "loss": 4.2061, "step": 1916 }, { "epoch": 0.7951882194337861, "grad_norm": 1.875, "learning_rate": 0.0001617608804402201, "loss": 4.4158, "step": 1917 }, { "epoch": 0.7956030281032873, "grad_norm": 2.109375, "learning_rate": 0.00016174087043521762, "loss": 4.1264, "step": 1918 }, { "epoch": 0.7960178367727886, "grad_norm": 1.8984375, "learning_rate": 0.0001617208604302151, "loss": 4.1823, "step": 1919 }, { "epoch": 0.7964326454422898, "grad_norm": 1.7578125, "learning_rate": 0.00016170085042521262, "loss": 4.2566, "step": 1920 }, { "epoch": 0.7968474541117909, "grad_norm": 1.828125, "learning_rate": 0.00016168084042021013, "loss": 4.405, "step": 1921 }, { "epoch": 0.7972622627812921, "grad_norm": 2.171875, "learning_rate": 0.00016166083041520762, "loss": 4.1805, "step": 1922 }, { "epoch": 0.7976770714507934, "grad_norm": 1.6328125, "learning_rate": 0.0001616408204102051, "loss": 4.3404, "step": 1923 }, { "epoch": 0.7980918801202945, "grad_norm": 1.8671875, "learning_rate": 0.0001616208104052026, "loss": 4.0248, "step": 1924 }, { "epoch": 0.7985066887897957, "grad_norm": 1.9140625, "learning_rate": 0.0001616008004002001, "loss": 4.155, "step": 1925 }, { "epoch": 0.7989214974592969, "grad_norm": 1.765625, "learning_rate": 0.0001615807903951976, "loss": 4.0772, "step": 1926 }, { "epoch": 0.7993363061287981, "grad_norm": 1.6328125, "learning_rate": 0.0001615607803901951, "loss": 4.2948, "step": 1927 }, { "epoch": 0.7997511147982993, "grad_norm": 1.71875, "learning_rate": 0.0001615407703851926, "loss": 4.3545, "step": 1928 }, { "epoch": 0.8001659234678005, "grad_norm": 1.9453125, "learning_rate": 0.0001615207603801901, "loss": 4.2569, "step": 1929 }, { "epoch": 0.8005807321373016, "grad_norm": 1.6171875, "learning_rate": 0.0001615007503751876, "loss": 4.3671, "step": 1930 }, { "epoch": 0.8009955408068029, "grad_norm": 1.7578125, "learning_rate": 0.00016148074037018511, "loss": 4.2905, "step": 1931 }, { "epoch": 0.8014103494763041, "grad_norm": 1.8515625, "learning_rate": 0.0001614607303651826, "loss": 4.2352, "step": 1932 }, { "epoch": 0.8018251581458052, "grad_norm": 1.828125, "learning_rate": 0.0001614407203601801, "loss": 4.4255, "step": 1933 }, { "epoch": 0.8022399668153064, "grad_norm": 1.8203125, "learning_rate": 0.0001614207103551776, "loss": 4.0116, "step": 1934 }, { "epoch": 0.8026547754848077, "grad_norm": 1.703125, "learning_rate": 0.0001614007003501751, "loss": 4.12, "step": 1935 }, { "epoch": 0.8030695841543088, "grad_norm": 1.9140625, "learning_rate": 0.0001613806903451726, "loss": 4.1637, "step": 1936 }, { "epoch": 0.80348439282381, "grad_norm": 1.8359375, "learning_rate": 0.0001613606803401701, "loss": 4.3536, "step": 1937 }, { "epoch": 0.8038992014933112, "grad_norm": 1.6328125, "learning_rate": 0.0001613406703351676, "loss": 4.267, "step": 1938 }, { "epoch": 0.8043140101628125, "grad_norm": 1.8046875, "learning_rate": 0.0001613206603301651, "loss": 4.4897, "step": 1939 }, { "epoch": 0.8047288188323136, "grad_norm": 1.6484375, "learning_rate": 0.00016130065032516258, "loss": 4.1714, "step": 1940 }, { "epoch": 0.8051436275018148, "grad_norm": 1.984375, "learning_rate": 0.00016128064032016007, "loss": 4.2259, "step": 1941 }, { "epoch": 0.805558436171316, "grad_norm": 1.5859375, "learning_rate": 0.00016126063031515758, "loss": 4.2209, "step": 1942 }, { "epoch": 0.8059732448408172, "grad_norm": 1.765625, "learning_rate": 0.00016124062031015507, "loss": 4.1714, "step": 1943 }, { "epoch": 0.8063880535103184, "grad_norm": 1.765625, "learning_rate": 0.00016122061030515258, "loss": 4.144, "step": 1944 }, { "epoch": 0.8068028621798196, "grad_norm": 1.921875, "learning_rate": 0.00016120060030015007, "loss": 4.06, "step": 1945 }, { "epoch": 0.8072176708493207, "grad_norm": 1.7421875, "learning_rate": 0.00016118059029514759, "loss": 4.2008, "step": 1946 }, { "epoch": 0.8076324795188219, "grad_norm": 2.3125, "learning_rate": 0.0001611605802901451, "loss": 4.2848, "step": 1947 }, { "epoch": 0.8080472881883232, "grad_norm": 2.296875, "learning_rate": 0.0001611405702851426, "loss": 4.2837, "step": 1948 }, { "epoch": 0.8084620968578243, "grad_norm": 2.0625, "learning_rate": 0.00016112056028014007, "loss": 4.3758, "step": 1949 }, { "epoch": 0.8088769055273255, "grad_norm": 1.734375, "learning_rate": 0.00016110055027513756, "loss": 4.3414, "step": 1950 }, { "epoch": 0.8092917141968267, "grad_norm": 1.8046875, "learning_rate": 0.00016108054027013508, "loss": 4.2916, "step": 1951 }, { "epoch": 0.8097065228663279, "grad_norm": 1.8359375, "learning_rate": 0.00016106053026513256, "loss": 4.1258, "step": 1952 }, { "epoch": 0.8101213315358291, "grad_norm": 1.703125, "learning_rate": 0.00016104052026013008, "loss": 4.5907, "step": 1953 }, { "epoch": 0.8105361402053303, "grad_norm": 1.8203125, "learning_rate": 0.00016102051025512757, "loss": 4.2004, "step": 1954 }, { "epoch": 0.8109509488748314, "grad_norm": 1.9765625, "learning_rate": 0.00016100050025012508, "loss": 4.0807, "step": 1955 }, { "epoch": 0.8113657575443327, "grad_norm": 1.890625, "learning_rate": 0.00016098049024512257, "loss": 4.2416, "step": 1956 }, { "epoch": 0.8117805662138339, "grad_norm": 1.703125, "learning_rate": 0.00016096048024012008, "loss": 4.5034, "step": 1957 }, { "epoch": 0.812195374883335, "grad_norm": 1.7890625, "learning_rate": 0.00016094047023511757, "loss": 3.9632, "step": 1958 }, { "epoch": 0.8126101835528362, "grad_norm": 1.6640625, "learning_rate": 0.00016092046023011506, "loss": 4.3644, "step": 1959 }, { "epoch": 0.8130249922223375, "grad_norm": 2.109375, "learning_rate": 0.00016090045022511254, "loss": 4.4008, "step": 1960 }, { "epoch": 0.8134398008918386, "grad_norm": 1.9609375, "learning_rate": 0.00016088044022011006, "loss": 4.2465, "step": 1961 }, { "epoch": 0.8138546095613398, "grad_norm": 1.84375, "learning_rate": 0.00016086043021510757, "loss": 4.267, "step": 1962 }, { "epoch": 0.814269418230841, "grad_norm": 1.921875, "learning_rate": 0.00016084042021010506, "loss": 4.1756, "step": 1963 }, { "epoch": 0.8146842269003423, "grad_norm": 1.828125, "learning_rate": 0.00016082041020510257, "loss": 4.1291, "step": 1964 }, { "epoch": 0.8150990355698434, "grad_norm": 1.8359375, "learning_rate": 0.00016080040020010006, "loss": 4.139, "step": 1965 }, { "epoch": 0.8155138442393446, "grad_norm": 1.796875, "learning_rate": 0.00016078039019509758, "loss": 4.1982, "step": 1966 }, { "epoch": 0.8159286529088458, "grad_norm": 1.671875, "learning_rate": 0.00016076038019009506, "loss": 4.0969, "step": 1967 }, { "epoch": 0.816343461578347, "grad_norm": 1.6171875, "learning_rate": 0.00016074037018509255, "loss": 4.2043, "step": 1968 }, { "epoch": 0.8167582702478482, "grad_norm": 1.953125, "learning_rate": 0.00016072036018009004, "loss": 4.1012, "step": 1969 }, { "epoch": 0.8171730789173494, "grad_norm": 1.8046875, "learning_rate": 0.00016070035017508755, "loss": 4.1507, "step": 1970 }, { "epoch": 0.8175878875868505, "grad_norm": 1.5625, "learning_rate": 0.00016068034017008504, "loss": 4.0963, "step": 1971 }, { "epoch": 0.8180026962563518, "grad_norm": 1.78125, "learning_rate": 0.00016066033016508255, "loss": 4.2856, "step": 1972 }, { "epoch": 0.818417504925853, "grad_norm": 1.7734375, "learning_rate": 0.00016064032016008004, "loss": 4.0937, "step": 1973 }, { "epoch": 0.8188323135953541, "grad_norm": 1.7265625, "learning_rate": 0.00016062031015507756, "loss": 4.0772, "step": 1974 }, { "epoch": 0.8192471222648553, "grad_norm": 1.796875, "learning_rate": 0.00016060030015007507, "loss": 3.9974, "step": 1975 }, { "epoch": 0.8196619309343566, "grad_norm": 1.8984375, "learning_rate": 0.00016058029014507253, "loss": 4.0982, "step": 1976 }, { "epoch": 0.8200767396038577, "grad_norm": 1.6875, "learning_rate": 0.00016056028014007004, "loss": 4.3984, "step": 1977 }, { "epoch": 0.8204915482733589, "grad_norm": 1.7890625, "learning_rate": 0.00016054027013506753, "loss": 4.0591, "step": 1978 }, { "epoch": 0.8209063569428601, "grad_norm": 1.859375, "learning_rate": 0.00016052026013006505, "loss": 4.1635, "step": 1979 }, { "epoch": 0.8213211656123613, "grad_norm": 1.8828125, "learning_rate": 0.00016050025012506253, "loss": 4.3204, "step": 1980 }, { "epoch": 0.8217359742818625, "grad_norm": 1.7734375, "learning_rate": 0.00016048024012006005, "loss": 4.1082, "step": 1981 }, { "epoch": 0.8221507829513637, "grad_norm": 1.8125, "learning_rate": 0.00016046023011505753, "loss": 4.1742, "step": 1982 }, { "epoch": 0.8225655916208648, "grad_norm": 1.6953125, "learning_rate": 0.00016044022011005505, "loss": 4.3782, "step": 1983 }, { "epoch": 0.8229804002903661, "grad_norm": 1.8984375, "learning_rate": 0.00016042021010505254, "loss": 4.2875, "step": 1984 }, { "epoch": 0.8233952089598673, "grad_norm": 1.9921875, "learning_rate": 0.00016040020010005002, "loss": 4.3425, "step": 1985 }, { "epoch": 0.8238100176293685, "grad_norm": 2.28125, "learning_rate": 0.0001603801900950475, "loss": 4.1819, "step": 1986 }, { "epoch": 0.8242248262988696, "grad_norm": 1.8515625, "learning_rate": 0.00016036018009004503, "loss": 4.197, "step": 1987 }, { "epoch": 0.8246396349683709, "grad_norm": 1.9609375, "learning_rate": 0.0001603401700850425, "loss": 4.2766, "step": 1988 }, { "epoch": 0.8250544436378721, "grad_norm": 2.15625, "learning_rate": 0.00016032016008004003, "loss": 4.2494, "step": 1989 }, { "epoch": 0.8254692523073732, "grad_norm": 2.0, "learning_rate": 0.00016030015007503754, "loss": 4.3987, "step": 1990 }, { "epoch": 0.8258840609768744, "grad_norm": 1.65625, "learning_rate": 0.00016028014007003503, "loss": 4.1686, "step": 1991 }, { "epoch": 0.8262988696463756, "grad_norm": 2.046875, "learning_rate": 0.00016026013006503254, "loss": 4.2965, "step": 1992 }, { "epoch": 0.8267136783158768, "grad_norm": 1.828125, "learning_rate": 0.00016024012006003003, "loss": 4.1832, "step": 1993 }, { "epoch": 0.827128486985378, "grad_norm": 1.9296875, "learning_rate": 0.00016022011005502752, "loss": 4.0664, "step": 1994 }, { "epoch": 0.8275432956548792, "grad_norm": 2.046875, "learning_rate": 0.000160200100050025, "loss": 4.1719, "step": 1995 }, { "epoch": 0.8279581043243803, "grad_norm": 1.9375, "learning_rate": 0.00016018009004502252, "loss": 4.3811, "step": 1996 }, { "epoch": 0.8283729129938816, "grad_norm": 1.890625, "learning_rate": 0.00016016008004002, "loss": 4.007, "step": 1997 }, { "epoch": 0.8287877216633828, "grad_norm": 1.6953125, "learning_rate": 0.00016014007003501752, "loss": 4.1535, "step": 1998 }, { "epoch": 0.8292025303328839, "grad_norm": 2.203125, "learning_rate": 0.000160120060030015, "loss": 4.1579, "step": 1999 }, { "epoch": 0.8296173390023851, "grad_norm": 1.8359375, "learning_rate": 0.00016010005002501252, "loss": 4.3744, "step": 2000 }, { "epoch": 0.8300321476718864, "grad_norm": 1.8671875, "learning_rate": 0.00016008004002001, "loss": 4.0944, "step": 2001 }, { "epoch": 0.8304469563413875, "grad_norm": 1.9375, "learning_rate": 0.00016006003001500752, "loss": 4.6294, "step": 2002 }, { "epoch": 0.8308617650108887, "grad_norm": 1.6953125, "learning_rate": 0.000160040020010005, "loss": 4.1584, "step": 2003 }, { "epoch": 0.8312765736803899, "grad_norm": 1.7109375, "learning_rate": 0.0001600200100050025, "loss": 4.3815, "step": 2004 }, { "epoch": 0.8316913823498912, "grad_norm": 1.78125, "learning_rate": 0.00016, "loss": 4.2994, "step": 2005 }, { "epoch": 0.8321061910193923, "grad_norm": 1.7734375, "learning_rate": 0.0001599799899949975, "loss": 4.2908, "step": 2006 }, { "epoch": 0.8325209996888935, "grad_norm": 1.9453125, "learning_rate": 0.00015995997998999501, "loss": 4.5768, "step": 2007 }, { "epoch": 0.8329358083583946, "grad_norm": 1.734375, "learning_rate": 0.0001599399699849925, "loss": 4.1997, "step": 2008 }, { "epoch": 0.8333506170278959, "grad_norm": 1.828125, "learning_rate": 0.00015991995997999002, "loss": 4.1174, "step": 2009 }, { "epoch": 0.8337654256973971, "grad_norm": 1.765625, "learning_rate": 0.0001598999499749875, "loss": 4.2547, "step": 2010 }, { "epoch": 0.8341802343668983, "grad_norm": 1.7890625, "learning_rate": 0.000159879939969985, "loss": 3.9529, "step": 2011 }, { "epoch": 0.8345950430363994, "grad_norm": 1.8828125, "learning_rate": 0.00015985992996498248, "loss": 4.4329, "step": 2012 }, { "epoch": 0.8350098517059007, "grad_norm": 2.03125, "learning_rate": 0.00015983991995998, "loss": 4.1626, "step": 2013 }, { "epoch": 0.8354246603754019, "grad_norm": 1.8515625, "learning_rate": 0.00015981990995497748, "loss": 4.3675, "step": 2014 }, { "epoch": 0.835839469044903, "grad_norm": 1.9765625, "learning_rate": 0.000159799899949975, "loss": 4.0634, "step": 2015 }, { "epoch": 0.8362542777144042, "grad_norm": 1.7734375, "learning_rate": 0.00015977988994497248, "loss": 4.0611, "step": 2016 }, { "epoch": 0.8366690863839055, "grad_norm": 1.8359375, "learning_rate": 0.00015975987993997, "loss": 4.2217, "step": 2017 }, { "epoch": 0.8370838950534066, "grad_norm": 1.859375, "learning_rate": 0.0001597398699349675, "loss": 4.1328, "step": 2018 }, { "epoch": 0.8374987037229078, "grad_norm": 1.671875, "learning_rate": 0.000159719859929965, "loss": 4.281, "step": 2019 }, { "epoch": 0.837913512392409, "grad_norm": 1.734375, "learning_rate": 0.00015969984992496248, "loss": 4.2259, "step": 2020 }, { "epoch": 0.8383283210619102, "grad_norm": 1.828125, "learning_rate": 0.00015967983991995997, "loss": 4.3036, "step": 2021 }, { "epoch": 0.8387431297314114, "grad_norm": 1.8671875, "learning_rate": 0.0001596598299149575, "loss": 4.236, "step": 2022 }, { "epoch": 0.8391579384009126, "grad_norm": 1.9375, "learning_rate": 0.00015963981990995497, "loss": 4.1976, "step": 2023 }, { "epoch": 0.8395727470704137, "grad_norm": 2.1875, "learning_rate": 0.0001596198099049525, "loss": 3.9924, "step": 2024 }, { "epoch": 0.839987555739915, "grad_norm": 1.7890625, "learning_rate": 0.00015959979989994998, "loss": 4.2626, "step": 2025 }, { "epoch": 0.8404023644094162, "grad_norm": 1.7890625, "learning_rate": 0.0001595797898949475, "loss": 4.1824, "step": 2026 }, { "epoch": 0.8408171730789173, "grad_norm": 1.859375, "learning_rate": 0.00015955977988994498, "loss": 4.1441, "step": 2027 }, { "epoch": 0.8412319817484185, "grad_norm": 1.9140625, "learning_rate": 0.0001595397698849425, "loss": 4.019, "step": 2028 }, { "epoch": 0.8416467904179198, "grad_norm": 2.140625, "learning_rate": 0.00015951975987993998, "loss": 4.3492, "step": 2029 }, { "epoch": 0.842061599087421, "grad_norm": 1.890625, "learning_rate": 0.00015949974987493747, "loss": 4.048, "step": 2030 }, { "epoch": 0.8424764077569221, "grad_norm": 1.734375, "learning_rate": 0.00015947973986993498, "loss": 4.0639, "step": 2031 }, { "epoch": 0.8428912164264233, "grad_norm": 1.9296875, "learning_rate": 0.00015945972986493247, "loss": 4.2477, "step": 2032 }, { "epoch": 0.8433060250959244, "grad_norm": 2.078125, "learning_rate": 0.00015943971985992998, "loss": 4.1823, "step": 2033 }, { "epoch": 0.8437208337654257, "grad_norm": 1.8671875, "learning_rate": 0.00015941970985492747, "loss": 4.299, "step": 2034 }, { "epoch": 0.8441356424349269, "grad_norm": 1.7734375, "learning_rate": 0.00015939969984992498, "loss": 4.2606, "step": 2035 }, { "epoch": 0.8445504511044281, "grad_norm": 1.875, "learning_rate": 0.00015937968984492247, "loss": 4.2716, "step": 2036 }, { "epoch": 0.8449652597739292, "grad_norm": 1.859375, "learning_rate": 0.00015935967983991999, "loss": 4.1566, "step": 2037 }, { "epoch": 0.8453800684434305, "grad_norm": 1.8671875, "learning_rate": 0.00015933966983491747, "loss": 4.3453, "step": 2038 }, { "epoch": 0.8457948771129317, "grad_norm": 1.828125, "learning_rate": 0.00015931965982991496, "loss": 4.0379, "step": 2039 }, { "epoch": 0.8462096857824328, "grad_norm": 1.8828125, "learning_rate": 0.00015929964982491245, "loss": 4.0374, "step": 2040 }, { "epoch": 0.846624494451934, "grad_norm": 1.921875, "learning_rate": 0.00015927963981990996, "loss": 4.2103, "step": 2041 }, { "epoch": 0.8470393031214353, "grad_norm": 2.0625, "learning_rate": 0.00015925962981490745, "loss": 4.2873, "step": 2042 }, { "epoch": 0.8474541117909364, "grad_norm": 1.875, "learning_rate": 0.00015923961980990496, "loss": 4.3596, "step": 2043 }, { "epoch": 0.8478689204604376, "grad_norm": 1.921875, "learning_rate": 0.00015921960980490248, "loss": 4.2258, "step": 2044 }, { "epoch": 0.8482837291299388, "grad_norm": 1.8671875, "learning_rate": 0.00015919959979989997, "loss": 4.416, "step": 2045 }, { "epoch": 0.84869853779944, "grad_norm": 1.96875, "learning_rate": 0.00015917958979489745, "loss": 4.1224, "step": 2046 }, { "epoch": 0.8491133464689412, "grad_norm": 1.8046875, "learning_rate": 0.00015915957978989494, "loss": 4.1571, "step": 2047 }, { "epoch": 0.8495281551384424, "grad_norm": 1.8046875, "learning_rate": 0.00015913956978489245, "loss": 4.0442, "step": 2048 }, { "epoch": 0.8499429638079435, "grad_norm": 1.640625, "learning_rate": 0.00015911955977988994, "loss": 4.2112, "step": 2049 }, { "epoch": 0.8503577724774448, "grad_norm": 1.7109375, "learning_rate": 0.00015909954977488746, "loss": 4.0259, "step": 2050 }, { "epoch": 0.850772581146946, "grad_norm": 1.8203125, "learning_rate": 0.00015907953976988494, "loss": 4.1872, "step": 2051 }, { "epoch": 0.8511873898164471, "grad_norm": 1.765625, "learning_rate": 0.00015905952976488246, "loss": 4.1497, "step": 2052 }, { "epoch": 0.8516021984859483, "grad_norm": 1.8515625, "learning_rate": 0.00015903951975987994, "loss": 4.0657, "step": 2053 }, { "epoch": 0.8520170071554496, "grad_norm": 1.9375, "learning_rate": 0.00015901950975487746, "loss": 4.0609, "step": 2054 }, { "epoch": 0.8524318158249508, "grad_norm": 2.09375, "learning_rate": 0.00015899949974987495, "loss": 3.9947, "step": 2055 }, { "epoch": 0.8528466244944519, "grad_norm": 1.8828125, "learning_rate": 0.00015897948974487243, "loss": 4.1609, "step": 2056 }, { "epoch": 0.8532614331639531, "grad_norm": 1.828125, "learning_rate": 0.00015895947973986992, "loss": 4.1743, "step": 2057 }, { "epoch": 0.8536762418334544, "grad_norm": 1.890625, "learning_rate": 0.00015893946973486744, "loss": 4.3007, "step": 2058 }, { "epoch": 0.8540910505029555, "grad_norm": 1.9296875, "learning_rate": 0.00015891945972986495, "loss": 4.3385, "step": 2059 }, { "epoch": 0.8545058591724567, "grad_norm": 1.75, "learning_rate": 0.00015889944972486244, "loss": 4.2264, "step": 2060 }, { "epoch": 0.8549206678419579, "grad_norm": 1.75, "learning_rate": 0.00015887943971985995, "loss": 4.3002, "step": 2061 }, { "epoch": 0.8553354765114591, "grad_norm": 2.140625, "learning_rate": 0.00015885942971485744, "loss": 4.4753, "step": 2062 }, { "epoch": 0.8557502851809603, "grad_norm": 1.796875, "learning_rate": 0.00015883941970985495, "loss": 4.255, "step": 2063 }, { "epoch": 0.8561650938504615, "grad_norm": 2.09375, "learning_rate": 0.00015881940970485244, "loss": 4.0755, "step": 2064 }, { "epoch": 0.8565799025199626, "grad_norm": 1.7421875, "learning_rate": 0.00015879939969984993, "loss": 4.3859, "step": 2065 }, { "epoch": 0.8569947111894639, "grad_norm": 1.7265625, "learning_rate": 0.00015877938969484741, "loss": 4.4064, "step": 2066 }, { "epoch": 0.8574095198589651, "grad_norm": 1.8515625, "learning_rate": 0.00015875937968984493, "loss": 4.0705, "step": 2067 }, { "epoch": 0.8578243285284662, "grad_norm": 1.9453125, "learning_rate": 0.00015873936968484242, "loss": 3.9955, "step": 2068 }, { "epoch": 0.8582391371979674, "grad_norm": 1.78125, "learning_rate": 0.00015871935967983993, "loss": 4.3657, "step": 2069 }, { "epoch": 0.8586539458674687, "grad_norm": 1.8359375, "learning_rate": 0.00015869934967483742, "loss": 4.3224, "step": 2070 }, { "epoch": 0.8590687545369698, "grad_norm": 1.7578125, "learning_rate": 0.00015867933966983493, "loss": 4.264, "step": 2071 }, { "epoch": 0.859483563206471, "grad_norm": 1.6640625, "learning_rate": 0.00015865932966483245, "loss": 4.0686, "step": 2072 }, { "epoch": 0.8598983718759722, "grad_norm": 2.125, "learning_rate": 0.00015863931965982993, "loss": 4.2345, "step": 2073 }, { "epoch": 0.8603131805454735, "grad_norm": 1.796875, "learning_rate": 0.00015861930965482742, "loss": 4.2295, "step": 2074 }, { "epoch": 0.8607279892149746, "grad_norm": 1.96875, "learning_rate": 0.0001585992996498249, "loss": 4.0095, "step": 2075 }, { "epoch": 0.8611427978844758, "grad_norm": 1.71875, "learning_rate": 0.00015857928964482242, "loss": 4.178, "step": 2076 }, { "epoch": 0.861557606553977, "grad_norm": 1.90625, "learning_rate": 0.0001585592796398199, "loss": 4.3847, "step": 2077 }, { "epoch": 0.8619724152234781, "grad_norm": 1.890625, "learning_rate": 0.00015853926963481742, "loss": 4.3103, "step": 2078 }, { "epoch": 0.8623872238929794, "grad_norm": 2.078125, "learning_rate": 0.0001585192596298149, "loss": 4.1957, "step": 2079 }, { "epoch": 0.8628020325624806, "grad_norm": 1.84375, "learning_rate": 0.00015849924962481243, "loss": 4.1953, "step": 2080 }, { "epoch": 0.8632168412319817, "grad_norm": 1.96875, "learning_rate": 0.00015847923961980991, "loss": 4.1814, "step": 2081 }, { "epoch": 0.8636316499014829, "grad_norm": 1.8046875, "learning_rate": 0.0001584592296148074, "loss": 4.0798, "step": 2082 }, { "epoch": 0.8640464585709842, "grad_norm": 1.7890625, "learning_rate": 0.0001584392196098049, "loss": 4.0589, "step": 2083 }, { "epoch": 0.8644612672404853, "grad_norm": 1.8515625, "learning_rate": 0.0001584192096048024, "loss": 4.1309, "step": 2084 }, { "epoch": 0.8648760759099865, "grad_norm": 1.8125, "learning_rate": 0.0001583991995997999, "loss": 4.1277, "step": 2085 }, { "epoch": 0.8652908845794877, "grad_norm": 1.8828125, "learning_rate": 0.0001583791895947974, "loss": 4.2242, "step": 2086 }, { "epoch": 0.8657056932489889, "grad_norm": 1.875, "learning_rate": 0.00015835917958979492, "loss": 4.2391, "step": 2087 }, { "epoch": 0.8661205019184901, "grad_norm": 2.078125, "learning_rate": 0.0001583391695847924, "loss": 4.2056, "step": 2088 }, { "epoch": 0.8665353105879913, "grad_norm": 1.8359375, "learning_rate": 0.00015831915957978992, "loss": 4.3799, "step": 2089 }, { "epoch": 0.8669501192574924, "grad_norm": 1.78125, "learning_rate": 0.0001582991495747874, "loss": 4.16, "step": 2090 }, { "epoch": 0.8673649279269937, "grad_norm": 1.90625, "learning_rate": 0.0001582791395697849, "loss": 4.5515, "step": 2091 }, { "epoch": 0.8677797365964949, "grad_norm": 1.875, "learning_rate": 0.00015825912956478238, "loss": 4.3274, "step": 2092 }, { "epoch": 0.868194545265996, "grad_norm": 1.7734375, "learning_rate": 0.0001582391195597799, "loss": 4.2869, "step": 2093 }, { "epoch": 0.8686093539354972, "grad_norm": 1.8046875, "learning_rate": 0.00015821910955477738, "loss": 4.2235, "step": 2094 }, { "epoch": 0.8690241626049985, "grad_norm": 1.8515625, "learning_rate": 0.0001581990995497749, "loss": 4.0367, "step": 2095 }, { "epoch": 0.8694389712744997, "grad_norm": 1.8515625, "learning_rate": 0.00015817908954477239, "loss": 4.1958, "step": 2096 }, { "epoch": 0.8698537799440008, "grad_norm": 1.7734375, "learning_rate": 0.0001581590795397699, "loss": 4.2112, "step": 2097 }, { "epoch": 0.870268588613502, "grad_norm": 1.90625, "learning_rate": 0.0001581390695347674, "loss": 4.2587, "step": 2098 }, { "epoch": 0.8706833972830033, "grad_norm": 1.671875, "learning_rate": 0.0001581190595297649, "loss": 4.1819, "step": 2099 }, { "epoch": 0.8710982059525044, "grad_norm": 1.953125, "learning_rate": 0.0001580990495247624, "loss": 4.3334, "step": 2100 }, { "epoch": 0.8715130146220056, "grad_norm": 1.8046875, "learning_rate": 0.00015807903951975988, "loss": 4.7363, "step": 2101 }, { "epoch": 0.8719278232915068, "grad_norm": 1.828125, "learning_rate": 0.0001580590295147574, "loss": 4.2494, "step": 2102 }, { "epoch": 0.872342631961008, "grad_norm": 1.859375, "learning_rate": 0.00015803901950975488, "loss": 4.2409, "step": 2103 }, { "epoch": 0.8727574406305092, "grad_norm": 1.75, "learning_rate": 0.0001580190095047524, "loss": 4.3603, "step": 2104 }, { "epoch": 0.8731722493000104, "grad_norm": 1.7421875, "learning_rate": 0.00015799899949974988, "loss": 4.0479, "step": 2105 }, { "epoch": 0.8735870579695115, "grad_norm": 1.84375, "learning_rate": 0.0001579789894947474, "loss": 4.1474, "step": 2106 }, { "epoch": 0.8740018666390128, "grad_norm": 1.84375, "learning_rate": 0.00015795897948974488, "loss": 4.5182, "step": 2107 }, { "epoch": 0.874416675308514, "grad_norm": 1.8828125, "learning_rate": 0.0001579389694847424, "loss": 4.2338, "step": 2108 }, { "epoch": 0.8748314839780151, "grad_norm": 1.984375, "learning_rate": 0.00015791895947973988, "loss": 4.1606, "step": 2109 }, { "epoch": 0.8752462926475163, "grad_norm": 1.7890625, "learning_rate": 0.00015789894947473737, "loss": 4.1812, "step": 2110 }, { "epoch": 0.8756611013170176, "grad_norm": 1.9765625, "learning_rate": 0.00015787893946973486, "loss": 4.228, "step": 2111 }, { "epoch": 0.8760759099865187, "grad_norm": 1.8203125, "learning_rate": 0.00015785892946473237, "loss": 4.0865, "step": 2112 }, { "epoch": 0.8764907186560199, "grad_norm": 1.828125, "learning_rate": 0.00015783891945972986, "loss": 4.292, "step": 2113 }, { "epoch": 0.8769055273255211, "grad_norm": 1.96875, "learning_rate": 0.00015781890945472737, "loss": 4.152, "step": 2114 }, { "epoch": 0.8773203359950223, "grad_norm": 1.8359375, "learning_rate": 0.0001577988994497249, "loss": 4.2904, "step": 2115 }, { "epoch": 0.8777351446645235, "grad_norm": 1.953125, "learning_rate": 0.00015777888944472238, "loss": 4.2578, "step": 2116 }, { "epoch": 0.8781499533340247, "grad_norm": 2.25, "learning_rate": 0.00015775887943971986, "loss": 4.2063, "step": 2117 }, { "epoch": 0.8785647620035258, "grad_norm": 1.8046875, "learning_rate": 0.00015773886943471735, "loss": 4.3394, "step": 2118 }, { "epoch": 0.878979570673027, "grad_norm": 1.8828125, "learning_rate": 0.00015771885942971486, "loss": 4.1386, "step": 2119 }, { "epoch": 0.8793943793425283, "grad_norm": 1.8984375, "learning_rate": 0.00015769884942471235, "loss": 4.02, "step": 2120 }, { "epoch": 0.8798091880120295, "grad_norm": 1.890625, "learning_rate": 0.00015767883941970987, "loss": 4.0521, "step": 2121 }, { "epoch": 0.8802239966815306, "grad_norm": 1.796875, "learning_rate": 0.00015765882941470735, "loss": 4.4281, "step": 2122 }, { "epoch": 0.8806388053510318, "grad_norm": 1.875, "learning_rate": 0.00015763881940970487, "loss": 4.2597, "step": 2123 }, { "epoch": 0.8810536140205331, "grad_norm": 2.0625, "learning_rate": 0.00015761880940470235, "loss": 4.1314, "step": 2124 }, { "epoch": 0.8814684226900342, "grad_norm": 2.03125, "learning_rate": 0.00015759879939969987, "loss": 4.4347, "step": 2125 }, { "epoch": 0.8818832313595354, "grad_norm": 2.3125, "learning_rate": 0.00015757878939469736, "loss": 4.3353, "step": 2126 }, { "epoch": 0.8822980400290366, "grad_norm": 1.921875, "learning_rate": 0.00015755877938969484, "loss": 4.2573, "step": 2127 }, { "epoch": 0.8827128486985378, "grad_norm": 1.8046875, "learning_rate": 0.00015753876938469236, "loss": 4.2421, "step": 2128 }, { "epoch": 0.883127657368039, "grad_norm": 1.7734375, "learning_rate": 0.00015751875937968985, "loss": 4.3355, "step": 2129 }, { "epoch": 0.8835424660375402, "grad_norm": 1.8125, "learning_rate": 0.00015749874937468736, "loss": 4.3497, "step": 2130 }, { "epoch": 0.8839572747070413, "grad_norm": 1.8828125, "learning_rate": 0.00015747873936968485, "loss": 4.2719, "step": 2131 }, { "epoch": 0.8843720833765426, "grad_norm": 1.8984375, "learning_rate": 0.00015745872936468236, "loss": 3.9974, "step": 2132 }, { "epoch": 0.8847868920460438, "grad_norm": 1.8984375, "learning_rate": 0.00015743871935967985, "loss": 4.3923, "step": 2133 }, { "epoch": 0.8852017007155449, "grad_norm": 1.859375, "learning_rate": 0.00015741870935467736, "loss": 4.1713, "step": 2134 }, { "epoch": 0.8856165093850461, "grad_norm": 1.875, "learning_rate": 0.00015739869934967485, "loss": 4.2646, "step": 2135 }, { "epoch": 0.8860313180545474, "grad_norm": 1.84375, "learning_rate": 0.00015737868934467234, "loss": 4.3139, "step": 2136 }, { "epoch": 0.8864461267240485, "grad_norm": 1.9375, "learning_rate": 0.00015735867933966982, "loss": 4.2384, "step": 2137 }, { "epoch": 0.8868609353935497, "grad_norm": 1.65625, "learning_rate": 0.00015733866933466734, "loss": 4.1046, "step": 2138 }, { "epoch": 0.8872757440630509, "grad_norm": 1.8046875, "learning_rate": 0.00015731865932966483, "loss": 4.1377, "step": 2139 }, { "epoch": 0.8876905527325522, "grad_norm": 1.78125, "learning_rate": 0.00015729864932466234, "loss": 4.2011, "step": 2140 }, { "epoch": 0.8881053614020533, "grad_norm": 1.828125, "learning_rate": 0.00015727863931965986, "loss": 4.4309, "step": 2141 }, { "epoch": 0.8885201700715545, "grad_norm": 2.015625, "learning_rate": 0.00015725862931465734, "loss": 4.3402, "step": 2142 }, { "epoch": 0.8889349787410556, "grad_norm": 1.625, "learning_rate": 0.00015723861930965486, "loss": 4.1873, "step": 2143 }, { "epoch": 0.8893497874105569, "grad_norm": 1.625, "learning_rate": 0.00015721860930465234, "loss": 4.2746, "step": 2144 }, { "epoch": 0.8897645960800581, "grad_norm": 1.7578125, "learning_rate": 0.00015719859929964983, "loss": 4.187, "step": 2145 }, { "epoch": 0.8901794047495593, "grad_norm": 1.9765625, "learning_rate": 0.00015717858929464732, "loss": 4.036, "step": 2146 }, { "epoch": 0.8905942134190604, "grad_norm": 1.8046875, "learning_rate": 0.00015715857928964483, "loss": 4.2291, "step": 2147 }, { "epoch": 0.8910090220885617, "grad_norm": 1.7890625, "learning_rate": 0.00015713856928464232, "loss": 4.0767, "step": 2148 }, { "epoch": 0.8914238307580629, "grad_norm": 1.7109375, "learning_rate": 0.00015711855927963983, "loss": 4.4093, "step": 2149 }, { "epoch": 0.891838639427564, "grad_norm": 1.8828125, "learning_rate": 0.00015709854927463732, "loss": 4.4395, "step": 2150 }, { "epoch": 0.8922534480970652, "grad_norm": 1.796875, "learning_rate": 0.00015707853926963484, "loss": 4.4243, "step": 2151 }, { "epoch": 0.8926682567665665, "grad_norm": 1.7734375, "learning_rate": 0.00015705852926463232, "loss": 4.4296, "step": 2152 }, { "epoch": 0.8930830654360676, "grad_norm": 1.6640625, "learning_rate": 0.0001570385192596298, "loss": 4.1258, "step": 2153 }, { "epoch": 0.8934978741055688, "grad_norm": 1.7890625, "learning_rate": 0.0001570185092546273, "loss": 4.2254, "step": 2154 }, { "epoch": 0.89391268277507, "grad_norm": 1.953125, "learning_rate": 0.0001569984992496248, "loss": 4.3687, "step": 2155 }, { "epoch": 0.8943274914445712, "grad_norm": 1.875, "learning_rate": 0.00015697848924462233, "loss": 4.0439, "step": 2156 }, { "epoch": 0.8947423001140724, "grad_norm": 1.890625, "learning_rate": 0.00015695847923961981, "loss": 4.4301, "step": 2157 }, { "epoch": 0.8951571087835736, "grad_norm": 2.015625, "learning_rate": 0.00015693846923461733, "loss": 4.1404, "step": 2158 }, { "epoch": 0.8955719174530747, "grad_norm": 1.953125, "learning_rate": 0.00015691845922961482, "loss": 4.0361, "step": 2159 }, { "epoch": 0.895986726122576, "grad_norm": 1.8671875, "learning_rate": 0.00015689844922461233, "loss": 4.05, "step": 2160 }, { "epoch": 0.8964015347920772, "grad_norm": 2.015625, "learning_rate": 0.00015687843921960982, "loss": 4.376, "step": 2161 }, { "epoch": 0.8968163434615783, "grad_norm": 2.109375, "learning_rate": 0.0001568584292146073, "loss": 4.2686, "step": 2162 }, { "epoch": 0.8972311521310795, "grad_norm": 2.015625, "learning_rate": 0.0001568384192096048, "loss": 4.1084, "step": 2163 }, { "epoch": 0.8976459608005807, "grad_norm": 1.8984375, "learning_rate": 0.0001568184092046023, "loss": 4.1774, "step": 2164 }, { "epoch": 0.898060769470082, "grad_norm": 1.921875, "learning_rate": 0.0001567983991995998, "loss": 4.4955, "step": 2165 }, { "epoch": 0.8984755781395831, "grad_norm": 1.8125, "learning_rate": 0.0001567783891945973, "loss": 4.3741, "step": 2166 }, { "epoch": 0.8988903868090843, "grad_norm": 1.7421875, "learning_rate": 0.0001567583791895948, "loss": 4.4769, "step": 2167 }, { "epoch": 0.8993051954785855, "grad_norm": 1.7734375, "learning_rate": 0.0001567383691845923, "loss": 4.2052, "step": 2168 }, { "epoch": 0.8997200041480867, "grad_norm": 1.828125, "learning_rate": 0.00015671835917958982, "loss": 4.1392, "step": 2169 }, { "epoch": 0.9001348128175879, "grad_norm": 1.9375, "learning_rate": 0.0001566983491745873, "loss": 4.179, "step": 2170 }, { "epoch": 0.9005496214870891, "grad_norm": 1.8828125, "learning_rate": 0.0001566783391695848, "loss": 4.6296, "step": 2171 }, { "epoch": 0.9009644301565902, "grad_norm": 2.046875, "learning_rate": 0.00015665832916458229, "loss": 4.2581, "step": 2172 }, { "epoch": 0.9013792388260915, "grad_norm": 1.765625, "learning_rate": 0.0001566383191595798, "loss": 4.3259, "step": 2173 }, { "epoch": 0.9017940474955927, "grad_norm": 1.8125, "learning_rate": 0.0001566183091545773, "loss": 3.9884, "step": 2174 }, { "epoch": 0.9022088561650938, "grad_norm": 1.6328125, "learning_rate": 0.0001565982991495748, "loss": 4.2128, "step": 2175 }, { "epoch": 0.902623664834595, "grad_norm": 1.8984375, "learning_rate": 0.0001565782891445723, "loss": 4.1028, "step": 2176 }, { "epoch": 0.9030384735040963, "grad_norm": 1.9140625, "learning_rate": 0.0001565582791395698, "loss": 4.3002, "step": 2177 }, { "epoch": 0.9034532821735974, "grad_norm": 1.7734375, "learning_rate": 0.0001565382691345673, "loss": 4.138, "step": 2178 }, { "epoch": 0.9038680908430986, "grad_norm": 1.7265625, "learning_rate": 0.0001565182591295648, "loss": 4.3235, "step": 2179 }, { "epoch": 0.9042828995125998, "grad_norm": 1.9921875, "learning_rate": 0.0001564982491245623, "loss": 4.2286, "step": 2180 }, { "epoch": 0.904697708182101, "grad_norm": 1.75, "learning_rate": 0.00015647823911955978, "loss": 4.1445, "step": 2181 }, { "epoch": 0.9051125168516022, "grad_norm": 2.0, "learning_rate": 0.00015645822911455727, "loss": 4.4364, "step": 2182 }, { "epoch": 0.9055273255211034, "grad_norm": 1.9296875, "learning_rate": 0.00015643821910955478, "loss": 4.3469, "step": 2183 }, { "epoch": 0.9059421341906045, "grad_norm": 1.9765625, "learning_rate": 0.0001564182091045523, "loss": 4.2046, "step": 2184 }, { "epoch": 0.9063569428601058, "grad_norm": 1.8359375, "learning_rate": 0.00015639819909954978, "loss": 4.147, "step": 2185 }, { "epoch": 0.906771751529607, "grad_norm": 1.8984375, "learning_rate": 0.0001563781890945473, "loss": 4.1118, "step": 2186 }, { "epoch": 0.9071865601991081, "grad_norm": 1.875, "learning_rate": 0.00015635817908954479, "loss": 4.2591, "step": 2187 }, { "epoch": 0.9076013688686093, "grad_norm": 1.703125, "learning_rate": 0.00015633816908454227, "loss": 4.0841, "step": 2188 }, { "epoch": 0.9080161775381106, "grad_norm": 1.75, "learning_rate": 0.00015631815907953976, "loss": 4.1434, "step": 2189 }, { "epoch": 0.9084309862076118, "grad_norm": 1.875, "learning_rate": 0.00015629814907453727, "loss": 4.0406, "step": 2190 }, { "epoch": 0.9088457948771129, "grad_norm": 1.78125, "learning_rate": 0.00015627813906953476, "loss": 4.3395, "step": 2191 }, { "epoch": 0.9092606035466141, "grad_norm": 1.7109375, "learning_rate": 0.00015625812906453228, "loss": 4.1554, "step": 2192 }, { "epoch": 0.9096754122161154, "grad_norm": 1.9453125, "learning_rate": 0.00015623811905952976, "loss": 4.0993, "step": 2193 }, { "epoch": 0.9100902208856165, "grad_norm": 1.875, "learning_rate": 0.00015621810905452728, "loss": 4.1924, "step": 2194 }, { "epoch": 0.9105050295551177, "grad_norm": 1.8046875, "learning_rate": 0.00015619809904952476, "loss": 4.1262, "step": 2195 }, { "epoch": 0.9109198382246189, "grad_norm": 1.8515625, "learning_rate": 0.00015617808904452228, "loss": 4.3138, "step": 2196 }, { "epoch": 0.9113346468941201, "grad_norm": 1.9140625, "learning_rate": 0.00015615807903951977, "loss": 4.3179, "step": 2197 }, { "epoch": 0.9117494555636213, "grad_norm": 1.78125, "learning_rate": 0.00015613806903451725, "loss": 4.3464, "step": 2198 }, { "epoch": 0.9121642642331225, "grad_norm": 1.8671875, "learning_rate": 0.00015611805902951477, "loss": 4.3695, "step": 2199 }, { "epoch": 0.9125790729026236, "grad_norm": 1.8828125, "learning_rate": 0.00015609804902451226, "loss": 4.3916, "step": 2200 }, { "epoch": 0.9129938815721249, "grad_norm": 1.78125, "learning_rate": 0.00015607803901950977, "loss": 3.9858, "step": 2201 }, { "epoch": 0.9134086902416261, "grad_norm": 1.8203125, "learning_rate": 0.00015605802901450726, "loss": 4.3782, "step": 2202 }, { "epoch": 0.9138234989111272, "grad_norm": 1.8203125, "learning_rate": 0.00015603801900950477, "loss": 4.2365, "step": 2203 }, { "epoch": 0.9142383075806284, "grad_norm": 1.8125, "learning_rate": 0.00015601800900450226, "loss": 4.1364, "step": 2204 }, { "epoch": 0.9146531162501297, "grad_norm": 1.78125, "learning_rate": 0.00015599799899949977, "loss": 4.2844, "step": 2205 }, { "epoch": 0.9150679249196308, "grad_norm": 1.8359375, "learning_rate": 0.00015597798899449726, "loss": 4.2396, "step": 2206 }, { "epoch": 0.915482733589132, "grad_norm": 1.953125, "learning_rate": 0.00015595797898949475, "loss": 4.4181, "step": 2207 }, { "epoch": 0.9158975422586332, "grad_norm": 1.9453125, "learning_rate": 0.00015593796898449223, "loss": 3.9374, "step": 2208 }, { "epoch": 0.9163123509281343, "grad_norm": 1.9609375, "learning_rate": 0.00015591795897948975, "loss": 4.2968, "step": 2209 }, { "epoch": 0.9167271595976356, "grad_norm": 2.09375, "learning_rate": 0.00015589794897448724, "loss": 4.04, "step": 2210 }, { "epoch": 0.9171419682671368, "grad_norm": 1.7890625, "learning_rate": 0.00015587793896948475, "loss": 4.2759, "step": 2211 }, { "epoch": 0.917556776936638, "grad_norm": 1.9453125, "learning_rate": 0.00015585792896448227, "loss": 4.1548, "step": 2212 }, { "epoch": 0.9179715856061391, "grad_norm": 1.96875, "learning_rate": 0.00015583791895947975, "loss": 4.3143, "step": 2213 }, { "epoch": 0.9183863942756404, "grad_norm": 1.7890625, "learning_rate": 0.00015581790895447727, "loss": 4.1779, "step": 2214 }, { "epoch": 0.9188012029451416, "grad_norm": 1.71875, "learning_rate": 0.00015579789894947475, "loss": 4.2725, "step": 2215 }, { "epoch": 0.9192160116146427, "grad_norm": 1.5703125, "learning_rate": 0.00015577788894447224, "loss": 4.1206, "step": 2216 }, { "epoch": 0.9196308202841439, "grad_norm": 1.84375, "learning_rate": 0.00015575787893946973, "loss": 4.2409, "step": 2217 }, { "epoch": 0.9200456289536452, "grad_norm": 1.75, "learning_rate": 0.00015573786893446724, "loss": 4.2366, "step": 2218 }, { "epoch": 0.9204604376231463, "grad_norm": 1.8046875, "learning_rate": 0.00015571785892946473, "loss": 4.2977, "step": 2219 }, { "epoch": 0.9208752462926475, "grad_norm": 1.921875, "learning_rate": 0.00015569784892446224, "loss": 4.2987, "step": 2220 }, { "epoch": 0.9212900549621487, "grad_norm": 1.828125, "learning_rate": 0.00015567783891945973, "loss": 4.036, "step": 2221 }, { "epoch": 0.9217048636316499, "grad_norm": 1.8828125, "learning_rate": 0.00015565782891445725, "loss": 4.118, "step": 2222 }, { "epoch": 0.9221196723011511, "grad_norm": 1.9765625, "learning_rate": 0.00015563781890945473, "loss": 4.3668, "step": 2223 }, { "epoch": 0.9225344809706523, "grad_norm": 1.828125, "learning_rate": 0.00015561780890445222, "loss": 4.1182, "step": 2224 }, { "epoch": 0.9229492896401534, "grad_norm": 1.7578125, "learning_rate": 0.00015559779889944974, "loss": 4.0947, "step": 2225 }, { "epoch": 0.9233640983096547, "grad_norm": 1.6875, "learning_rate": 0.00015557778889444722, "loss": 4.1627, "step": 2226 }, { "epoch": 0.9237789069791559, "grad_norm": 1.796875, "learning_rate": 0.00015555777888944474, "loss": 3.9763, "step": 2227 }, { "epoch": 0.924193715648657, "grad_norm": 1.828125, "learning_rate": 0.00015553776888444222, "loss": 4.2452, "step": 2228 }, { "epoch": 0.9246085243181582, "grad_norm": 1.828125, "learning_rate": 0.00015551775887943974, "loss": 4.3029, "step": 2229 }, { "epoch": 0.9250233329876595, "grad_norm": 1.84375, "learning_rate": 0.00015549774887443723, "loss": 4.1845, "step": 2230 }, { "epoch": 0.9254381416571607, "grad_norm": 1.8359375, "learning_rate": 0.00015547773886943474, "loss": 4.3291, "step": 2231 }, { "epoch": 0.9258529503266618, "grad_norm": 1.796875, "learning_rate": 0.00015545772886443223, "loss": 4.3184, "step": 2232 }, { "epoch": 0.926267758996163, "grad_norm": 1.9140625, "learning_rate": 0.00015543771885942972, "loss": 4.3399, "step": 2233 }, { "epoch": 0.9266825676656643, "grad_norm": 2.03125, "learning_rate": 0.0001554177088544272, "loss": 4.0807, "step": 2234 }, { "epoch": 0.9270973763351654, "grad_norm": 1.8515625, "learning_rate": 0.00015539769884942472, "loss": 4.1234, "step": 2235 }, { "epoch": 0.9275121850046666, "grad_norm": 1.8671875, "learning_rate": 0.0001553776888444222, "loss": 4.0685, "step": 2236 }, { "epoch": 0.9279269936741678, "grad_norm": 1.828125, "learning_rate": 0.00015535767883941972, "loss": 4.2384, "step": 2237 }, { "epoch": 0.928341802343669, "grad_norm": 1.9453125, "learning_rate": 0.00015533766883441723, "loss": 4.3294, "step": 2238 }, { "epoch": 0.9287566110131702, "grad_norm": 1.8515625, "learning_rate": 0.00015531765882941472, "loss": 4.3791, "step": 2239 }, { "epoch": 0.9291714196826714, "grad_norm": 1.8203125, "learning_rate": 0.00015529764882441223, "loss": 4.1268, "step": 2240 }, { "epoch": 0.9295862283521725, "grad_norm": 1.90625, "learning_rate": 0.00015527763881940972, "loss": 3.8973, "step": 2241 }, { "epoch": 0.9300010370216738, "grad_norm": 1.9453125, "learning_rate": 0.0001552576288144072, "loss": 4.3121, "step": 2242 }, { "epoch": 0.930415845691175, "grad_norm": 2.265625, "learning_rate": 0.0001552376188094047, "loss": 4.2304, "step": 2243 }, { "epoch": 0.9308306543606761, "grad_norm": 1.8046875, "learning_rate": 0.0001552176088044022, "loss": 4.0923, "step": 2244 }, { "epoch": 0.9312454630301773, "grad_norm": 1.8203125, "learning_rate": 0.0001551975987993997, "loss": 4.1484, "step": 2245 }, { "epoch": 0.9316602716996786, "grad_norm": 1.890625, "learning_rate": 0.0001551775887943972, "loss": 4.2686, "step": 2246 }, { "epoch": 0.9320750803691797, "grad_norm": 2.046875, "learning_rate": 0.0001551575787893947, "loss": 3.8836, "step": 2247 }, { "epoch": 0.9324898890386809, "grad_norm": 1.5625, "learning_rate": 0.00015513756878439221, "loss": 4.2454, "step": 2248 }, { "epoch": 0.9329046977081821, "grad_norm": 1.9765625, "learning_rate": 0.0001551175587793897, "loss": 4.4171, "step": 2249 }, { "epoch": 0.9333195063776832, "grad_norm": 1.65625, "learning_rate": 0.00015509754877438722, "loss": 4.119, "step": 2250 }, { "epoch": 0.9337343150471845, "grad_norm": 1.6875, "learning_rate": 0.00015507753876938468, "loss": 4.1932, "step": 2251 }, { "epoch": 0.9341491237166857, "grad_norm": 1.921875, "learning_rate": 0.0001550575287643822, "loss": 4.1594, "step": 2252 }, { "epoch": 0.9345639323861868, "grad_norm": 1.6640625, "learning_rate": 0.0001550375187593797, "loss": 4.173, "step": 2253 }, { "epoch": 0.934978741055688, "grad_norm": 1.7734375, "learning_rate": 0.0001550175087543772, "loss": 4.092, "step": 2254 }, { "epoch": 0.9353935497251893, "grad_norm": 1.8203125, "learning_rate": 0.0001549974987493747, "loss": 4.3293, "step": 2255 }, { "epoch": 0.9358083583946905, "grad_norm": 2.03125, "learning_rate": 0.0001549774887443722, "loss": 4.0989, "step": 2256 }, { "epoch": 0.9362231670641916, "grad_norm": 1.8359375, "learning_rate": 0.0001549574787393697, "loss": 4.1777, "step": 2257 }, { "epoch": 0.9366379757336928, "grad_norm": 1.6640625, "learning_rate": 0.0001549374687343672, "loss": 4.3114, "step": 2258 }, { "epoch": 0.9370527844031941, "grad_norm": 1.78125, "learning_rate": 0.00015491745872936468, "loss": 4.0765, "step": 2259 }, { "epoch": 0.9374675930726952, "grad_norm": 1.84375, "learning_rate": 0.00015489744872436217, "loss": 4.0534, "step": 2260 }, { "epoch": 0.9378824017421964, "grad_norm": 1.828125, "learning_rate": 0.00015487743871935968, "loss": 4.0581, "step": 2261 }, { "epoch": 0.9382972104116976, "grad_norm": 1.9296875, "learning_rate": 0.00015485742871435717, "loss": 4.1622, "step": 2262 }, { "epoch": 0.9387120190811988, "grad_norm": 1.75, "learning_rate": 0.00015483741870935469, "loss": 3.9718, "step": 2263 }, { "epoch": 0.9391268277507, "grad_norm": 1.78125, "learning_rate": 0.00015481740870435217, "loss": 4.1838, "step": 2264 }, { "epoch": 0.9395416364202012, "grad_norm": 2.015625, "learning_rate": 0.0001547973986993497, "loss": 4.2526, "step": 2265 }, { "epoch": 0.9399564450897023, "grad_norm": 2.21875, "learning_rate": 0.0001547773886943472, "loss": 4.1919, "step": 2266 }, { "epoch": 0.9403712537592036, "grad_norm": 1.890625, "learning_rate": 0.0001547573786893447, "loss": 4.2269, "step": 2267 }, { "epoch": 0.9407860624287048, "grad_norm": 1.9296875, "learning_rate": 0.00015473736868434218, "loss": 4.4348, "step": 2268 }, { "epoch": 0.9412008710982059, "grad_norm": 2.03125, "learning_rate": 0.00015471735867933966, "loss": 4.2872, "step": 2269 }, { "epoch": 0.9416156797677071, "grad_norm": 1.8515625, "learning_rate": 0.00015469734867433718, "loss": 4.0856, "step": 2270 }, { "epoch": 0.9420304884372084, "grad_norm": 1.9765625, "learning_rate": 0.00015467733866933467, "loss": 4.2502, "step": 2271 }, { "epoch": 0.9424452971067095, "grad_norm": 1.734375, "learning_rate": 0.00015465732866433218, "loss": 4.1272, "step": 2272 }, { "epoch": 0.9428601057762107, "grad_norm": 1.734375, "learning_rate": 0.00015463731865932967, "loss": 4.0924, "step": 2273 }, { "epoch": 0.9432749144457119, "grad_norm": 1.859375, "learning_rate": 0.00015461730865432718, "loss": 4.2098, "step": 2274 }, { "epoch": 0.9436897231152132, "grad_norm": 1.9453125, "learning_rate": 0.00015459729864932467, "loss": 4.3583, "step": 2275 }, { "epoch": 0.9441045317847143, "grad_norm": 1.9765625, "learning_rate": 0.00015457728864432218, "loss": 4.2959, "step": 2276 }, { "epoch": 0.9445193404542155, "grad_norm": 2.078125, "learning_rate": 0.00015455727863931967, "loss": 3.9754, "step": 2277 }, { "epoch": 0.9449341491237166, "grad_norm": 1.9921875, "learning_rate": 0.00015453726863431716, "loss": 4.0492, "step": 2278 }, { "epoch": 0.9453489577932179, "grad_norm": 1.7734375, "learning_rate": 0.00015451725862931464, "loss": 4.2916, "step": 2279 }, { "epoch": 0.9457637664627191, "grad_norm": 1.78125, "learning_rate": 0.00015449724862431216, "loss": 4.2627, "step": 2280 }, { "epoch": 0.9461785751322203, "grad_norm": 1.859375, "learning_rate": 0.00015447723861930967, "loss": 4.201, "step": 2281 }, { "epoch": 0.9465933838017214, "grad_norm": 1.6953125, "learning_rate": 0.00015445722861430716, "loss": 3.9571, "step": 2282 }, { "epoch": 0.9470081924712227, "grad_norm": 1.796875, "learning_rate": 0.00015443721860930468, "loss": 4.2158, "step": 2283 }, { "epoch": 0.9474230011407239, "grad_norm": 1.859375, "learning_rate": 0.00015441720860430216, "loss": 4.2579, "step": 2284 }, { "epoch": 0.947837809810225, "grad_norm": 1.9609375, "learning_rate": 0.00015439719859929968, "loss": 4.0209, "step": 2285 }, { "epoch": 0.9482526184797262, "grad_norm": 1.7734375, "learning_rate": 0.00015437718859429716, "loss": 4.2051, "step": 2286 }, { "epoch": 0.9486674271492275, "grad_norm": 1.703125, "learning_rate": 0.00015435717858929465, "loss": 4.1164, "step": 2287 }, { "epoch": 0.9490822358187286, "grad_norm": 1.890625, "learning_rate": 0.00015433716858429214, "loss": 4.2679, "step": 2288 }, { "epoch": 0.9494970444882298, "grad_norm": 1.71875, "learning_rate": 0.00015431715857928965, "loss": 4.3908, "step": 2289 }, { "epoch": 0.949911853157731, "grad_norm": 1.9453125, "learning_rate": 0.00015429714857428714, "loss": 4.2127, "step": 2290 }, { "epoch": 0.9503266618272322, "grad_norm": 1.8046875, "learning_rate": 0.00015427713856928465, "loss": 4.0933, "step": 2291 }, { "epoch": 0.9507414704967334, "grad_norm": 1.6640625, "learning_rate": 0.00015425712856428214, "loss": 4.2816, "step": 2292 }, { "epoch": 0.9511562791662346, "grad_norm": 1.984375, "learning_rate": 0.00015423711855927966, "loss": 4.1509, "step": 2293 }, { "epoch": 0.9515710878357357, "grad_norm": 1.8828125, "learning_rate": 0.00015421710855427714, "loss": 4.1942, "step": 2294 }, { "epoch": 0.9519858965052369, "grad_norm": 1.8203125, "learning_rate": 0.00015419709854927463, "loss": 4.2518, "step": 2295 }, { "epoch": 0.9524007051747382, "grad_norm": 1.9296875, "learning_rate": 0.00015417708854427215, "loss": 4.1259, "step": 2296 }, { "epoch": 0.9528155138442393, "grad_norm": 2.1875, "learning_rate": 0.00015415707853926963, "loss": 4.0836, "step": 2297 }, { "epoch": 0.9532303225137405, "grad_norm": 1.90625, "learning_rate": 0.00015413706853426715, "loss": 4.1742, "step": 2298 }, { "epoch": 0.9536451311832417, "grad_norm": 1.8828125, "learning_rate": 0.00015411705852926463, "loss": 4.094, "step": 2299 }, { "epoch": 0.954059939852743, "grad_norm": 1.7890625, "learning_rate": 0.00015409704852426215, "loss": 3.98, "step": 2300 }, { "epoch": 0.9544747485222441, "grad_norm": 1.78125, "learning_rate": 0.00015407703851925964, "loss": 4.2881, "step": 2301 }, { "epoch": 0.9548895571917453, "grad_norm": 1.703125, "learning_rate": 0.00015405702851425715, "loss": 4.2774, "step": 2302 }, { "epoch": 0.9553043658612465, "grad_norm": 1.765625, "learning_rate": 0.00015403701850925464, "loss": 4.0926, "step": 2303 }, { "epoch": 0.9557191745307477, "grad_norm": 1.78125, "learning_rate": 0.00015401700850425213, "loss": 4.1852, "step": 2304 }, { "epoch": 0.9561339832002489, "grad_norm": 1.7109375, "learning_rate": 0.0001539969984992496, "loss": 4.1277, "step": 2305 }, { "epoch": 0.9565487918697501, "grad_norm": 2.109375, "learning_rate": 0.00015397698849424713, "loss": 4.3584, "step": 2306 }, { "epoch": 0.9569636005392512, "grad_norm": 1.828125, "learning_rate": 0.00015395697848924461, "loss": 4.4178, "step": 2307 }, { "epoch": 0.9573784092087525, "grad_norm": 1.8046875, "learning_rate": 0.00015393696848424213, "loss": 4.0052, "step": 2308 }, { "epoch": 0.9577932178782537, "grad_norm": 2.015625, "learning_rate": 0.00015391695847923964, "loss": 4.411, "step": 2309 }, { "epoch": 0.9582080265477548, "grad_norm": 1.8359375, "learning_rate": 0.00015389694847423713, "loss": 4.3911, "step": 2310 }, { "epoch": 0.958622835217256, "grad_norm": 1.828125, "learning_rate": 0.00015387693846923464, "loss": 4.2802, "step": 2311 }, { "epoch": 0.9590376438867573, "grad_norm": 1.703125, "learning_rate": 0.00015385692846423213, "loss": 4.3074, "step": 2312 }, { "epoch": 0.9594524525562584, "grad_norm": 1.796875, "learning_rate": 0.00015383691845922962, "loss": 4.3487, "step": 2313 }, { "epoch": 0.9598672612257596, "grad_norm": 1.7890625, "learning_rate": 0.0001538169084542271, "loss": 4.3998, "step": 2314 }, { "epoch": 0.9602820698952608, "grad_norm": 1.8203125, "learning_rate": 0.00015379689844922462, "loss": 4.5377, "step": 2315 }, { "epoch": 0.960696878564762, "grad_norm": 1.75, "learning_rate": 0.0001537768884442221, "loss": 4.2207, "step": 2316 }, { "epoch": 0.9611116872342632, "grad_norm": 1.8203125, "learning_rate": 0.00015375687843921962, "loss": 4.1297, "step": 2317 }, { "epoch": 0.9615264959037644, "grad_norm": 1.828125, "learning_rate": 0.0001537368684342171, "loss": 4.065, "step": 2318 }, { "epoch": 0.9619413045732655, "grad_norm": 1.7265625, "learning_rate": 0.00015371685842921462, "loss": 3.9849, "step": 2319 }, { "epoch": 0.9623561132427668, "grad_norm": 2.03125, "learning_rate": 0.0001536968484242121, "loss": 4.3975, "step": 2320 }, { "epoch": 0.962770921912268, "grad_norm": 1.6953125, "learning_rate": 0.00015367683841920963, "loss": 4.2253, "step": 2321 }, { "epoch": 0.9631857305817692, "grad_norm": 1.9453125, "learning_rate": 0.0001536568284142071, "loss": 4.2342, "step": 2322 }, { "epoch": 0.9636005392512703, "grad_norm": 1.75, "learning_rate": 0.0001536368184092046, "loss": 4.0354, "step": 2323 }, { "epoch": 0.9640153479207716, "grad_norm": 1.8046875, "learning_rate": 0.00015361680840420211, "loss": 4.0316, "step": 2324 }, { "epoch": 0.9644301565902728, "grad_norm": 1.9296875, "learning_rate": 0.0001535967983991996, "loss": 4.399, "step": 2325 }, { "epoch": 0.9648449652597739, "grad_norm": 1.8046875, "learning_rate": 0.00015357678839419712, "loss": 4.1505, "step": 2326 }, { "epoch": 0.9652597739292751, "grad_norm": 1.859375, "learning_rate": 0.0001535567783891946, "loss": 4.1503, "step": 2327 }, { "epoch": 0.9656745825987764, "grad_norm": 1.765625, "learning_rate": 0.00015353676838419212, "loss": 4.267, "step": 2328 }, { "epoch": 0.9660893912682775, "grad_norm": 1.875, "learning_rate": 0.0001535167583791896, "loss": 4.1939, "step": 2329 }, { "epoch": 0.9665041999377787, "grad_norm": 1.78125, "learning_rate": 0.0001534967483741871, "loss": 4.135, "step": 2330 }, { "epoch": 0.9669190086072799, "grad_norm": 1.875, "learning_rate": 0.00015347673836918458, "loss": 4.2507, "step": 2331 }, { "epoch": 0.9673338172767811, "grad_norm": 1.875, "learning_rate": 0.0001534567283641821, "loss": 4.4498, "step": 2332 }, { "epoch": 0.9677486259462823, "grad_norm": 1.828125, "learning_rate": 0.00015343671835917958, "loss": 4.318, "step": 2333 }, { "epoch": 0.9681634346157835, "grad_norm": 2.0, "learning_rate": 0.0001534167083541771, "loss": 4.2349, "step": 2334 }, { "epoch": 0.9685782432852846, "grad_norm": 1.890625, "learning_rate": 0.0001533966983491746, "loss": 4.2108, "step": 2335 }, { "epoch": 0.9689930519547858, "grad_norm": 1.8046875, "learning_rate": 0.0001533766883441721, "loss": 4.0123, "step": 2336 }, { "epoch": 0.9694078606242871, "grad_norm": 1.90625, "learning_rate": 0.0001533566783391696, "loss": 4.2305, "step": 2337 }, { "epoch": 0.9698226692937882, "grad_norm": 1.734375, "learning_rate": 0.0001533366683341671, "loss": 4.3404, "step": 2338 }, { "epoch": 0.9702374779632894, "grad_norm": 1.796875, "learning_rate": 0.00015331665832916459, "loss": 4.412, "step": 2339 }, { "epoch": 0.9706522866327906, "grad_norm": 2.046875, "learning_rate": 0.00015329664832416207, "loss": 4.2519, "step": 2340 }, { "epoch": 0.9710670953022918, "grad_norm": 1.765625, "learning_rate": 0.0001532766383191596, "loss": 3.9364, "step": 2341 }, { "epoch": 0.971481903971793, "grad_norm": 1.890625, "learning_rate": 0.00015325662831415708, "loss": 4.1125, "step": 2342 }, { "epoch": 0.9718967126412942, "grad_norm": 1.859375, "learning_rate": 0.0001532366183091546, "loss": 4.1954, "step": 2343 }, { "epoch": 0.9723115213107953, "grad_norm": 1.8671875, "learning_rate": 0.00015321660830415208, "loss": 4.2875, "step": 2344 }, { "epoch": 0.9727263299802966, "grad_norm": 1.6640625, "learning_rate": 0.0001531965982991496, "loss": 4.016, "step": 2345 }, { "epoch": 0.9731411386497978, "grad_norm": 1.7421875, "learning_rate": 0.00015317658829414708, "loss": 4.1595, "step": 2346 }, { "epoch": 0.973555947319299, "grad_norm": 1.953125, "learning_rate": 0.0001531565782891446, "loss": 4.259, "step": 2347 }, { "epoch": 0.9739707559888001, "grad_norm": 1.7734375, "learning_rate": 0.00015313656828414208, "loss": 4.106, "step": 2348 }, { "epoch": 0.9743855646583014, "grad_norm": 1.7421875, "learning_rate": 0.00015311655827913957, "loss": 4.2606, "step": 2349 }, { "epoch": 0.9748003733278026, "grad_norm": 1.828125, "learning_rate": 0.00015309654827413708, "loss": 4.1439, "step": 2350 }, { "epoch": 0.9752151819973037, "grad_norm": 1.9375, "learning_rate": 0.00015307653826913457, "loss": 4.4001, "step": 2351 }, { "epoch": 0.9756299906668049, "grad_norm": 1.734375, "learning_rate": 0.00015305652826413208, "loss": 4.2097, "step": 2352 }, { "epoch": 0.9760447993363062, "grad_norm": 2.015625, "learning_rate": 0.00015303651825912957, "loss": 4.1315, "step": 2353 }, { "epoch": 0.9764596080058073, "grad_norm": 1.7734375, "learning_rate": 0.00015301650825412709, "loss": 4.1145, "step": 2354 }, { "epoch": 0.9768744166753085, "grad_norm": 1.984375, "learning_rate": 0.00015299649824912457, "loss": 4.2575, "step": 2355 }, { "epoch": 0.9772892253448097, "grad_norm": 1.953125, "learning_rate": 0.0001529764882441221, "loss": 4.0654, "step": 2356 }, { "epoch": 0.9777040340143109, "grad_norm": 1.9140625, "learning_rate": 0.00015295647823911955, "loss": 4.0221, "step": 2357 }, { "epoch": 0.9781188426838121, "grad_norm": 1.9296875, "learning_rate": 0.00015293646823411706, "loss": 4.157, "step": 2358 }, { "epoch": 0.9785336513533133, "grad_norm": 1.6640625, "learning_rate": 0.00015291645822911455, "loss": 4.2064, "step": 2359 }, { "epoch": 0.9789484600228144, "grad_norm": 1.8046875, "learning_rate": 0.00015289644822411206, "loss": 4.0656, "step": 2360 }, { "epoch": 0.9793632686923157, "grad_norm": 1.7421875, "learning_rate": 0.00015287643821910955, "loss": 3.9975, "step": 2361 }, { "epoch": 0.9797780773618169, "grad_norm": 1.546875, "learning_rate": 0.00015285642821410706, "loss": 4.0994, "step": 2362 }, { "epoch": 0.980192886031318, "grad_norm": 1.9296875, "learning_rate": 0.00015283641820910458, "loss": 4.0238, "step": 2363 }, { "epoch": 0.9806076947008192, "grad_norm": 1.7890625, "learning_rate": 0.00015281640820410207, "loss": 3.9868, "step": 2364 }, { "epoch": 0.9810225033703205, "grad_norm": 1.671875, "learning_rate": 0.00015279639819909955, "loss": 4.239, "step": 2365 }, { "epoch": 0.9814373120398217, "grad_norm": 1.9453125, "learning_rate": 0.00015277638819409704, "loss": 4.4004, "step": 2366 }, { "epoch": 0.9818521207093228, "grad_norm": 1.7890625, "learning_rate": 0.00015275637818909456, "loss": 4.4064, "step": 2367 }, { "epoch": 0.982266929378824, "grad_norm": 1.9765625, "learning_rate": 0.00015273636818409204, "loss": 4.2464, "step": 2368 }, { "epoch": 0.9826817380483253, "grad_norm": 1.890625, "learning_rate": 0.00015271635817908956, "loss": 4.0766, "step": 2369 }, { "epoch": 0.9830965467178264, "grad_norm": 1.671875, "learning_rate": 0.00015269634817408704, "loss": 4.2185, "step": 2370 }, { "epoch": 0.9835113553873276, "grad_norm": 1.8125, "learning_rate": 0.00015267633816908456, "loss": 3.8992, "step": 2371 }, { "epoch": 0.9839261640568288, "grad_norm": 1.75, "learning_rate": 0.00015265632816408205, "loss": 4.202, "step": 2372 }, { "epoch": 0.98434097272633, "grad_norm": 1.90625, "learning_rate": 0.00015263631815907956, "loss": 4.0973, "step": 2373 }, { "epoch": 0.9847557813958312, "grad_norm": 2.265625, "learning_rate": 0.00015261630815407705, "loss": 4.4208, "step": 2374 }, { "epoch": 0.9851705900653324, "grad_norm": 1.90625, "learning_rate": 0.00015259629814907454, "loss": 3.9895, "step": 2375 }, { "epoch": 0.9855853987348335, "grad_norm": 1.7578125, "learning_rate": 0.00015257628814407202, "loss": 4.0862, "step": 2376 }, { "epoch": 0.9860002074043348, "grad_norm": 1.71875, "learning_rate": 0.00015255627813906954, "loss": 4.3058, "step": 2377 }, { "epoch": 0.986415016073836, "grad_norm": 1.7578125, "learning_rate": 0.00015253626813406705, "loss": 4.2387, "step": 2378 }, { "epoch": 0.9868298247433371, "grad_norm": 2.40625, "learning_rate": 0.00015251625812906454, "loss": 4.4021, "step": 2379 }, { "epoch": 0.9872446334128383, "grad_norm": 1.765625, "learning_rate": 0.00015249624812406205, "loss": 4.4758, "step": 2380 }, { "epoch": 0.9876594420823395, "grad_norm": 1.765625, "learning_rate": 0.00015247623811905954, "loss": 3.9762, "step": 2381 }, { "epoch": 0.9880742507518407, "grad_norm": 1.78125, "learning_rate": 0.00015245622811405705, "loss": 4.1106, "step": 2382 }, { "epoch": 0.9884890594213419, "grad_norm": 1.875, "learning_rate": 0.00015243621810905454, "loss": 4.2533, "step": 2383 }, { "epoch": 0.9889038680908431, "grad_norm": 1.8359375, "learning_rate": 0.00015241620810405203, "loss": 4.3298, "step": 2384 }, { "epoch": 0.9893186767603442, "grad_norm": 1.9765625, "learning_rate": 0.00015239619809904952, "loss": 4.374, "step": 2385 }, { "epoch": 0.9897334854298455, "grad_norm": 1.859375, "learning_rate": 0.00015237618809404703, "loss": 4.0625, "step": 2386 }, { "epoch": 0.9901482940993467, "grad_norm": 1.8125, "learning_rate": 0.00015235617808904452, "loss": 4.0474, "step": 2387 }, { "epoch": 0.9905631027688478, "grad_norm": 2.109375, "learning_rate": 0.00015233616808404203, "loss": 4.1933, "step": 2388 }, { "epoch": 0.990977911438349, "grad_norm": 1.7109375, "learning_rate": 0.00015231615807903952, "loss": 4.2599, "step": 2389 }, { "epoch": 0.9913927201078503, "grad_norm": 2.046875, "learning_rate": 0.00015229614807403703, "loss": 4.3878, "step": 2390 }, { "epoch": 0.9918075287773515, "grad_norm": 1.875, "learning_rate": 0.00015227613806903455, "loss": 4.0606, "step": 2391 }, { "epoch": 0.9922223374468526, "grad_norm": 1.8203125, "learning_rate": 0.000152256128064032, "loss": 4.2195, "step": 2392 }, { "epoch": 0.9926371461163538, "grad_norm": 1.859375, "learning_rate": 0.00015223611805902952, "loss": 4.1313, "step": 2393 }, { "epoch": 0.9930519547858551, "grad_norm": 1.78125, "learning_rate": 0.000152216108054027, "loss": 4.4139, "step": 2394 }, { "epoch": 0.9934667634553562, "grad_norm": 1.9609375, "learning_rate": 0.00015219609804902452, "loss": 3.9348, "step": 2395 }, { "epoch": 0.9938815721248574, "grad_norm": 1.859375, "learning_rate": 0.000152176088044022, "loss": 4.1242, "step": 2396 }, { "epoch": 0.9942963807943586, "grad_norm": 1.8984375, "learning_rate": 0.00015215607803901953, "loss": 4.3554, "step": 2397 }, { "epoch": 0.9947111894638598, "grad_norm": 1.828125, "learning_rate": 0.000152136068034017, "loss": 4.188, "step": 2398 }, { "epoch": 0.995125998133361, "grad_norm": 1.828125, "learning_rate": 0.00015211605802901453, "loss": 4.1448, "step": 2399 }, { "epoch": 0.9955408068028622, "grad_norm": 2.0625, "learning_rate": 0.00015209604802401202, "loss": 4.0907, "step": 2400 }, { "epoch": 0.9959556154723633, "grad_norm": 1.765625, "learning_rate": 0.0001520760380190095, "loss": 4.1027, "step": 2401 }, { "epoch": 0.9963704241418646, "grad_norm": 1.734375, "learning_rate": 0.000152056028014007, "loss": 4.0463, "step": 2402 }, { "epoch": 0.9967852328113658, "grad_norm": 1.7734375, "learning_rate": 0.0001520360180090045, "loss": 4.2325, "step": 2403 }, { "epoch": 0.9972000414808669, "grad_norm": 1.9453125, "learning_rate": 0.000152016008004002, "loss": 4.2699, "step": 2404 }, { "epoch": 0.9976148501503681, "grad_norm": 1.7890625, "learning_rate": 0.0001519959979989995, "loss": 4.1393, "step": 2405 }, { "epoch": 0.9980296588198694, "grad_norm": 1.828125, "learning_rate": 0.00015197598799399702, "loss": 4.2185, "step": 2406 }, { "epoch": 0.9984444674893705, "grad_norm": 1.8828125, "learning_rate": 0.0001519559779889945, "loss": 4.4126, "step": 2407 }, { "epoch": 0.9988592761588717, "grad_norm": 1.75, "learning_rate": 0.00015193596798399202, "loss": 4.2729, "step": 2408 }, { "epoch": 0.9992740848283729, "grad_norm": 1.7890625, "learning_rate": 0.0001519159579789895, "loss": 4.0373, "step": 2409 }, { "epoch": 0.9996888934978742, "grad_norm": 1.8203125, "learning_rate": 0.000151895947973987, "loss": 4.192, "step": 2410 }, { "epoch": 1.0, "grad_norm": 2.171875, "learning_rate": 0.00015187593796898448, "loss": 4.1616, "step": 2411 }, { "epoch": 1.0004148086695013, "grad_norm": 1.953125, "learning_rate": 0.000151855927963982, "loss": 3.5347, "step": 2412 }, { "epoch": 1.0008296173390023, "grad_norm": 1.828125, "learning_rate": 0.00015183591795897949, "loss": 3.755, "step": 2413 }, { "epoch": 1.0012444260085036, "grad_norm": 1.8359375, "learning_rate": 0.000151815907953977, "loss": 3.7922, "step": 2414 }, { "epoch": 1.0016592346780047, "grad_norm": 1.8828125, "learning_rate": 0.0001517958979489745, "loss": 3.4733, "step": 2415 }, { "epoch": 1.002074043347506, "grad_norm": 1.9140625, "learning_rate": 0.000151775887943972, "loss": 3.7274, "step": 2416 }, { "epoch": 1.0024888520170072, "grad_norm": 1.8125, "learning_rate": 0.0001517558779389695, "loss": 3.3831, "step": 2417 }, { "epoch": 1.0029036606865083, "grad_norm": 1.8125, "learning_rate": 0.000151735867933967, "loss": 3.8185, "step": 2418 }, { "epoch": 1.0033184693560095, "grad_norm": 2.046875, "learning_rate": 0.0001517158579289645, "loss": 3.9623, "step": 2419 }, { "epoch": 1.0037332780255108, "grad_norm": 1.9609375, "learning_rate": 0.00015169584792396198, "loss": 3.8141, "step": 2420 }, { "epoch": 1.0041480866950119, "grad_norm": 1.9296875, "learning_rate": 0.0001516758379189595, "loss": 3.6663, "step": 2421 }, { "epoch": 1.0045628953645132, "grad_norm": 1.953125, "learning_rate": 0.00015165582791395698, "loss": 3.6595, "step": 2422 }, { "epoch": 1.0049777040340142, "grad_norm": 1.9765625, "learning_rate": 0.0001516358179089545, "loss": 3.7838, "step": 2423 }, { "epoch": 1.0053925127035155, "grad_norm": 1.9296875, "learning_rate": 0.00015161580790395198, "loss": 3.6254, "step": 2424 }, { "epoch": 1.0058073213730168, "grad_norm": 1.875, "learning_rate": 0.0001515957978989495, "loss": 3.5359, "step": 2425 }, { "epoch": 1.0062221300425178, "grad_norm": 1.8046875, "learning_rate": 0.00015157578789394698, "loss": 3.6811, "step": 2426 }, { "epoch": 1.006636938712019, "grad_norm": 2.09375, "learning_rate": 0.00015155577788894447, "loss": 3.8427, "step": 2427 }, { "epoch": 1.0070517473815204, "grad_norm": 2.140625, "learning_rate": 0.00015153576788394196, "loss": 3.5824, "step": 2428 }, { "epoch": 1.0074665560510214, "grad_norm": 2.078125, "learning_rate": 0.00015151575787893947, "loss": 3.9875, "step": 2429 }, { "epoch": 1.0078813647205227, "grad_norm": 2.046875, "learning_rate": 0.00015149574787393696, "loss": 3.4021, "step": 2430 }, { "epoch": 1.0082961733900238, "grad_norm": 1.9375, "learning_rate": 0.00015147573786893447, "loss": 3.5041, "step": 2431 }, { "epoch": 1.008710982059525, "grad_norm": 2.109375, "learning_rate": 0.000151455727863932, "loss": 3.7729, "step": 2432 }, { "epoch": 1.0091257907290263, "grad_norm": 2.046875, "learning_rate": 0.00015143571785892947, "loss": 3.9331, "step": 2433 }, { "epoch": 1.0095405993985274, "grad_norm": 1.953125, "learning_rate": 0.000151415707853927, "loss": 3.4512, "step": 2434 }, { "epoch": 1.0099554080680286, "grad_norm": 1.7421875, "learning_rate": 0.00015139569784892448, "loss": 3.7014, "step": 2435 }, { "epoch": 1.01037021673753, "grad_norm": 2.078125, "learning_rate": 0.00015137568784392196, "loss": 3.9404, "step": 2436 }, { "epoch": 1.010785025407031, "grad_norm": 2.046875, "learning_rate": 0.00015135567783891945, "loss": 3.4264, "step": 2437 }, { "epoch": 1.0111998340765322, "grad_norm": 1.84375, "learning_rate": 0.00015133566783391697, "loss": 3.5706, "step": 2438 }, { "epoch": 1.0116146427460333, "grad_norm": 1.7734375, "learning_rate": 0.00015131565782891445, "loss": 3.6831, "step": 2439 }, { "epoch": 1.0120294514155346, "grad_norm": 1.9765625, "learning_rate": 0.00015129564782391197, "loss": 3.5959, "step": 2440 }, { "epoch": 1.0124442600850359, "grad_norm": 1.71875, "learning_rate": 0.00015127563781890945, "loss": 3.8055, "step": 2441 }, { "epoch": 1.012859068754537, "grad_norm": 1.9296875, "learning_rate": 0.00015125562781390697, "loss": 3.9808, "step": 2442 }, { "epoch": 1.0132738774240382, "grad_norm": 1.796875, "learning_rate": 0.00015123561780890446, "loss": 3.4725, "step": 2443 }, { "epoch": 1.0136886860935395, "grad_norm": 1.9453125, "learning_rate": 0.00015121560780390197, "loss": 3.681, "step": 2444 }, { "epoch": 1.0141034947630405, "grad_norm": 1.8203125, "learning_rate": 0.00015119559779889946, "loss": 3.4493, "step": 2445 }, { "epoch": 1.0145183034325418, "grad_norm": 2.046875, "learning_rate": 0.00015117558779389695, "loss": 3.4333, "step": 2446 }, { "epoch": 1.0149331121020428, "grad_norm": 1.984375, "learning_rate": 0.00015115557778889446, "loss": 3.4997, "step": 2447 }, { "epoch": 1.0153479207715441, "grad_norm": 1.8359375, "learning_rate": 0.00015113556778389195, "loss": 3.416, "step": 2448 }, { "epoch": 1.0157627294410454, "grad_norm": 1.8984375, "learning_rate": 0.00015111555777888946, "loss": 3.5624, "step": 2449 }, { "epoch": 1.0161775381105465, "grad_norm": 2.046875, "learning_rate": 0.00015109554777388695, "loss": 3.7515, "step": 2450 }, { "epoch": 1.0165923467800477, "grad_norm": 2.046875, "learning_rate": 0.00015107553776888446, "loss": 3.3465, "step": 2451 }, { "epoch": 1.0170071554495488, "grad_norm": 2.046875, "learning_rate": 0.00015105552776388195, "loss": 3.4543, "step": 2452 }, { "epoch": 1.01742196411905, "grad_norm": 2.046875, "learning_rate": 0.00015103551775887946, "loss": 3.7298, "step": 2453 }, { "epoch": 1.0178367727885513, "grad_norm": 2.046875, "learning_rate": 0.00015101550775387695, "loss": 3.5135, "step": 2454 }, { "epoch": 1.0182515814580524, "grad_norm": 1.8515625, "learning_rate": 0.00015099549774887444, "loss": 3.6175, "step": 2455 }, { "epoch": 1.0186663901275537, "grad_norm": 2.03125, "learning_rate": 0.00015097548774387193, "loss": 3.8001, "step": 2456 }, { "epoch": 1.019081198797055, "grad_norm": 1.8671875, "learning_rate": 0.00015095547773886944, "loss": 3.6726, "step": 2457 }, { "epoch": 1.019496007466556, "grad_norm": 1.8828125, "learning_rate": 0.00015093546773386693, "loss": 3.7125, "step": 2458 }, { "epoch": 1.0199108161360573, "grad_norm": 1.765625, "learning_rate": 0.00015091545772886444, "loss": 3.2787, "step": 2459 }, { "epoch": 1.0203256248055583, "grad_norm": 1.8671875, "learning_rate": 0.00015089544772386196, "loss": 3.5944, "step": 2460 }, { "epoch": 1.0207404334750596, "grad_norm": 1.9453125, "learning_rate": 0.00015087543771885944, "loss": 3.6335, "step": 2461 }, { "epoch": 1.0211552421445609, "grad_norm": 1.9140625, "learning_rate": 0.00015085542771385696, "loss": 3.6647, "step": 2462 }, { "epoch": 1.021570050814062, "grad_norm": 1.9296875, "learning_rate": 0.00015083541770885442, "loss": 3.7441, "step": 2463 }, { "epoch": 1.0219848594835632, "grad_norm": 1.8359375, "learning_rate": 0.00015081540770385193, "loss": 3.7554, "step": 2464 }, { "epoch": 1.0223996681530645, "grad_norm": 1.7890625, "learning_rate": 0.00015079539769884942, "loss": 3.6766, "step": 2465 }, { "epoch": 1.0228144768225655, "grad_norm": 1.9609375, "learning_rate": 0.00015077538769384693, "loss": 3.7817, "step": 2466 }, { "epoch": 1.0232292854920668, "grad_norm": 1.640625, "learning_rate": 0.00015075537768884442, "loss": 3.6867, "step": 2467 }, { "epoch": 1.0236440941615679, "grad_norm": 2.046875, "learning_rate": 0.00015073536768384194, "loss": 3.6887, "step": 2468 }, { "epoch": 1.0240589028310692, "grad_norm": 1.9140625, "learning_rate": 0.00015071535767883942, "loss": 3.6777, "step": 2469 }, { "epoch": 1.0244737115005704, "grad_norm": 2.0625, "learning_rate": 0.00015069534767383694, "loss": 4.0965, "step": 2470 }, { "epoch": 1.0248885201700715, "grad_norm": 1.796875, "learning_rate": 0.00015067533766883443, "loss": 3.6502, "step": 2471 }, { "epoch": 1.0253033288395728, "grad_norm": 1.9453125, "learning_rate": 0.0001506553276638319, "loss": 3.7523, "step": 2472 }, { "epoch": 1.025718137509074, "grad_norm": 1.8671875, "learning_rate": 0.0001506353176588294, "loss": 3.4445, "step": 2473 }, { "epoch": 1.026132946178575, "grad_norm": 1.8515625, "learning_rate": 0.00015061530765382691, "loss": 3.548, "step": 2474 }, { "epoch": 1.0265477548480764, "grad_norm": 2.046875, "learning_rate": 0.00015059529764882443, "loss": 3.5629, "step": 2475 }, { "epoch": 1.0269625635175774, "grad_norm": 1.953125, "learning_rate": 0.00015057528764382192, "loss": 3.5515, "step": 2476 }, { "epoch": 1.0273773721870787, "grad_norm": 2.015625, "learning_rate": 0.00015055527763881943, "loss": 3.7584, "step": 2477 }, { "epoch": 1.02779218085658, "grad_norm": 2.109375, "learning_rate": 0.00015053526763381692, "loss": 3.6684, "step": 2478 }, { "epoch": 1.028206989526081, "grad_norm": 1.953125, "learning_rate": 0.00015051525762881443, "loss": 3.6705, "step": 2479 }, { "epoch": 1.0286217981955823, "grad_norm": 1.890625, "learning_rate": 0.00015049524762381192, "loss": 3.9502, "step": 2480 }, { "epoch": 1.0290366068650836, "grad_norm": 1.7109375, "learning_rate": 0.0001504752376188094, "loss": 3.5354, "step": 2481 }, { "epoch": 1.0294514155345846, "grad_norm": 2.046875, "learning_rate": 0.0001504552276138069, "loss": 3.6834, "step": 2482 }, { "epoch": 1.029866224204086, "grad_norm": 1.8125, "learning_rate": 0.0001504352176088044, "loss": 3.3595, "step": 2483 }, { "epoch": 1.030281032873587, "grad_norm": 2.0625, "learning_rate": 0.0001504152076038019, "loss": 3.7037, "step": 2484 }, { "epoch": 1.0306958415430882, "grad_norm": 1.7734375, "learning_rate": 0.0001503951975987994, "loss": 3.6547, "step": 2485 }, { "epoch": 1.0311106502125895, "grad_norm": 1.953125, "learning_rate": 0.0001503751875937969, "loss": 3.5773, "step": 2486 }, { "epoch": 1.0315254588820906, "grad_norm": 1.8984375, "learning_rate": 0.0001503551775887944, "loss": 3.5001, "step": 2487 }, { "epoch": 1.0319402675515918, "grad_norm": 1.875, "learning_rate": 0.00015033516758379193, "loss": 3.7381, "step": 2488 }, { "epoch": 1.0323550762210931, "grad_norm": 2.015625, "learning_rate": 0.0001503151575787894, "loss": 3.3491, "step": 2489 }, { "epoch": 1.0327698848905942, "grad_norm": 2.03125, "learning_rate": 0.0001502951475737869, "loss": 3.5707, "step": 2490 }, { "epoch": 1.0331846935600955, "grad_norm": 1.8203125, "learning_rate": 0.0001502751375687844, "loss": 3.824, "step": 2491 }, { "epoch": 1.0335995022295965, "grad_norm": 1.921875, "learning_rate": 0.0001502551275637819, "loss": 3.6082, "step": 2492 }, { "epoch": 1.0340143108990978, "grad_norm": 1.828125, "learning_rate": 0.0001502351175587794, "loss": 3.4404, "step": 2493 }, { "epoch": 1.034429119568599, "grad_norm": 1.984375, "learning_rate": 0.0001502151075537769, "loss": 3.4452, "step": 2494 }, { "epoch": 1.0348439282381001, "grad_norm": 1.8828125, "learning_rate": 0.0001501950975487744, "loss": 3.6358, "step": 2495 }, { "epoch": 1.0352587369076014, "grad_norm": 1.953125, "learning_rate": 0.0001501750875437719, "loss": 3.6787, "step": 2496 }, { "epoch": 1.0356735455771027, "grad_norm": 2.015625, "learning_rate": 0.0001501550775387694, "loss": 3.4537, "step": 2497 }, { "epoch": 1.0360883542466037, "grad_norm": 2.0625, "learning_rate": 0.00015013506753376688, "loss": 3.4812, "step": 2498 }, { "epoch": 1.036503162916105, "grad_norm": 1.71875, "learning_rate": 0.00015011505752876437, "loss": 3.7389, "step": 2499 }, { "epoch": 1.036917971585606, "grad_norm": 2.015625, "learning_rate": 0.00015009504752376188, "loss": 3.5596, "step": 2500 }, { "epoch": 1.0373327802551073, "grad_norm": 1.9765625, "learning_rate": 0.00015007503751875937, "loss": 3.3716, "step": 2501 }, { "epoch": 1.0377475889246086, "grad_norm": 2.046875, "learning_rate": 0.00015005502751375688, "loss": 3.4179, "step": 2502 }, { "epoch": 1.0381623975941097, "grad_norm": 2.09375, "learning_rate": 0.0001500350175087544, "loss": 3.8151, "step": 2503 }, { "epoch": 1.038577206263611, "grad_norm": 2.109375, "learning_rate": 0.00015001500750375188, "loss": 3.7528, "step": 2504 }, { "epoch": 1.038992014933112, "grad_norm": 1.9140625, "learning_rate": 0.0001499949974987494, "loss": 3.6275, "step": 2505 }, { "epoch": 1.0394068236026133, "grad_norm": 1.8984375, "learning_rate": 0.0001499749874937469, "loss": 3.3978, "step": 2506 }, { "epoch": 1.0398216322721145, "grad_norm": 1.9609375, "learning_rate": 0.00014995497748874437, "loss": 3.5794, "step": 2507 }, { "epoch": 1.0402364409416156, "grad_norm": 1.84375, "learning_rate": 0.00014993496748374186, "loss": 3.8088, "step": 2508 }, { "epoch": 1.0406512496111169, "grad_norm": 1.9921875, "learning_rate": 0.00014991495747873938, "loss": 3.5888, "step": 2509 }, { "epoch": 1.0410660582806182, "grad_norm": 1.8984375, "learning_rate": 0.00014989494747373686, "loss": 3.5038, "step": 2510 }, { "epoch": 1.0414808669501192, "grad_norm": 1.8515625, "learning_rate": 0.00014987493746873438, "loss": 3.6577, "step": 2511 }, { "epoch": 1.0418956756196205, "grad_norm": 2.078125, "learning_rate": 0.00014985492746373186, "loss": 3.6019, "step": 2512 }, { "epoch": 1.0423104842891215, "grad_norm": 2.125, "learning_rate": 0.00014983491745872938, "loss": 3.6782, "step": 2513 }, { "epoch": 1.0427252929586228, "grad_norm": 1.8671875, "learning_rate": 0.00014981490745372687, "loss": 3.4082, "step": 2514 }, { "epoch": 1.043140101628124, "grad_norm": 1.9453125, "learning_rate": 0.00014979489744872438, "loss": 3.4208, "step": 2515 }, { "epoch": 1.0435549102976251, "grad_norm": 1.90625, "learning_rate": 0.00014977488744372187, "loss": 3.6975, "step": 2516 }, { "epoch": 1.0439697189671264, "grad_norm": 1.8984375, "learning_rate": 0.00014975487743871936, "loss": 3.6911, "step": 2517 }, { "epoch": 1.0443845276366277, "grad_norm": 2.09375, "learning_rate": 0.00014973486743371687, "loss": 3.7987, "step": 2518 }, { "epoch": 1.0447993363061288, "grad_norm": 1.9609375, "learning_rate": 0.00014971485742871436, "loss": 3.6926, "step": 2519 }, { "epoch": 1.04521414497563, "grad_norm": 2.078125, "learning_rate": 0.00014969484742371187, "loss": 3.6167, "step": 2520 }, { "epoch": 1.045628953645131, "grad_norm": 1.9453125, "learning_rate": 0.00014967483741870936, "loss": 3.6158, "step": 2521 }, { "epoch": 1.0460437623146324, "grad_norm": 1.8359375, "learning_rate": 0.00014965482741370687, "loss": 3.7021, "step": 2522 }, { "epoch": 1.0464585709841336, "grad_norm": 1.8828125, "learning_rate": 0.00014963481740870436, "loss": 3.5105, "step": 2523 }, { "epoch": 1.0468733796536347, "grad_norm": 1.9140625, "learning_rate": 0.00014961480740370187, "loss": 3.6337, "step": 2524 }, { "epoch": 1.047288188323136, "grad_norm": 1.7890625, "learning_rate": 0.00014959479739869936, "loss": 3.7465, "step": 2525 }, { "epoch": 1.0477029969926372, "grad_norm": 1.828125, "learning_rate": 0.00014957478739369685, "loss": 3.6691, "step": 2526 }, { "epoch": 1.0481178056621383, "grad_norm": 2.046875, "learning_rate": 0.00014955477738869434, "loss": 3.7112, "step": 2527 }, { "epoch": 1.0485326143316396, "grad_norm": 1.8828125, "learning_rate": 0.00014953476738369185, "loss": 3.2331, "step": 2528 }, { "epoch": 1.0489474230011406, "grad_norm": 1.9921875, "learning_rate": 0.00014951475737868934, "loss": 3.7657, "step": 2529 }, { "epoch": 1.049362231670642, "grad_norm": 2.03125, "learning_rate": 0.00014949474737368685, "loss": 3.7667, "step": 2530 }, { "epoch": 1.0497770403401432, "grad_norm": 1.90625, "learning_rate": 0.00014947473736868437, "loss": 3.4443, "step": 2531 }, { "epoch": 1.0501918490096442, "grad_norm": 1.8671875, "learning_rate": 0.00014945472736368185, "loss": 3.4903, "step": 2532 }, { "epoch": 1.0506066576791455, "grad_norm": 1.9296875, "learning_rate": 0.00014943471735867934, "loss": 3.6053, "step": 2533 }, { "epoch": 1.0510214663486468, "grad_norm": 1.96875, "learning_rate": 0.00014941470735367683, "loss": 3.6947, "step": 2534 }, { "epoch": 1.0514362750181478, "grad_norm": 1.8046875, "learning_rate": 0.00014939469734867434, "loss": 3.6316, "step": 2535 }, { "epoch": 1.0518510836876491, "grad_norm": 1.875, "learning_rate": 0.00014937468734367183, "loss": 3.5421, "step": 2536 }, { "epoch": 1.0522658923571502, "grad_norm": 2.1875, "learning_rate": 0.00014935467733866934, "loss": 3.5858, "step": 2537 }, { "epoch": 1.0526807010266515, "grad_norm": 2.171875, "learning_rate": 0.00014933466733366683, "loss": 3.7442, "step": 2538 }, { "epoch": 1.0530955096961527, "grad_norm": 1.828125, "learning_rate": 0.00014931465732866435, "loss": 3.5059, "step": 2539 }, { "epoch": 1.0535103183656538, "grad_norm": 1.9296875, "learning_rate": 0.00014929464732366183, "loss": 3.8674, "step": 2540 }, { "epoch": 1.053925127035155, "grad_norm": 1.9453125, "learning_rate": 0.00014927463731865935, "loss": 3.4539, "step": 2541 }, { "epoch": 1.0543399357046561, "grad_norm": 1.8359375, "learning_rate": 0.00014925462731365684, "loss": 3.4168, "step": 2542 }, { "epoch": 1.0547547443741574, "grad_norm": 1.8359375, "learning_rate": 0.00014923461730865432, "loss": 3.7141, "step": 2543 }, { "epoch": 1.0551695530436587, "grad_norm": 1.9453125, "learning_rate": 0.00014921460730365184, "loss": 3.7519, "step": 2544 }, { "epoch": 1.0555843617131597, "grad_norm": 1.9765625, "learning_rate": 0.00014919459729864932, "loss": 3.9559, "step": 2545 }, { "epoch": 1.055999170382661, "grad_norm": 1.859375, "learning_rate": 0.00014917458729364684, "loss": 3.5136, "step": 2546 }, { "epoch": 1.0564139790521623, "grad_norm": 1.8984375, "learning_rate": 0.00014915457728864433, "loss": 3.7584, "step": 2547 }, { "epoch": 1.0568287877216633, "grad_norm": 2.078125, "learning_rate": 0.00014913456728364184, "loss": 3.5063, "step": 2548 }, { "epoch": 1.0572435963911646, "grad_norm": 1.765625, "learning_rate": 0.00014911455727863933, "loss": 3.6587, "step": 2549 }, { "epoch": 1.0576584050606657, "grad_norm": 2.078125, "learning_rate": 0.00014909454727363684, "loss": 3.8086, "step": 2550 }, { "epoch": 1.058073213730167, "grad_norm": 2.1875, "learning_rate": 0.00014907453726863433, "loss": 3.7092, "step": 2551 }, { "epoch": 1.0584880223996682, "grad_norm": 2.09375, "learning_rate": 0.00014905452726363182, "loss": 3.6432, "step": 2552 }, { "epoch": 1.0589028310691693, "grad_norm": 2.015625, "learning_rate": 0.0001490345172586293, "loss": 3.5431, "step": 2553 }, { "epoch": 1.0593176397386705, "grad_norm": 1.9453125, "learning_rate": 0.00014901450725362682, "loss": 3.7514, "step": 2554 }, { "epoch": 1.0597324484081718, "grad_norm": 1.9375, "learning_rate": 0.0001489944972486243, "loss": 3.669, "step": 2555 }, { "epoch": 1.0601472570776729, "grad_norm": 2.015625, "learning_rate": 0.00014897448724362182, "loss": 3.5218, "step": 2556 }, { "epoch": 1.0605620657471742, "grad_norm": 1.75, "learning_rate": 0.00014895447723861933, "loss": 3.4142, "step": 2557 }, { "epoch": 1.0609768744166752, "grad_norm": 1.7734375, "learning_rate": 0.00014893446723361682, "loss": 3.5798, "step": 2558 }, { "epoch": 1.0613916830861765, "grad_norm": 1.9765625, "learning_rate": 0.00014891445722861434, "loss": 3.8852, "step": 2559 }, { "epoch": 1.0618064917556778, "grad_norm": 1.890625, "learning_rate": 0.00014889444722361182, "loss": 3.5586, "step": 2560 }, { "epoch": 1.0622213004251788, "grad_norm": 1.9375, "learning_rate": 0.0001488744372186093, "loss": 3.5303, "step": 2561 }, { "epoch": 1.06263610909468, "grad_norm": 1.8203125, "learning_rate": 0.0001488544272136068, "loss": 3.5196, "step": 2562 }, { "epoch": 1.0630509177641814, "grad_norm": 1.9609375, "learning_rate": 0.0001488344172086043, "loss": 3.4405, "step": 2563 }, { "epoch": 1.0634657264336824, "grad_norm": 2.015625, "learning_rate": 0.0001488144072036018, "loss": 3.6592, "step": 2564 }, { "epoch": 1.0638805351031837, "grad_norm": 1.828125, "learning_rate": 0.00014879439719859931, "loss": 3.4422, "step": 2565 }, { "epoch": 1.0642953437726848, "grad_norm": 2.09375, "learning_rate": 0.0001487743871935968, "loss": 3.6584, "step": 2566 }, { "epoch": 1.064710152442186, "grad_norm": 1.953125, "learning_rate": 0.00014875437718859432, "loss": 3.6786, "step": 2567 }, { "epoch": 1.0651249611116873, "grad_norm": 1.953125, "learning_rate": 0.0001487343671835918, "loss": 3.6906, "step": 2568 }, { "epoch": 1.0655397697811884, "grad_norm": 1.921875, "learning_rate": 0.0001487143571785893, "loss": 3.7105, "step": 2569 }, { "epoch": 1.0659545784506896, "grad_norm": 1.8359375, "learning_rate": 0.00014869434717358678, "loss": 3.3378, "step": 2570 }, { "epoch": 1.066369387120191, "grad_norm": 2.09375, "learning_rate": 0.0001486743371685843, "loss": 3.6973, "step": 2571 }, { "epoch": 1.066784195789692, "grad_norm": 1.734375, "learning_rate": 0.0001486543271635818, "loss": 3.9785, "step": 2572 }, { "epoch": 1.0671990044591932, "grad_norm": 1.7421875, "learning_rate": 0.0001486343171585793, "loss": 3.5402, "step": 2573 }, { "epoch": 1.0676138131286943, "grad_norm": 1.8671875, "learning_rate": 0.0001486143071535768, "loss": 3.3775, "step": 2574 }, { "epoch": 1.0680286217981956, "grad_norm": 1.9765625, "learning_rate": 0.0001485942971485743, "loss": 3.7391, "step": 2575 }, { "epoch": 1.0684434304676969, "grad_norm": 1.8828125, "learning_rate": 0.0001485742871435718, "loss": 3.6208, "step": 2576 }, { "epoch": 1.068858239137198, "grad_norm": 1.84375, "learning_rate": 0.0001485542771385693, "loss": 3.5635, "step": 2577 }, { "epoch": 1.0692730478066992, "grad_norm": 1.859375, "learning_rate": 0.00014853426713356678, "loss": 3.4331, "step": 2578 }, { "epoch": 1.0696878564762002, "grad_norm": 1.8984375, "learning_rate": 0.00014851425712856427, "loss": 3.5328, "step": 2579 }, { "epoch": 1.0701026651457015, "grad_norm": 2.03125, "learning_rate": 0.00014849424712356179, "loss": 3.7104, "step": 2580 }, { "epoch": 1.0705174738152028, "grad_norm": 1.9453125, "learning_rate": 0.00014847423711855927, "loss": 3.6166, "step": 2581 }, { "epoch": 1.0709322824847038, "grad_norm": 2.078125, "learning_rate": 0.0001484542271135568, "loss": 3.4971, "step": 2582 }, { "epoch": 1.0713470911542051, "grad_norm": 1.8828125, "learning_rate": 0.00014843421710855427, "loss": 3.8875, "step": 2583 }, { "epoch": 1.0717618998237064, "grad_norm": 1.8515625, "learning_rate": 0.0001484142071035518, "loss": 3.7037, "step": 2584 }, { "epoch": 1.0721767084932075, "grad_norm": 2.015625, "learning_rate": 0.0001483941970985493, "loss": 3.4992, "step": 2585 }, { "epoch": 1.0725915171627087, "grad_norm": 1.9921875, "learning_rate": 0.0001483741870935468, "loss": 3.6193, "step": 2586 }, { "epoch": 1.07300632583221, "grad_norm": 1.9921875, "learning_rate": 0.00014835417708854428, "loss": 4.1132, "step": 2587 }, { "epoch": 1.073421134501711, "grad_norm": 1.8984375, "learning_rate": 0.00014833416708354177, "loss": 3.606, "step": 2588 }, { "epoch": 1.0738359431712123, "grad_norm": 1.875, "learning_rate": 0.00014831415707853928, "loss": 3.7583, "step": 2589 }, { "epoch": 1.0742507518407134, "grad_norm": 1.890625, "learning_rate": 0.00014829414707353677, "loss": 3.4787, "step": 2590 }, { "epoch": 1.0746655605102147, "grad_norm": 1.71875, "learning_rate": 0.00014827413706853428, "loss": 3.5306, "step": 2591 }, { "epoch": 1.075080369179716, "grad_norm": 1.8359375, "learning_rate": 0.00014825412706353177, "loss": 3.8874, "step": 2592 }, { "epoch": 1.075495177849217, "grad_norm": 2.0625, "learning_rate": 0.00014823411705852928, "loss": 3.5383, "step": 2593 }, { "epoch": 1.0759099865187183, "grad_norm": 2.03125, "learning_rate": 0.00014821410705352677, "loss": 3.7357, "step": 2594 }, { "epoch": 1.0763247951882193, "grad_norm": 1.796875, "learning_rate": 0.00014819409704852428, "loss": 3.8247, "step": 2595 }, { "epoch": 1.0767396038577206, "grad_norm": 1.9609375, "learning_rate": 0.00014817408704352177, "loss": 3.6684, "step": 2596 }, { "epoch": 1.0771544125272219, "grad_norm": 1.984375, "learning_rate": 0.00014815407703851926, "loss": 3.7438, "step": 2597 }, { "epoch": 1.077569221196723, "grad_norm": 1.7890625, "learning_rate": 0.00014813406703351675, "loss": 3.5572, "step": 2598 }, { "epoch": 1.0779840298662242, "grad_norm": 1.921875, "learning_rate": 0.00014811405702851426, "loss": 3.5262, "step": 2599 }, { "epoch": 1.0783988385357255, "grad_norm": 1.9375, "learning_rate": 0.00014809404702351178, "loss": 3.8041, "step": 2600 }, { "epoch": 1.0788136472052265, "grad_norm": 2.046875, "learning_rate": 0.00014807403701850926, "loss": 3.6961, "step": 2601 }, { "epoch": 1.0792284558747278, "grad_norm": 2.125, "learning_rate": 0.00014805402701350678, "loss": 3.6977, "step": 2602 }, { "epoch": 1.0796432645442289, "grad_norm": 2.046875, "learning_rate": 0.00014803401700850426, "loss": 3.6741, "step": 2603 }, { "epoch": 1.0800580732137302, "grad_norm": 2.109375, "learning_rate": 0.00014801400700350175, "loss": 3.5935, "step": 2604 }, { "epoch": 1.0804728818832314, "grad_norm": 1.8984375, "learning_rate": 0.00014799399699849924, "loss": 3.6541, "step": 2605 }, { "epoch": 1.0808876905527325, "grad_norm": 1.9296875, "learning_rate": 0.00014797398699349675, "loss": 3.6662, "step": 2606 }, { "epoch": 1.0813024992222338, "grad_norm": 1.921875, "learning_rate": 0.00014795397698849424, "loss": 3.4872, "step": 2607 }, { "epoch": 1.081717307891735, "grad_norm": 2.078125, "learning_rate": 0.00014793396698349175, "loss": 3.7747, "step": 2608 }, { "epoch": 1.082132116561236, "grad_norm": 1.9140625, "learning_rate": 0.00014791395697848924, "loss": 4.0554, "step": 2609 }, { "epoch": 1.0825469252307374, "grad_norm": 2.0625, "learning_rate": 0.00014789394697348676, "loss": 3.4036, "step": 2610 }, { "epoch": 1.0829617339002384, "grad_norm": 1.796875, "learning_rate": 0.00014787393696848424, "loss": 3.3755, "step": 2611 }, { "epoch": 1.0833765425697397, "grad_norm": 1.953125, "learning_rate": 0.00014785392696348176, "loss": 3.5424, "step": 2612 }, { "epoch": 1.083791351239241, "grad_norm": 1.9140625, "learning_rate": 0.00014783391695847925, "loss": 3.6625, "step": 2613 }, { "epoch": 1.084206159908742, "grad_norm": 1.96875, "learning_rate": 0.00014781390695347673, "loss": 3.3482, "step": 2614 }, { "epoch": 1.0846209685782433, "grad_norm": 1.8984375, "learning_rate": 0.00014779389694847425, "loss": 3.7874, "step": 2615 }, { "epoch": 1.0850357772477446, "grad_norm": 1.8984375, "learning_rate": 0.00014777388694347173, "loss": 3.4656, "step": 2616 }, { "epoch": 1.0854505859172456, "grad_norm": 2.0625, "learning_rate": 0.00014775387693846925, "loss": 3.6858, "step": 2617 }, { "epoch": 1.085865394586747, "grad_norm": 2.09375, "learning_rate": 0.00014773386693346674, "loss": 3.5341, "step": 2618 }, { "epoch": 1.086280203256248, "grad_norm": 2.078125, "learning_rate": 0.00014771385692846425, "loss": 3.4877, "step": 2619 }, { "epoch": 1.0866950119257492, "grad_norm": 1.9140625, "learning_rate": 0.00014769384692346174, "loss": 3.6685, "step": 2620 }, { "epoch": 1.0871098205952505, "grad_norm": 2.046875, "learning_rate": 0.00014767383691845925, "loss": 3.5171, "step": 2621 }, { "epoch": 1.0875246292647516, "grad_norm": 2.078125, "learning_rate": 0.00014765382691345674, "loss": 3.4801, "step": 2622 }, { "epoch": 1.0879394379342529, "grad_norm": 1.9375, "learning_rate": 0.00014763381690845423, "loss": 3.7098, "step": 2623 }, { "epoch": 1.0883542466037541, "grad_norm": 1.9921875, "learning_rate": 0.00014761380690345171, "loss": 3.8358, "step": 2624 }, { "epoch": 1.0887690552732552, "grad_norm": 1.953125, "learning_rate": 0.00014759379689844923, "loss": 3.8276, "step": 2625 }, { "epoch": 1.0891838639427565, "grad_norm": 1.8359375, "learning_rate": 0.00014757378689344672, "loss": 3.6157, "step": 2626 }, { "epoch": 1.0895986726122575, "grad_norm": 1.921875, "learning_rate": 0.00014755377688844423, "loss": 3.4464, "step": 2627 }, { "epoch": 1.0900134812817588, "grad_norm": 1.828125, "learning_rate": 0.00014753376688344174, "loss": 3.6227, "step": 2628 }, { "epoch": 1.09042828995126, "grad_norm": 2.03125, "learning_rate": 0.00014751375687843923, "loss": 3.7227, "step": 2629 }, { "epoch": 1.0908430986207611, "grad_norm": 1.8359375, "learning_rate": 0.00014749374687343675, "loss": 3.7065, "step": 2630 }, { "epoch": 1.0912579072902624, "grad_norm": 1.8671875, "learning_rate": 0.00014747373686843423, "loss": 3.4889, "step": 2631 }, { "epoch": 1.0916727159597635, "grad_norm": 2.125, "learning_rate": 0.00014745372686343172, "loss": 3.7095, "step": 2632 }, { "epoch": 1.0920875246292647, "grad_norm": 2.109375, "learning_rate": 0.0001474337168584292, "loss": 3.5487, "step": 2633 }, { "epoch": 1.092502333298766, "grad_norm": 1.8515625, "learning_rate": 0.00014741370685342672, "loss": 3.5447, "step": 2634 }, { "epoch": 1.092917141968267, "grad_norm": 2.046875, "learning_rate": 0.0001473936968484242, "loss": 3.8792, "step": 2635 }, { "epoch": 1.0933319506377683, "grad_norm": 1.75, "learning_rate": 0.00014737368684342172, "loss": 3.748, "step": 2636 }, { "epoch": 1.0937467593072696, "grad_norm": 1.84375, "learning_rate": 0.0001473536768384192, "loss": 3.657, "step": 2637 }, { "epoch": 1.0941615679767707, "grad_norm": 2.046875, "learning_rate": 0.00014733366683341673, "loss": 3.8029, "step": 2638 }, { "epoch": 1.094576376646272, "grad_norm": 1.8515625, "learning_rate": 0.0001473136568284142, "loss": 3.6401, "step": 2639 }, { "epoch": 1.094991185315773, "grad_norm": 1.8671875, "learning_rate": 0.0001472936468234117, "loss": 3.655, "step": 2640 }, { "epoch": 1.0954059939852743, "grad_norm": 1.9296875, "learning_rate": 0.00014727363681840921, "loss": 3.62, "step": 2641 }, { "epoch": 1.0958208026547755, "grad_norm": 1.9140625, "learning_rate": 0.0001472536268134067, "loss": 3.7285, "step": 2642 }, { "epoch": 1.0962356113242766, "grad_norm": 2.015625, "learning_rate": 0.00014723361680840422, "loss": 3.6061, "step": 2643 }, { "epoch": 1.0966504199937779, "grad_norm": 1.96875, "learning_rate": 0.0001472136068034017, "loss": 3.5672, "step": 2644 }, { "epoch": 1.0970652286632792, "grad_norm": 1.9296875, "learning_rate": 0.00014719359679839922, "loss": 3.5496, "step": 2645 }, { "epoch": 1.0974800373327802, "grad_norm": 1.7578125, "learning_rate": 0.0001471735867933967, "loss": 3.7, "step": 2646 }, { "epoch": 1.0978948460022815, "grad_norm": 1.8046875, "learning_rate": 0.00014715357678839422, "loss": 3.5113, "step": 2647 }, { "epoch": 1.0983096546717825, "grad_norm": 1.8515625, "learning_rate": 0.0001471335667833917, "loss": 3.6778, "step": 2648 }, { "epoch": 1.0987244633412838, "grad_norm": 2.015625, "learning_rate": 0.0001471135567783892, "loss": 3.8986, "step": 2649 }, { "epoch": 1.099139272010785, "grad_norm": 1.90625, "learning_rate": 0.00014709354677338668, "loss": 3.8822, "step": 2650 }, { "epoch": 1.0995540806802861, "grad_norm": 1.75, "learning_rate": 0.0001470735367683842, "loss": 3.6831, "step": 2651 }, { "epoch": 1.0999688893497874, "grad_norm": 2.0, "learning_rate": 0.00014705352676338168, "loss": 3.8942, "step": 2652 }, { "epoch": 1.1003836980192887, "grad_norm": 1.8984375, "learning_rate": 0.0001470335167583792, "loss": 3.3791, "step": 2653 }, { "epoch": 1.1007985066887898, "grad_norm": 1.828125, "learning_rate": 0.0001470135067533767, "loss": 3.6833, "step": 2654 }, { "epoch": 1.101213315358291, "grad_norm": 1.8046875, "learning_rate": 0.0001469934967483742, "loss": 3.7328, "step": 2655 }, { "epoch": 1.101628124027792, "grad_norm": 1.828125, "learning_rate": 0.0001469734867433717, "loss": 3.9325, "step": 2656 }, { "epoch": 1.1020429326972934, "grad_norm": 1.9921875, "learning_rate": 0.0001469534767383692, "loss": 3.8382, "step": 2657 }, { "epoch": 1.1024577413667946, "grad_norm": 1.9609375, "learning_rate": 0.0001469334667333667, "loss": 3.6612, "step": 2658 }, { "epoch": 1.1028725500362957, "grad_norm": 2.46875, "learning_rate": 0.00014691345672836418, "loss": 3.4903, "step": 2659 }, { "epoch": 1.103287358705797, "grad_norm": 1.921875, "learning_rate": 0.0001468934467233617, "loss": 3.4603, "step": 2660 }, { "epoch": 1.1037021673752982, "grad_norm": 1.9296875, "learning_rate": 0.00014687343671835918, "loss": 3.4904, "step": 2661 }, { "epoch": 1.1041169760447993, "grad_norm": 1.7734375, "learning_rate": 0.0001468534267133567, "loss": 3.65, "step": 2662 }, { "epoch": 1.1045317847143006, "grad_norm": 1.9609375, "learning_rate": 0.00014683341670835418, "loss": 3.7065, "step": 2663 }, { "epoch": 1.1049465933838016, "grad_norm": 1.9453125, "learning_rate": 0.0001468134067033517, "loss": 3.7929, "step": 2664 }, { "epoch": 1.105361402053303, "grad_norm": 1.78125, "learning_rate": 0.00014679339669834918, "loss": 3.8289, "step": 2665 }, { "epoch": 1.1057762107228042, "grad_norm": 2.03125, "learning_rate": 0.0001467733866933467, "loss": 3.5784, "step": 2666 }, { "epoch": 1.1061910193923052, "grad_norm": 1.7421875, "learning_rate": 0.00014675337668834418, "loss": 3.4601, "step": 2667 }, { "epoch": 1.1066058280618065, "grad_norm": 2.0625, "learning_rate": 0.00014673336668334167, "loss": 3.7158, "step": 2668 }, { "epoch": 1.1070206367313076, "grad_norm": 1.921875, "learning_rate": 0.00014671335667833918, "loss": 3.5047, "step": 2669 }, { "epoch": 1.1074354454008088, "grad_norm": 1.8828125, "learning_rate": 0.00014669334667333667, "loss": 3.3853, "step": 2670 }, { "epoch": 1.1078502540703101, "grad_norm": 1.890625, "learning_rate": 0.00014667333666833419, "loss": 3.7692, "step": 2671 }, { "epoch": 1.1082650627398112, "grad_norm": 1.9765625, "learning_rate": 0.00014665332666333167, "loss": 3.847, "step": 2672 }, { "epoch": 1.1086798714093125, "grad_norm": 1.9140625, "learning_rate": 0.0001466333166583292, "loss": 3.5746, "step": 2673 }, { "epoch": 1.1090946800788137, "grad_norm": 1.96875, "learning_rate": 0.00014661330665332667, "loss": 3.6637, "step": 2674 }, { "epoch": 1.1095094887483148, "grad_norm": 2.046875, "learning_rate": 0.00014659329664832416, "loss": 3.5462, "step": 2675 }, { "epoch": 1.109924297417816, "grad_norm": 1.875, "learning_rate": 0.00014657328664332165, "loss": 3.5583, "step": 2676 }, { "epoch": 1.1103391060873173, "grad_norm": 1.9921875, "learning_rate": 0.00014655327663831916, "loss": 3.3509, "step": 2677 }, { "epoch": 1.1107539147568184, "grad_norm": 1.78125, "learning_rate": 0.00014653326663331665, "loss": 3.7002, "step": 2678 }, { "epoch": 1.1111687234263197, "grad_norm": 1.890625, "learning_rate": 0.00014651325662831416, "loss": 3.6219, "step": 2679 }, { "epoch": 1.1115835320958207, "grad_norm": 1.953125, "learning_rate": 0.00014649324662331165, "loss": 3.5907, "step": 2680 }, { "epoch": 1.111998340765322, "grad_norm": 1.890625, "learning_rate": 0.00014647323661830917, "loss": 3.3256, "step": 2681 }, { "epoch": 1.1124131494348233, "grad_norm": 1.8125, "learning_rate": 0.00014645322661330668, "loss": 3.5689, "step": 2682 }, { "epoch": 1.1128279581043243, "grad_norm": 2.046875, "learning_rate": 0.00014643321660830417, "loss": 3.6709, "step": 2683 }, { "epoch": 1.1132427667738256, "grad_norm": 2.046875, "learning_rate": 0.00014641320660330166, "loss": 3.7691, "step": 2684 }, { "epoch": 1.1136575754433267, "grad_norm": 2.015625, "learning_rate": 0.00014639319659829914, "loss": 3.7233, "step": 2685 }, { "epoch": 1.114072384112828, "grad_norm": 1.75, "learning_rate": 0.00014637318659329666, "loss": 3.8131, "step": 2686 }, { "epoch": 1.1144871927823292, "grad_norm": 1.9375, "learning_rate": 0.00014635317658829414, "loss": 3.6411, "step": 2687 }, { "epoch": 1.1149020014518303, "grad_norm": 2.015625, "learning_rate": 0.00014633316658329166, "loss": 3.7486, "step": 2688 }, { "epoch": 1.1153168101213315, "grad_norm": 1.8828125, "learning_rate": 0.00014631315657828915, "loss": 3.5778, "step": 2689 }, { "epoch": 1.1157316187908328, "grad_norm": 2.09375, "learning_rate": 0.00014629314657328666, "loss": 3.745, "step": 2690 }, { "epoch": 1.1161464274603339, "grad_norm": 1.953125, "learning_rate": 0.00014627313656828415, "loss": 3.7167, "step": 2691 }, { "epoch": 1.1165612361298352, "grad_norm": 1.984375, "learning_rate": 0.00014625312656328166, "loss": 3.7319, "step": 2692 }, { "epoch": 1.1169760447993362, "grad_norm": 1.96875, "learning_rate": 0.00014623311655827915, "loss": 3.723, "step": 2693 }, { "epoch": 1.1173908534688375, "grad_norm": 2.046875, "learning_rate": 0.00014621310655327664, "loss": 3.4696, "step": 2694 }, { "epoch": 1.1178056621383388, "grad_norm": 2.015625, "learning_rate": 0.00014619309654827412, "loss": 3.9351, "step": 2695 }, { "epoch": 1.1182204708078398, "grad_norm": 1.7890625, "learning_rate": 0.00014617308654327164, "loss": 3.6838, "step": 2696 }, { "epoch": 1.118635279477341, "grad_norm": 1.9921875, "learning_rate": 0.00014615307653826915, "loss": 3.913, "step": 2697 }, { "epoch": 1.1190500881468424, "grad_norm": 1.90625, "learning_rate": 0.00014613306653326664, "loss": 3.3328, "step": 2698 }, { "epoch": 1.1194648968163434, "grad_norm": 1.90625, "learning_rate": 0.00014611305652826415, "loss": 3.5048, "step": 2699 }, { "epoch": 1.1198797054858447, "grad_norm": 2.0, "learning_rate": 0.00014609304652326164, "loss": 3.604, "step": 2700 }, { "epoch": 1.1202945141553458, "grad_norm": 2.09375, "learning_rate": 0.00014607303651825916, "loss": 3.781, "step": 2701 }, { "epoch": 1.120709322824847, "grad_norm": 1.9453125, "learning_rate": 0.00014605302651325664, "loss": 3.7573, "step": 2702 }, { "epoch": 1.1211241314943483, "grad_norm": 1.984375, "learning_rate": 0.00014603301650825413, "loss": 3.7391, "step": 2703 }, { "epoch": 1.1215389401638494, "grad_norm": 1.9921875, "learning_rate": 0.00014601300650325162, "loss": 3.7634, "step": 2704 }, { "epoch": 1.1219537488333506, "grad_norm": 1.9375, "learning_rate": 0.00014599299649824913, "loss": 3.6542, "step": 2705 }, { "epoch": 1.1223685575028517, "grad_norm": 1.8828125, "learning_rate": 0.00014597298649324662, "loss": 3.7114, "step": 2706 }, { "epoch": 1.122783366172353, "grad_norm": 2.03125, "learning_rate": 0.00014595297648824413, "loss": 3.4745, "step": 2707 }, { "epoch": 1.1231981748418542, "grad_norm": 1.796875, "learning_rate": 0.00014593296648324162, "loss": 3.5491, "step": 2708 }, { "epoch": 1.1236129835113553, "grad_norm": 1.9140625, "learning_rate": 0.00014591295647823914, "loss": 3.7213, "step": 2709 }, { "epoch": 1.1240277921808566, "grad_norm": 1.78125, "learning_rate": 0.00014589294647323662, "loss": 3.7674, "step": 2710 }, { "epoch": 1.1244426008503579, "grad_norm": 1.8359375, "learning_rate": 0.0001458729364682341, "loss": 3.7574, "step": 2711 }, { "epoch": 1.124857409519859, "grad_norm": 1.765625, "learning_rate": 0.00014585292646323162, "loss": 3.9064, "step": 2712 }, { "epoch": 1.1252722181893602, "grad_norm": 1.8359375, "learning_rate": 0.0001458329164582291, "loss": 3.8359, "step": 2713 }, { "epoch": 1.1256870268588615, "grad_norm": 1.890625, "learning_rate": 0.00014581290645322663, "loss": 3.3738, "step": 2714 }, { "epoch": 1.1261018355283625, "grad_norm": 1.828125, "learning_rate": 0.0001457928964482241, "loss": 3.9012, "step": 2715 }, { "epoch": 1.1265166441978638, "grad_norm": 2.171875, "learning_rate": 0.00014577288644322163, "loss": 3.6165, "step": 2716 }, { "epoch": 1.1269314528673648, "grad_norm": 1.96875, "learning_rate": 0.00014575287643821912, "loss": 3.5677, "step": 2717 }, { "epoch": 1.1273462615368661, "grad_norm": 1.953125, "learning_rate": 0.00014573286643321663, "loss": 3.5323, "step": 2718 }, { "epoch": 1.1277610702063674, "grad_norm": 2.265625, "learning_rate": 0.00014571285642821412, "loss": 3.5342, "step": 2719 }, { "epoch": 1.1281758788758685, "grad_norm": 1.953125, "learning_rate": 0.0001456928464232116, "loss": 3.8583, "step": 2720 }, { "epoch": 1.1285906875453697, "grad_norm": 2.125, "learning_rate": 0.0001456728364182091, "loss": 3.8198, "step": 2721 }, { "epoch": 1.1290054962148708, "grad_norm": 1.7421875, "learning_rate": 0.0001456528264132066, "loss": 3.9549, "step": 2722 }, { "epoch": 1.129420304884372, "grad_norm": 1.8125, "learning_rate": 0.0001456328164082041, "loss": 3.7338, "step": 2723 }, { "epoch": 1.1298351135538733, "grad_norm": 2.09375, "learning_rate": 0.0001456128064032016, "loss": 3.9352, "step": 2724 }, { "epoch": 1.1302499222233744, "grad_norm": 1.8203125, "learning_rate": 0.00014559279639819912, "loss": 3.5162, "step": 2725 }, { "epoch": 1.1306647308928757, "grad_norm": 1.7421875, "learning_rate": 0.0001455727863931966, "loss": 3.7076, "step": 2726 }, { "epoch": 1.131079539562377, "grad_norm": 1.890625, "learning_rate": 0.00014555277638819412, "loss": 3.5999, "step": 2727 }, { "epoch": 1.131494348231878, "grad_norm": 1.78125, "learning_rate": 0.0001455327663831916, "loss": 3.6754, "step": 2728 }, { "epoch": 1.1319091569013793, "grad_norm": 1.9375, "learning_rate": 0.0001455127563781891, "loss": 3.4896, "step": 2729 }, { "epoch": 1.1323239655708806, "grad_norm": 1.828125, "learning_rate": 0.00014549274637318659, "loss": 3.5255, "step": 2730 }, { "epoch": 1.1327387742403816, "grad_norm": 1.8515625, "learning_rate": 0.0001454727363681841, "loss": 3.728, "step": 2731 }, { "epoch": 1.1331535829098829, "grad_norm": 2.28125, "learning_rate": 0.0001454527263631816, "loss": 3.6991, "step": 2732 }, { "epoch": 1.133568391579384, "grad_norm": 1.9375, "learning_rate": 0.0001454327163581791, "loss": 3.6486, "step": 2733 }, { "epoch": 1.1339832002488852, "grad_norm": 1.8046875, "learning_rate": 0.0001454127063531766, "loss": 3.9643, "step": 2734 }, { "epoch": 1.1343980089183865, "grad_norm": 1.9765625, "learning_rate": 0.0001453926963481741, "loss": 3.6587, "step": 2735 }, { "epoch": 1.1348128175878875, "grad_norm": 1.765625, "learning_rate": 0.0001453726863431716, "loss": 3.5688, "step": 2736 }, { "epoch": 1.1352276262573888, "grad_norm": 1.90625, "learning_rate": 0.0001453526763381691, "loss": 3.779, "step": 2737 }, { "epoch": 1.1356424349268899, "grad_norm": 2.0625, "learning_rate": 0.0001453326663331666, "loss": 3.5628, "step": 2738 }, { "epoch": 1.1360572435963912, "grad_norm": 1.984375, "learning_rate": 0.00014531265632816408, "loss": 3.9479, "step": 2739 }, { "epoch": 1.1364720522658924, "grad_norm": 1.953125, "learning_rate": 0.0001452926463231616, "loss": 3.5297, "step": 2740 }, { "epoch": 1.1368868609353935, "grad_norm": 2.03125, "learning_rate": 0.00014527263631815908, "loss": 3.6331, "step": 2741 }, { "epoch": 1.1373016696048948, "grad_norm": 1.671875, "learning_rate": 0.0001452526263131566, "loss": 3.782, "step": 2742 }, { "epoch": 1.1377164782743958, "grad_norm": 1.90625, "learning_rate": 0.00014523261630815408, "loss": 3.6592, "step": 2743 }, { "epoch": 1.138131286943897, "grad_norm": 2.09375, "learning_rate": 0.0001452126063031516, "loss": 3.7172, "step": 2744 }, { "epoch": 1.1385460956133984, "grad_norm": 1.7734375, "learning_rate": 0.00014519259629814908, "loss": 3.7093, "step": 2745 }, { "epoch": 1.1389609042828994, "grad_norm": 1.8125, "learning_rate": 0.00014517258629314657, "loss": 3.89, "step": 2746 }, { "epoch": 1.1393757129524007, "grad_norm": 1.8828125, "learning_rate": 0.00014515257628814406, "loss": 3.6732, "step": 2747 }, { "epoch": 1.139790521621902, "grad_norm": 1.859375, "learning_rate": 0.00014513256628314157, "loss": 3.7199, "step": 2748 }, { "epoch": 1.140205330291403, "grad_norm": 1.9453125, "learning_rate": 0.00014511255627813906, "loss": 3.7086, "step": 2749 }, { "epoch": 1.1406201389609043, "grad_norm": 1.7421875, "learning_rate": 0.00014509254627313657, "loss": 3.6315, "step": 2750 }, { "epoch": 1.1410349476304056, "grad_norm": 1.6953125, "learning_rate": 0.0001450725362681341, "loss": 3.7849, "step": 2751 }, { "epoch": 1.1414497562999066, "grad_norm": 1.9296875, "learning_rate": 0.00014505252626313158, "loss": 3.6623, "step": 2752 }, { "epoch": 1.141864564969408, "grad_norm": 1.84375, "learning_rate": 0.0001450325162581291, "loss": 3.6009, "step": 2753 }, { "epoch": 1.142279373638909, "grad_norm": 1.8671875, "learning_rate": 0.00014501250625312658, "loss": 3.7548, "step": 2754 }, { "epoch": 1.1426941823084102, "grad_norm": 1.8203125, "learning_rate": 0.00014499249624812407, "loss": 3.5654, "step": 2755 }, { "epoch": 1.1431089909779115, "grad_norm": 1.90625, "learning_rate": 0.00014497248624312155, "loss": 3.4343, "step": 2756 }, { "epoch": 1.1435237996474126, "grad_norm": 1.7265625, "learning_rate": 0.00014495247623811907, "loss": 3.7575, "step": 2757 }, { "epoch": 1.1439386083169139, "grad_norm": 1.6796875, "learning_rate": 0.00014493246623311655, "loss": 3.4625, "step": 2758 }, { "epoch": 1.144353416986415, "grad_norm": 2.03125, "learning_rate": 0.00014491245622811407, "loss": 3.5805, "step": 2759 }, { "epoch": 1.1447682256559162, "grad_norm": 1.96875, "learning_rate": 0.00014489244622311156, "loss": 3.584, "step": 2760 }, { "epoch": 1.1451830343254175, "grad_norm": 2.078125, "learning_rate": 0.00014487243621810907, "loss": 3.7252, "step": 2761 }, { "epoch": 1.1455978429949185, "grad_norm": 1.8359375, "learning_rate": 0.00014485242621310656, "loss": 3.7268, "step": 2762 }, { "epoch": 1.1460126516644198, "grad_norm": 2.109375, "learning_rate": 0.00014483241620810407, "loss": 3.491, "step": 2763 }, { "epoch": 1.146427460333921, "grad_norm": 1.8203125, "learning_rate": 0.00014481240620310156, "loss": 3.5642, "step": 2764 }, { "epoch": 1.1468422690034221, "grad_norm": 1.8671875, "learning_rate": 0.00014479239619809905, "loss": 3.6324, "step": 2765 }, { "epoch": 1.1472570776729234, "grad_norm": 1.828125, "learning_rate": 0.00014477238619309656, "loss": 3.5035, "step": 2766 }, { "epoch": 1.1476718863424247, "grad_norm": 2.0, "learning_rate": 0.00014475237618809405, "loss": 3.5, "step": 2767 }, { "epoch": 1.1480866950119257, "grad_norm": 1.7421875, "learning_rate": 0.00014473236618309156, "loss": 3.5603, "step": 2768 }, { "epoch": 1.148501503681427, "grad_norm": 1.8984375, "learning_rate": 0.00014471235617808905, "loss": 3.9204, "step": 2769 }, { "epoch": 1.148916312350928, "grad_norm": 1.828125, "learning_rate": 0.00014469234617308656, "loss": 3.6896, "step": 2770 }, { "epoch": 1.1493311210204293, "grad_norm": 1.9140625, "learning_rate": 0.00014467233616808405, "loss": 3.5482, "step": 2771 }, { "epoch": 1.1497459296899306, "grad_norm": 1.71875, "learning_rate": 0.00014465232616308157, "loss": 3.6724, "step": 2772 }, { "epoch": 1.1501607383594317, "grad_norm": 1.8203125, "learning_rate": 0.00014463231615807905, "loss": 3.7472, "step": 2773 }, { "epoch": 1.150575547028933, "grad_norm": 1.8984375, "learning_rate": 0.00014461230615307654, "loss": 3.5753, "step": 2774 }, { "epoch": 1.150990355698434, "grad_norm": 2.046875, "learning_rate": 0.00014459229614807403, "loss": 3.5203, "step": 2775 }, { "epoch": 1.1514051643679353, "grad_norm": 1.984375, "learning_rate": 0.00014457228614307154, "loss": 3.6933, "step": 2776 }, { "epoch": 1.1518199730374366, "grad_norm": 1.75, "learning_rate": 0.00014455227613806903, "loss": 3.6061, "step": 2777 }, { "epoch": 1.1522347817069376, "grad_norm": 1.8046875, "learning_rate": 0.00014453226613306654, "loss": 3.6256, "step": 2778 }, { "epoch": 1.1526495903764389, "grad_norm": 1.8046875, "learning_rate": 0.00014451225612806406, "loss": 3.5995, "step": 2779 }, { "epoch": 1.15306439904594, "grad_norm": 2.03125, "learning_rate": 0.00014449224612306155, "loss": 3.6167, "step": 2780 }, { "epoch": 1.1534792077154412, "grad_norm": 1.9296875, "learning_rate": 0.00014447223611805903, "loss": 3.4645, "step": 2781 }, { "epoch": 1.1538940163849425, "grad_norm": 1.953125, "learning_rate": 0.00014445222611305652, "loss": 3.4478, "step": 2782 }, { "epoch": 1.1543088250544435, "grad_norm": 1.8828125, "learning_rate": 0.00014443221610805403, "loss": 3.6721, "step": 2783 }, { "epoch": 1.1547236337239448, "grad_norm": 1.9765625, "learning_rate": 0.00014441220610305152, "loss": 3.5365, "step": 2784 }, { "epoch": 1.155138442393446, "grad_norm": 1.96875, "learning_rate": 0.00014439219609804904, "loss": 3.7075, "step": 2785 }, { "epoch": 1.1555532510629472, "grad_norm": 1.859375, "learning_rate": 0.00014437218609304652, "loss": 3.6159, "step": 2786 }, { "epoch": 1.1559680597324484, "grad_norm": 1.9609375, "learning_rate": 0.00014435217608804404, "loss": 3.4795, "step": 2787 }, { "epoch": 1.1563828684019497, "grad_norm": 1.984375, "learning_rate": 0.00014433216608304153, "loss": 3.5723, "step": 2788 }, { "epoch": 1.1567976770714508, "grad_norm": 1.75, "learning_rate": 0.00014431215607803904, "loss": 3.5442, "step": 2789 }, { "epoch": 1.157212485740952, "grad_norm": 2.03125, "learning_rate": 0.00014429214607303653, "loss": 3.7566, "step": 2790 }, { "epoch": 1.157627294410453, "grad_norm": 1.96875, "learning_rate": 0.00014427213606803401, "loss": 3.4855, "step": 2791 }, { "epoch": 1.1580421030799544, "grad_norm": 1.7890625, "learning_rate": 0.0001442521260630315, "loss": 3.7645, "step": 2792 }, { "epoch": 1.1584569117494556, "grad_norm": 1.90625, "learning_rate": 0.00014423211605802902, "loss": 3.753, "step": 2793 }, { "epoch": 1.1588717204189567, "grad_norm": 1.875, "learning_rate": 0.00014421210605302653, "loss": 3.71, "step": 2794 }, { "epoch": 1.159286529088458, "grad_norm": 1.84375, "learning_rate": 0.00014419209604802402, "loss": 3.8298, "step": 2795 }, { "epoch": 1.159701337757959, "grad_norm": 1.7265625, "learning_rate": 0.00014417208604302153, "loss": 3.7213, "step": 2796 }, { "epoch": 1.1601161464274603, "grad_norm": 1.984375, "learning_rate": 0.00014415207603801902, "loss": 3.8175, "step": 2797 }, { "epoch": 1.1605309550969616, "grad_norm": 1.828125, "learning_rate": 0.00014413206603301653, "loss": 3.6552, "step": 2798 }, { "epoch": 1.1609457637664626, "grad_norm": 1.875, "learning_rate": 0.00014411205602801402, "loss": 3.5121, "step": 2799 }, { "epoch": 1.161360572435964, "grad_norm": 1.8828125, "learning_rate": 0.0001440920460230115, "loss": 3.5569, "step": 2800 }, { "epoch": 1.1617753811054652, "grad_norm": 1.828125, "learning_rate": 0.000144072036018009, "loss": 3.5649, "step": 2801 }, { "epoch": 1.1621901897749662, "grad_norm": 1.921875, "learning_rate": 0.0001440520260130065, "loss": 3.5324, "step": 2802 }, { "epoch": 1.1626049984444675, "grad_norm": 1.921875, "learning_rate": 0.000144032016008004, "loss": 3.702, "step": 2803 }, { "epoch": 1.1630198071139688, "grad_norm": 1.8828125, "learning_rate": 0.0001440120060030015, "loss": 3.7457, "step": 2804 }, { "epoch": 1.1634346157834698, "grad_norm": 1.9375, "learning_rate": 0.000143991995997999, "loss": 3.4864, "step": 2805 }, { "epoch": 1.1638494244529711, "grad_norm": 1.9765625, "learning_rate": 0.0001439719859929965, "loss": 3.7031, "step": 2806 }, { "epoch": 1.1642642331224722, "grad_norm": 1.8125, "learning_rate": 0.00014395197598799403, "loss": 3.7062, "step": 2807 }, { "epoch": 1.1646790417919735, "grad_norm": 1.796875, "learning_rate": 0.00014393196598299151, "loss": 3.8515, "step": 2808 }, { "epoch": 1.1650938504614747, "grad_norm": 1.875, "learning_rate": 0.000143911955977989, "loss": 3.6796, "step": 2809 }, { "epoch": 1.1655086591309758, "grad_norm": 1.8046875, "learning_rate": 0.0001438919459729865, "loss": 3.6652, "step": 2810 }, { "epoch": 1.165923467800477, "grad_norm": 1.9609375, "learning_rate": 0.000143871935967984, "loss": 3.7132, "step": 2811 }, { "epoch": 1.1663382764699781, "grad_norm": 1.875, "learning_rate": 0.0001438519259629815, "loss": 3.6851, "step": 2812 }, { "epoch": 1.1667530851394794, "grad_norm": 1.875, "learning_rate": 0.000143831915957979, "loss": 3.4704, "step": 2813 }, { "epoch": 1.1671678938089807, "grad_norm": 1.828125, "learning_rate": 0.0001438119059529765, "loss": 3.6679, "step": 2814 }, { "epoch": 1.1675827024784817, "grad_norm": 1.84375, "learning_rate": 0.000143791895947974, "loss": 3.7961, "step": 2815 }, { "epoch": 1.167997511147983, "grad_norm": 1.9296875, "learning_rate": 0.0001437718859429715, "loss": 3.585, "step": 2816 }, { "epoch": 1.1684123198174843, "grad_norm": 1.6796875, "learning_rate": 0.00014375187593796898, "loss": 3.5079, "step": 2817 }, { "epoch": 1.1688271284869853, "grad_norm": 1.859375, "learning_rate": 0.00014373186593296647, "loss": 3.4433, "step": 2818 }, { "epoch": 1.1692419371564866, "grad_norm": 1.90625, "learning_rate": 0.00014371185592796398, "loss": 3.7011, "step": 2819 }, { "epoch": 1.1696567458259879, "grad_norm": 2.203125, "learning_rate": 0.00014369184592296147, "loss": 3.6654, "step": 2820 }, { "epoch": 1.170071554495489, "grad_norm": 1.8828125, "learning_rate": 0.00014367183591795898, "loss": 3.5881, "step": 2821 }, { "epoch": 1.1704863631649902, "grad_norm": 1.859375, "learning_rate": 0.0001436518259129565, "loss": 3.6436, "step": 2822 }, { "epoch": 1.1709011718344913, "grad_norm": 1.9453125, "learning_rate": 0.00014363181590795399, "loss": 3.565, "step": 2823 }, { "epoch": 1.1713159805039925, "grad_norm": 1.7734375, "learning_rate": 0.0001436118059029515, "loss": 3.6787, "step": 2824 }, { "epoch": 1.1717307891734938, "grad_norm": 1.703125, "learning_rate": 0.000143591795897949, "loss": 3.6812, "step": 2825 }, { "epoch": 1.1721455978429949, "grad_norm": 1.9609375, "learning_rate": 0.00014357178589294648, "loss": 3.7807, "step": 2826 }, { "epoch": 1.1725604065124962, "grad_norm": 1.8828125, "learning_rate": 0.00014355177588794396, "loss": 3.8135, "step": 2827 }, { "epoch": 1.1729752151819972, "grad_norm": 1.9765625, "learning_rate": 0.00014353176588294148, "loss": 3.6971, "step": 2828 }, { "epoch": 1.1733900238514985, "grad_norm": 1.7890625, "learning_rate": 0.00014351175587793896, "loss": 3.6359, "step": 2829 }, { "epoch": 1.1738048325209998, "grad_norm": 1.8984375, "learning_rate": 0.00014349174587293648, "loss": 3.7153, "step": 2830 }, { "epoch": 1.1742196411905008, "grad_norm": 1.9609375, "learning_rate": 0.00014347173586793397, "loss": 3.7404, "step": 2831 }, { "epoch": 1.174634449860002, "grad_norm": 1.859375, "learning_rate": 0.00014345172586293148, "loss": 4.034, "step": 2832 }, { "epoch": 1.1750492585295031, "grad_norm": 1.7890625, "learning_rate": 0.00014343171585792897, "loss": 3.5267, "step": 2833 }, { "epoch": 1.1754640671990044, "grad_norm": 1.8671875, "learning_rate": 0.00014341170585292648, "loss": 3.5541, "step": 2834 }, { "epoch": 1.1758788758685057, "grad_norm": 1.8671875, "learning_rate": 0.00014339169584792397, "loss": 3.5256, "step": 2835 }, { "epoch": 1.1762936845380068, "grad_norm": 1.9296875, "learning_rate": 0.00014337168584292146, "loss": 3.7221, "step": 2836 }, { "epoch": 1.176708493207508, "grad_norm": 1.765625, "learning_rate": 0.00014335167583791897, "loss": 3.7609, "step": 2837 }, { "epoch": 1.1771233018770093, "grad_norm": 1.9609375, "learning_rate": 0.00014333166583291646, "loss": 3.7796, "step": 2838 }, { "epoch": 1.1775381105465104, "grad_norm": 1.890625, "learning_rate": 0.00014331165582791397, "loss": 3.4764, "step": 2839 }, { "epoch": 1.1779529192160116, "grad_norm": 2.0, "learning_rate": 0.00014329164582291146, "loss": 3.695, "step": 2840 }, { "epoch": 1.178367727885513, "grad_norm": 1.796875, "learning_rate": 0.00014327163581790897, "loss": 3.5043, "step": 2841 }, { "epoch": 1.178782536555014, "grad_norm": 2.015625, "learning_rate": 0.00014325162581290646, "loss": 3.6607, "step": 2842 }, { "epoch": 1.1791973452245152, "grad_norm": 1.875, "learning_rate": 0.00014323161580790398, "loss": 3.7645, "step": 2843 }, { "epoch": 1.1796121538940163, "grad_norm": 2.046875, "learning_rate": 0.00014321160580290144, "loss": 3.5352, "step": 2844 }, { "epoch": 1.1800269625635176, "grad_norm": 1.7734375, "learning_rate": 0.00014319159579789895, "loss": 3.6365, "step": 2845 }, { "epoch": 1.1804417712330189, "grad_norm": 1.8359375, "learning_rate": 0.00014317158579289644, "loss": 3.7357, "step": 2846 }, { "epoch": 1.18085657990252, "grad_norm": 1.9921875, "learning_rate": 0.00014315157578789395, "loss": 3.5689, "step": 2847 }, { "epoch": 1.1812713885720212, "grad_norm": 1.8359375, "learning_rate": 0.00014313156578289147, "loss": 3.5948, "step": 2848 }, { "epoch": 1.1816861972415222, "grad_norm": 1.84375, "learning_rate": 0.00014311155577788895, "loss": 3.5355, "step": 2849 }, { "epoch": 1.1821010059110235, "grad_norm": 2.0625, "learning_rate": 0.00014309154577288647, "loss": 3.7361, "step": 2850 }, { "epoch": 1.1825158145805248, "grad_norm": 1.8828125, "learning_rate": 0.00014307153576788396, "loss": 3.5139, "step": 2851 }, { "epoch": 1.1829306232500258, "grad_norm": 1.9296875, "learning_rate": 0.00014305152576288144, "loss": 3.7594, "step": 2852 }, { "epoch": 1.1833454319195271, "grad_norm": 1.9921875, "learning_rate": 0.00014303151575787893, "loss": 3.4869, "step": 2853 }, { "epoch": 1.1837602405890284, "grad_norm": 1.890625, "learning_rate": 0.00014301150575287644, "loss": 3.4872, "step": 2854 }, { "epoch": 1.1841750492585295, "grad_norm": 1.8671875, "learning_rate": 0.00014299149574787393, "loss": 3.7279, "step": 2855 }, { "epoch": 1.1845898579280307, "grad_norm": 1.9765625, "learning_rate": 0.00014297148574287145, "loss": 3.6318, "step": 2856 }, { "epoch": 1.185004666597532, "grad_norm": 1.8671875, "learning_rate": 0.00014295147573786893, "loss": 3.6339, "step": 2857 }, { "epoch": 1.185419475267033, "grad_norm": 2.046875, "learning_rate": 0.00014293146573286645, "loss": 3.7374, "step": 2858 }, { "epoch": 1.1858342839365343, "grad_norm": 1.9921875, "learning_rate": 0.00014291145572786394, "loss": 3.6323, "step": 2859 }, { "epoch": 1.1862490926060354, "grad_norm": 2.140625, "learning_rate": 0.00014289144572286145, "loss": 3.7485, "step": 2860 }, { "epoch": 1.1866639012755367, "grad_norm": 1.953125, "learning_rate": 0.00014287143571785894, "loss": 3.6543, "step": 2861 }, { "epoch": 1.187078709945038, "grad_norm": 1.8515625, "learning_rate": 0.00014285142571285642, "loss": 3.4706, "step": 2862 }, { "epoch": 1.187493518614539, "grad_norm": 1.765625, "learning_rate": 0.00014283141570785394, "loss": 3.791, "step": 2863 }, { "epoch": 1.1879083272840403, "grad_norm": 1.96875, "learning_rate": 0.00014281140570285143, "loss": 3.546, "step": 2864 }, { "epoch": 1.1883231359535413, "grad_norm": 1.84375, "learning_rate": 0.00014279139569784894, "loss": 3.4693, "step": 2865 }, { "epoch": 1.1887379446230426, "grad_norm": 1.84375, "learning_rate": 0.00014277138569284643, "loss": 3.6377, "step": 2866 }, { "epoch": 1.1891527532925439, "grad_norm": 2.09375, "learning_rate": 0.00014275137568784394, "loss": 3.828, "step": 2867 }, { "epoch": 1.189567561962045, "grad_norm": 1.8828125, "learning_rate": 0.00014273136568284143, "loss": 3.5923, "step": 2868 }, { "epoch": 1.1899823706315462, "grad_norm": 1.765625, "learning_rate": 0.00014271135567783894, "loss": 3.7468, "step": 2869 }, { "epoch": 1.1903971793010473, "grad_norm": 1.796875, "learning_rate": 0.00014269134567283643, "loss": 3.5415, "step": 2870 }, { "epoch": 1.1908119879705485, "grad_norm": 1.75, "learning_rate": 0.00014267133566783392, "loss": 3.4995, "step": 2871 }, { "epoch": 1.1912267966400498, "grad_norm": 1.84375, "learning_rate": 0.0001426513256628314, "loss": 3.5286, "step": 2872 }, { "epoch": 1.1916416053095509, "grad_norm": 1.75, "learning_rate": 0.00014263131565782892, "loss": 3.585, "step": 2873 }, { "epoch": 1.1920564139790522, "grad_norm": 1.9375, "learning_rate": 0.0001426113056528264, "loss": 3.2943, "step": 2874 }, { "epoch": 1.1924712226485534, "grad_norm": 1.7578125, "learning_rate": 0.00014259129564782392, "loss": 3.7593, "step": 2875 }, { "epoch": 1.1928860313180545, "grad_norm": 1.8671875, "learning_rate": 0.00014257128564282144, "loss": 3.7729, "step": 2876 }, { "epoch": 1.1933008399875558, "grad_norm": 1.890625, "learning_rate": 0.00014255127563781892, "loss": 3.5906, "step": 2877 }, { "epoch": 1.193715648657057, "grad_norm": 1.9296875, "learning_rate": 0.00014253126563281644, "loss": 3.616, "step": 2878 }, { "epoch": 1.194130457326558, "grad_norm": 1.828125, "learning_rate": 0.0001425112556278139, "loss": 3.5913, "step": 2879 }, { "epoch": 1.1945452659960594, "grad_norm": 2.03125, "learning_rate": 0.0001424912456228114, "loss": 3.5686, "step": 2880 }, { "epoch": 1.1949600746655604, "grad_norm": 1.9609375, "learning_rate": 0.0001424712356178089, "loss": 3.692, "step": 2881 }, { "epoch": 1.1953748833350617, "grad_norm": 1.921875, "learning_rate": 0.00014245122561280641, "loss": 3.6567, "step": 2882 }, { "epoch": 1.195789692004563, "grad_norm": 1.6640625, "learning_rate": 0.0001424312156078039, "loss": 3.5626, "step": 2883 }, { "epoch": 1.196204500674064, "grad_norm": 1.9453125, "learning_rate": 0.00014241120560280142, "loss": 3.7859, "step": 2884 }, { "epoch": 1.1966193093435653, "grad_norm": 1.9140625, "learning_rate": 0.0001423911955977989, "loss": 3.8272, "step": 2885 }, { "epoch": 1.1970341180130664, "grad_norm": 1.859375, "learning_rate": 0.00014237118559279642, "loss": 3.4493, "step": 2886 }, { "epoch": 1.1974489266825676, "grad_norm": 1.9921875, "learning_rate": 0.0001423511755877939, "loss": 3.817, "step": 2887 }, { "epoch": 1.197863735352069, "grad_norm": 1.8515625, "learning_rate": 0.0001423311655827914, "loss": 3.6408, "step": 2888 }, { "epoch": 1.19827854402157, "grad_norm": 1.8671875, "learning_rate": 0.00014231115557778888, "loss": 3.658, "step": 2889 }, { "epoch": 1.1986933526910712, "grad_norm": 1.84375, "learning_rate": 0.0001422911455727864, "loss": 3.5641, "step": 2890 }, { "epoch": 1.1991081613605725, "grad_norm": 2.15625, "learning_rate": 0.0001422711355677839, "loss": 3.5071, "step": 2891 }, { "epoch": 1.1995229700300736, "grad_norm": 1.9140625, "learning_rate": 0.0001422511255627814, "loss": 3.721, "step": 2892 }, { "epoch": 1.1999377786995749, "grad_norm": 1.8515625, "learning_rate": 0.0001422311155577789, "loss": 3.5271, "step": 2893 }, { "epoch": 1.2003525873690761, "grad_norm": 2.046875, "learning_rate": 0.0001422111055527764, "loss": 3.8319, "step": 2894 }, { "epoch": 1.2007673960385772, "grad_norm": 1.8125, "learning_rate": 0.0001421910955477739, "loss": 3.7065, "step": 2895 }, { "epoch": 1.2011822047080785, "grad_norm": 1.875, "learning_rate": 0.0001421710855427714, "loss": 3.4674, "step": 2896 }, { "epoch": 1.2015970133775795, "grad_norm": 1.9375, "learning_rate": 0.00014215107553776889, "loss": 3.5969, "step": 2897 }, { "epoch": 1.2020118220470808, "grad_norm": 1.859375, "learning_rate": 0.00014213106553276637, "loss": 3.7263, "step": 2898 }, { "epoch": 1.202426630716582, "grad_norm": 1.9921875, "learning_rate": 0.0001421110555277639, "loss": 3.6903, "step": 2899 }, { "epoch": 1.2028414393860831, "grad_norm": 1.828125, "learning_rate": 0.00014209104552276137, "loss": 3.595, "step": 2900 }, { "epoch": 1.2032562480555844, "grad_norm": 1.90625, "learning_rate": 0.0001420710355177589, "loss": 3.6022, "step": 2901 }, { "epoch": 1.2036710567250855, "grad_norm": 1.8671875, "learning_rate": 0.00014205102551275638, "loss": 3.6199, "step": 2902 }, { "epoch": 1.2040858653945867, "grad_norm": 1.96875, "learning_rate": 0.0001420310155077539, "loss": 3.4031, "step": 2903 }, { "epoch": 1.204500674064088, "grad_norm": 1.78125, "learning_rate": 0.0001420110055027514, "loss": 3.8017, "step": 2904 }, { "epoch": 1.204915482733589, "grad_norm": 2.109375, "learning_rate": 0.0001419909954977489, "loss": 3.6897, "step": 2905 }, { "epoch": 1.2053302914030903, "grad_norm": 2.109375, "learning_rate": 0.00014197098549274638, "loss": 3.3942, "step": 2906 }, { "epoch": 1.2057451000725916, "grad_norm": 1.9140625, "learning_rate": 0.00014195097548774387, "loss": 3.6161, "step": 2907 }, { "epoch": 1.2061599087420927, "grad_norm": 1.9375, "learning_rate": 0.00014193096548274138, "loss": 3.5552, "step": 2908 }, { "epoch": 1.206574717411594, "grad_norm": 2.046875, "learning_rate": 0.00014191095547773887, "loss": 3.7692, "step": 2909 }, { "epoch": 1.206989526081095, "grad_norm": 1.8984375, "learning_rate": 0.00014189094547273638, "loss": 3.459, "step": 2910 }, { "epoch": 1.2074043347505963, "grad_norm": 1.921875, "learning_rate": 0.00014187093546773387, "loss": 3.419, "step": 2911 }, { "epoch": 1.2078191434200976, "grad_norm": 1.8984375, "learning_rate": 0.00014185092546273138, "loss": 3.4951, "step": 2912 }, { "epoch": 1.2082339520895986, "grad_norm": 2.046875, "learning_rate": 0.00014183091545772887, "loss": 3.7703, "step": 2913 }, { "epoch": 1.2086487607590999, "grad_norm": 2.140625, "learning_rate": 0.00014181090545272639, "loss": 3.6549, "step": 2914 }, { "epoch": 1.2090635694286012, "grad_norm": 2.015625, "learning_rate": 0.00014179089544772385, "loss": 3.5655, "step": 2915 }, { "epoch": 1.2094783780981022, "grad_norm": 1.953125, "learning_rate": 0.00014177088544272136, "loss": 3.7126, "step": 2916 }, { "epoch": 1.2098931867676035, "grad_norm": 1.9296875, "learning_rate": 0.00014175087543771885, "loss": 3.5264, "step": 2917 }, { "epoch": 1.2103079954371045, "grad_norm": 2.03125, "learning_rate": 0.00014173086543271636, "loss": 3.6943, "step": 2918 }, { "epoch": 1.2107228041066058, "grad_norm": 1.9765625, "learning_rate": 0.00014171085542771388, "loss": 3.7252, "step": 2919 }, { "epoch": 1.211137612776107, "grad_norm": 1.7890625, "learning_rate": 0.00014169084542271136, "loss": 3.74, "step": 2920 }, { "epoch": 1.2115524214456082, "grad_norm": 1.8671875, "learning_rate": 0.00014167083541770888, "loss": 3.5888, "step": 2921 }, { "epoch": 1.2119672301151094, "grad_norm": 1.8671875, "learning_rate": 0.00014165082541270637, "loss": 3.7725, "step": 2922 }, { "epoch": 1.2123820387846105, "grad_norm": 1.9609375, "learning_rate": 0.00014163081540770385, "loss": 3.5492, "step": 2923 }, { "epoch": 1.2127968474541118, "grad_norm": 2.15625, "learning_rate": 0.00014161080540270134, "loss": 3.6502, "step": 2924 }, { "epoch": 1.213211656123613, "grad_norm": 1.9140625, "learning_rate": 0.00014159079539769885, "loss": 3.7871, "step": 2925 }, { "epoch": 1.213626464793114, "grad_norm": 1.7109375, "learning_rate": 0.00014157078539269634, "loss": 3.6848, "step": 2926 }, { "epoch": 1.2140412734626154, "grad_norm": 1.9375, "learning_rate": 0.00014155077538769386, "loss": 4.0056, "step": 2927 }, { "epoch": 1.2144560821321166, "grad_norm": 1.640625, "learning_rate": 0.00014153076538269134, "loss": 3.8078, "step": 2928 }, { "epoch": 1.2148708908016177, "grad_norm": 1.8828125, "learning_rate": 0.00014151075537768886, "loss": 3.8713, "step": 2929 }, { "epoch": 1.215285699471119, "grad_norm": 1.84375, "learning_rate": 0.00014149074537268635, "loss": 3.6605, "step": 2930 }, { "epoch": 1.2157005081406203, "grad_norm": 1.7421875, "learning_rate": 0.00014147073536768386, "loss": 3.781, "step": 2931 }, { "epoch": 1.2161153168101213, "grad_norm": 1.84375, "learning_rate": 0.00014145072536268135, "loss": 3.714, "step": 2932 }, { "epoch": 1.2165301254796226, "grad_norm": 1.8828125, "learning_rate": 0.00014143071535767883, "loss": 3.6549, "step": 2933 }, { "epoch": 1.2169449341491236, "grad_norm": 1.953125, "learning_rate": 0.00014141070535267635, "loss": 3.6088, "step": 2934 }, { "epoch": 1.217359742818625, "grad_norm": 1.8203125, "learning_rate": 0.00014139069534767384, "loss": 3.8472, "step": 2935 }, { "epoch": 1.2177745514881262, "grad_norm": 1.75, "learning_rate": 0.00014137068534267135, "loss": 3.8285, "step": 2936 }, { "epoch": 1.2181893601576272, "grad_norm": 1.9609375, "learning_rate": 0.00014135067533766884, "loss": 3.4787, "step": 2937 }, { "epoch": 1.2186041688271285, "grad_norm": 1.7578125, "learning_rate": 0.00014133066533266635, "loss": 3.783, "step": 2938 }, { "epoch": 1.2190189774966296, "grad_norm": 1.7421875, "learning_rate": 0.00014131065532766384, "loss": 3.6051, "step": 2939 }, { "epoch": 1.2194337861661309, "grad_norm": 1.71875, "learning_rate": 0.00014129064532266135, "loss": 3.6421, "step": 2940 }, { "epoch": 1.2198485948356321, "grad_norm": 1.9375, "learning_rate": 0.00014127063531765884, "loss": 3.4257, "step": 2941 }, { "epoch": 1.2202634035051332, "grad_norm": 2.0, "learning_rate": 0.00014125062531265633, "loss": 3.6652, "step": 2942 }, { "epoch": 1.2206782121746345, "grad_norm": 1.890625, "learning_rate": 0.00014123061530765382, "loss": 3.73, "step": 2943 }, { "epoch": 1.2210930208441357, "grad_norm": 1.8046875, "learning_rate": 0.00014121060530265133, "loss": 3.5192, "step": 2944 }, { "epoch": 1.2215078295136368, "grad_norm": 1.8984375, "learning_rate": 0.00014119059529764884, "loss": 3.4722, "step": 2945 }, { "epoch": 1.221922638183138, "grad_norm": 1.9609375, "learning_rate": 0.00014117058529264633, "loss": 3.7632, "step": 2946 }, { "epoch": 1.2223374468526393, "grad_norm": 1.7421875, "learning_rate": 0.00014115057528764385, "loss": 3.5233, "step": 2947 }, { "epoch": 1.2227522555221404, "grad_norm": 2.046875, "learning_rate": 0.00014113056528264133, "loss": 3.795, "step": 2948 }, { "epoch": 1.2231670641916417, "grad_norm": 1.9765625, "learning_rate": 0.00014111055527763885, "loss": 3.5851, "step": 2949 }, { "epoch": 1.2235818728611427, "grad_norm": 2.046875, "learning_rate": 0.0001410905452726363, "loss": 3.6206, "step": 2950 }, { "epoch": 1.223996681530644, "grad_norm": 1.6875, "learning_rate": 0.00014107053526763382, "loss": 3.6176, "step": 2951 }, { "epoch": 1.2244114902001453, "grad_norm": 1.8359375, "learning_rate": 0.0001410505252626313, "loss": 3.5821, "step": 2952 }, { "epoch": 1.2248262988696463, "grad_norm": 1.984375, "learning_rate": 0.00014103051525762882, "loss": 3.6583, "step": 2953 }, { "epoch": 1.2252411075391476, "grad_norm": 1.8046875, "learning_rate": 0.0001410105052526263, "loss": 3.6434, "step": 2954 }, { "epoch": 1.2256559162086487, "grad_norm": 1.9765625, "learning_rate": 0.00014099049524762383, "loss": 3.7295, "step": 2955 }, { "epoch": 1.22607072487815, "grad_norm": 1.859375, "learning_rate": 0.0001409704852426213, "loss": 3.6551, "step": 2956 }, { "epoch": 1.2264855335476512, "grad_norm": 1.6796875, "learning_rate": 0.00014095047523761883, "loss": 3.6274, "step": 2957 }, { "epoch": 1.2269003422171523, "grad_norm": 1.8515625, "learning_rate": 0.00014093046523261631, "loss": 3.9299, "step": 2958 }, { "epoch": 1.2273151508866535, "grad_norm": 1.90625, "learning_rate": 0.0001409104552276138, "loss": 3.6454, "step": 2959 }, { "epoch": 1.2277299595561546, "grad_norm": 1.875, "learning_rate": 0.00014089044522261132, "loss": 3.8022, "step": 2960 }, { "epoch": 1.2281447682256559, "grad_norm": 2.109375, "learning_rate": 0.0001408704352176088, "loss": 3.6751, "step": 2961 }, { "epoch": 1.2285595768951572, "grad_norm": 1.734375, "learning_rate": 0.00014085042521260632, "loss": 3.8049, "step": 2962 }, { "epoch": 1.2289743855646582, "grad_norm": 2.046875, "learning_rate": 0.0001408304152076038, "loss": 3.6189, "step": 2963 }, { "epoch": 1.2293891942341595, "grad_norm": 1.8125, "learning_rate": 0.00014081040520260132, "loss": 3.7181, "step": 2964 }, { "epoch": 1.2298040029036608, "grad_norm": 1.796875, "learning_rate": 0.0001407903951975988, "loss": 3.55, "step": 2965 }, { "epoch": 1.2302188115731618, "grad_norm": 2.015625, "learning_rate": 0.00014077038519259632, "loss": 3.6247, "step": 2966 }, { "epoch": 1.230633620242663, "grad_norm": 2.015625, "learning_rate": 0.0001407503751875938, "loss": 3.5042, "step": 2967 }, { "epoch": 1.2310484289121644, "grad_norm": 1.8828125, "learning_rate": 0.0001407303651825913, "loss": 3.583, "step": 2968 }, { "epoch": 1.2314632375816654, "grad_norm": 1.828125, "learning_rate": 0.00014071035517758878, "loss": 3.5858, "step": 2969 }, { "epoch": 1.2318780462511667, "grad_norm": 1.7890625, "learning_rate": 0.0001406903451725863, "loss": 3.6822, "step": 2970 }, { "epoch": 1.2322928549206678, "grad_norm": 1.859375, "learning_rate": 0.00014067033516758378, "loss": 3.7763, "step": 2971 }, { "epoch": 1.232707663590169, "grad_norm": 1.8046875, "learning_rate": 0.0001406503251625813, "loss": 3.6066, "step": 2972 }, { "epoch": 1.2331224722596703, "grad_norm": 1.953125, "learning_rate": 0.0001406303151575788, "loss": 3.5753, "step": 2973 }, { "epoch": 1.2335372809291714, "grad_norm": 1.8828125, "learning_rate": 0.0001406103051525763, "loss": 3.7046, "step": 2974 }, { "epoch": 1.2339520895986726, "grad_norm": 2.125, "learning_rate": 0.00014059029514757381, "loss": 3.8291, "step": 2975 }, { "epoch": 1.2343668982681737, "grad_norm": 1.9921875, "learning_rate": 0.0001405702851425713, "loss": 3.7256, "step": 2976 }, { "epoch": 1.234781706937675, "grad_norm": 1.8515625, "learning_rate": 0.0001405502751375688, "loss": 3.5583, "step": 2977 }, { "epoch": 1.2351965156071762, "grad_norm": 2.0, "learning_rate": 0.00014053026513256628, "loss": 3.5365, "step": 2978 }, { "epoch": 1.2356113242766773, "grad_norm": 1.8046875, "learning_rate": 0.0001405102551275638, "loss": 3.8073, "step": 2979 }, { "epoch": 1.2360261329461786, "grad_norm": 1.953125, "learning_rate": 0.00014049024512256128, "loss": 3.7875, "step": 2980 }, { "epoch": 1.2364409416156799, "grad_norm": 1.765625, "learning_rate": 0.0001404702351175588, "loss": 3.6355, "step": 2981 }, { "epoch": 1.236855750285181, "grad_norm": 1.796875, "learning_rate": 0.00014045022511255628, "loss": 3.5576, "step": 2982 }, { "epoch": 1.2372705589546822, "grad_norm": 1.921875, "learning_rate": 0.0001404302151075538, "loss": 3.7807, "step": 2983 }, { "epoch": 1.2376853676241835, "grad_norm": 1.78125, "learning_rate": 0.00014041020510255128, "loss": 3.7402, "step": 2984 }, { "epoch": 1.2381001762936845, "grad_norm": 1.9609375, "learning_rate": 0.00014039019509754877, "loss": 3.6976, "step": 2985 }, { "epoch": 1.2385149849631858, "grad_norm": 1.8359375, "learning_rate": 0.00014037018509254626, "loss": 3.6573, "step": 2986 }, { "epoch": 1.2389297936326868, "grad_norm": 2.015625, "learning_rate": 0.00014035017508754377, "loss": 3.6681, "step": 2987 }, { "epoch": 1.2393446023021881, "grad_norm": 1.90625, "learning_rate": 0.00014033016508254128, "loss": 3.6611, "step": 2988 }, { "epoch": 1.2397594109716894, "grad_norm": 1.8203125, "learning_rate": 0.00014031015507753877, "loss": 3.686, "step": 2989 }, { "epoch": 1.2401742196411905, "grad_norm": 2.078125, "learning_rate": 0.0001402901450725363, "loss": 3.6735, "step": 2990 }, { "epoch": 1.2405890283106917, "grad_norm": 1.890625, "learning_rate": 0.00014027013506753377, "loss": 3.6497, "step": 2991 }, { "epoch": 1.2410038369801928, "grad_norm": 1.9609375, "learning_rate": 0.0001402501250625313, "loss": 3.6557, "step": 2992 }, { "epoch": 1.241418645649694, "grad_norm": 1.96875, "learning_rate": 0.00014023011505752878, "loss": 3.5023, "step": 2993 }, { "epoch": 1.2418334543191953, "grad_norm": 2.0625, "learning_rate": 0.00014021010505252626, "loss": 3.6123, "step": 2994 }, { "epoch": 1.2422482629886964, "grad_norm": 2.046875, "learning_rate": 0.00014019009504752375, "loss": 3.6946, "step": 2995 }, { "epoch": 1.2426630716581977, "grad_norm": 1.8828125, "learning_rate": 0.00014017008504252126, "loss": 3.8907, "step": 2996 }, { "epoch": 1.2430778803276987, "grad_norm": 1.8359375, "learning_rate": 0.00014015007503751875, "loss": 3.8092, "step": 2997 }, { "epoch": 1.2434926889972, "grad_norm": 1.7890625, "learning_rate": 0.00014013006503251627, "loss": 3.7444, "step": 2998 }, { "epoch": 1.2439074976667013, "grad_norm": 1.859375, "learning_rate": 0.00014011005502751375, "loss": 3.5605, "step": 2999 }, { "epoch": 1.2443223063362023, "grad_norm": 1.7890625, "learning_rate": 0.00014009004502251127, "loss": 3.705, "step": 3000 } ], "logging_steps": 1, "max_steps": 10000, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 9.966856822078464e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }