rabiulawal's picture
Add files using upload-large-folder tool
acd74f4 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 6.866113574595909,
"eval_steps": 400,
"global_step": 24000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.014304105278214848,
"grad_norm": 3.5871307485399613,
"learning_rate": 1.6666666666666667e-06,
"loss": 4.2171,
"step": 50
},
{
"epoch": 0.028608210556429696,
"grad_norm": 8.578020193760794,
"learning_rate": 3.3333333333333333e-06,
"loss": 3.5086,
"step": 100
},
{
"epoch": 0.04291231583464454,
"grad_norm": 3.453360153921811,
"learning_rate": 5e-06,
"loss": 2.7411,
"step": 150
},
{
"epoch": 0.05721642111285939,
"grad_norm": 6.637419959801433,
"learning_rate": 6.666666666666667e-06,
"loss": 2.5587,
"step": 200
},
{
"epoch": 0.07152052639107424,
"grad_norm": 5.39103050730964,
"learning_rate": 8.333333333333334e-06,
"loss": 2.4626,
"step": 250
},
{
"epoch": 0.08582463166928908,
"grad_norm": 10.145361416938202,
"learning_rate": 1e-05,
"loss": 2.3934,
"step": 300
},
{
"epoch": 0.10012873694750393,
"grad_norm": 7.783722807777708,
"learning_rate": 9.999953760295448e-06,
"loss": 2.272,
"step": 350
},
{
"epoch": 0.11443284222571878,
"grad_norm": 3.65791345167478,
"learning_rate": 9.999815042132062e-06,
"loss": 2.2033,
"step": 400
},
{
"epoch": 0.11443284222571878,
"eval_loss": 2.1356396675109863,
"eval_runtime": 13.2089,
"eval_samples_per_second": 75.707,
"eval_steps_per_second": 2.423,
"step": 400
},
{
"epoch": 0.12873694750393364,
"grad_norm": 8.621408307551274,
"learning_rate": 9.999583848360633e-06,
"loss": 2.1457,
"step": 450
},
{
"epoch": 0.1430410527821485,
"grad_norm": 6.2605099034671,
"learning_rate": 9.999260183732424e-06,
"loss": 2.1022,
"step": 500
},
{
"epoch": 0.1573451580603633,
"grad_norm": 7.684157630671387,
"learning_rate": 9.998844054899058e-06,
"loss": 2.0793,
"step": 550
},
{
"epoch": 0.17164926333857816,
"grad_norm": 1.7771086513044794,
"learning_rate": 9.998335470412393e-06,
"loss": 2.0783,
"step": 600
},
{
"epoch": 0.185953368616793,
"grad_norm": 4.839674439593447,
"learning_rate": 9.997734440724333e-06,
"loss": 2.0455,
"step": 650
},
{
"epoch": 0.20025747389500786,
"grad_norm": 1.7547598316419108,
"learning_rate": 9.997040978186633e-06,
"loss": 2.0206,
"step": 700
},
{
"epoch": 0.21456157917322272,
"grad_norm": 2.5514452897242035,
"learning_rate": 9.996255097050624e-06,
"loss": 2.003,
"step": 750
},
{
"epoch": 0.22886568445143757,
"grad_norm": 3.9407907366389976,
"learning_rate": 9.995376813466934e-06,
"loss": 1.9948,
"step": 800
},
{
"epoch": 0.22886568445143757,
"eval_loss": 1.9659738540649414,
"eval_runtime": 13.202,
"eval_samples_per_second": 75.746,
"eval_steps_per_second": 2.424,
"step": 800
},
{
"epoch": 0.24316978972965242,
"grad_norm": 1.7301759518540019,
"learning_rate": 9.994406145485151e-06,
"loss": 1.9861,
"step": 850
},
{
"epoch": 0.25747389500786727,
"grad_norm": 4.606477014527872,
"learning_rate": 9.993343113053454e-06,
"loss": 1.9713,
"step": 900
},
{
"epoch": 0.2717780002860821,
"grad_norm": 2.8102943343276183,
"learning_rate": 9.992187738018203e-06,
"loss": 1.9675,
"step": 950
},
{
"epoch": 0.286082105564297,
"grad_norm": 1.6938293305410583,
"learning_rate": 9.99094004412348e-06,
"loss": 1.9556,
"step": 1000
},
{
"epoch": 0.3003862108425118,
"grad_norm": 3.4489466081512044,
"learning_rate": 9.989600057010625e-06,
"loss": 1.9417,
"step": 1050
},
{
"epoch": 0.3146903161207266,
"grad_norm": 1.454314969912362,
"learning_rate": 9.988167804217682e-06,
"loss": 1.9361,
"step": 1100
},
{
"epoch": 0.3289944213989415,
"grad_norm": 1.168052123243911,
"learning_rate": 9.986643315178848e-06,
"loss": 1.9264,
"step": 1150
},
{
"epoch": 0.3432985266771563,
"grad_norm": 1.7334339924315771,
"learning_rate": 9.98502662122387e-06,
"loss": 1.9308,
"step": 1200
},
{
"epoch": 0.3432985266771563,
"eval_loss": 1.8975478410720825,
"eval_runtime": 13.3147,
"eval_samples_per_second": 75.105,
"eval_steps_per_second": 2.403,
"step": 1200
},
{
"epoch": 0.3576026319553712,
"grad_norm": 1.5051784354611382,
"learning_rate": 9.983317755577392e-06,
"loss": 1.9208,
"step": 1250
},
{
"epoch": 0.371906737233586,
"grad_norm": 1.5926471600228091,
"learning_rate": 9.981516753358274e-06,
"loss": 1.9156,
"step": 1300
},
{
"epoch": 0.3862108425118009,
"grad_norm": 1.4680846710500992,
"learning_rate": 9.979623651578881e-06,
"loss": 1.9003,
"step": 1350
},
{
"epoch": 0.40051494779001573,
"grad_norm": 1.3763604778464051,
"learning_rate": 9.977638489144308e-06,
"loss": 1.9069,
"step": 1400
},
{
"epoch": 0.4148190530682306,
"grad_norm": 1.524436140637499,
"learning_rate": 9.975561306851585e-06,
"loss": 1.8973,
"step": 1450
},
{
"epoch": 0.42912315834644543,
"grad_norm": 1.6595586978206922,
"learning_rate": 9.973392147388847e-06,
"loss": 1.8994,
"step": 1500
},
{
"epoch": 0.4434272636246603,
"grad_norm": 1.5658370421245245,
"learning_rate": 9.971131055334445e-06,
"loss": 1.8875,
"step": 1550
},
{
"epoch": 0.45773136890287514,
"grad_norm": 1.9874689334717497,
"learning_rate": 9.968778077156035e-06,
"loss": 1.8885,
"step": 1600
},
{
"epoch": 0.45773136890287514,
"eval_loss": 1.8600050210952759,
"eval_runtime": 13.4365,
"eval_samples_per_second": 74.424,
"eval_steps_per_second": 2.382,
"step": 1600
},
{
"epoch": 0.47203547418109,
"grad_norm": 1.391686515937597,
"learning_rate": 9.966333261209625e-06,
"loss": 1.8825,
"step": 1650
},
{
"epoch": 0.48633957945930484,
"grad_norm": 1.250031324432951,
"learning_rate": 9.96379665773858e-06,
"loss": 1.8812,
"step": 1700
},
{
"epoch": 0.5006436847375196,
"grad_norm": 0.937500185481468,
"learning_rate": 9.961168318872583e-06,
"loss": 1.873,
"step": 1750
},
{
"epoch": 0.5149477900157345,
"grad_norm": 1.3831225145690145,
"learning_rate": 9.958448298626576e-06,
"loss": 1.8733,
"step": 1800
},
{
"epoch": 0.5292518952939493,
"grad_norm": 2.8853368812546845,
"learning_rate": 9.95563665289964e-06,
"loss": 1.8607,
"step": 1850
},
{
"epoch": 0.5435560005721642,
"grad_norm": 1.8333114897013831,
"learning_rate": 9.952733439473847e-06,
"loss": 1.867,
"step": 1900
},
{
"epoch": 0.557860105850379,
"grad_norm": 1.4955344425097794,
"learning_rate": 9.94973871801308e-06,
"loss": 1.8668,
"step": 1950
},
{
"epoch": 0.572164211128594,
"grad_norm": 1.058766324989062,
"learning_rate": 9.946652550061798e-06,
"loss": 1.8468,
"step": 2000
},
{
"epoch": 0.572164211128594,
"eval_loss": 1.8319525718688965,
"eval_runtime": 13.2128,
"eval_samples_per_second": 75.684,
"eval_steps_per_second": 2.422,
"step": 2000
},
{
"epoch": 0.5864683164068087,
"grad_norm": 1.061225891590699,
"learning_rate": 9.943474999043775e-06,
"loss": 1.8513,
"step": 2050
},
{
"epoch": 0.6007724216850236,
"grad_norm": 1.8307541279709998,
"learning_rate": 9.9402061302608e-06,
"loss": 1.8489,
"step": 2100
},
{
"epoch": 0.6150765269632384,
"grad_norm": 1.3493002065305932,
"learning_rate": 9.93684601089133e-06,
"loss": 1.8428,
"step": 2150
},
{
"epoch": 0.6293806322414532,
"grad_norm": 1.3269702867770274,
"learning_rate": 9.933394709989109e-06,
"loss": 1.8485,
"step": 2200
},
{
"epoch": 0.6436847375196681,
"grad_norm": 1.3587430620093466,
"learning_rate": 9.92985229848175e-06,
"loss": 1.8323,
"step": 2250
},
{
"epoch": 0.657988842797883,
"grad_norm": 1.6220456911234815,
"learning_rate": 9.926218849169284e-06,
"loss": 1.8417,
"step": 2300
},
{
"epoch": 0.6722929480760979,
"grad_norm": 1.3306733931984986,
"learning_rate": 9.922494436722653e-06,
"loss": 1.8319,
"step": 2350
},
{
"epoch": 0.6865970533543126,
"grad_norm": 1.841456096268586,
"learning_rate": 9.91867913768218e-06,
"loss": 1.8221,
"step": 2400
},
{
"epoch": 0.6865970533543126,
"eval_loss": 1.8098633289337158,
"eval_runtime": 13.2499,
"eval_samples_per_second": 75.472,
"eval_steps_per_second": 2.415,
"step": 2400
},
{
"epoch": 0.7009011586325276,
"grad_norm": 1.0743552491621962,
"learning_rate": 9.914773030456001e-06,
"loss": 1.8264,
"step": 2450
},
{
"epoch": 0.7152052639107424,
"grad_norm": 0.9817511549306134,
"learning_rate": 9.910776195318448e-06,
"loss": 1.8339,
"step": 2500
},
{
"epoch": 0.7295093691889573,
"grad_norm": 2.5586651091316166,
"learning_rate": 9.906688714408396e-06,
"loss": 1.8361,
"step": 2550
},
{
"epoch": 0.743813474467172,
"grad_norm": 1.3521713422261818,
"learning_rate": 9.902510671727583e-06,
"loss": 1.8172,
"step": 2600
},
{
"epoch": 0.758117579745387,
"grad_norm": 1.196264818533523,
"learning_rate": 9.898242153138882e-06,
"loss": 1.815,
"step": 2650
},
{
"epoch": 0.7724216850236018,
"grad_norm": 1.1236416672692373,
"learning_rate": 9.89388324636453e-06,
"loss": 1.8217,
"step": 2700
},
{
"epoch": 0.7867257903018167,
"grad_norm": 1.0482615125286014,
"learning_rate": 9.889434040984333e-06,
"loss": 1.8111,
"step": 2750
},
{
"epoch": 0.8010298955800315,
"grad_norm": 1.2877895138097581,
"learning_rate": 9.88489462843382e-06,
"loss": 1.8139,
"step": 2800
},
{
"epoch": 0.8010298955800315,
"eval_loss": 1.7917475700378418,
"eval_runtime": 13.2172,
"eval_samples_per_second": 75.659,
"eval_steps_per_second": 2.421,
"step": 2800
},
{
"epoch": 0.8153340008582464,
"grad_norm": 1.3263953998902482,
"learning_rate": 9.880265102002369e-06,
"loss": 1.7969,
"step": 2850
},
{
"epoch": 0.8296381061364612,
"grad_norm": 0.8011721107369111,
"learning_rate": 9.875545556831283e-06,
"loss": 1.8123,
"step": 2900
},
{
"epoch": 0.843942211414676,
"grad_norm": 1.452056398499391,
"learning_rate": 9.870736089911836e-06,
"loss": 1.8137,
"step": 2950
},
{
"epoch": 0.8582463166928909,
"grad_norm": 1.1618490357992515,
"learning_rate": 9.865836800083291e-06,
"loss": 1.812,
"step": 3000
},
{
"epoch": 0.8725504219711057,
"grad_norm": 1.2693408191050113,
"learning_rate": 9.860847788030852e-06,
"loss": 1.7989,
"step": 3050
},
{
"epoch": 0.8868545272493206,
"grad_norm": 1.1394593474503365,
"learning_rate": 9.855769156283604e-06,
"loss": 1.8068,
"step": 3100
},
{
"epoch": 0.9011586325275354,
"grad_norm": 1.739305216161417,
"learning_rate": 9.850601009212408e-06,
"loss": 1.8026,
"step": 3150
},
{
"epoch": 0.9154627378057503,
"grad_norm": 0.9509459350795362,
"learning_rate": 9.845343453027747e-06,
"loss": 1.8055,
"step": 3200
},
{
"epoch": 0.9154627378057503,
"eval_loss": 1.7801611423492432,
"eval_runtime": 13.218,
"eval_samples_per_second": 75.654,
"eval_steps_per_second": 2.421,
"step": 3200
},
{
"epoch": 0.9297668430839651,
"grad_norm": 1.279124709454881,
"learning_rate": 9.839996595777552e-06,
"loss": 1.7968,
"step": 3250
},
{
"epoch": 0.94407094836218,
"grad_norm": 1.092437156034456,
"learning_rate": 9.83456054734498e-06,
"loss": 1.7892,
"step": 3300
},
{
"epoch": 0.9583750536403948,
"grad_norm": 0.9619111877602855,
"learning_rate": 9.829035419446156e-06,
"loss": 1.7951,
"step": 3350
},
{
"epoch": 0.9726791589186097,
"grad_norm": 0.9476479377652314,
"learning_rate": 9.823421325627865e-06,
"loss": 1.8003,
"step": 3400
},
{
"epoch": 0.9869832641968245,
"grad_norm": 0.7761637732992411,
"learning_rate": 9.81771838126524e-06,
"loss": 1.7852,
"step": 3450
},
{
"epoch": 1.0012873694750393,
"grad_norm": 1.6427439742947594,
"learning_rate": 9.811926703559374e-06,
"loss": 1.7813,
"step": 3500
},
{
"epoch": 1.0155914747532542,
"grad_norm": 1.1995527718721077,
"learning_rate": 9.806046411534916e-06,
"loss": 1.7693,
"step": 3550
},
{
"epoch": 1.029895580031469,
"grad_norm": 0.6804399746581633,
"learning_rate": 9.800077626037633e-06,
"loss": 1.7779,
"step": 3600
},
{
"epoch": 1.029895580031469,
"eval_loss": 1.770551323890686,
"eval_runtime": 13.1907,
"eval_samples_per_second": 75.811,
"eval_steps_per_second": 2.426,
"step": 3600
},
{
"epoch": 1.044199685309684,
"grad_norm": 0.9739432549705861,
"learning_rate": 9.794020469731915e-06,
"loss": 1.7844,
"step": 3650
},
{
"epoch": 1.0585037905878987,
"grad_norm": 1.24072763349112,
"learning_rate": 9.787875067098257e-06,
"loss": 1.7826,
"step": 3700
},
{
"epoch": 1.0728078958661136,
"grad_norm": 1.0497808533387654,
"learning_rate": 9.781641544430703e-06,
"loss": 1.7875,
"step": 3750
},
{
"epoch": 1.0871120011443285,
"grad_norm": 1.608878092364201,
"learning_rate": 9.775320029834255e-06,
"loss": 1.7676,
"step": 3800
},
{
"epoch": 1.1014161064225432,
"grad_norm": 0.8265375857322264,
"learning_rate": 9.76891065322223e-06,
"loss": 1.7758,
"step": 3850
},
{
"epoch": 1.115720211700758,
"grad_norm": 0.8074509140095618,
"learning_rate": 9.762413546313597e-06,
"loss": 1.773,
"step": 3900
},
{
"epoch": 1.130024316978973,
"grad_norm": 1.2955529307297018,
"learning_rate": 9.755828842630269e-06,
"loss": 1.7653,
"step": 3950
},
{
"epoch": 1.144328422257188,
"grad_norm": 0.9723230317765355,
"learning_rate": 9.749156677494357e-06,
"loss": 1.7813,
"step": 4000
},
{
"epoch": 1.144328422257188,
"eval_loss": 1.7602086067199707,
"eval_runtime": 13.0715,
"eval_samples_per_second": 76.502,
"eval_steps_per_second": 2.448,
"step": 4000
},
{
"epoch": 1.1587755685881849,
"grad_norm": 1.6203576789230651,
"learning_rate": 9.742397188025394e-06,
"loss": 1.7709,
"step": 4050
},
{
"epoch": 1.1730796738663996,
"grad_norm": 1.1690135821413727,
"learning_rate": 9.735550513137513e-06,
"loss": 1.7583,
"step": 4100
},
{
"epoch": 1.1873837791446145,
"grad_norm": 1.2584964373503669,
"learning_rate": 9.728616793536588e-06,
"loss": 1.7653,
"step": 4150
},
{
"epoch": 1.2016878844228294,
"grad_norm": 1.022986798804273,
"learning_rate": 9.721596171717352e-06,
"loss": 1.7661,
"step": 4200
},
{
"epoch": 1.2159919897010443,
"grad_norm": 0.9715095488392225,
"learning_rate": 9.714488791960463e-06,
"loss": 1.7745,
"step": 4250
},
{
"epoch": 1.230296094979259,
"grad_norm": 1.1370122008173429,
"learning_rate": 9.707294800329536e-06,
"loss": 1.7684,
"step": 4300
},
{
"epoch": 1.2446002002574739,
"grad_norm": 0.9536116406695613,
"learning_rate": 9.700014344668152e-06,
"loss": 1.7606,
"step": 4350
},
{
"epoch": 1.2589043055356888,
"grad_norm": 0.7353885681465049,
"learning_rate": 9.692647574596803e-06,
"loss": 1.7633,
"step": 4400
},
{
"epoch": 1.2589043055356888,
"eval_loss": 1.752835988998413,
"eval_runtime": 13.2443,
"eval_samples_per_second": 75.504,
"eval_steps_per_second": 2.416,
"step": 4400
},
{
"epoch": 1.2732084108139037,
"grad_norm": 1.1802783991524877,
"learning_rate": 9.685194641509837e-06,
"loss": 1.7686,
"step": 4450
},
{
"epoch": 1.2875125160921184,
"grad_norm": 0.9318456023565893,
"learning_rate": 9.677655698572326e-06,
"loss": 1.7556,
"step": 4500
},
{
"epoch": 1.3018166213703333,
"grad_norm": 1.0437155245378358,
"learning_rate": 9.670030900716941e-06,
"loss": 1.7571,
"step": 4550
},
{
"epoch": 1.3161207266485482,
"grad_norm": 1.0912943460238356,
"learning_rate": 9.662320404640743e-06,
"loss": 1.7546,
"step": 4600
},
{
"epoch": 1.3304248319267629,
"grad_norm": 1.0063174963239234,
"learning_rate": 9.654524368801982e-06,
"loss": 1.7644,
"step": 4650
},
{
"epoch": 1.3447289372049778,
"grad_norm": 1.0417193249219145,
"learning_rate": 9.646642953416835e-06,
"loss": 1.7525,
"step": 4700
},
{
"epoch": 1.3590330424831927,
"grad_norm": 0.698422675064761,
"learning_rate": 9.638676320456109e-06,
"loss": 1.7586,
"step": 4750
},
{
"epoch": 1.3733371477614076,
"grad_norm": 1.1910335445073554,
"learning_rate": 9.630624633641918e-06,
"loss": 1.7528,
"step": 4800
},
{
"epoch": 1.3733371477614076,
"eval_loss": 1.7467565536499023,
"eval_runtime": 13.1696,
"eval_samples_per_second": 75.932,
"eval_steps_per_second": 2.43,
"step": 4800
},
{
"epoch": 1.3876412530396225,
"grad_norm": 0.860266713830524,
"learning_rate": 9.622488058444313e-06,
"loss": 1.7526,
"step": 4850
},
{
"epoch": 1.4019453583178372,
"grad_norm": 0.6381947527857678,
"learning_rate": 9.614266762077891e-06,
"loss": 1.7548,
"step": 4900
},
{
"epoch": 1.416249463596052,
"grad_norm": 0.7554075411255639,
"learning_rate": 9.605960913498342e-06,
"loss": 1.7481,
"step": 4950
},
{
"epoch": 1.430553568874267,
"grad_norm": 1.2154500150186984,
"learning_rate": 9.597570683398996e-06,
"loss": 1.7584,
"step": 5000
},
{
"epoch": 1.4448576741524817,
"grad_norm": 0.6439874478006935,
"learning_rate": 9.5890962442073e-06,
"loss": 1.7445,
"step": 5050
},
{
"epoch": 1.4591617794306966,
"grad_norm": 1.1984560510642397,
"learning_rate": 9.580537770081285e-06,
"loss": 1.7442,
"step": 5100
},
{
"epoch": 1.4734658847089115,
"grad_norm": 1.1871050477497047,
"learning_rate": 9.57189543690598e-06,
"loss": 1.7498,
"step": 5150
},
{
"epoch": 1.4877699899871262,
"grad_norm": 0.7133766609729304,
"learning_rate": 9.563169422289798e-06,
"loss": 1.7423,
"step": 5200
},
{
"epoch": 1.4877699899871262,
"eval_loss": 1.7412633895874023,
"eval_runtime": 13.1818,
"eval_samples_per_second": 75.862,
"eval_steps_per_second": 2.428,
"step": 5200
},
{
"epoch": 1.5020740952653413,
"grad_norm": 0.7612313099084432,
"learning_rate": 9.554359905560887e-06,
"loss": 1.748,
"step": 5250
},
{
"epoch": 1.516378200543556,
"grad_norm": 0.750224706069401,
"learning_rate": 9.54546706776345e-06,
"loss": 1.7435,
"step": 5300
},
{
"epoch": 1.530682305821771,
"grad_norm": 0.9331760005679197,
"learning_rate": 9.536491091654018e-06,
"loss": 1.7324,
"step": 5350
},
{
"epoch": 1.5449864110999858,
"grad_norm": 0.8613210624917503,
"learning_rate": 9.527432161697696e-06,
"loss": 1.751,
"step": 5400
},
{
"epoch": 1.5592905163782005,
"grad_norm": 4.862833366342641,
"learning_rate": 9.518290464064365e-06,
"loss": 1.7365,
"step": 5450
},
{
"epoch": 1.5735946216564154,
"grad_norm": 0.7838749485611254,
"learning_rate": 9.509066186624872e-06,
"loss": 1.7399,
"step": 5500
},
{
"epoch": 1.5878987269346303,
"grad_norm": 0.798441318649249,
"learning_rate": 9.499759518947156e-06,
"loss": 1.737,
"step": 5550
},
{
"epoch": 1.602202832212845,
"grad_norm": 0.7078515946606195,
"learning_rate": 9.490370652292357e-06,
"loss": 1.7412,
"step": 5600
},
{
"epoch": 1.602202832212845,
"eval_loss": 1.7350177764892578,
"eval_runtime": 13.1662,
"eval_samples_per_second": 75.952,
"eval_steps_per_second": 2.43,
"step": 5600
},
{
"epoch": 1.61650693749106,
"grad_norm": 0.8939699763780663,
"learning_rate": 9.480899779610883e-06,
"loss": 1.7485,
"step": 5650
},
{
"epoch": 1.6308110427692748,
"grad_norm": 0.7241948247907901,
"learning_rate": 9.471347095538448e-06,
"loss": 1.7338,
"step": 5700
},
{
"epoch": 1.6451151480474895,
"grad_norm": 0.7753555958595959,
"learning_rate": 9.461712796392067e-06,
"loss": 1.7423,
"step": 5750
},
{
"epoch": 1.6594192533257046,
"grad_norm": 0.9453545668822089,
"learning_rate": 9.45199708016603e-06,
"loss": 1.7411,
"step": 5800
},
{
"epoch": 1.6737233586039193,
"grad_norm": 0.8069354542585767,
"learning_rate": 9.442200146527824e-06,
"loss": 1.7341,
"step": 5850
},
{
"epoch": 1.6880274638821342,
"grad_norm": 0.9570095002200463,
"learning_rate": 9.432322196814032e-06,
"loss": 1.7309,
"step": 5900
},
{
"epoch": 1.7023315691603491,
"grad_norm": 0.778079342980594,
"learning_rate": 9.422363434026205e-06,
"loss": 1.7331,
"step": 5950
},
{
"epoch": 1.7166356744385638,
"grad_norm": 0.8582782820893037,
"learning_rate": 9.41232406282667e-06,
"loss": 1.7375,
"step": 6000
},
{
"epoch": 1.7166356744385638,
"eval_loss": 1.7310324907302856,
"eval_runtime": 13.1951,
"eval_samples_per_second": 75.786,
"eval_steps_per_second": 2.425,
"step": 6000
},
{
"epoch": 1.7309397797167787,
"grad_norm": 0.657112371502439,
"learning_rate": 9.402204289534344e-06,
"loss": 1.725,
"step": 6050
},
{
"epoch": 1.7452438849949936,
"grad_norm": 1.1123814501158173,
"learning_rate": 9.392004322120484e-06,
"loss": 1.7303,
"step": 6100
},
{
"epoch": 1.7595479902732083,
"grad_norm": 0.9499171364049573,
"learning_rate": 9.381724370204414e-06,
"loss": 1.7203,
"step": 6150
},
{
"epoch": 1.7738520955514234,
"grad_norm": 1.095712190889447,
"learning_rate": 9.371364645049216e-06,
"loss": 1.7291,
"step": 6200
},
{
"epoch": 1.788156200829638,
"grad_norm": 1.0199371888229702,
"learning_rate": 9.360925359557397e-06,
"loss": 1.7155,
"step": 6250
},
{
"epoch": 1.8024603061078528,
"grad_norm": 0.827904086276593,
"learning_rate": 9.3504067282665e-06,
"loss": 1.7404,
"step": 6300
},
{
"epoch": 1.816764411386068,
"grad_norm": 0.8252103968718575,
"learning_rate": 9.339808967344701e-06,
"loss": 1.7275,
"step": 6350
},
{
"epoch": 1.8310685166642826,
"grad_norm": 0.8608721928591242,
"learning_rate": 9.329132294586374e-06,
"loss": 1.7257,
"step": 6400
},
{
"epoch": 1.8310685166642826,
"eval_loss": 1.7276860475540161,
"eval_runtime": 13.1438,
"eval_samples_per_second": 76.082,
"eval_steps_per_second": 2.435,
"step": 6400
},
{
"epoch": 1.8453726219424975,
"grad_norm": 1.0655198477534613,
"learning_rate": 9.318376929407606e-06,
"loss": 1.7219,
"step": 6450
},
{
"epoch": 1.8596767272207124,
"grad_norm": 0.8202058864526146,
"learning_rate": 9.307543092841688e-06,
"loss": 1.7219,
"step": 6500
},
{
"epoch": 1.873980832498927,
"grad_norm": 0.9119156406563158,
"learning_rate": 9.296631007534576e-06,
"loss": 1.7232,
"step": 6550
},
{
"epoch": 1.888284937777142,
"grad_norm": 0.8561460550361113,
"learning_rate": 9.285640897740316e-06,
"loss": 1.718,
"step": 6600
},
{
"epoch": 1.902589043055357,
"grad_norm": 0.9549020817113293,
"learning_rate": 9.27457298931643e-06,
"loss": 1.7348,
"step": 6650
},
{
"epoch": 1.9168931483335716,
"grad_norm": 0.6729589207861246,
"learning_rate": 9.263427509719287e-06,
"loss": 1.7175,
"step": 6700
},
{
"epoch": 1.9311972536117867,
"grad_norm": 0.709786130314777,
"learning_rate": 9.252204687999401e-06,
"loss": 1.7287,
"step": 6750
},
{
"epoch": 1.9455013588900014,
"grad_norm": 0.9827727167496969,
"learning_rate": 9.240904754796767e-06,
"loss": 1.7241,
"step": 6800
},
{
"epoch": 1.9455013588900014,
"eval_loss": 1.7200103998184204,
"eval_runtime": 13.113,
"eval_samples_per_second": 76.26,
"eval_steps_per_second": 2.44,
"step": 6800
},
{
"epoch": 1.9598054641682163,
"grad_norm": 0.8218927694104888,
"learning_rate": 9.22952794233608e-06,
"loss": 1.7334,
"step": 6850
},
{
"epoch": 1.9741095694464312,
"grad_norm": 0.9202146449681148,
"learning_rate": 9.218074484421977e-06,
"loss": 1.7176,
"step": 6900
},
{
"epoch": 1.988413674724646,
"grad_norm": 0.9757466555134802,
"learning_rate": 9.206544616434249e-06,
"loss": 1.716,
"step": 6950
},
{
"epoch": 2.0027177800028606,
"grad_norm": 0.9665391581844135,
"learning_rate": 9.194938575322973e-06,
"loss": 1.7104,
"step": 7000
},
{
"epoch": 2.0170218852810757,
"grad_norm": 0.747601621726994,
"learning_rate": 9.183256599603672e-06,
"loss": 1.7045,
"step": 7050
},
{
"epoch": 2.0313259905592904,
"grad_norm": 0.590573479620199,
"learning_rate": 9.171498929352388e-06,
"loss": 1.7012,
"step": 7100
},
{
"epoch": 2.0456300958375055,
"grad_norm": 0.7932008874801502,
"learning_rate": 9.159665806200766e-06,
"loss": 1.7017,
"step": 7150
},
{
"epoch": 2.0599342011157202,
"grad_norm": 0.824718775314277,
"learning_rate": 9.147757473331082e-06,
"loss": 1.7019,
"step": 7200
},
{
"epoch": 2.0599342011157202,
"eval_loss": 1.7177115678787231,
"eval_runtime": 13.1014,
"eval_samples_per_second": 76.328,
"eval_steps_per_second": 2.442,
"step": 7200
},
{
"epoch": 2.074238306393935,
"grad_norm": 1.080101464877567,
"learning_rate": 9.135774175471244e-06,
"loss": 1.7056,
"step": 7250
},
{
"epoch": 2.08854241167215,
"grad_norm": 0.6161967559479131,
"learning_rate": 9.123716158889765e-06,
"loss": 1.6892,
"step": 7300
},
{
"epoch": 2.1028465169503647,
"grad_norm": 0.6259364978680784,
"learning_rate": 9.111583671390697e-06,
"loss": 1.6876,
"step": 7350
},
{
"epoch": 2.1171506222285794,
"grad_norm": 0.9054346826665011,
"learning_rate": 9.09937696230855e-06,
"loss": 1.7004,
"step": 7400
},
{
"epoch": 2.1314547275067945,
"grad_norm": 0.8115910709185014,
"learning_rate": 9.087096282503152e-06,
"loss": 1.6951,
"step": 7450
},
{
"epoch": 2.1457588327850092,
"grad_norm": 0.939188480700838,
"learning_rate": 9.074741884354507e-06,
"loss": 1.698,
"step": 7500
},
{
"epoch": 2.1600629380632244,
"grad_norm": 0.5461813129993677,
"learning_rate": 9.062314021757603e-06,
"loss": 1.6953,
"step": 7550
},
{
"epoch": 2.174367043341439,
"grad_norm": 0.8915398126133827,
"learning_rate": 9.049812950117191e-06,
"loss": 1.6838,
"step": 7600
},
{
"epoch": 2.174367043341439,
"eval_loss": 1.7138569355010986,
"eval_runtime": 13.1252,
"eval_samples_per_second": 76.189,
"eval_steps_per_second": 2.438,
"step": 7600
},
{
"epoch": 2.1886711486196537,
"grad_norm": 0.8119718442892704,
"learning_rate": 9.037238926342544e-06,
"loss": 1.7012,
"step": 7650
},
{
"epoch": 2.202975253897869,
"grad_norm": 0.6869354062120884,
"learning_rate": 9.02459220884217e-06,
"loss": 1.6918,
"step": 7700
},
{
"epoch": 2.2172793591760835,
"grad_norm": 0.7341604750585072,
"learning_rate": 9.011873057518503e-06,
"loss": 1.699,
"step": 7750
},
{
"epoch": 2.2315834644542982,
"grad_norm": 0.8261410417964352,
"learning_rate": 8.999081733762568e-06,
"loss": 1.6993,
"step": 7800
},
{
"epoch": 2.2458875697325134,
"grad_norm": 0.7611803854476097,
"learning_rate": 8.986218500448598e-06,
"loss": 1.7069,
"step": 7850
},
{
"epoch": 2.260191675010728,
"grad_norm": 0.771284372878959,
"learning_rate": 8.973283621928644e-06,
"loss": 1.7018,
"step": 7900
},
{
"epoch": 2.2744957802889427,
"grad_norm": 0.6069248728630589,
"learning_rate": 8.96027736402713e-06,
"loss": 1.6894,
"step": 7950
},
{
"epoch": 2.288799885567158,
"grad_norm": 0.6935253796751046,
"learning_rate": 8.947199994035402e-06,
"loss": 1.6857,
"step": 8000
},
{
"epoch": 2.288799885567158,
"eval_loss": 1.709199070930481,
"eval_runtime": 13.105,
"eval_samples_per_second": 76.307,
"eval_steps_per_second": 2.442,
"step": 8000
},
{
"epoch": 2.3031039908453725,
"grad_norm": 0.6395079401863663,
"learning_rate": 8.934051780706226e-06,
"loss": 1.6867,
"step": 8050
},
{
"epoch": 2.3174080961235877,
"grad_norm": 0.8808561738320105,
"learning_rate": 8.920832994248268e-06,
"loss": 1.6947,
"step": 8100
},
{
"epoch": 2.3317122014018024,
"grad_norm": 0.6888071930360784,
"learning_rate": 8.907543906320542e-06,
"loss": 1.7005,
"step": 8150
},
{
"epoch": 2.346016306680017,
"grad_norm": 0.6515296032830782,
"learning_rate": 8.894184790026823e-06,
"loss": 1.686,
"step": 8200
},
{
"epoch": 2.360320411958232,
"grad_norm": 0.7401595322241383,
"learning_rate": 8.880755919910048e-06,
"loss": 1.6865,
"step": 8250
},
{
"epoch": 2.374624517236447,
"grad_norm": 0.6000159117604718,
"learning_rate": 8.867257571946646e-06,
"loss": 1.6996,
"step": 8300
},
{
"epoch": 2.3889286225146615,
"grad_norm": 0.6577366719572108,
"learning_rate": 8.853690023540898e-06,
"loss": 1.6929,
"step": 8350
},
{
"epoch": 2.4032327277928767,
"grad_norm": 0.7392515614712054,
"learning_rate": 8.840053553519216e-06,
"loss": 1.6848,
"step": 8400
},
{
"epoch": 2.4032327277928767,
"eval_loss": 1.7050341367721558,
"eval_runtime": 13.1691,
"eval_samples_per_second": 75.935,
"eval_steps_per_second": 2.43,
"step": 8400
},
{
"epoch": 2.4175368330710914,
"grad_norm": 0.9889920645015057,
"learning_rate": 8.82634844212442e-06,
"loss": 1.6893,
"step": 8450
},
{
"epoch": 2.431840938349306,
"grad_norm": 0.9590160412926145,
"learning_rate": 8.81257497100998e-06,
"loss": 1.6846,
"step": 8500
},
{
"epoch": 2.446145043627521,
"grad_norm": 0.8138219488107618,
"learning_rate": 8.79873342323422e-06,
"loss": 1.689,
"step": 8550
},
{
"epoch": 2.460449148905736,
"grad_norm": 0.6771189511098283,
"learning_rate": 8.78482408325451e-06,
"loss": 1.684,
"step": 8600
},
{
"epoch": 2.474753254183951,
"grad_norm": 0.637569459443737,
"learning_rate": 8.770847236921412e-06,
"loss": 1.6893,
"step": 8650
},
{
"epoch": 2.4890573594621657,
"grad_norm": 0.8520093583138391,
"learning_rate": 8.756803171472817e-06,
"loss": 1.6858,
"step": 8700
},
{
"epoch": 2.5033614647403803,
"grad_norm": 1.0853751747133211,
"learning_rate": 8.742692175528027e-06,
"loss": 1.6929,
"step": 8750
},
{
"epoch": 2.5176655700185955,
"grad_norm": 0.5845518675223795,
"learning_rate": 8.728514539081837e-06,
"loss": 1.6795,
"step": 8800
},
{
"epoch": 2.5176655700185955,
"eval_loss": 1.6996084451675415,
"eval_runtime": 13.1313,
"eval_samples_per_second": 76.154,
"eval_steps_per_second": 2.437,
"step": 8800
},
{
"epoch": 2.53196967529681,
"grad_norm": 0.6987736172220224,
"learning_rate": 8.714270553498567e-06,
"loss": 1.6786,
"step": 8850
},
{
"epoch": 2.5462737805750253,
"grad_norm": 0.6845251902273746,
"learning_rate": 8.699960511506077e-06,
"loss": 1.6915,
"step": 8900
},
{
"epoch": 2.56057788585324,
"grad_norm": 0.5813491957603439,
"learning_rate": 8.685584707189749e-06,
"loss": 1.6864,
"step": 8950
},
{
"epoch": 2.5748819911314547,
"grad_norm": 0.7758297556506881,
"learning_rate": 8.671143435986447e-06,
"loss": 1.6853,
"step": 9000
},
{
"epoch": 2.5891860964096693,
"grad_norm": 0.6480784306311093,
"learning_rate": 8.656636994678447e-06,
"loss": 1.6945,
"step": 9050
},
{
"epoch": 2.6034902016878845,
"grad_norm": 0.752842227645809,
"learning_rate": 8.642065681387329e-06,
"loss": 1.6856,
"step": 9100
},
{
"epoch": 2.617794306966099,
"grad_norm": 0.8563670508754716,
"learning_rate": 8.627429795567858e-06,
"loss": 1.6813,
"step": 9150
},
{
"epoch": 2.6320984122443143,
"grad_norm": 0.649858543338678,
"learning_rate": 8.61272963800183e-06,
"loss": 1.6807,
"step": 9200
},
{
"epoch": 2.6320984122443143,
"eval_loss": 1.6953411102294922,
"eval_runtime": 13.1331,
"eval_samples_per_second": 76.143,
"eval_steps_per_second": 2.437,
"step": 9200
},
{
"epoch": 2.646402517522529,
"grad_norm": 0.6485355884487186,
"learning_rate": 8.597965510791883e-06,
"loss": 1.691,
"step": 9250
},
{
"epoch": 2.6607066228007437,
"grad_norm": 0.8331706908409144,
"learning_rate": 8.5831377173553e-06,
"loss": 1.6757,
"step": 9300
},
{
"epoch": 2.675010728078959,
"grad_norm": 0.7237339174259747,
"learning_rate": 8.568246562417762e-06,
"loss": 1.6838,
"step": 9350
},
{
"epoch": 2.6893148333571735,
"grad_norm": 0.6551466282285732,
"learning_rate": 8.553292352007096e-06,
"loss": 1.6815,
"step": 9400
},
{
"epoch": 2.7036189386353886,
"grad_norm": 0.6601355008192104,
"learning_rate": 8.538275393446976e-06,
"loss": 1.6857,
"step": 9450
},
{
"epoch": 2.7179230439136033,
"grad_norm": 0.6614316555641063,
"learning_rate": 8.523195995350613e-06,
"loss": 1.6823,
"step": 9500
},
{
"epoch": 2.732227149191818,
"grad_norm": 0.9043872683705563,
"learning_rate": 8.508054467614417e-06,
"loss": 1.6807,
"step": 9550
},
{
"epoch": 2.7465312544700327,
"grad_norm": 0.6024321505835114,
"learning_rate": 8.492851121411614e-06,
"loss": 1.6775,
"step": 9600
},
{
"epoch": 2.7465312544700327,
"eval_loss": 1.6916736364364624,
"eval_runtime": 13.1667,
"eval_samples_per_second": 75.949,
"eval_steps_per_second": 2.43,
"step": 9600
},
{
"epoch": 2.760835359748248,
"grad_norm": 0.5885102755974863,
"learning_rate": 8.477586269185868e-06,
"loss": 1.6783,
"step": 9650
},
{
"epoch": 2.7751394650264625,
"grad_norm": 0.5923048948697456,
"learning_rate": 8.462260224644848e-06,
"loss": 1.6754,
"step": 9700
},
{
"epoch": 2.7894435703046776,
"grad_norm": 0.658689052183787,
"learning_rate": 8.446873302753783e-06,
"loss": 1.6879,
"step": 9750
},
{
"epoch": 2.8037476755828923,
"grad_norm": 0.6839558742874721,
"learning_rate": 8.431425819728998e-06,
"loss": 1.6833,
"step": 9800
},
{
"epoch": 2.818051780861107,
"grad_norm": 0.563076108100303,
"learning_rate": 8.415918093031403e-06,
"loss": 1.6746,
"step": 9850
},
{
"epoch": 2.832355886139322,
"grad_norm": 0.6412513978425523,
"learning_rate": 8.400350441359976e-06,
"loss": 1.6673,
"step": 9900
},
{
"epoch": 2.846659991417537,
"grad_norm": 0.6509884557319101,
"learning_rate": 8.384723184645211e-06,
"loss": 1.6736,
"step": 9950
},
{
"epoch": 2.860964096695752,
"grad_norm": 0.6795285605676114,
"learning_rate": 8.369036644042546e-06,
"loss": 1.6602,
"step": 10000
},
{
"epoch": 2.860964096695752,
"eval_loss": 1.6880682706832886,
"eval_runtime": 13.1264,
"eval_samples_per_second": 76.182,
"eval_steps_per_second": 2.438,
"step": 10000
},
{
"epoch": 2.8752682019739666,
"grad_norm": 0.6331270907140948,
"learning_rate": 8.353291141925763e-06,
"loss": 1.6678,
"step": 10050
},
{
"epoch": 2.8895723072521813,
"grad_norm": 0.6165703754864105,
"learning_rate": 8.337487001880353e-06,
"loss": 1.6779,
"step": 10100
},
{
"epoch": 2.903876412530396,
"grad_norm": 0.9180760124515276,
"learning_rate": 8.32162454869688e-06,
"loss": 1.6709,
"step": 10150
},
{
"epoch": 2.918180517808611,
"grad_norm": 0.6067014553628142,
"learning_rate": 8.305704108364301e-06,
"loss": 1.6748,
"step": 10200
},
{
"epoch": 2.932484623086826,
"grad_norm": 0.6822167283880958,
"learning_rate": 8.289726008063265e-06,
"loss": 1.6734,
"step": 10250
},
{
"epoch": 2.946788728365041,
"grad_norm": 0.5414334397184379,
"learning_rate": 8.273690576159383e-06,
"loss": 1.674,
"step": 10300
},
{
"epoch": 2.9610928336432556,
"grad_norm": 0.6960756922615432,
"learning_rate": 8.257598142196496e-06,
"loss": 1.6611,
"step": 10350
},
{
"epoch": 2.9753969389214703,
"grad_norm": 0.6604617349648914,
"learning_rate": 8.241449036889892e-06,
"loss": 1.6676,
"step": 10400
},
{
"epoch": 2.9753969389214703,
"eval_loss": 1.6850733757019043,
"eval_runtime": 13.109,
"eval_samples_per_second": 76.284,
"eval_steps_per_second": 2.441,
"step": 10400
},
{
"epoch": 2.9897010441996854,
"grad_norm": 0.803147857087257,
"learning_rate": 8.225243592119501e-06,
"loss": 1.6742,
"step": 10450
},
{
"epoch": 3.0040051494779,
"grad_norm": 0.7661508794366779,
"learning_rate": 8.208982140923095e-06,
"loss": 1.6643,
"step": 10500
},
{
"epoch": 3.0183092547561152,
"grad_norm": 0.9234438361457962,
"learning_rate": 8.192665017489431e-06,
"loss": 1.6358,
"step": 10550
},
{
"epoch": 3.03261336003433,
"grad_norm": 0.6961614825939386,
"learning_rate": 8.17629255715138e-06,
"loss": 1.6545,
"step": 10600
},
{
"epoch": 3.0469174653125446,
"grad_norm": 0.6895369107938377,
"learning_rate": 8.159865096379046e-06,
"loss": 1.6321,
"step": 10650
},
{
"epoch": 3.0612215705907597,
"grad_norm": 0.7078335937537477,
"learning_rate": 8.14338297277284e-06,
"loss": 1.6349,
"step": 10700
},
{
"epoch": 3.0755256758689744,
"grad_norm": 0.6076483839068002,
"learning_rate": 8.126846525056555e-06,
"loss": 1.6365,
"step": 10750
},
{
"epoch": 3.089829781147189,
"grad_norm": 0.6898999480835771,
"learning_rate": 8.110256093070393e-06,
"loss": 1.6546,
"step": 10800
},
{
"epoch": 3.089829781147189,
"eval_loss": 1.6832627058029175,
"eval_runtime": 13.1347,
"eval_samples_per_second": 76.134,
"eval_steps_per_second": 2.436,
"step": 10800
},
{
"epoch": 3.104133886425404,
"grad_norm": 0.7013516335683451,
"learning_rate": 8.093612017763986e-06,
"loss": 1.639,
"step": 10850
},
{
"epoch": 3.118437991703619,
"grad_norm": 0.6585546257998172,
"learning_rate": 8.076914641189388e-06,
"loss": 1.649,
"step": 10900
},
{
"epoch": 3.1327420969818336,
"grad_norm": 0.6634919864983012,
"learning_rate": 8.060164306494052e-06,
"loss": 1.6349,
"step": 10950
},
{
"epoch": 3.1470462022600487,
"grad_norm": 0.6535125345700776,
"learning_rate": 8.043361357913763e-06,
"loss": 1.6354,
"step": 11000
},
{
"epoch": 3.1613503075382634,
"grad_norm": 0.6365757392209358,
"learning_rate": 8.026506140765581e-06,
"loss": 1.6411,
"step": 11050
},
{
"epoch": 3.1756544128164785,
"grad_norm": 0.7613496780923112,
"learning_rate": 8.009599001440733e-06,
"loss": 1.6443,
"step": 11100
},
{
"epoch": 3.189958518094693,
"grad_norm": 0.6788371997973981,
"learning_rate": 7.992640287397498e-06,
"loss": 1.6406,
"step": 11150
},
{
"epoch": 3.204262623372908,
"grad_norm": 0.7267161462233287,
"learning_rate": 7.975630347154062e-06,
"loss": 1.6416,
"step": 11200
},
{
"epoch": 3.204262623372908,
"eval_loss": 1.6800661087036133,
"eval_runtime": 13.1106,
"eval_samples_per_second": 76.274,
"eval_steps_per_second": 2.441,
"step": 11200
},
{
"epoch": 3.218566728651123,
"grad_norm": 0.6610572852986926,
"learning_rate": 7.958569530281369e-06,
"loss": 1.637,
"step": 11250
},
{
"epoch": 3.2328708339293377,
"grad_norm": 0.9622885989079347,
"learning_rate": 7.941458187395918e-06,
"loss": 1.6356,
"step": 11300
},
{
"epoch": 3.2471749392075524,
"grad_norm": 0.6376730332996383,
"learning_rate": 7.924296670152573e-06,
"loss": 1.6353,
"step": 11350
},
{
"epoch": 3.2614790444857675,
"grad_norm": 0.7223900847461898,
"learning_rate": 7.907085331237328e-06,
"loss": 1.6484,
"step": 11400
},
{
"epoch": 3.275783149763982,
"grad_norm": 0.5989178515182939,
"learning_rate": 7.889824524360058e-06,
"loss": 1.6451,
"step": 11450
},
{
"epoch": 3.290087255042197,
"grad_norm": 0.6779901779346886,
"learning_rate": 7.872514604247261e-06,
"loss": 1.6453,
"step": 11500
},
{
"epoch": 3.304391360320412,
"grad_norm": 0.7005880085935425,
"learning_rate": 7.855155926634755e-06,
"loss": 1.6472,
"step": 11550
},
{
"epoch": 3.3186954655986267,
"grad_norm": 0.7609081934908348,
"learning_rate": 7.837748848260372e-06,
"loss": 1.6465,
"step": 11600
},
{
"epoch": 3.3186954655986267,
"eval_loss": 1.6766809225082397,
"eval_runtime": 13.1249,
"eval_samples_per_second": 76.191,
"eval_steps_per_second": 2.438,
"step": 11600
},
{
"epoch": 3.332999570876842,
"grad_norm": 0.6431999289309001,
"learning_rate": 7.820293726856625e-06,
"loss": 1.6363,
"step": 11650
},
{
"epoch": 3.3473036761550565,
"grad_norm": 0.7971053094613834,
"learning_rate": 7.802790921143367e-06,
"loss": 1.642,
"step": 11700
},
{
"epoch": 3.361607781433271,
"grad_norm": 0.8315194233152097,
"learning_rate": 7.785240790820403e-06,
"loss": 1.6356,
"step": 11750
},
{
"epoch": 3.3759118867114863,
"grad_norm": 0.7233317425806471,
"learning_rate": 7.767643696560103e-06,
"loss": 1.625,
"step": 11800
},
{
"epoch": 3.390215991989701,
"grad_norm": 0.7043201889466395,
"learning_rate": 7.75e-06,
"loss": 1.6454,
"step": 11850
},
{
"epoch": 3.4045200972679157,
"grad_norm": 0.8514114238482182,
"learning_rate": 7.732310063735346e-06,
"loss": 1.6283,
"step": 11900
},
{
"epoch": 3.418824202546131,
"grad_norm": 0.8115066197624821,
"learning_rate": 7.71457425131166e-06,
"loss": 1.6235,
"step": 11950
},
{
"epoch": 3.4331283078243455,
"grad_norm": 0.6131161672110997,
"learning_rate": 7.696792927217266e-06,
"loss": 1.6332,
"step": 12000
},
{
"epoch": 3.4331283078243455,
"eval_loss": 1.6723066568374634,
"eval_runtime": 13.1167,
"eval_samples_per_second": 76.238,
"eval_steps_per_second": 2.44,
"step": 12000
},
{
"epoch": 3.44743241310256,
"grad_norm": 0.5768237911834109,
"learning_rate": 7.6789664568758e-06,
"loss": 1.6232,
"step": 12050
},
{
"epoch": 3.4617365183807753,
"grad_norm": 0.6357616712470612,
"learning_rate": 7.661095206638688e-06,
"loss": 1.6384,
"step": 12100
},
{
"epoch": 3.47604062365899,
"grad_norm": 0.6064381660716113,
"learning_rate": 7.643179543777636e-06,
"loss": 1.6423,
"step": 12150
},
{
"epoch": 3.490344728937205,
"grad_norm": 0.6129739275970794,
"learning_rate": 7.625219836477073e-06,
"loss": 1.6376,
"step": 12200
},
{
"epoch": 3.50464883421542,
"grad_norm": 0.6413234223207633,
"learning_rate": 7.607216453826575e-06,
"loss": 1.6342,
"step": 12250
},
{
"epoch": 3.5189529394936345,
"grad_norm": 0.649929569744414,
"learning_rate": 7.589169765813298e-06,
"loss": 1.6418,
"step": 12300
},
{
"epoch": 3.5332570447718497,
"grad_norm": 0.5694768483225782,
"learning_rate": 7.571080143314362e-06,
"loss": 1.6242,
"step": 12350
},
{
"epoch": 3.5475611500500643,
"grad_norm": 0.7350828307091747,
"learning_rate": 7.552947958089234e-06,
"loss": 1.6217,
"step": 12400
},
{
"epoch": 3.5475611500500643,
"eval_loss": 1.6681584119796753,
"eval_runtime": 13.1321,
"eval_samples_per_second": 76.149,
"eval_steps_per_second": 2.437,
"step": 12400
},
{
"epoch": 3.5618652553282795,
"grad_norm": 0.8136490657582578,
"learning_rate": 7.534773582772087e-06,
"loss": 1.6303,
"step": 12450
},
{
"epoch": 3.576169360606494,
"grad_norm": 0.7180717788221607,
"learning_rate": 7.51655739086414e-06,
"loss": 1.6361,
"step": 12500
},
{
"epoch": 3.590473465884709,
"grad_norm": 0.7947804203928884,
"learning_rate": 7.498299756725984e-06,
"loss": 1.6387,
"step": 12550
},
{
"epoch": 3.6047775711629235,
"grad_norm": 0.560362351616168,
"learning_rate": 7.480001055569892e-06,
"loss": 1.6309,
"step": 12600
},
{
"epoch": 3.6190816764411387,
"grad_norm": 0.6426607371565076,
"learning_rate": 7.4616616634521e-06,
"loss": 1.633,
"step": 12650
},
{
"epoch": 3.6333857817193533,
"grad_norm": 0.6171738461490535,
"learning_rate": 7.443281957265086e-06,
"loss": 1.6281,
"step": 12700
},
{
"epoch": 3.6476898869975685,
"grad_norm": 0.6363040797185718,
"learning_rate": 7.424862314729819e-06,
"loss": 1.6311,
"step": 12750
},
{
"epoch": 3.661993992275783,
"grad_norm": 0.7596986784676188,
"learning_rate": 7.406403114388003e-06,
"loss": 1.6292,
"step": 12800
},
{
"epoch": 3.661993992275783,
"eval_loss": 1.6635466814041138,
"eval_runtime": 13.15,
"eval_samples_per_second": 76.046,
"eval_steps_per_second": 2.433,
"step": 12800
},
{
"epoch": 3.676298097553998,
"grad_norm": 0.7293094685670296,
"learning_rate": 7.387904735594291e-06,
"loss": 1.6287,
"step": 12850
},
{
"epoch": 3.690602202832213,
"grad_norm": 0.6158921819942098,
"learning_rate": 7.36936755850849e-06,
"loss": 1.6385,
"step": 12900
},
{
"epoch": 3.7049063081104276,
"grad_norm": 0.5784944167280377,
"learning_rate": 7.3507919640877535e-06,
"loss": 1.6328,
"step": 12950
},
{
"epoch": 3.7192104133886428,
"grad_norm": 0.6275769116215365,
"learning_rate": 7.332178334078746e-06,
"loss": 1.6346,
"step": 13000
},
{
"epoch": 3.7335145186668575,
"grad_norm": 0.7049335766852525,
"learning_rate": 7.313527051009803e-06,
"loss": 1.6314,
"step": 13050
},
{
"epoch": 3.747818623945072,
"grad_norm": 0.5885503989540276,
"learning_rate": 7.2948384981830655e-06,
"loss": 1.6276,
"step": 13100
},
{
"epoch": 3.762122729223287,
"grad_norm": 0.6144133824628834,
"learning_rate": 7.2761130596666045e-06,
"loss": 1.6194,
"step": 13150
},
{
"epoch": 3.776426834501502,
"grad_norm": 0.6880469614329583,
"learning_rate": 7.25735112028653e-06,
"loss": 1.6263,
"step": 13200
},
{
"epoch": 3.776426834501502,
"eval_loss": 1.6603440046310425,
"eval_runtime": 13.1562,
"eval_samples_per_second": 76.01,
"eval_steps_per_second": 2.432,
"step": 13200
},
{
"epoch": 3.7907309397797166,
"grad_norm": 0.6695854088226235,
"learning_rate": 7.2385530656190785e-06,
"loss": 1.6163,
"step": 13250
},
{
"epoch": 3.8050350450579318,
"grad_norm": 0.7403722736764125,
"learning_rate": 7.219719281982694e-06,
"loss": 1.6223,
"step": 13300
},
{
"epoch": 3.8193391503361465,
"grad_norm": 0.6035992360261074,
"learning_rate": 7.20085015643008e-06,
"loss": 1.6177,
"step": 13350
},
{
"epoch": 3.833643255614361,
"grad_norm": 0.5706930216009009,
"learning_rate": 7.181946076740257e-06,
"loss": 1.6177,
"step": 13400
},
{
"epoch": 3.8479473608925763,
"grad_norm": 0.6806370138912894,
"learning_rate": 7.163007431410583e-06,
"loss": 1.6262,
"step": 13450
},
{
"epoch": 3.862251466170791,
"grad_norm": 0.6234185420830436,
"learning_rate": 7.144034609648779e-06,
"loss": 1.6331,
"step": 13500
},
{
"epoch": 3.876555571449006,
"grad_norm": 0.7677470872346758,
"learning_rate": 7.125028001364918e-06,
"loss": 1.6197,
"step": 13550
},
{
"epoch": 3.8908596767272208,
"grad_norm": 0.6034269149098646,
"learning_rate": 7.105987997163424e-06,
"loss": 1.6266,
"step": 13600
},
{
"epoch": 3.8908596767272208,
"eval_loss": 1.6566662788391113,
"eval_runtime": 13.1267,
"eval_samples_per_second": 76.181,
"eval_steps_per_second": 2.438,
"step": 13600
},
{
"epoch": 3.9051637820054355,
"grad_norm": 0.8302705314191118,
"learning_rate": 7.086914988335039e-06,
"loss": 1.6269,
"step": 13650
},
{
"epoch": 3.91946788728365,
"grad_norm": 0.6389864823326697,
"learning_rate": 7.0678093668487836e-06,
"loss": 1.6323,
"step": 13700
},
{
"epoch": 3.9337719925618653,
"grad_norm": 0.7866068542682738,
"learning_rate": 7.048671525343898e-06,
"loss": 1.6106,
"step": 13750
},
{
"epoch": 3.94807609784008,
"grad_norm": 0.7727876700796928,
"learning_rate": 7.029501857121776e-06,
"loss": 1.6173,
"step": 13800
},
{
"epoch": 3.962380203118295,
"grad_norm": 0.6170837296024364,
"learning_rate": 7.010300756137882e-06,
"loss": 1.6109,
"step": 13850
},
{
"epoch": 3.9766843083965098,
"grad_norm": 0.6847504318409611,
"learning_rate": 6.991068616993655e-06,
"loss": 1.6208,
"step": 13900
},
{
"epoch": 3.9909884136747245,
"grad_norm": 0.7026084916595097,
"learning_rate": 6.971805834928399e-06,
"loss": 1.6147,
"step": 13950
},
{
"epoch": 4.005292518952939,
"grad_norm": 0.5565822000879449,
"learning_rate": 6.952512805811156e-06,
"loss": 1.6238,
"step": 14000
},
{
"epoch": 4.005292518952939,
"eval_loss": 1.6554406881332397,
"eval_runtime": 13.1672,
"eval_samples_per_second": 75.947,
"eval_steps_per_second": 2.43,
"step": 14000
},
{
"epoch": 4.019596624231155,
"grad_norm": 0.7906915205693054,
"learning_rate": 6.933189926132581e-06,
"loss": 1.5887,
"step": 14050
},
{
"epoch": 4.033900729509369,
"grad_norm": 0.6014668104450175,
"learning_rate": 6.913837592996783e-06,
"loss": 1.5824,
"step": 14100
},
{
"epoch": 4.048204834787584,
"grad_norm": 0.6349783814707618,
"learning_rate": 6.894456204113167e-06,
"loss": 1.5993,
"step": 14150
},
{
"epoch": 4.062508940065799,
"grad_norm": 0.7004788731024523,
"learning_rate": 6.875046157788267e-06,
"loss": 1.5856,
"step": 14200
},
{
"epoch": 4.0768130453440135,
"grad_norm": 0.5890601619700472,
"learning_rate": 6.855607852917555e-06,
"loss": 1.5992,
"step": 14250
},
{
"epoch": 4.091117150622228,
"grad_norm": 0.7378157690547319,
"learning_rate": 6.836141688977238e-06,
"loss": 1.5933,
"step": 14300
},
{
"epoch": 4.105421255900444,
"grad_norm": 0.5985785926592089,
"learning_rate": 6.816648066016059e-06,
"loss": 1.6059,
"step": 14350
},
{
"epoch": 4.119725361178658,
"grad_norm": 0.7254510710383179,
"learning_rate": 6.7971273846470696e-06,
"loss": 1.5755,
"step": 14400
},
{
"epoch": 4.119725361178658,
"eval_loss": 1.652251124382019,
"eval_runtime": 13.1229,
"eval_samples_per_second": 76.203,
"eval_steps_per_second": 2.438,
"step": 14400
},
{
"epoch": 4.134029466456873,
"grad_norm": 0.701638747657545,
"learning_rate": 6.777580046039399e-06,
"loss": 1.5878,
"step": 14450
},
{
"epoch": 4.148333571735088,
"grad_norm": 0.6156043257629281,
"learning_rate": 6.758006451910008e-06,
"loss": 1.5888,
"step": 14500
},
{
"epoch": 4.1626376770133025,
"grad_norm": 0.6311348402263556,
"learning_rate": 6.738407004515434e-06,
"loss": 1.5906,
"step": 14550
},
{
"epoch": 4.176941782291518,
"grad_norm": 0.7457367222382035,
"learning_rate": 6.718782106643524e-06,
"loss": 1.5935,
"step": 14600
},
{
"epoch": 4.191245887569733,
"grad_norm": 0.6497146588667823,
"learning_rate": 6.699132161605158e-06,
"loss": 1.5944,
"step": 14650
},
{
"epoch": 4.205549992847947,
"grad_norm": 0.6021866479049087,
"learning_rate": 6.679457573225961e-06,
"loss": 1.5872,
"step": 14700
},
{
"epoch": 4.219854098126162,
"grad_norm": 0.606706719077781,
"learning_rate": 6.659758745837998e-06,
"loss": 1.5892,
"step": 14750
},
{
"epoch": 4.234158203404377,
"grad_norm": 0.796088041146077,
"learning_rate": 6.640036084271477e-06,
"loss": 1.5778,
"step": 14800
},
{
"epoch": 4.234158203404377,
"eval_loss": 1.652685523033142,
"eval_runtime": 13.11,
"eval_samples_per_second": 76.278,
"eval_steps_per_second": 2.441,
"step": 14800
},
{
"epoch": 4.248462308682592,
"grad_norm": 0.5720799505387223,
"learning_rate": 6.620289993846416e-06,
"loss": 1.5853,
"step": 14850
},
{
"epoch": 4.262766413960807,
"grad_norm": 0.638641651597387,
"learning_rate": 6.600520880364318e-06,
"loss": 1.5821,
"step": 14900
},
{
"epoch": 4.277070519239022,
"grad_norm": 0.6826954900918857,
"learning_rate": 6.5807291500998385e-06,
"loss": 1.5926,
"step": 14950
},
{
"epoch": 4.291374624517236,
"grad_norm": 0.7142773516133535,
"learning_rate": 6.560915209792424e-06,
"loss": 1.5756,
"step": 15000
},
{
"epoch": 4.305678729795451,
"grad_norm": 0.6893665096075864,
"learning_rate": 6.541079466637962e-06,
"loss": 1.5898,
"step": 15050
},
{
"epoch": 4.319982835073666,
"grad_norm": 0.6934025133503401,
"learning_rate": 6.52122232828041e-06,
"loss": 1.5906,
"step": 15100
},
{
"epoch": 4.334286940351881,
"grad_norm": 0.5946314963781922,
"learning_rate": 6.501344202803415e-06,
"loss": 1.5876,
"step": 15150
},
{
"epoch": 4.348591045630096,
"grad_norm": 0.6765765096436505,
"learning_rate": 6.4814454987219355e-06,
"loss": 1.5876,
"step": 15200
},
{
"epoch": 4.348591045630096,
"eval_loss": 1.6479697227478027,
"eval_runtime": 13.1313,
"eval_samples_per_second": 76.154,
"eval_steps_per_second": 2.437,
"step": 15200
},
{
"epoch": 4.362895150908311,
"grad_norm": 0.6577195429173401,
"learning_rate": 6.461526624973836e-06,
"loss": 1.5899,
"step": 15250
},
{
"epoch": 4.377199256186525,
"grad_norm": 0.8277675923235659,
"learning_rate": 6.441587990911489e-06,
"loss": 1.5887,
"step": 15300
},
{
"epoch": 4.39150336146474,
"grad_norm": 0.7292088895911856,
"learning_rate": 6.421630006293359e-06,
"loss": 1.5728,
"step": 15350
},
{
"epoch": 4.405807466742955,
"grad_norm": 0.7886180995867164,
"learning_rate": 6.401653081275586e-06,
"loss": 1.5765,
"step": 15400
},
{
"epoch": 4.42011157202117,
"grad_norm": 0.6450755047638138,
"learning_rate": 6.38165762640355e-06,
"loss": 1.5831,
"step": 15450
},
{
"epoch": 4.434415677299385,
"grad_norm": 0.6698809351304791,
"learning_rate": 6.361644052603445e-06,
"loss": 1.5795,
"step": 15500
},
{
"epoch": 4.4487197825776,
"grad_norm": 0.7804327125701724,
"learning_rate": 6.341612771173817e-06,
"loss": 1.5896,
"step": 15550
},
{
"epoch": 4.463023887855814,
"grad_norm": 0.5570965590960153,
"learning_rate": 6.321564193777129e-06,
"loss": 1.5778,
"step": 15600
},
{
"epoch": 4.463023887855814,
"eval_loss": 1.6438684463500977,
"eval_runtime": 13.1716,
"eval_samples_per_second": 75.921,
"eval_steps_per_second": 2.429,
"step": 15600
},
{
"epoch": 4.477327993134029,
"grad_norm": 0.81125064044786,
"learning_rate": 6.301498732431287e-06,
"loss": 1.5898,
"step": 15650
},
{
"epoch": 4.491632098412245,
"grad_norm": 0.7306624425115567,
"learning_rate": 6.281416799501188e-06,
"loss": 1.5775,
"step": 15700
},
{
"epoch": 4.505936203690459,
"grad_norm": 0.7181992700870559,
"learning_rate": 6.261318807690223e-06,
"loss": 1.5844,
"step": 15750
},
{
"epoch": 4.520240308968674,
"grad_norm": 0.7748503653032427,
"learning_rate": 6.24120517003182e-06,
"loss": 1.5807,
"step": 15800
},
{
"epoch": 4.534544414246889,
"grad_norm": 0.692890389924739,
"learning_rate": 6.221076299880939e-06,
"loss": 1.5779,
"step": 15850
},
{
"epoch": 4.548848519525103,
"grad_norm": 0.750802300795554,
"learning_rate": 6.200932610905584e-06,
"loss": 1.5914,
"step": 15900
},
{
"epoch": 4.563152624803319,
"grad_norm": 0.6119826848690993,
"learning_rate": 6.180774517078301e-06,
"loss": 1.5745,
"step": 15950
},
{
"epoch": 4.577456730081534,
"grad_norm": 0.6472801196876264,
"learning_rate": 6.160602432667668e-06,
"loss": 1.5737,
"step": 16000
},
{
"epoch": 4.577456730081534,
"eval_loss": 1.6402463912963867,
"eval_runtime": 13.0995,
"eval_samples_per_second": 76.339,
"eval_steps_per_second": 2.443,
"step": 16000
},
{
"epoch": 4.591760835359748,
"grad_norm": 0.6082312742368647,
"learning_rate": 6.140416772229785e-06,
"loss": 1.5843,
"step": 16050
},
{
"epoch": 4.606064940637963,
"grad_norm": 0.6109383078063736,
"learning_rate": 6.12021795059975e-06,
"loss": 1.5795,
"step": 16100
},
{
"epoch": 4.620369045916178,
"grad_norm": 0.5808974596081403,
"learning_rate": 6.10000638288314e-06,
"loss": 1.5834,
"step": 16150
},
{
"epoch": 4.634673151194393,
"grad_norm": 0.6471974483578133,
"learning_rate": 6.079782484447475e-06,
"loss": 1.5685,
"step": 16200
},
{
"epoch": 4.648977256472608,
"grad_norm": 0.6706099898660832,
"learning_rate": 6.059546670913684e-06,
"loss": 1.582,
"step": 16250
},
{
"epoch": 4.663281361750823,
"grad_norm": 0.634560651660765,
"learning_rate": 6.03929935814756e-06,
"loss": 1.572,
"step": 16300
},
{
"epoch": 4.677585467029037,
"grad_norm": 0.6380102332948996,
"learning_rate": 6.01904096225122e-06,
"loss": 1.5715,
"step": 16350
},
{
"epoch": 4.691889572307252,
"grad_norm": 0.6355595383367377,
"learning_rate": 5.998771899554551e-06,
"loss": 1.5724,
"step": 16400
},
{
"epoch": 4.691889572307252,
"eval_loss": 1.6363039016723633,
"eval_runtime": 13.141,
"eval_samples_per_second": 76.098,
"eval_steps_per_second": 2.435,
"step": 16400
},
{
"epoch": 4.706193677585467,
"grad_norm": 0.6132759482920135,
"learning_rate": 5.978492586606647e-06,
"loss": 1.5769,
"step": 16450
},
{
"epoch": 4.720497782863681,
"grad_norm": 0.6639573383492796,
"learning_rate": 5.958203440167261e-06,
"loss": 1.5671,
"step": 16500
},
{
"epoch": 4.734801888141897,
"grad_norm": 0.5417322612862875,
"learning_rate": 5.93790487719823e-06,
"loss": 1.5755,
"step": 16550
},
{
"epoch": 4.749105993420112,
"grad_norm": 0.6771963599708853,
"learning_rate": 5.917597314854914e-06,
"loss": 1.569,
"step": 16600
},
{
"epoch": 4.763410098698326,
"grad_norm": 0.7200514277275499,
"learning_rate": 5.897281170477614e-06,
"loss": 1.577,
"step": 16650
},
{
"epoch": 4.777714203976541,
"grad_norm": 0.5811950543492076,
"learning_rate": 5.876956861583007e-06,
"loss": 1.5622,
"step": 16700
},
{
"epoch": 4.792018309254756,
"grad_norm": 0.7104533054634696,
"learning_rate": 5.856624805855548e-06,
"loss": 1.579,
"step": 16750
},
{
"epoch": 4.806322414532971,
"grad_norm": 0.655887536338841,
"learning_rate": 5.83628542113891e-06,
"loss": 1.5857,
"step": 16800
},
{
"epoch": 4.806322414532971,
"eval_loss": 1.633513331413269,
"eval_runtime": 13.1532,
"eval_samples_per_second": 76.027,
"eval_steps_per_second": 2.433,
"step": 16800
},
{
"epoch": 4.820626519811186,
"grad_norm": 0.6455948477676112,
"learning_rate": 5.815939125427373e-06,
"loss": 1.5707,
"step": 16850
},
{
"epoch": 4.834930625089401,
"grad_norm": 0.6271361263073431,
"learning_rate": 5.795586336857253e-06,
"loss": 1.563,
"step": 16900
},
{
"epoch": 4.849234730367615,
"grad_norm": 0.7674869603862556,
"learning_rate": 5.775227473698294e-06,
"loss": 1.5779,
"step": 16950
},
{
"epoch": 4.86353883564583,
"grad_norm": 0.637190923985461,
"learning_rate": 5.754862954345081e-06,
"loss": 1.5713,
"step": 17000
},
{
"epoch": 4.877842940924046,
"grad_norm": 0.6651093545639066,
"learning_rate": 5.734493197308442e-06,
"loss": 1.5752,
"step": 17050
},
{
"epoch": 4.89214704620226,
"grad_norm": 0.7868806203083217,
"learning_rate": 5.714118621206843e-06,
"loss": 1.5692,
"step": 17100
},
{
"epoch": 4.906451151480475,
"grad_norm": 0.8067206283766785,
"learning_rate": 5.693739644757781e-06,
"loss": 1.5824,
"step": 17150
},
{
"epoch": 4.92075525675869,
"grad_norm": 0.6156907735488538,
"learning_rate": 5.673356686769194e-06,
"loss": 1.5582,
"step": 17200
},
{
"epoch": 4.92075525675869,
"eval_loss": 1.6318423748016357,
"eval_runtime": 13.1193,
"eval_samples_per_second": 76.224,
"eval_steps_per_second": 2.439,
"step": 17200
},
{
"epoch": 4.935059362036904,
"grad_norm": 0.7876895181186726,
"learning_rate": 5.6529701661308415e-06,
"loss": 1.5689,
"step": 17250
},
{
"epoch": 4.94936346731512,
"grad_norm": 0.6292685628972012,
"learning_rate": 5.632580501805692e-06,
"loss": 1.5739,
"step": 17300
},
{
"epoch": 4.963667572593335,
"grad_norm": 0.7149239853712192,
"learning_rate": 5.612188112821328e-06,
"loss": 1.5797,
"step": 17350
},
{
"epoch": 4.977971677871549,
"grad_norm": 0.834110019775529,
"learning_rate": 5.591793418261326e-06,
"loss": 1.5763,
"step": 17400
},
{
"epoch": 4.992275783149764,
"grad_norm": 0.7309326453697327,
"learning_rate": 5.571396837256637e-06,
"loss": 1.5726,
"step": 17450
},
{
"epoch": 5.006579888427979,
"grad_norm": 0.6697280370753163,
"learning_rate": 5.550998788976988e-06,
"loss": 1.5634,
"step": 17500
},
{
"epoch": 5.020883993706193,
"grad_norm": 0.6174166960990144,
"learning_rate": 5.530599692622257e-06,
"loss": 1.5467,
"step": 17550
},
{
"epoch": 5.035188098984409,
"grad_norm": 0.6241514857747369,
"learning_rate": 5.510199967413862e-06,
"loss": 1.5508,
"step": 17600
},
{
"epoch": 5.035188098984409,
"eval_loss": 1.6312223672866821,
"eval_runtime": 13.1239,
"eval_samples_per_second": 76.197,
"eval_steps_per_second": 2.438,
"step": 17600
},
{
"epoch": 5.049492204262624,
"grad_norm": 0.607902431110512,
"learning_rate": 5.489800032586141e-06,
"loss": 1.5467,
"step": 17650
},
{
"epoch": 5.063796309540838,
"grad_norm": 0.6361444457465594,
"learning_rate": 5.4694003073777446e-06,
"loss": 1.5323,
"step": 17700
},
{
"epoch": 5.078100414819053,
"grad_norm": 0.7502195043466207,
"learning_rate": 5.449001211023014e-06,
"loss": 1.5372,
"step": 17750
},
{
"epoch": 5.092404520097268,
"grad_norm": 0.6959223099372485,
"learning_rate": 5.428603162743365e-06,
"loss": 1.5507,
"step": 17800
},
{
"epoch": 5.106708625375482,
"grad_norm": 0.8282524586776218,
"learning_rate": 5.408206581738677e-06,
"loss": 1.5511,
"step": 17850
},
{
"epoch": 5.121012730653698,
"grad_norm": 0.6152478153542872,
"learning_rate": 5.387811887178673e-06,
"loss": 1.5454,
"step": 17900
},
{
"epoch": 5.135316835931913,
"grad_norm": 0.7178652735808533,
"learning_rate": 5.367419498194309e-06,
"loss": 1.5328,
"step": 17950
},
{
"epoch": 5.149620941210127,
"grad_norm": 0.6199866059077553,
"learning_rate": 5.347029833869161e-06,
"loss": 1.5323,
"step": 18000
},
{
"epoch": 5.149620941210127,
"eval_loss": 1.6294597387313843,
"eval_runtime": 13.0999,
"eval_samples_per_second": 76.336,
"eval_steps_per_second": 2.443,
"step": 18000
},
{
"epoch": 5.163925046488342,
"grad_norm": 0.5994719638332001,
"learning_rate": 5.326643313230806e-06,
"loss": 1.544,
"step": 18050
},
{
"epoch": 5.178229151766557,
"grad_norm": 0.6101948560685385,
"learning_rate": 5.306260355242221e-06,
"loss": 1.5393,
"step": 18100
},
{
"epoch": 5.192533257044772,
"grad_norm": 0.6358629359192572,
"learning_rate": 5.2858813787931605e-06,
"loss": 1.5418,
"step": 18150
},
{
"epoch": 5.206837362322987,
"grad_norm": 0.5992154701681206,
"learning_rate": 5.26550680269156e-06,
"loss": 1.5399,
"step": 18200
},
{
"epoch": 5.221141467601202,
"grad_norm": 0.7386890859479155,
"learning_rate": 5.24513704565492e-06,
"loss": 1.5541,
"step": 18250
},
{
"epoch": 5.235445572879416,
"grad_norm": 0.6887954344111628,
"learning_rate": 5.224772526301709e-06,
"loss": 1.5342,
"step": 18300
},
{
"epoch": 5.249749678157631,
"grad_norm": 0.6961019703931222,
"learning_rate": 5.20441366314275e-06,
"loss": 1.5463,
"step": 18350
},
{
"epoch": 5.2640537834358465,
"grad_norm": 0.7142803422247822,
"learning_rate": 5.184060874572628e-06,
"loss": 1.5415,
"step": 18400
},
{
"epoch": 5.2640537834358465,
"eval_loss": 1.6250876188278198,
"eval_runtime": 13.1209,
"eval_samples_per_second": 76.214,
"eval_steps_per_second": 2.439,
"step": 18400
},
{
"epoch": 5.278357888714061,
"grad_norm": 0.7209586773593375,
"learning_rate": 5.163714578861091e-06,
"loss": 1.5432,
"step": 18450
},
{
"epoch": 5.292661993992276,
"grad_norm": 0.8558312401002982,
"learning_rate": 5.143375194144452e-06,
"loss": 1.5396,
"step": 18500
},
{
"epoch": 5.306966099270491,
"grad_norm": 0.6140496298664848,
"learning_rate": 5.123043138416996e-06,
"loss": 1.5347,
"step": 18550
},
{
"epoch": 5.321270204548705,
"grad_norm": 0.584034537933172,
"learning_rate": 5.102718829522387e-06,
"loss": 1.5455,
"step": 18600
},
{
"epoch": 5.33557430982692,
"grad_norm": 0.7567323256836789,
"learning_rate": 5.082402685145088e-06,
"loss": 1.5373,
"step": 18650
},
{
"epoch": 5.3498784151051355,
"grad_norm": 0.6914309741643679,
"learning_rate": 5.062095122801771e-06,
"loss": 1.5353,
"step": 18700
},
{
"epoch": 5.36418252038335,
"grad_norm": 0.5912773883840152,
"learning_rate": 5.041796559832742e-06,
"loss": 1.5339,
"step": 18750
},
{
"epoch": 5.378486625661565,
"grad_norm": 0.6263016315822758,
"learning_rate": 5.021507413393355e-06,
"loss": 1.5409,
"step": 18800
},
{
"epoch": 5.378486625661565,
"eval_loss": 1.6224277019500732,
"eval_runtime": 13.1785,
"eval_samples_per_second": 75.881,
"eval_steps_per_second": 2.428,
"step": 18800
},
{
"epoch": 5.39279073093978,
"grad_norm": 0.5697288065062571,
"learning_rate": 5.001228100445451e-06,
"loss": 1.5362,
"step": 18850
},
{
"epoch": 5.407094836217994,
"grad_norm": 0.67253268999925,
"learning_rate": 4.9809590377487795e-06,
"loss": 1.5245,
"step": 18900
},
{
"epoch": 5.42139894149621,
"grad_norm": 0.6736080205591396,
"learning_rate": 4.960700641852442e-06,
"loss": 1.5263,
"step": 18950
},
{
"epoch": 5.4357030467744245,
"grad_norm": 0.7729780534104512,
"learning_rate": 4.9404533290863186e-06,
"loss": 1.5297,
"step": 19000
},
{
"epoch": 5.450007152052639,
"grad_norm": 0.6214293118686126,
"learning_rate": 4.920217515552526e-06,
"loss": 1.5397,
"step": 19050
},
{
"epoch": 5.464311257330854,
"grad_norm": 0.6343711973214149,
"learning_rate": 4.8999936171168615e-06,
"loss": 1.5439,
"step": 19100
},
{
"epoch": 5.478615362609069,
"grad_norm": 0.7291330555122011,
"learning_rate": 4.879782049400251e-06,
"loss": 1.5336,
"step": 19150
},
{
"epoch": 5.492919467887283,
"grad_norm": 0.6712685546736542,
"learning_rate": 4.8595832277702175e-06,
"loss": 1.5329,
"step": 19200
},
{
"epoch": 5.492919467887283,
"eval_loss": 1.622052788734436,
"eval_runtime": 13.144,
"eval_samples_per_second": 76.08,
"eval_steps_per_second": 2.435,
"step": 19200
},
{
"epoch": 5.507223573165499,
"grad_norm": 0.5709944496814116,
"learning_rate": 4.839397567332334e-06,
"loss": 1.5404,
"step": 19250
},
{
"epoch": 5.5215276784437135,
"grad_norm": 0.683232233789078,
"learning_rate": 4.8192254829217e-06,
"loss": 1.5422,
"step": 19300
},
{
"epoch": 5.535831783721928,
"grad_norm": 0.6192373192804173,
"learning_rate": 4.799067389094416e-06,
"loss": 1.5443,
"step": 19350
},
{
"epoch": 5.550135889000143,
"grad_norm": 0.7502591192271298,
"learning_rate": 4.7789237001190624e-06,
"loss": 1.5296,
"step": 19400
},
{
"epoch": 5.564439994278358,
"grad_norm": 0.6351610614349129,
"learning_rate": 4.758794829968181e-06,
"loss": 1.5348,
"step": 19450
},
{
"epoch": 5.578744099556573,
"grad_norm": 0.6108826056364483,
"learning_rate": 4.738681192309778e-06,
"loss": 1.5279,
"step": 19500
},
{
"epoch": 5.593048204834788,
"grad_norm": 0.6437182525691378,
"learning_rate": 4.718583200498814e-06,
"loss": 1.5366,
"step": 19550
},
{
"epoch": 5.6073523101130025,
"grad_norm": 0.639815861886649,
"learning_rate": 4.698501267568713e-06,
"loss": 1.5357,
"step": 19600
},
{
"epoch": 5.6073523101130025,
"eval_loss": 1.6180142164230347,
"eval_runtime": 13.085,
"eval_samples_per_second": 76.423,
"eval_steps_per_second": 2.446,
"step": 19600
},
{
"epoch": 5.621656415391217,
"grad_norm": 0.7200835319066389,
"learning_rate": 4.678435806222873e-06,
"loss": 1.5321,
"step": 19650
},
{
"epoch": 5.635960520669432,
"grad_norm": 0.783772141557461,
"learning_rate": 4.658387228826185e-06,
"loss": 1.5222,
"step": 19700
},
{
"epoch": 5.6502646259476474,
"grad_norm": 0.6132596760513117,
"learning_rate": 4.638355947396557e-06,
"loss": 1.5297,
"step": 19750
},
{
"epoch": 5.664568731225862,
"grad_norm": 0.6283378075518518,
"learning_rate": 4.61834237359645e-06,
"loss": 1.5295,
"step": 19800
},
{
"epoch": 5.678872836504077,
"grad_norm": 0.6326692578607431,
"learning_rate": 4.598346918724417e-06,
"loss": 1.5351,
"step": 19850
},
{
"epoch": 5.6931769417822915,
"grad_norm": 0.5940063235489941,
"learning_rate": 4.578369993706643e-06,
"loss": 1.5308,
"step": 19900
},
{
"epoch": 5.707481047060506,
"grad_norm": 0.6349874592624309,
"learning_rate": 4.5584120090885125e-06,
"loss": 1.5214,
"step": 19950
},
{
"epoch": 5.721785152338721,
"grad_norm": 0.7828872685863243,
"learning_rate": 4.538473375026164e-06,
"loss": 1.5428,
"step": 20000
},
{
"epoch": 5.721785152338721,
"eval_loss": 1.6172399520874023,
"eval_runtime": 13.1288,
"eval_samples_per_second": 76.168,
"eval_steps_per_second": 2.437,
"step": 20000
},
{
"epoch": 5.736089257616936,
"grad_norm": 0.6299009394488376,
"learning_rate": 4.518554501278064e-06,
"loss": 1.5295,
"step": 20050
},
{
"epoch": 5.750393362895151,
"grad_norm": 0.6258430789114957,
"learning_rate": 4.498655797196586e-06,
"loss": 1.5329,
"step": 20100
},
{
"epoch": 5.764697468173366,
"grad_norm": 0.696293581732556,
"learning_rate": 4.478777671719593e-06,
"loss": 1.5413,
"step": 20150
},
{
"epoch": 5.7790015734515805,
"grad_norm": 0.7124246186685519,
"learning_rate": 4.458920533362039e-06,
"loss": 1.5228,
"step": 20200
},
{
"epoch": 5.793305678729795,
"grad_norm": 0.5760090289969003,
"learning_rate": 4.439084790207577e-06,
"loss": 1.5353,
"step": 20250
},
{
"epoch": 5.80760978400801,
"grad_norm": 0.6449974415534122,
"learning_rate": 4.419270849900164e-06,
"loss": 1.5357,
"step": 20300
},
{
"epoch": 5.821913889286225,
"grad_norm": 0.6282375399916217,
"learning_rate": 4.399479119635683e-06,
"loss": 1.5321,
"step": 20350
},
{
"epoch": 5.83621799456444,
"grad_norm": 0.5919862101266721,
"learning_rate": 4.3797100061535856e-06,
"loss": 1.526,
"step": 20400
},
{
"epoch": 5.83621799456444,
"eval_loss": 1.6131237745285034,
"eval_runtime": 13.1139,
"eval_samples_per_second": 76.255,
"eval_steps_per_second": 2.44,
"step": 20400
},
{
"epoch": 5.850522099842655,
"grad_norm": 0.6029658262074238,
"learning_rate": 4.359963915728523e-06,
"loss": 1.5247,
"step": 20450
},
{
"epoch": 5.8648262051208695,
"grad_norm": 0.670869434044021,
"learning_rate": 4.340241254162004e-06,
"loss": 1.525,
"step": 20500
},
{
"epoch": 5.879130310399084,
"grad_norm": 0.6367180654906258,
"learning_rate": 4.320542426774042e-06,
"loss": 1.5269,
"step": 20550
},
{
"epoch": 5.8934344156773,
"grad_norm": 0.5920409715922154,
"learning_rate": 4.300867838394843e-06,
"loss": 1.5241,
"step": 20600
},
{
"epoch": 5.907738520955514,
"grad_norm": 0.5951037902934461,
"learning_rate": 4.281217893356478e-06,
"loss": 1.5404,
"step": 20650
},
{
"epoch": 5.922042626233729,
"grad_norm": 0.7114294289873735,
"learning_rate": 4.261592995484567e-06,
"loss": 1.5329,
"step": 20700
},
{
"epoch": 5.936346731511944,
"grad_norm": 0.5783828576596121,
"learning_rate": 4.241993548089994e-06,
"loss": 1.533,
"step": 20750
},
{
"epoch": 5.9506508367901585,
"grad_norm": 0.6407329854514554,
"learning_rate": 4.2224199539606026e-06,
"loss": 1.5145,
"step": 20800
},
{
"epoch": 5.9506508367901585,
"eval_loss": 1.6121162176132202,
"eval_runtime": 13.1147,
"eval_samples_per_second": 76.25,
"eval_steps_per_second": 2.44,
"step": 20800
},
{
"epoch": 5.964954942068374,
"grad_norm": 0.634888254231562,
"learning_rate": 4.202872615352932e-06,
"loss": 1.5212,
"step": 20850
},
{
"epoch": 5.979259047346589,
"grad_norm": 0.6229293024525429,
"learning_rate": 4.183351933983942e-06,
"loss": 1.5226,
"step": 20900
},
{
"epoch": 5.993563152624803,
"grad_norm": 0.8314381666700691,
"learning_rate": 4.163858311022765e-06,
"loss": 1.5387,
"step": 20950
},
{
"epoch": 6.007867257903018,
"grad_norm": 0.59693291322645,
"learning_rate": 4.144392147082447e-06,
"loss": 1.5043,
"step": 21000
},
{
"epoch": 6.022171363181233,
"grad_norm": 0.665194604166561,
"learning_rate": 4.124953842211733e-06,
"loss": 1.5002,
"step": 21050
},
{
"epoch": 6.0364754684594475,
"grad_norm": 0.6456678113415395,
"learning_rate": 4.105543795886834e-06,
"loss": 1.5004,
"step": 21100
},
{
"epoch": 6.050779573737663,
"grad_norm": 0.642020830639423,
"learning_rate": 4.086162407003219e-06,
"loss": 1.499,
"step": 21150
},
{
"epoch": 6.065083679015878,
"grad_norm": 0.6641186183494755,
"learning_rate": 4.066810073867421e-06,
"loss": 1.4973,
"step": 21200
},
{
"epoch": 6.065083679015878,
"eval_loss": 1.6121020317077637,
"eval_runtime": 13.0899,
"eval_samples_per_second": 76.395,
"eval_steps_per_second": 2.445,
"step": 21200
},
{
"epoch": 6.079387784294092,
"grad_norm": 0.634272316306304,
"learning_rate": 4.047487194188846e-06,
"loss": 1.5132,
"step": 21250
},
{
"epoch": 6.093691889572307,
"grad_norm": 0.6748754827417714,
"learning_rate": 4.028194165071603e-06,
"loss": 1.5064,
"step": 21300
},
{
"epoch": 6.107995994850522,
"grad_norm": 0.7061607465519819,
"learning_rate": 4.008931383006345e-06,
"loss": 1.505,
"step": 21350
},
{
"epoch": 6.122300100128737,
"grad_norm": 0.662154780004609,
"learning_rate": 3.9896992438621195e-06,
"loss": 1.5005,
"step": 21400
},
{
"epoch": 6.136604205406952,
"grad_norm": 0.6380113889018068,
"learning_rate": 3.9704981428782255e-06,
"loss": 1.5061,
"step": 21450
},
{
"epoch": 6.150908310685167,
"grad_norm": 0.6382768302952283,
"learning_rate": 3.951328474656103e-06,
"loss": 1.4931,
"step": 21500
},
{
"epoch": 6.165212415963381,
"grad_norm": 0.6767629241771801,
"learning_rate": 3.932190633151217e-06,
"loss": 1.5037,
"step": 21550
},
{
"epoch": 6.179516521241596,
"grad_norm": 0.6282216484213683,
"learning_rate": 3.913085011664961e-06,
"loss": 1.5036,
"step": 21600
},
{
"epoch": 6.179516521241596,
"eval_loss": 1.6118288040161133,
"eval_runtime": 13.1481,
"eval_samples_per_second": 76.057,
"eval_steps_per_second": 2.434,
"step": 21600
},
{
"epoch": 6.193820626519811,
"grad_norm": 0.6004617238206564,
"learning_rate": 3.894012002836578e-06,
"loss": 1.5117,
"step": 21650
},
{
"epoch": 6.208124731798026,
"grad_norm": 0.7154680728784966,
"learning_rate": 3.874971998635083e-06,
"loss": 1.5095,
"step": 21700
},
{
"epoch": 6.222428837076241,
"grad_norm": 0.6725566004485558,
"learning_rate": 3.855965390351222e-06,
"loss": 1.4909,
"step": 21750
},
{
"epoch": 6.236732942354456,
"grad_norm": 0.5965877047430266,
"learning_rate": 3.836992568589417e-06,
"loss": 1.4975,
"step": 21800
},
{
"epoch": 6.25103704763267,
"grad_norm": 0.6027834656468215,
"learning_rate": 3.818053923259745e-06,
"loss": 1.5007,
"step": 21850
},
{
"epoch": 6.265341152910885,
"grad_norm": 0.6663203163783424,
"learning_rate": 3.7991498435699213e-06,
"loss": 1.5051,
"step": 21900
},
{
"epoch": 6.279645258189101,
"grad_norm": 0.7125353969852749,
"learning_rate": 3.780280718017307e-06,
"loss": 1.5012,
"step": 21950
},
{
"epoch": 6.293949363467315,
"grad_norm": 0.6482680068197426,
"learning_rate": 3.7614469343809208e-06,
"loss": 1.5049,
"step": 22000
},
{
"epoch": 6.293949363467315,
"eval_loss": 1.6089048385620117,
"eval_runtime": 13.1237,
"eval_samples_per_second": 76.198,
"eval_steps_per_second": 2.438,
"step": 22000
},
{
"epoch": 6.30825346874553,
"grad_norm": 0.7881816467240778,
"learning_rate": 3.7426488797134706e-06,
"loss": 1.5095,
"step": 22050
},
{
"epoch": 6.322557574023745,
"grad_norm": 0.6541787463398775,
"learning_rate": 3.7238869403333965e-06,
"loss": 1.504,
"step": 22100
},
{
"epoch": 6.336861679301959,
"grad_norm": 0.664079330402415,
"learning_rate": 3.7051615018169363e-06,
"loss": 1.5066,
"step": 22150
},
{
"epoch": 6.351165784580174,
"grad_norm": 0.5916075057540389,
"learning_rate": 3.686472948990198e-06,
"loss": 1.5086,
"step": 22200
},
{
"epoch": 6.36546988985839,
"grad_norm": 0.7325464747774915,
"learning_rate": 3.6678216659212546e-06,
"loss": 1.4992,
"step": 22250
},
{
"epoch": 6.379773995136604,
"grad_norm": 0.7007083689793716,
"learning_rate": 3.6492080359122487e-06,
"loss": 1.5,
"step": 22300
},
{
"epoch": 6.394078100414819,
"grad_norm": 0.6739401831383736,
"learning_rate": 3.630632441491512e-06,
"loss": 1.4995,
"step": 22350
},
{
"epoch": 6.408382205693034,
"grad_norm": 0.6863089101260594,
"learning_rate": 3.6120952644057104e-06,
"loss": 1.5017,
"step": 22400
},
{
"epoch": 6.408382205693034,
"eval_loss": 1.606228232383728,
"eval_runtime": 13.1397,
"eval_samples_per_second": 76.105,
"eval_steps_per_second": 2.435,
"step": 22400
},
{
"epoch": 6.422686310971248,
"grad_norm": 0.6147754073523626,
"learning_rate": 3.593596885611997e-06,
"loss": 1.4926,
"step": 22450
},
{
"epoch": 6.436990416249464,
"grad_norm": 0.5871499975242569,
"learning_rate": 3.575137685270182e-06,
"loss": 1.4983,
"step": 22500
},
{
"epoch": 6.451294521527679,
"grad_norm": 0.5786188568209956,
"learning_rate": 3.556718042734915e-06,
"loss": 1.4949,
"step": 22550
},
{
"epoch": 6.465598626805893,
"grad_norm": 0.6071462043583132,
"learning_rate": 3.538338336547902e-06,
"loss": 1.4987,
"step": 22600
},
{
"epoch": 6.479902732084108,
"grad_norm": 0.5785217797393283,
"learning_rate": 3.5199989444301082e-06,
"loss": 1.4931,
"step": 22650
},
{
"epoch": 6.494206837362323,
"grad_norm": 0.623076022970078,
"learning_rate": 3.5017002432740164e-06,
"loss": 1.4981,
"step": 22700
},
{
"epoch": 6.508510942640537,
"grad_norm": 0.73033655866744,
"learning_rate": 3.483442609135862e-06,
"loss": 1.5008,
"step": 22750
},
{
"epoch": 6.522815047918753,
"grad_norm": 0.5934926417919788,
"learning_rate": 3.4652264172279153e-06,
"loss": 1.4881,
"step": 22800
},
{
"epoch": 6.522815047918753,
"eval_loss": 1.604162335395813,
"eval_runtime": 13.1475,
"eval_samples_per_second": 76.06,
"eval_steps_per_second": 2.434,
"step": 22800
},
{
"epoch": 6.537119153196968,
"grad_norm": 0.6069986350354387,
"learning_rate": 3.4470520419107667e-06,
"loss": 1.4879,
"step": 22850
},
{
"epoch": 6.551423258475182,
"grad_norm": 0.7399652835966712,
"learning_rate": 3.4289198566856396e-06,
"loss": 1.4942,
"step": 22900
},
{
"epoch": 6.565727363753397,
"grad_norm": 0.6114058208337956,
"learning_rate": 3.4108302341867045e-06,
"loss": 1.4961,
"step": 22950
},
{
"epoch": 6.580031469031612,
"grad_norm": 0.6007896515222266,
"learning_rate": 3.392783546173426e-06,
"loss": 1.4929,
"step": 23000
},
{
"epoch": 6.594335574309827,
"grad_norm": 0.7491177037600624,
"learning_rate": 3.374780163522929e-06,
"loss": 1.4955,
"step": 23050
},
{
"epoch": 6.608639679588042,
"grad_norm": 0.6917907740940776,
"learning_rate": 3.356820456222362e-06,
"loss": 1.5053,
"step": 23100
},
{
"epoch": 6.622943784866257,
"grad_norm": 0.5797836988455244,
"learning_rate": 3.338904793361314e-06,
"loss": 1.4974,
"step": 23150
},
{
"epoch": 6.637247890144471,
"grad_norm": 0.6589102327372058,
"learning_rate": 3.321033543124202e-06,
"loss": 1.4969,
"step": 23200
},
{
"epoch": 6.637247890144471,
"eval_loss": 1.6024819612503052,
"eval_runtime": 13.1262,
"eval_samples_per_second": 76.183,
"eval_steps_per_second": 2.438,
"step": 23200
},
{
"epoch": 6.651551995422686,
"grad_norm": 0.67042895269073,
"learning_rate": 3.3032070727827353e-06,
"loss": 1.4931,
"step": 23250
},
{
"epoch": 6.665856100700902,
"grad_norm": 0.6465762171859113,
"learning_rate": 3.285425748688341e-06,
"loss": 1.5047,
"step": 23300
},
{
"epoch": 6.680160205979116,
"grad_norm": 0.5928906211207966,
"learning_rate": 3.2676899362646563e-06,
"loss": 1.5009,
"step": 23350
},
{
"epoch": 6.694464311257331,
"grad_norm": 0.6204319354437976,
"learning_rate": 3.2500000000000015e-06,
"loss": 1.4904,
"step": 23400
},
{
"epoch": 6.708768416535546,
"grad_norm": 0.6370248234668529,
"learning_rate": 3.232356303439897e-06,
"loss": 1.4962,
"step": 23450
},
{
"epoch": 6.72307252181376,
"grad_norm": 0.5797494709987538,
"learning_rate": 3.2147592091795992e-06,
"loss": 1.4938,
"step": 23500
},
{
"epoch": 6.737376627091975,
"grad_norm": 0.5800297263686035,
"learning_rate": 3.1972090788566325e-06,
"loss": 1.4852,
"step": 23550
},
{
"epoch": 6.751680732370191,
"grad_norm": 0.6400211162161145,
"learning_rate": 3.1797062731433764e-06,
"loss": 1.4978,
"step": 23600
},
{
"epoch": 6.751680732370191,
"eval_loss": 1.6009289026260376,
"eval_runtime": 13.1205,
"eval_samples_per_second": 76.216,
"eval_steps_per_second": 2.439,
"step": 23600
},
{
"epoch": 6.765984837648405,
"grad_norm": 0.6047950853143369,
"learning_rate": 3.1622511517396297e-06,
"loss": 1.4931,
"step": 23650
},
{
"epoch": 6.78028894292662,
"grad_norm": 0.6976178519904986,
"learning_rate": 3.144844073365247e-06,
"loss": 1.4996,
"step": 23700
},
{
"epoch": 6.794593048204835,
"grad_norm": 0.6200936341605483,
"learning_rate": 3.127485395752739e-06,
"loss": 1.4874,
"step": 23750
},
{
"epoch": 6.808897153483049,
"grad_norm": 0.6678212868779091,
"learning_rate": 3.1101754756399427e-06,
"loss": 1.4951,
"step": 23800
},
{
"epoch": 6.823201258761264,
"grad_norm": 0.6777775649765228,
"learning_rate": 3.0929146687626756e-06,
"loss": 1.4936,
"step": 23850
},
{
"epoch": 6.83750536403948,
"grad_norm": 0.5932024233832688,
"learning_rate": 3.0757033298474274e-06,
"loss": 1.4908,
"step": 23900
},
{
"epoch": 6.851809469317694,
"grad_norm": 0.6344747435348402,
"learning_rate": 3.0585418126040832e-06,
"loss": 1.5015,
"step": 23950
},
{
"epoch": 6.866113574595909,
"grad_norm": 0.629407558903598,
"learning_rate": 3.0414304697186313e-06,
"loss": 1.4797,
"step": 24000
},
{
"epoch": 6.866113574595909,
"eval_loss": 1.5976924896240234,
"eval_runtime": 13.1566,
"eval_samples_per_second": 76.008,
"eval_steps_per_second": 2.432,
"step": 24000
}
],
"logging_steps": 50,
"max_steps": 34950,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 800,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.1342699751735296e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}