|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.5003700962250185, |
|
"eval_steps": 507, |
|
"global_step": 507, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 92.56075816528, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 1.6849, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"eval_loss": 1.7294352054595947, |
|
"eval_runtime": 1884.169, |
|
"eval_samples_per_second": 0.701, |
|
"eval_steps_per_second": 0.078, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 62.47252473867764, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 1.4808, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 86.16163524072095, |
|
"learning_rate": 1.5e-06, |
|
"loss": 1.622, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 60.0955549964575, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.5835, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 29.984043062746373, |
|
"learning_rate": 2.5e-06, |
|
"loss": 1.5242, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 301.1618364383654, |
|
"learning_rate": 3e-06, |
|
"loss": 1.2748, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 109.29021464248359, |
|
"learning_rate": 3.5e-06, |
|
"loss": 1.1112, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 10.444328634170152, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.9269, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 6.523925938807769, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.8644, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 17.340636604619814, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8207, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 44.61126816135743, |
|
"learning_rate": 4.9999969645113385e-06, |
|
"loss": 1.1849, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 4.3252416457498315, |
|
"learning_rate": 4.999987858052726e-06, |
|
"loss": 0.885, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.9355463597411195, |
|
"learning_rate": 4.999972680646276e-06, |
|
"loss": 0.8329, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.0811257726357524, |
|
"learning_rate": 4.999951432328846e-06, |
|
"loss": 0.7869, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.9698959240524336, |
|
"learning_rate": 4.999924113152034e-06, |
|
"loss": 0.7948, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.6266575755898287, |
|
"learning_rate": 4.999890723182183e-06, |
|
"loss": 0.8005, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.419625838626686, |
|
"learning_rate": 4.999851262500375e-06, |
|
"loss": 0.7565, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.4352187159776242, |
|
"learning_rate": 4.999805731202437e-06, |
|
"loss": 0.7602, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.2355497382357647, |
|
"learning_rate": 4.999754129398938e-06, |
|
"loss": 0.7564, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.2700121185974393, |
|
"learning_rate": 4.999696457215186e-06, |
|
"loss": 0.8125, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1702941851369728, |
|
"learning_rate": 4.9996327147912315e-06, |
|
"loss": 0.7382, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1798174730304447, |
|
"learning_rate": 4.999562902281866e-06, |
|
"loss": 0.7742, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1364491950081737, |
|
"learning_rate": 4.999487019856623e-06, |
|
"loss": 0.7282, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1131547945883817, |
|
"learning_rate": 4.999405067699773e-06, |
|
"loss": 0.7475, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1506149455536359, |
|
"learning_rate": 4.9993170460103295e-06, |
|
"loss": 0.7072, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.2889319229905287, |
|
"learning_rate": 4.999222955002041e-06, |
|
"loss": 0.8064, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.059562137107165, |
|
"learning_rate": 4.999122794903399e-06, |
|
"loss": 0.7147, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.149120672076531, |
|
"learning_rate": 4.999016565957633e-06, |
|
"loss": 0.7792, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.1596428841302342, |
|
"learning_rate": 4.998904268422705e-06, |
|
"loss": 0.7586, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.980994033140509, |
|
"learning_rate": 4.998785902571319e-06, |
|
"loss": 0.7721, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.9969455208276253, |
|
"learning_rate": 4.998661468690914e-06, |
|
"loss": 0.7482, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.2088730805217034, |
|
"learning_rate": 4.998530967083664e-06, |
|
"loss": 0.7687, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.1772131571685314, |
|
"learning_rate": 4.998394398066477e-06, |
|
"loss": 0.7336, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.0095424896414942, |
|
"learning_rate": 4.998251761970997e-06, |
|
"loss": 0.7454, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.058399487048206, |
|
"learning_rate": 4.998103059143599e-06, |
|
"loss": 0.7424, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0553895642417541, |
|
"learning_rate": 4.997948289945392e-06, |
|
"loss": 0.7034, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.9737916869473869, |
|
"learning_rate": 4.997787454752217e-06, |
|
"loss": 0.7225, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0497906961287207, |
|
"learning_rate": 4.997620553954645e-06, |
|
"loss": 0.7545, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0351771937970027, |
|
"learning_rate": 4.9974475879579745e-06, |
|
"loss": 0.6999, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.025178950396183, |
|
"learning_rate": 4.9972685571822355e-06, |
|
"loss": 0.6836, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.8990499683406185, |
|
"learning_rate": 4.997083462062185e-06, |
|
"loss": 0.749, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.1934033398652506, |
|
"learning_rate": 4.996892303047306e-06, |
|
"loss": 0.7118, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0093747627659915, |
|
"learning_rate": 4.9966950806018075e-06, |
|
"loss": 0.6931, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.1259320175035152, |
|
"learning_rate": 4.996491795204623e-06, |
|
"loss": 0.7011, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.9340587879454492, |
|
"learning_rate": 4.996282447349408e-06, |
|
"loss": 0.7031, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.093497345220761, |
|
"learning_rate": 4.996067037544542e-06, |
|
"loss": 0.6984, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.9253425670099831, |
|
"learning_rate": 4.9958455663131235e-06, |
|
"loss": 0.7211, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.1321089016179777, |
|
"learning_rate": 4.99561803419297e-06, |
|
"loss": 0.6809, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.9225335387284348, |
|
"learning_rate": 4.995384441736622e-06, |
|
"loss": 0.6926, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.03537386787448, |
|
"learning_rate": 4.9951447895113305e-06, |
|
"loss": 0.6868, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.9586247752600078, |
|
"learning_rate": 4.994899078099064e-06, |
|
"loss": 0.6973, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.164879075983742, |
|
"learning_rate": 4.994647308096509e-06, |
|
"loss": 0.6976, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.007151099232745, |
|
"learning_rate": 4.994389480115059e-06, |
|
"loss": 0.6728, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.194002279613801, |
|
"learning_rate": 4.9941255947808226e-06, |
|
"loss": 0.7386, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.2903234642197774, |
|
"learning_rate": 4.993855652734616e-06, |
|
"loss": 0.6884, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9696246122793456, |
|
"learning_rate": 4.993579654631963e-06, |
|
"loss": 0.7175, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.1155328763962573, |
|
"learning_rate": 4.993297601143095e-06, |
|
"loss": 0.7146, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9599001918058567, |
|
"learning_rate": 4.993009492952951e-06, |
|
"loss": 0.7054, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9180819748743216, |
|
"learning_rate": 4.992715330761167e-06, |
|
"loss": 0.6984, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9088984413752035, |
|
"learning_rate": 4.992415115282085e-06, |
|
"loss": 0.7001, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.1141716959997088, |
|
"learning_rate": 4.992108847244746e-06, |
|
"loss": 0.6756, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.8636787369748702, |
|
"learning_rate": 4.991796527392888e-06, |
|
"loss": 0.7548, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.0780460703120494, |
|
"learning_rate": 4.991478156484945e-06, |
|
"loss": 0.7039, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9164733455688905, |
|
"learning_rate": 4.9911537352940485e-06, |
|
"loss": 0.7385, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.0268015729747888, |
|
"learning_rate": 4.990823264608017e-06, |
|
"loss": 0.7283, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.9924581532223246, |
|
"learning_rate": 4.990486745229364e-06, |
|
"loss": 0.7315, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.953526608505078, |
|
"learning_rate": 4.99014417797529e-06, |
|
"loss": 0.7194, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.9721089885130799, |
|
"learning_rate": 4.9897955636776824e-06, |
|
"loss": 0.6786, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.095580212055755, |
|
"learning_rate": 4.989440903183112e-06, |
|
"loss": 0.6547, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.0176000454229972, |
|
"learning_rate": 4.989080197352834e-06, |
|
"loss": 0.7024, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.9764964284876755, |
|
"learning_rate": 4.988713447062784e-06, |
|
"loss": 0.6897, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.9720697960414958, |
|
"learning_rate": 4.9883406532035735e-06, |
|
"loss": 0.7368, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.8924650786548242, |
|
"learning_rate": 4.987961816680493e-06, |
|
"loss": 0.6986, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.0583230570989346, |
|
"learning_rate": 4.987576938413504e-06, |
|
"loss": 0.6488, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.8532905196520026, |
|
"learning_rate": 4.9871860193372424e-06, |
|
"loss": 0.6545, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.110205997163892, |
|
"learning_rate": 4.9867890604010125e-06, |
|
"loss": 0.7296, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.8585453329590255, |
|
"learning_rate": 4.986386062568787e-06, |
|
"loss": 0.6478, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.072721980278167, |
|
"learning_rate": 4.9859770268192e-06, |
|
"loss": 0.683, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9114010594463168, |
|
"learning_rate": 4.985561954145552e-06, |
|
"loss": 0.7278, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9407855861796833, |
|
"learning_rate": 4.985140845555799e-06, |
|
"loss": 0.6767, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9716353726299576, |
|
"learning_rate": 4.984713702072561e-06, |
|
"loss": 0.7058, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9289919706551977, |
|
"learning_rate": 4.984280524733107e-06, |
|
"loss": 0.7465, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9719373864834964, |
|
"learning_rate": 4.9838413145893615e-06, |
|
"loss": 0.7173, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.8656006720690229, |
|
"learning_rate": 4.983396072707898e-06, |
|
"loss": 0.7138, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9405397793417023, |
|
"learning_rate": 4.982944800169939e-06, |
|
"loss": 0.7172, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.8968931033898299, |
|
"learning_rate": 4.982487498071349e-06, |
|
"loss": 0.6959, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.969922689772122, |
|
"learning_rate": 4.982024167522638e-06, |
|
"loss": 0.706, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8691971891272707, |
|
"learning_rate": 4.981554809648952e-06, |
|
"loss": 0.7156, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9182420056058627, |
|
"learning_rate": 4.981079425590078e-06, |
|
"loss": 0.6694, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9014669660783321, |
|
"learning_rate": 4.980598016500431e-06, |
|
"loss": 0.6842, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8598791805359901, |
|
"learning_rate": 4.980110583549062e-06, |
|
"loss": 0.6707, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8569795805490359, |
|
"learning_rate": 4.97961712791965e-06, |
|
"loss": 0.6637, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8939370143328529, |
|
"learning_rate": 4.979117650810495e-06, |
|
"loss": 0.6927, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9152075194521869, |
|
"learning_rate": 4.978612153434527e-06, |
|
"loss": 0.6803, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9298563182624136, |
|
"learning_rate": 4.978100637019287e-06, |
|
"loss": 0.6776, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9132188893536833, |
|
"learning_rate": 4.97758310280694e-06, |
|
"loss": 0.7032, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9574161232781394, |
|
"learning_rate": 4.97705955205426e-06, |
|
"loss": 0.6731, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.8327779171567138, |
|
"learning_rate": 4.976529986032632e-06, |
|
"loss": 0.6754, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.1175664554073028, |
|
"learning_rate": 4.975994406028051e-06, |
|
"loss": 0.6774, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9122529617157105, |
|
"learning_rate": 4.9754528133411144e-06, |
|
"loss": 0.6905, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9495918841010983, |
|
"learning_rate": 4.97490520928702e-06, |
|
"loss": 0.6683, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9199809745368361, |
|
"learning_rate": 4.9743515951955655e-06, |
|
"loss": 0.6512, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9371967382154123, |
|
"learning_rate": 4.973791972411142e-06, |
|
"loss": 0.7004, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.0249455893660548, |
|
"learning_rate": 4.973226342292732e-06, |
|
"loss": 0.6902, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.007634999853075, |
|
"learning_rate": 4.972654706213906e-06, |
|
"loss": 0.712, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9151412270511807, |
|
"learning_rate": 4.9720770655628216e-06, |
|
"loss": 0.689, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9263042922044208, |
|
"learning_rate": 4.971493421742215e-06, |
|
"loss": 0.7188, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9875225370080567, |
|
"learning_rate": 4.970903776169403e-06, |
|
"loss": 0.7313, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9329951904410932, |
|
"learning_rate": 4.970308130276273e-06, |
|
"loss": 0.6473, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.952122809501081, |
|
"learning_rate": 4.969706485509287e-06, |
|
"loss": 0.6866, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.8675487151632736, |
|
"learning_rate": 4.969098843329475e-06, |
|
"loss": 0.6621, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.0124220925484857, |
|
"learning_rate": 4.968485205212429e-06, |
|
"loss": 0.696, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.8692279553117837, |
|
"learning_rate": 4.967865572648302e-06, |
|
"loss": 0.6503, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.8684082104461566, |
|
"learning_rate": 4.967239947141803e-06, |
|
"loss": 0.717, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9363131782505817, |
|
"learning_rate": 4.966608330212198e-06, |
|
"loss": 0.67, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9483280670290893, |
|
"learning_rate": 4.965970723393297e-06, |
|
"loss": 0.6645, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9645550818323707, |
|
"learning_rate": 4.965327128233462e-06, |
|
"loss": 0.6402, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9020494921501199, |
|
"learning_rate": 4.96467754629559e-06, |
|
"loss": 0.6848, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9273157548895928, |
|
"learning_rate": 4.964021979157122e-06, |
|
"loss": 0.6888, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0083484650433756, |
|
"learning_rate": 4.963360428410031e-06, |
|
"loss": 0.7223, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9556311862868435, |
|
"learning_rate": 4.96269289566082e-06, |
|
"loss": 0.6924, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.8565613256126575, |
|
"learning_rate": 4.962019382530521e-06, |
|
"loss": 0.6551, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0082880819703033, |
|
"learning_rate": 4.961339890654685e-06, |
|
"loss": 0.6208, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0159617953712363, |
|
"learning_rate": 4.9606544216833865e-06, |
|
"loss": 0.7223, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9941249295005459, |
|
"learning_rate": 4.95996297728121e-06, |
|
"loss": 0.6853, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.980987656479997, |
|
"learning_rate": 4.959265559127253e-06, |
|
"loss": 0.6531, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.8906826655489122, |
|
"learning_rate": 4.958562168915121e-06, |
|
"loss": 0.6822, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9447717033103807, |
|
"learning_rate": 4.95785280835292e-06, |
|
"loss": 0.6735, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9322226016156214, |
|
"learning_rate": 4.957137479163253e-06, |
|
"loss": 0.6853, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9016538169637961, |
|
"learning_rate": 4.9564161830832214e-06, |
|
"loss": 0.6546, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9425807696655617, |
|
"learning_rate": 4.955688921864411e-06, |
|
"loss": 0.6874, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.939091235820591, |
|
"learning_rate": 4.954955697272899e-06, |
|
"loss": 0.7002, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.8787416221103219, |
|
"learning_rate": 4.954216511089242e-06, |
|
"loss": 0.7502, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9742122954687594, |
|
"learning_rate": 4.95347136510847e-06, |
|
"loss": 0.6657, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.8801095585358768, |
|
"learning_rate": 4.95272026114009e-06, |
|
"loss": 0.6555, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9481198118358727, |
|
"learning_rate": 4.9519632010080765e-06, |
|
"loss": 0.6737, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9764499477724559, |
|
"learning_rate": 4.951200186550868e-06, |
|
"loss": 0.6616, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9677532996854814, |
|
"learning_rate": 4.9504312196213596e-06, |
|
"loss": 0.6604, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9262840329326826, |
|
"learning_rate": 4.949656302086907e-06, |
|
"loss": 0.6634, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9913169871400999, |
|
"learning_rate": 4.94887543582931e-06, |
|
"loss": 0.6834, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.037107833577886, |
|
"learning_rate": 4.948088622744819e-06, |
|
"loss": 0.7047, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.0270598791002519, |
|
"learning_rate": 4.947295864744121e-06, |
|
"loss": 0.6637, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9780188865838979, |
|
"learning_rate": 4.9464971637523465e-06, |
|
"loss": 0.7196, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9564318039750372, |
|
"learning_rate": 4.94569252170905e-06, |
|
"loss": 0.661, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.070366815703729, |
|
"learning_rate": 4.944881940568219e-06, |
|
"loss": 0.7288, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.951311961745425, |
|
"learning_rate": 4.944065422298262e-06, |
|
"loss": 0.6846, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9924454965931206, |
|
"learning_rate": 4.943242968882002e-06, |
|
"loss": 0.6382, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.899515518807678, |
|
"learning_rate": 4.942414582316679e-06, |
|
"loss": 0.6949, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9872873524132176, |
|
"learning_rate": 4.94158026461394e-06, |
|
"loss": 0.6765, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.8805014666801008, |
|
"learning_rate": 4.9407400177998335e-06, |
|
"loss": 0.6288, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9546823658223654, |
|
"learning_rate": 4.939893843914808e-06, |
|
"loss": 0.6749, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9963934911162335, |
|
"learning_rate": 4.939041745013703e-06, |
|
"loss": 0.6433, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9708720605788677, |
|
"learning_rate": 4.93818372316575e-06, |
|
"loss": 0.6774, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9473055224950518, |
|
"learning_rate": 4.937319780454559e-06, |
|
"loss": 0.685, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9819312296607249, |
|
"learning_rate": 4.936449918978123e-06, |
|
"loss": 0.6611, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9603793426190534, |
|
"learning_rate": 4.9355741408488045e-06, |
|
"loss": 0.6939, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.123192378969838, |
|
"learning_rate": 4.9346924481933345e-06, |
|
"loss": 0.6596, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.049989264664389, |
|
"learning_rate": 4.933804843152808e-06, |
|
"loss": 0.6664, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9287929732517263, |
|
"learning_rate": 4.932911327882679e-06, |
|
"loss": 0.6401, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.913529924198974, |
|
"learning_rate": 4.932011904552749e-06, |
|
"loss": 0.6525, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8908499663624638, |
|
"learning_rate": 4.931106575347171e-06, |
|
"loss": 0.6781, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8871778748426131, |
|
"learning_rate": 4.930195342464437e-06, |
|
"loss": 0.6816, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8819674351946986, |
|
"learning_rate": 4.929278208117378e-06, |
|
"loss": 0.6153, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9892043785009771, |
|
"learning_rate": 4.928355174533153e-06, |
|
"loss": 0.6716, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9215453869699415, |
|
"learning_rate": 4.927426243953252e-06, |
|
"loss": 0.6784, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9652774517156478, |
|
"learning_rate": 4.9264914186334775e-06, |
|
"loss": 0.7372, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8716382127920056, |
|
"learning_rate": 4.925550700843953e-06, |
|
"loss": 0.6594, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.02563864851558, |
|
"learning_rate": 4.924604092869109e-06, |
|
"loss": 0.6678, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.980206197810093, |
|
"learning_rate": 4.923651597007679e-06, |
|
"loss": 0.7016, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9257361907115577, |
|
"learning_rate": 4.922693215572695e-06, |
|
"loss": 0.6537, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9407972324104933, |
|
"learning_rate": 4.9217289508914836e-06, |
|
"loss": 0.6906, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.953792733030727, |
|
"learning_rate": 4.920758805305654e-06, |
|
"loss": 0.6874, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9679520423391571, |
|
"learning_rate": 4.919782781171101e-06, |
|
"loss": 0.7057, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.8643570019367653, |
|
"learning_rate": 4.918800880857991e-06, |
|
"loss": 0.6224, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9696150692028724, |
|
"learning_rate": 4.917813106750763e-06, |
|
"loss": 0.6623, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9068337402094646, |
|
"learning_rate": 4.916819461248119e-06, |
|
"loss": 0.6756, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9032967004326324, |
|
"learning_rate": 4.915819946763017e-06, |
|
"loss": 0.6624, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.9042931017160107, |
|
"learning_rate": 4.914814565722671e-06, |
|
"loss": 0.6711, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8611902344265154, |
|
"learning_rate": 4.913803320568538e-06, |
|
"loss": 0.6652, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.1208348114642743, |
|
"learning_rate": 4.912786213756316e-06, |
|
"loss": 0.7054, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8514149510371133, |
|
"learning_rate": 4.91176324775594e-06, |
|
"loss": 0.6777, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.0707648844359314, |
|
"learning_rate": 4.91073442505157e-06, |
|
"loss": 0.6793, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.007444952147854, |
|
"learning_rate": 4.9096997481415885e-06, |
|
"loss": 0.6532, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8430194319590454, |
|
"learning_rate": 4.908659219538598e-06, |
|
"loss": 0.6532, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.9867088678023405, |
|
"learning_rate": 4.907612841769407e-06, |
|
"loss": 0.7107, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.9369756600022385, |
|
"learning_rate": 4.90656061737503e-06, |
|
"loss": 0.7306, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8672758137612632, |
|
"learning_rate": 4.905502548910681e-06, |
|
"loss": 0.6412, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8706341698734509, |
|
"learning_rate": 4.904438638945761e-06, |
|
"loss": 0.7022, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9114920575560276, |
|
"learning_rate": 4.903368890063861e-06, |
|
"loss": 0.6328, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.0338522059220963, |
|
"learning_rate": 4.9022933048627496e-06, |
|
"loss": 0.7328, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8878198226644961, |
|
"learning_rate": 4.901211885954367e-06, |
|
"loss": 0.6687, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9423243153918482, |
|
"learning_rate": 4.900124635964823e-06, |
|
"loss": 0.6604, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9129637670735964, |
|
"learning_rate": 4.899031557534383e-06, |
|
"loss": 0.6731, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9896776529578333, |
|
"learning_rate": 4.8979326533174696e-06, |
|
"loss": 0.6608, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8896434182833951, |
|
"learning_rate": 4.896827925982654e-06, |
|
"loss": 0.7074, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8194553474739297, |
|
"learning_rate": 4.895717378212644e-06, |
|
"loss": 0.6335, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9346190741848343, |
|
"learning_rate": 4.894601012704284e-06, |
|
"loss": 0.7068, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.874199819168889, |
|
"learning_rate": 4.893478832168545e-06, |
|
"loss": 0.6641, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9159502482028847, |
|
"learning_rate": 4.8923508393305224e-06, |
|
"loss": 0.6731, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8558744309352562, |
|
"learning_rate": 4.891217036929422e-06, |
|
"loss": 0.6083, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.936605889796404, |
|
"learning_rate": 4.89007742771856e-06, |
|
"loss": 0.6863, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8664253450652424, |
|
"learning_rate": 4.8889320144653525e-06, |
|
"loss": 0.6693, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.859861919600542, |
|
"learning_rate": 4.887780799951311e-06, |
|
"loss": 0.63, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8390583343121363, |
|
"learning_rate": 4.8866237869720334e-06, |
|
"loss": 0.7058, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9197736229674498, |
|
"learning_rate": 4.885460978337201e-06, |
|
"loss": 0.6782, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9686243956595285, |
|
"learning_rate": 4.884292376870567e-06, |
|
"loss": 0.6619, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8454295473139757, |
|
"learning_rate": 4.883117985409954e-06, |
|
"loss": 0.6551, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.9658575404117379, |
|
"learning_rate": 4.881937806807241e-06, |
|
"loss": 0.647, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8533364242044601, |
|
"learning_rate": 4.880751843928365e-06, |
|
"loss": 0.6563, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.9045931704682527, |
|
"learning_rate": 4.879560099653306e-06, |
|
"loss": 0.6667, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8723782063352707, |
|
"learning_rate": 4.8783625768760865e-06, |
|
"loss": 0.6425, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8891929748147736, |
|
"learning_rate": 4.877159278504759e-06, |
|
"loss": 0.638, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8842932376273528, |
|
"learning_rate": 4.875950207461403e-06, |
|
"loss": 0.6437, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8634008607871101, |
|
"learning_rate": 4.8747353666821155e-06, |
|
"loss": 0.6229, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8749107029341481, |
|
"learning_rate": 4.873514759117004e-06, |
|
"loss": 0.6785, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8531375678504259, |
|
"learning_rate": 4.872288387730182e-06, |
|
"loss": 0.637, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8582175373389599, |
|
"learning_rate": 4.871056255499758e-06, |
|
"loss": 0.6444, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.0656353331347395, |
|
"learning_rate": 4.86981836541783e-06, |
|
"loss": 0.6756, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.8937004109826505, |
|
"learning_rate": 4.8685747204904796e-06, |
|
"loss": 0.652, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9024529450351501, |
|
"learning_rate": 4.867325323737765e-06, |
|
"loss": 0.6606, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.847448497225188, |
|
"learning_rate": 4.866070178193707e-06, |
|
"loss": 0.6445, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9790412024382392, |
|
"learning_rate": 4.86480928690629e-06, |
|
"loss": 0.6377, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9166672541084665, |
|
"learning_rate": 4.863542652937453e-06, |
|
"loss": 0.6568, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9053859567083279, |
|
"learning_rate": 4.862270279363076e-06, |
|
"loss": 0.6729, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.8963082010943075, |
|
"learning_rate": 4.860992169272981e-06, |
|
"loss": 0.7187, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.8625928467278408, |
|
"learning_rate": 4.859708325770919e-06, |
|
"loss": 0.6569, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9572711715249694, |
|
"learning_rate": 4.858418751974564e-06, |
|
"loss": 0.6858, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9116206947499983, |
|
"learning_rate": 4.857123451015503e-06, |
|
"loss": 0.6341, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9469293485174248, |
|
"learning_rate": 4.855822426039236e-06, |
|
"loss": 0.6475, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9320786521604542, |
|
"learning_rate": 4.854515680205159e-06, |
|
"loss": 0.7021, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.038184288949622, |
|
"learning_rate": 4.853203216686562e-06, |
|
"loss": 0.5999, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.904022305585284, |
|
"learning_rate": 4.851885038670618e-06, |
|
"loss": 0.6555, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.323477712215313, |
|
"learning_rate": 4.8505611493583815e-06, |
|
"loss": 0.6297, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.2176660817908431, |
|
"learning_rate": 4.849231551964771e-06, |
|
"loss": 0.6771, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.8930100103847493, |
|
"learning_rate": 4.84789624971857e-06, |
|
"loss": 0.6296, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.0104643222544036, |
|
"learning_rate": 4.846555245862413e-06, |
|
"loss": 0.6329, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.0006315975644342, |
|
"learning_rate": 4.845208543652783e-06, |
|
"loss": 0.6448, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9719470767684136, |
|
"learning_rate": 4.843856146359999e-06, |
|
"loss": 0.6538, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 1.025734064898448, |
|
"learning_rate": 4.842498057268209e-06, |
|
"loss": 0.6335, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8825687655634383, |
|
"learning_rate": 4.841134279675386e-06, |
|
"loss": 0.6728, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.9411597091872842, |
|
"learning_rate": 4.839764816893315e-06, |
|
"loss": 0.6756, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8838664041310971, |
|
"learning_rate": 4.838389672247585e-06, |
|
"loss": 0.6593, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8745260448073756, |
|
"learning_rate": 4.837008849077588e-06, |
|
"loss": 0.6397, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8847557689287797, |
|
"learning_rate": 4.835622350736499e-06, |
|
"loss": 0.659, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.875366668905105, |
|
"learning_rate": 4.8342301805912815e-06, |
|
"loss": 0.627, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8537326681265562, |
|
"learning_rate": 4.832832342022666e-06, |
|
"loss": 0.6477, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8593654316451845, |
|
"learning_rate": 4.831428838425153e-06, |
|
"loss": 0.6698, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.9220753709893383, |
|
"learning_rate": 4.830019673206997e-06, |
|
"loss": 0.6477, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8570602383594389, |
|
"learning_rate": 4.828604849790201e-06, |
|
"loss": 0.6815, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.9015787695209869, |
|
"learning_rate": 4.827184371610511e-06, |
|
"loss": 0.6293, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.900429307913389, |
|
"learning_rate": 4.8257582421174025e-06, |
|
"loss": 0.6799, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8833032292011391, |
|
"learning_rate": 4.824326464774076e-06, |
|
"loss": 0.6665, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.9806435575510756, |
|
"learning_rate": 4.822889043057446e-06, |
|
"loss": 0.6828, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8273111947815865, |
|
"learning_rate": 4.821445980458134e-06, |
|
"loss": 0.6265, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.9867943428090185, |
|
"learning_rate": 4.8199972804804615e-06, |
|
"loss": 0.6856, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.941670204697157, |
|
"learning_rate": 4.8185429466424375e-06, |
|
"loss": 0.6241, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.950190491805473, |
|
"learning_rate": 4.817082982475753e-06, |
|
"loss": 0.6733, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.9059424297489737, |
|
"learning_rate": 4.815617391525772e-06, |
|
"loss": 0.6717, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.833767481386628, |
|
"learning_rate": 4.814146177351523e-06, |
|
"loss": 0.6167, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9120074119227388, |
|
"learning_rate": 4.812669343525688e-06, |
|
"loss": 0.6383, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.0846687157648525, |
|
"learning_rate": 4.811186893634597e-06, |
|
"loss": 0.6542, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9273661121621178, |
|
"learning_rate": 4.809698831278217e-06, |
|
"loss": 0.6658, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8614935020984111, |
|
"learning_rate": 4.808205160070147e-06, |
|
"loss": 0.6632, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8640864376511351, |
|
"learning_rate": 4.806705883637604e-06, |
|
"loss": 0.6714, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8588959960471779, |
|
"learning_rate": 4.8052010056214184e-06, |
|
"loss": 0.6365, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9735734771646465, |
|
"learning_rate": 4.80369052967602e-06, |
|
"loss": 0.6939, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9404781469926247, |
|
"learning_rate": 4.802174459469435e-06, |
|
"loss": 0.6291, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9663046015849157, |
|
"learning_rate": 4.800652798683277e-06, |
|
"loss": 0.6516, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.8807991294468791, |
|
"learning_rate": 4.799125551012731e-06, |
|
"loss": 0.6364, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9527629477737198, |
|
"learning_rate": 4.7975927201665515e-06, |
|
"loss": 0.6298, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.8525983935565186, |
|
"learning_rate": 4.796054309867053e-06, |
|
"loss": 0.6371, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9405625850498484, |
|
"learning_rate": 4.794510323850096e-06, |
|
"loss": 0.6264, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9478829826012205, |
|
"learning_rate": 4.79296076586508e-06, |
|
"loss": 0.6328, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9599057130243777, |
|
"learning_rate": 4.791405639674941e-06, |
|
"loss": 0.6151, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.959409086045542, |
|
"learning_rate": 4.789844949056131e-06, |
|
"loss": 0.6369, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.8492616257179605, |
|
"learning_rate": 4.788278697798619e-06, |
|
"loss": 0.6554, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9806877124101798, |
|
"learning_rate": 4.7867068897058725e-06, |
|
"loss": 0.6579, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9000450903810971, |
|
"learning_rate": 4.785129528594858e-06, |
|
"loss": 0.6235, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9010530779092166, |
|
"learning_rate": 4.783546618296025e-06, |
|
"loss": 0.658, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9784681165218017, |
|
"learning_rate": 4.781958162653298e-06, |
|
"loss": 0.6676, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9576730133216942, |
|
"learning_rate": 4.780364165524067e-06, |
|
"loss": 0.6228, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.0020359961640437, |
|
"learning_rate": 4.778764630779184e-06, |
|
"loss": 0.647, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.8688473783118098, |
|
"learning_rate": 4.77715956230294e-06, |
|
"loss": 0.652, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9278357926734184, |
|
"learning_rate": 4.775548963993072e-06, |
|
"loss": 0.6291, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9690649138149913, |
|
"learning_rate": 4.77393283976074e-06, |
|
"loss": 0.6509, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9407962797966478, |
|
"learning_rate": 4.7723111935305275e-06, |
|
"loss": 0.6197, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9768285903734765, |
|
"learning_rate": 4.770684029240425e-06, |
|
"loss": 0.6518, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.0077965365925552, |
|
"learning_rate": 4.769051350841822e-06, |
|
"loss": 0.6195, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.9474683243935313, |
|
"learning_rate": 4.767413162299501e-06, |
|
"loss": 0.6504, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8585900243531269, |
|
"learning_rate": 4.765769467591626e-06, |
|
"loss": 0.6539, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8904877736571806, |
|
"learning_rate": 4.764120270709727e-06, |
|
"loss": 0.6321, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8628282782617729, |
|
"learning_rate": 4.7624655756587e-06, |
|
"loss": 0.648, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 2.669416127497495, |
|
"learning_rate": 4.760805386456793e-06, |
|
"loss": 0.6594, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8981381784754892, |
|
"learning_rate": 4.759139707135592e-06, |
|
"loss": 0.6769, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.7879080017640494, |
|
"learning_rate": 4.757468541740019e-06, |
|
"loss": 0.6281, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8871289305729931, |
|
"learning_rate": 4.755791894328317e-06, |
|
"loss": 0.6321, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8299591322141577, |
|
"learning_rate": 4.75410976897204e-06, |
|
"loss": 0.6505, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8909590264299827, |
|
"learning_rate": 4.752422169756048e-06, |
|
"loss": 0.6811, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.9254976675249351, |
|
"learning_rate": 4.75072910077849e-06, |
|
"loss": 0.6545, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8588476509107275, |
|
"learning_rate": 4.7490305661508006e-06, |
|
"loss": 0.667, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8412766854913993, |
|
"learning_rate": 4.747326569997684e-06, |
|
"loss": 0.6003, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8760795055942986, |
|
"learning_rate": 4.74561711645711e-06, |
|
"loss": 0.6364, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.83956409107638, |
|
"learning_rate": 4.7439022096803024e-06, |
|
"loss": 0.6575, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8380872276529072, |
|
"learning_rate": 4.742181853831721e-06, |
|
"loss": 0.6424, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8125631213297304, |
|
"learning_rate": 4.740456053089065e-06, |
|
"loss": 0.6663, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8823004853380154, |
|
"learning_rate": 4.7387248116432524e-06, |
|
"loss": 0.6301, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8640770572532905, |
|
"learning_rate": 4.736988133698416e-06, |
|
"loss": 0.6406, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8823776031811036, |
|
"learning_rate": 4.735246023471885e-06, |
|
"loss": 0.6115, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.993013360555103, |
|
"learning_rate": 4.733498485194188e-06, |
|
"loss": 0.6643, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8904288915714781, |
|
"learning_rate": 4.731745523109029e-06, |
|
"loss": 0.6555, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8926370506160105, |
|
"learning_rate": 4.729987141473286e-06, |
|
"loss": 0.6601, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.953062196989694, |
|
"learning_rate": 4.728223344556996e-06, |
|
"loss": 0.6809, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8761001219987443, |
|
"learning_rate": 4.7264541366433495e-06, |
|
"loss": 0.6833, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.9566539767394971, |
|
"learning_rate": 4.724679522028672e-06, |
|
"loss": 0.6385, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.9891726403966534, |
|
"learning_rate": 4.722899505022424e-06, |
|
"loss": 0.6999, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8037801938204335, |
|
"learning_rate": 4.721114089947181e-06, |
|
"loss": 0.5879, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.9130617389052504, |
|
"learning_rate": 4.71932328113863e-06, |
|
"loss": 0.6373, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8259173345318234, |
|
"learning_rate": 4.717527082945555e-06, |
|
"loss": 0.6019, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8371433538017989, |
|
"learning_rate": 4.715725499729826e-06, |
|
"loss": 0.6329, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.9057753463438688, |
|
"learning_rate": 4.713918535866392e-06, |
|
"loss": 0.6281, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.9029311795134953, |
|
"learning_rate": 4.712106195743269e-06, |
|
"loss": 0.6324, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8858712104718339, |
|
"learning_rate": 4.710288483761524e-06, |
|
"loss": 0.6032, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8645912795937802, |
|
"learning_rate": 4.708465404335277e-06, |
|
"loss": 0.6165, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8006257669108777, |
|
"learning_rate": 4.706636961891673e-06, |
|
"loss": 0.6829, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8617001997741787, |
|
"learning_rate": 4.704803160870888e-06, |
|
"loss": 0.6338, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8966975029556187, |
|
"learning_rate": 4.702964005726106e-06, |
|
"loss": 0.6447, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 1.0966911622953233, |
|
"learning_rate": 4.701119500923516e-06, |
|
"loss": 0.6652, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.882038752989712, |
|
"learning_rate": 4.699269650942296e-06, |
|
"loss": 0.6022, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.9811354762670036, |
|
"learning_rate": 4.697414460274605e-06, |
|
"loss": 0.6672, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8694666503441387, |
|
"learning_rate": 4.6955539334255714e-06, |
|
"loss": 0.6327, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8671894922067316, |
|
"learning_rate": 4.693688074913282e-06, |
|
"loss": 0.6306, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8737657642067588, |
|
"learning_rate": 4.69181688926877e-06, |
|
"loss": 0.6059, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9300407192703628, |
|
"learning_rate": 4.689940381036005e-06, |
|
"loss": 0.6382, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.836435132546773, |
|
"learning_rate": 4.6880585547718845e-06, |
|
"loss": 0.6493, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.858710237692544, |
|
"learning_rate": 4.686171415046217e-06, |
|
"loss": 0.6392, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9070208021531481, |
|
"learning_rate": 4.684278966441716e-06, |
|
"loss": 0.6074, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8430640165966782, |
|
"learning_rate": 4.682381213553986e-06, |
|
"loss": 0.6386, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9181203953874308, |
|
"learning_rate": 4.680478160991514e-06, |
|
"loss": 0.6246, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8224529815938104, |
|
"learning_rate": 4.678569813375654e-06, |
|
"loss": 0.6367, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8840051978175106, |
|
"learning_rate": 4.676656175340621e-06, |
|
"loss": 0.6802, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.851500472699274, |
|
"learning_rate": 4.674737251533476e-06, |
|
"loss": 0.6415, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.9237880941549678, |
|
"learning_rate": 4.672813046614116e-06, |
|
"loss": 0.6324, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.835807900767003, |
|
"learning_rate": 4.670883565255264e-06, |
|
"loss": 0.6196, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8636603022537266, |
|
"learning_rate": 4.668948812142454e-06, |
|
"loss": 0.6054, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8467987581796347, |
|
"learning_rate": 4.6670087919740224e-06, |
|
"loss": 0.6364, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.7976168655689396, |
|
"learning_rate": 4.665063509461098e-06, |
|
"loss": 0.6407, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8787272073060952, |
|
"learning_rate": 4.663112969327584e-06, |
|
"loss": 0.6275, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.9257599968777892, |
|
"learning_rate": 4.661157176310159e-06, |
|
"loss": 0.6561, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8310789526292252, |
|
"learning_rate": 4.659196135158251e-06, |
|
"loss": 0.6718, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.909951412942146, |
|
"learning_rate": 4.657229850634033e-06, |
|
"loss": 0.6861, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8697647332483404, |
|
"learning_rate": 4.6552583275124145e-06, |
|
"loss": 0.6131, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.9004276489872411, |
|
"learning_rate": 4.653281570581023e-06, |
|
"loss": 0.6187, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8571109763430029, |
|
"learning_rate": 4.651299584640198e-06, |
|
"loss": 0.6096, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8734397889718317, |
|
"learning_rate": 4.6493123745029765e-06, |
|
"loss": 0.6915, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8669799668424895, |
|
"learning_rate": 4.64731994499508e-06, |
|
"loss": 0.6401, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8542352737214168, |
|
"learning_rate": 4.645322300954908e-06, |
|
"loss": 0.6559, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.870542275037437, |
|
"learning_rate": 4.643319447233521e-06, |
|
"loss": 0.6826, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8583311322550404, |
|
"learning_rate": 4.641311388694629e-06, |
|
"loss": 0.6258, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8737060383296704, |
|
"learning_rate": 4.639298130214585e-06, |
|
"loss": 0.6339, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8912967573704136, |
|
"learning_rate": 4.637279676682367e-06, |
|
"loss": 0.6469, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8715196658497849, |
|
"learning_rate": 4.635256032999569e-06, |
|
"loss": 0.6646, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8662783788798808, |
|
"learning_rate": 4.633227204080389e-06, |
|
"loss": 0.6338, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8927387239503449, |
|
"learning_rate": 4.631193194851617e-06, |
|
"loss": 0.6251, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.9498780462874753, |
|
"learning_rate": 4.629154010252624e-06, |
|
"loss": 0.615, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.956056661851013, |
|
"learning_rate": 4.627109655235345e-06, |
|
"loss": 0.6302, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.9067649800180307, |
|
"learning_rate": 4.625060134764273e-06, |
|
"loss": 0.6604, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 1.027192573699629, |
|
"learning_rate": 4.623005453816447e-06, |
|
"loss": 0.5859, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.9203828984363205, |
|
"learning_rate": 4.620945617381435e-06, |
|
"loss": 0.6202, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.951292364398945, |
|
"learning_rate": 4.618880630461324e-06, |
|
"loss": 0.65, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9118476668711323, |
|
"learning_rate": 4.6168104980707105e-06, |
|
"loss": 0.6487, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.878750842802124, |
|
"learning_rate": 4.614735225236685e-06, |
|
"loss": 0.615, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9060696025734105, |
|
"learning_rate": 4.612654816998821e-06, |
|
"loss": 0.6248, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9609216437070113, |
|
"learning_rate": 4.610569278409164e-06, |
|
"loss": 0.6778, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.898695374253286, |
|
"learning_rate": 4.608478614532215e-06, |
|
"loss": 0.6336, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9341703247271076, |
|
"learning_rate": 4.606382830444925e-06, |
|
"loss": 0.7034, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9888477139866373, |
|
"learning_rate": 4.604281931236675e-06, |
|
"loss": 0.6101, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.8782761928393398, |
|
"learning_rate": 4.602175922009272e-06, |
|
"loss": 0.5853, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9482746932782271, |
|
"learning_rate": 4.600064807876929e-06, |
|
"loss": 0.6652, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.8564815677643716, |
|
"learning_rate": 4.597948593966256e-06, |
|
"loss": 0.6782, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9240455443174742, |
|
"learning_rate": 4.595827285416248e-06, |
|
"loss": 0.6275, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8929439029031102, |
|
"learning_rate": 4.59370088737827e-06, |
|
"loss": 0.6398, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9148342501219822, |
|
"learning_rate": 4.59156940501605e-06, |
|
"loss": 0.6359, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.946705923671764, |
|
"learning_rate": 4.589432843505659e-06, |
|
"loss": 0.6128, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9114594357972271, |
|
"learning_rate": 4.587291208035504e-06, |
|
"loss": 0.6329, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8637275857494376, |
|
"learning_rate": 4.585144503806312e-06, |
|
"loss": 0.6444, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9422654099393534, |
|
"learning_rate": 4.5829927360311224e-06, |
|
"loss": 0.619, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8652896213966386, |
|
"learning_rate": 4.5808359099352675e-06, |
|
"loss": 0.5979, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 1.088012826550748, |
|
"learning_rate": 4.578674030756364e-06, |
|
"loss": 0.6621, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8470033759027679, |
|
"learning_rate": 4.576507103744299e-06, |
|
"loss": 0.6089, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8949204387411863, |
|
"learning_rate": 4.574335134161219e-06, |
|
"loss": 0.6306, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9395950836852421, |
|
"learning_rate": 4.572158127281516e-06, |
|
"loss": 0.6559, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8035578632803461, |
|
"learning_rate": 4.569976088391813e-06, |
|
"loss": 0.6262, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8860765726891832, |
|
"learning_rate": 4.567789022790953e-06, |
|
"loss": 0.6469, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9318594734495474, |
|
"learning_rate": 4.565596935789987e-06, |
|
"loss": 0.6153, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8747678520637625, |
|
"learning_rate": 4.5633998327121595e-06, |
|
"loss": 0.6666, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8349863455158287, |
|
"learning_rate": 4.561197718892896e-06, |
|
"loss": 0.6017, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8835445705707418, |
|
"learning_rate": 4.558990599679787e-06, |
|
"loss": 0.6418, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8929109099342171, |
|
"learning_rate": 4.556778480432584e-06, |
|
"loss": 0.6326, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8364797596101992, |
|
"learning_rate": 4.554561366523176e-06, |
|
"loss": 0.6782, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9260600125305487, |
|
"learning_rate": 4.552339263335581e-06, |
|
"loss": 0.595, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8599195336450693, |
|
"learning_rate": 4.550112176265937e-06, |
|
"loss": 0.5968, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9826419950054966, |
|
"learning_rate": 4.54788011072248e-06, |
|
"loss": 0.6456, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9203378835364787, |
|
"learning_rate": 4.5456430721255384e-06, |
|
"loss": 0.7066, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8939684740473081, |
|
"learning_rate": 4.5434010659075165e-06, |
|
"loss": 0.647, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8917246609121712, |
|
"learning_rate": 4.541154097512881e-06, |
|
"loss": 0.6568, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9075876860734314, |
|
"learning_rate": 4.538902172398151e-06, |
|
"loss": 0.6798, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8110519815726133, |
|
"learning_rate": 4.53664529603188e-06, |
|
"loss": 0.5882, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8925518187452417, |
|
"learning_rate": 4.534383473894646e-06, |
|
"loss": 0.6181, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9114329206202244, |
|
"learning_rate": 4.532116711479039e-06, |
|
"loss": 0.6026, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.846641811272952, |
|
"learning_rate": 4.529845014289642e-06, |
|
"loss": 0.5825, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8740877163769825, |
|
"learning_rate": 4.527568387843025e-06, |
|
"loss": 0.6989, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.9447837582033838, |
|
"learning_rate": 4.525286837667726e-06, |
|
"loss": 0.6582, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8674554766425348, |
|
"learning_rate": 4.523000369304244e-06, |
|
"loss": 0.6217, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8500815565912371, |
|
"learning_rate": 4.520708988305014e-06, |
|
"loss": 0.5941, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.0164281137951803, |
|
"learning_rate": 4.518412700234407e-06, |
|
"loss": 0.6469, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8420383857000564, |
|
"learning_rate": 4.516111510668707e-06, |
|
"loss": 0.6237, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.9412530438844999, |
|
"learning_rate": 4.513805425196103e-06, |
|
"loss": 0.6386, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.9175391505475148, |
|
"learning_rate": 4.511494449416671e-06, |
|
"loss": 0.6372, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8965680547901378, |
|
"learning_rate": 4.509178588942365e-06, |
|
"loss": 0.6398, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.161349181913427, |
|
"learning_rate": 4.506857849396998e-06, |
|
"loss": 0.7087, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9562290751417603, |
|
"learning_rate": 4.504532236416234e-06, |
|
"loss": 0.6969, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9026029147011317, |
|
"learning_rate": 4.502201755647571e-06, |
|
"loss": 0.6637, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9230372085626913, |
|
"learning_rate": 4.499866412750324e-06, |
|
"loss": 0.618, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9413375641434438, |
|
"learning_rate": 4.4975262133956235e-06, |
|
"loss": 0.6054, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9455820405339215, |
|
"learning_rate": 4.495181163266384e-06, |
|
"loss": 0.6606, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9648271805379507, |
|
"learning_rate": 4.492831268057307e-06, |
|
"loss": 0.6676, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8351069990624824, |
|
"learning_rate": 4.490476533474854e-06, |
|
"loss": 0.6479, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9101496071500632, |
|
"learning_rate": 4.488116965237244e-06, |
|
"loss": 0.6413, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8280285227695013, |
|
"learning_rate": 4.485752569074429e-06, |
|
"loss": 0.6223, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8992598915107023, |
|
"learning_rate": 4.4833833507280884e-06, |
|
"loss": 0.6742, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8923843348896123, |
|
"learning_rate": 4.48100931595161e-06, |
|
"loss": 0.5947, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8664845993947492, |
|
"learning_rate": 4.478630470510078e-06, |
|
"loss": 0.6369, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8726591582806938, |
|
"learning_rate": 4.476246820180259e-06, |
|
"loss": 0.6554, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8410498499068318, |
|
"learning_rate": 4.473858370750589e-06, |
|
"loss": 0.6293, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8755653995170062, |
|
"learning_rate": 4.4714651280211555e-06, |
|
"loss": 0.634, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.9524417549442475, |
|
"learning_rate": 4.469067097803689e-06, |
|
"loss": 0.6435, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.7960663615043987, |
|
"learning_rate": 4.466664285921543e-06, |
|
"loss": 0.6331, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8888628596240454, |
|
"learning_rate": 4.464256698209685e-06, |
|
"loss": 0.6552, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.9118498141803711, |
|
"learning_rate": 4.461844340514678e-06, |
|
"loss": 0.5826, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.854812295494578, |
|
"learning_rate": 4.459427218694671e-06, |
|
"loss": 0.6156, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8894677182707545, |
|
"learning_rate": 4.457005338619379e-06, |
|
"loss": 0.6316, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.92320969218264, |
|
"learning_rate": 4.454578706170075e-06, |
|
"loss": 0.6139, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9227943336374008, |
|
"learning_rate": 4.452147327239571e-06, |
|
"loss": 0.6377, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8967111083779129, |
|
"learning_rate": 4.4497112077322045e-06, |
|
"loss": 0.6004, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8705464350911942, |
|
"learning_rate": 4.447270353563828e-06, |
|
"loss": 0.7149, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9047682842604887, |
|
"learning_rate": 4.444824770661788e-06, |
|
"loss": 0.6214, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8332571186935797, |
|
"learning_rate": 4.442374464964916e-06, |
|
"loss": 0.61, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9191644952904162, |
|
"learning_rate": 4.439919442423513e-06, |
|
"loss": 0.6974, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8351147721256271, |
|
"learning_rate": 4.437459708999332e-06, |
|
"loss": 0.6633, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.892336463583514, |
|
"learning_rate": 4.434995270665569e-06, |
|
"loss": 0.6406, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8650896389402535, |
|
"learning_rate": 4.432526133406843e-06, |
|
"loss": 0.6345, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8153513627930788, |
|
"learning_rate": 4.430052303219185e-06, |
|
"loss": 0.6188, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8565464920766526, |
|
"learning_rate": 4.42757378611002e-06, |
|
"loss": 0.631, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8399364160683044, |
|
"learning_rate": 4.425090588098158e-06, |
|
"loss": 0.6577, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8912303770048037, |
|
"learning_rate": 4.422602715213774e-06, |
|
"loss": 0.6249, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8665388252830907, |
|
"learning_rate": 4.4201101734983965e-06, |
|
"loss": 0.6273, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8464141917671866, |
|
"learning_rate": 4.4176129690048905e-06, |
|
"loss": 0.5851, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8811314655875526, |
|
"learning_rate": 4.415111107797445e-06, |
|
"loss": 0.6214, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8083079113454086, |
|
"learning_rate": 4.412604595951558e-06, |
|
"loss": 0.6344, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8720341388764445, |
|
"learning_rate": 4.410093439554019e-06, |
|
"loss": 0.632, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8502476104259754, |
|
"learning_rate": 4.407577644702897e-06, |
|
"loss": 0.6184, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8813994548710394, |
|
"learning_rate": 4.405057217507527e-06, |
|
"loss": 0.637, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.917114660031616, |
|
"learning_rate": 4.40253216408849e-06, |
|
"loss": 0.7224, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8092450819851889, |
|
"learning_rate": 4.400002490577604e-06, |
|
"loss": 0.6247, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.9134141894112806, |
|
"learning_rate": 4.397468203117905e-06, |
|
"loss": 0.6813, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8877869537323849, |
|
"learning_rate": 4.394929307863633e-06, |
|
"loss": 0.6205, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.866374769727955, |
|
"learning_rate": 4.392385810980218e-06, |
|
"loss": 0.6463, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.071041707971838, |
|
"learning_rate": 4.3898377186442665e-06, |
|
"loss": 0.6571, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8411856812284965, |
|
"learning_rate": 4.38728503704354e-06, |
|
"loss": 0.6411, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.851584061852996, |
|
"learning_rate": 4.38472777237695e-06, |
|
"loss": 0.6966, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8291807888706711, |
|
"learning_rate": 4.382165930854534e-06, |
|
"loss": 0.6025, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8079509788666187, |
|
"learning_rate": 4.379599518697444e-06, |
|
"loss": 0.6461, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8116984175253492, |
|
"learning_rate": 4.377028542137933e-06, |
|
"loss": 0.6172, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8194670399241157, |
|
"learning_rate": 4.374453007419336e-06, |
|
"loss": 0.6244, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8544501253810839, |
|
"learning_rate": 4.371872920796059e-06, |
|
"loss": 0.6085, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8597333545854512, |
|
"learning_rate": 4.369288288533561e-06, |
|
"loss": 0.6757, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.912646868045953, |
|
"learning_rate": 4.366699116908339e-06, |
|
"loss": 0.6487, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8476447032290744, |
|
"learning_rate": 4.364105412207914e-06, |
|
"loss": 0.5805, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.9241438674933707, |
|
"learning_rate": 4.3615071807308165e-06, |
|
"loss": 0.6758, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8122777487524169, |
|
"learning_rate": 4.358904428786565e-06, |
|
"loss": 0.6153, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.9018439615133529, |
|
"learning_rate": 4.356297162695662e-06, |
|
"loss": 0.6453, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.9627678054923827, |
|
"learning_rate": 4.353685388789567e-06, |
|
"loss": 0.6282, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8406101285686195, |
|
"learning_rate": 4.351069113410689e-06, |
|
"loss": 0.6078, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.835904176007674, |
|
"learning_rate": 4.348448342912365e-06, |
|
"loss": 0.6436, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8568187537534087, |
|
"learning_rate": 4.345823083658855e-06, |
|
"loss": 0.6677, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8347696865421858, |
|
"learning_rate": 4.34319334202531e-06, |
|
"loss": 0.6171, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8348209799902971, |
|
"learning_rate": 4.340559124397774e-06, |
|
"loss": 0.6239, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8852176634720593, |
|
"learning_rate": 4.3379204371731555e-06, |
|
"loss": 0.6573, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8637190549045177, |
|
"learning_rate": 4.335277286759218e-06, |
|
"loss": 0.6394, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8678111098963714, |
|
"learning_rate": 4.332629679574566e-06, |
|
"loss": 0.6011, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.9549679439132955, |
|
"learning_rate": 4.3299776220486235e-06, |
|
"loss": 0.6205, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.834971592773752, |
|
"learning_rate": 4.3273211206216235e-06, |
|
"loss": 0.6121, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8863673003507534, |
|
"learning_rate": 4.324660181744589e-06, |
|
"loss": 0.6118, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8984092199591801, |
|
"learning_rate": 4.321994811879321e-06, |
|
"loss": 0.6897, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8880703411725283, |
|
"learning_rate": 4.319325017498379e-06, |
|
"loss": 0.5674, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8213603864968535, |
|
"learning_rate": 4.316650805085068e-06, |
|
"loss": 0.5959, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8521624706722543, |
|
"learning_rate": 4.31397218113342e-06, |
|
"loss": 0.6439, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8407539759191036, |
|
"learning_rate": 4.311289152148182e-06, |
|
"loss": 0.6495, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8832549384112272, |
|
"learning_rate": 4.308601724644797e-06, |
|
"loss": 0.6291, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.9007160602566869, |
|
"learning_rate": 4.305909905149389e-06, |
|
"loss": 0.6219, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8798855352674151, |
|
"learning_rate": 4.303213700198749e-06, |
|
"loss": 0.6365, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.841611556568396, |
|
"learning_rate": 4.300513116340317e-06, |
|
"loss": 0.6464, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.9363258404112693, |
|
"learning_rate": 4.297808160132165e-06, |
|
"loss": 0.6181, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8390395946351076, |
|
"learning_rate": 4.295098838142985e-06, |
|
"loss": 0.6186, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8925228288644416, |
|
"learning_rate": 4.292385156952069e-06, |
|
"loss": 0.5934, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8557639299914508, |
|
"learning_rate": 4.289667123149296e-06, |
|
"loss": 0.6458, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8078616638133473, |
|
"learning_rate": 4.2869447433351165e-06, |
|
"loss": 0.6045, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 0.6127398610115051, |
|
"eval_runtime": 1883.8371, |
|
"eval_samples_per_second": 0.701, |
|
"eval_steps_per_second": 0.078, |
|
"step": 507 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 2026, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 507, |
|
"total_flos": 238732066160640.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|