finetune_demo / trainer_state.json
kguo2's picture
Model save
16e2fd1 verified
raw
history blame
7.92 kB
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 170,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.029411764705882353,
"grad_norm": 4.553732872009277,
"learning_rate": 2.777777777777778e-05,
"loss": 0.7808,
"num_tokens": 163560.0,
"step": 5
},
{
"epoch": 0.058823529411764705,
"grad_norm": 2.3360044956207275,
"learning_rate": 4.9995716618706634e-05,
"loss": 0.1826,
"num_tokens": 327236.0,
"step": 10
},
{
"epoch": 0.08823529411764706,
"grad_norm": 0.659660279750824,
"learning_rate": 4.9845969445888354e-05,
"loss": 0.0853,
"num_tokens": 490970.0,
"step": 15
},
{
"epoch": 0.11764705882352941,
"grad_norm": 0.6894950866699219,
"learning_rate": 4.948368129547296e-05,
"loss": 0.0705,
"num_tokens": 654151.0,
"step": 20
},
{
"epoch": 0.14705882352941177,
"grad_norm": 0.47198617458343506,
"learning_rate": 4.891229802725401e-05,
"loss": 0.0624,
"num_tokens": 817991.0,
"step": 25
},
{
"epoch": 0.17647058823529413,
"grad_norm": 0.4311853051185608,
"learning_rate": 4.8137254283872696e-05,
"loss": 0.0577,
"num_tokens": 981831.0,
"step": 30
},
{
"epoch": 0.20588235294117646,
"grad_norm": 0.37233826518058777,
"learning_rate": 4.7165921799873716e-05,
"loss": 0.0542,
"num_tokens": 1145572.0,
"step": 35
},
{
"epoch": 0.23529411764705882,
"grad_norm": 0.32413461804389954,
"learning_rate": 4.6007539286346375e-05,
"loss": 0.0539,
"num_tokens": 1309412.0,
"step": 40
},
{
"epoch": 0.2647058823529412,
"grad_norm": 0.27661165595054626,
"learning_rate": 4.467312455804482e-05,
"loss": 0.0528,
"num_tokens": 1473224.0,
"step": 45
},
{
"epoch": 0.29411764705882354,
"grad_norm": 0.25189208984375,
"learning_rate": 4.317536973877955e-05,
"loss": 0.0528,
"num_tokens": 1637064.0,
"step": 50
},
{
"epoch": 0.3235294117647059,
"grad_norm": 0.258693665266037,
"learning_rate": 4.1528520541821506e-05,
"loss": 0.0506,
"num_tokens": 1800823.0,
"step": 55
},
{
"epoch": 0.35294117647058826,
"grad_norm": 0.22541610896587372,
"learning_rate": 3.974824077352845e-05,
"loss": 0.0504,
"num_tokens": 1964243.0,
"step": 60
},
{
"epoch": 0.38235294117647056,
"grad_norm": 0.31295299530029297,
"learning_rate": 3.785146334895093e-05,
"loss": 0.0504,
"num_tokens": 2127639.0,
"step": 65
},
{
"epoch": 0.4117647058823529,
"grad_norm": 0.24782125651836395,
"learning_rate": 3.58562292364649e-05,
"loss": 0.0489,
"num_tokens": 2291479.0,
"step": 70
},
{
"epoch": 0.4411764705882353,
"grad_norm": 0.20237773656845093,
"learning_rate": 3.378151586328963e-05,
"loss": 0.0489,
"num_tokens": 2455319.0,
"step": 75
},
{
"epoch": 0.47058823529411764,
"grad_norm": 0.15972378849983215,
"learning_rate": 3.164705661399079e-05,
"loss": 0.0489,
"num_tokens": 2619081.0,
"step": 80
},
{
"epoch": 0.5,
"grad_norm": 0.18197670578956604,
"learning_rate": 2.947315313878701e-05,
"loss": 0.0496,
"num_tokens": 2782547.0,
"step": 85
},
{
"epoch": 0.5294117647058824,
"grad_norm": 0.6986818909645081,
"learning_rate": 2.7280482256866697e-05,
"loss": 0.0478,
"num_tokens": 2946387.0,
"step": 90
},
{
"epoch": 0.5588235294117647,
"grad_norm": 0.22870764136314392,
"learning_rate": 2.508989929133051e-05,
"loss": 0.0474,
"num_tokens": 3109604.0,
"step": 95
},
{
"epoch": 0.5882352941176471,
"grad_norm": 0.17916245758533478,
"learning_rate": 2.2922239706315745e-05,
"loss": 0.0461,
"num_tokens": 3273444.0,
"step": 100
},
{
"epoch": 0.6176470588235294,
"grad_norm": 0.3453664779663086,
"learning_rate": 2.079812093300668e-05,
"loss": 0.046,
"num_tokens": 3435972.0,
"step": 105
},
{
"epoch": 0.6470588235294118,
"grad_norm": 0.26898708939552307,
"learning_rate": 1.8737746269439006e-05,
"loss": 0.0452,
"num_tokens": 3599812.0,
"step": 110
},
{
"epoch": 0.6764705882352942,
"grad_norm": 0.2884495258331299,
"learning_rate": 1.6760712719281375e-05,
"loss": 0.0459,
"num_tokens": 3763587.0,
"step": 115
},
{
"epoch": 0.7058823529411765,
"grad_norm": 0.25366732478141785,
"learning_rate": 1.4885824597312362e-05,
"loss": 0.0461,
"num_tokens": 3927092.0,
"step": 120
},
{
"epoch": 0.7352941176470589,
"grad_norm": 0.21009443700313568,
"learning_rate": 1.313091467446158e-05,
"loss": 0.0444,
"num_tokens": 4090932.0,
"step": 125
},
{
"epoch": 0.7647058823529411,
"grad_norm": 0.19418346881866455,
"learning_rate": 1.1512674563572253e-05,
"loss": 0.0459,
"num_tokens": 4254483.0,
"step": 130
},
{
"epoch": 0.7941176470588235,
"grad_norm": 0.19523853063583374,
"learning_rate": 1.0046495959150554e-05,
"loss": 0.0446,
"num_tokens": 4418323.0,
"step": 135
},
{
"epoch": 0.8235294117647058,
"grad_norm": 0.1927725076675415,
"learning_rate": 8.746324241130455e-06,
"loss": 0.0434,
"num_tokens": 4581587.0,
"step": 140
},
{
"epoch": 0.8529411764705882,
"grad_norm": 0.1469401717185974,
"learning_rate": 7.624525835084185e-06,
"loss": 0.0436,
"num_tokens": 4745427.0,
"step": 145
},
{
"epoch": 0.8823529411764706,
"grad_norm": 0.1357996016740799,
"learning_rate": 6.691770590465606e-06,
"loss": 0.0432,
"num_tokens": 4909267.0,
"step": 150
},
{
"epoch": 0.9117647058823529,
"grad_norm": 0.16276375949382782,
"learning_rate": 5.95693029563144e-06,
"loss": 0.0429,
"num_tokens": 5073107.0,
"step": 155
},
{
"epoch": 0.9411764705882353,
"grad_norm": 0.17533668875694275,
"learning_rate": 5.426994294902611e-06,
"loss": 0.0431,
"num_tokens": 5236905.0,
"step": 160
},
{
"epoch": 0.9705882352941176,
"grad_norm": 0.1326180398464203,
"learning_rate": 5.10700301026355e-06,
"loss": 0.0434,
"num_tokens": 5400452.0,
"step": 165
},
{
"epoch": 1.0,
"grad_norm": 0.19520561397075653,
"learning_rate": 5e-06,
"loss": 0.0428,
"num_tokens": 5563792.0,
"step": 170
},
{
"epoch": 1.0,
"step": 170,
"total_flos": 9.27422622990336e+16,
"train_loss": 0.0753641640438753,
"train_runtime": 542.7298,
"train_samples_per_second": 19.957,
"train_steps_per_second": 0.313
}
],
"logging_steps": 5,
"max_steps": 170,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 9.27422622990336e+16,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}