Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
py
TrainingArguments(, deepspeed="path/to/deepspeed_config.json")
py
ds_config_dict = dict(scheduler=scheduler_params, optimizer=optimizer_params)
args = TrainingArguments(, deepspeed=ds_config_dict)
trainer = Trainer(model, args, )
DeepSpeed and Trainer parameters
There are three types of configuration parameters:
Some of the configuration parameters are shared by [Trainer] and DeepSpeed, and it can be difficult to identify errors when there are conflicting definitions.