Skip to content

Commit

Permalink
align load_checkpoint args to save_checkpoint args
Browse files Browse the repository at this point in the history
  • Loading branch information
Samuel Weinbach committed Apr 30, 2021
1 parent 1d56953 commit 3f42973
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion megatron/checkpointing.py
Original file line number Diff line number Diff line change
Expand Up @@ -136,7 +136,7 @@ def save_checkpoint(neox_args, iteration, model, optimizer, lr_scheduler):
torch.distributed.barrier()


def load_checkpoint(model, optimizer, lr_scheduler, neox_args):
def load_checkpoint(neox_args, model, optimizer, lr_scheduler):
"""Load a model checkpoint and return the iteration."""

# Read the tracker file and set the iteration.
Expand Down
2 changes: 1 addition & 1 deletion megatron/training.py
Original file line number Diff line number Diff line change
Expand Up @@ -337,7 +337,7 @@ def setup_model_and_optimizer(neox_args, inference=False, get_key_value=True):
raise ValueError("Must be using deepspeed to run neox")

if neox_args.load is not None:
neox_args.iteration = load_checkpoint(model=model, optimizer=optimizer, lr_scheduler=lr_scheduler, neox_args=neox_args)
neox_args.iteration = load_checkpoint(neox_args=neox_args, model=model, optimizer=optimizer, lr_scheduler=lr_scheduler)
print_rank_0(f'Loading checkpoint and starting from iteration {neox_args.iteration}')
else:
neox_args.iteration = 0
Expand Down

0 comments on commit 3f42973

Please sign in to comment.