Skip to content

Commit

Permalink
Merge branch 'jbarker/github_issues_071123' into 'main'
Browse files Browse the repository at this point in the history
Fix variable name typo

See merge request ADLR/megatron-lm!671
  • Loading branch information
jon-barker committed Jul 11, 2023
2 parents 7809781 + 7e0558b commit 37b5490
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions megatron/optimizer/distrib_optimizer.py
Original file line number Diff line number Diff line change
Expand Up @@ -334,7 +334,7 @@ def build_model_and_main_param_groups(cls,
'torch.cuda.FloatTensor, '
'torch.cuda.HalfTensor, or '
'torch.cuda.BFloat16Tensor. '
'Received {}'.format(param.type()))
'Received {}'.format(model_param.type()))

# Update optimizer's params.
group_range["orig_group"]["params"] = [
Expand Down Expand Up @@ -386,7 +386,7 @@ def __init__(self, optimizer, clip_grad, log_num_zeros_in_grad,
self.model_param_group_index_map, self.opt_group_ranges = \
self.build_optimizer_group_ranges(self.optimizer.param_groups,
self.model_gbuf_ranges)

# Allocate main param shards.
(
self.model_float16_groups,
Expand Down Expand Up @@ -630,7 +630,7 @@ def save_parameter_state(self, filename):
# Gather contiguous shards on DP rank 0.
world_tensors = {}
for key, send_tensor in local_shards.items():

# Gather tensor list.
if data_parallel_rank == 0:
recv_tensors = [torch.empty((gbuf_local_numel,),
Expand Down Expand Up @@ -700,7 +700,7 @@ def load_parameter_state(self, filename):

# Scatter local shards from DP rank 0.
for key, recv_tensor in local_shards.items():

# Scatter tensor list.
if data_parallel_rank == 0:
world_tensor = loaded_state[model_idx][dtype][key]
Expand Down

0 comments on commit 37b5490

Please sign in to comment.