Skip to content

Commit

Permalink
move hint to correct location
Browse files Browse the repository at this point in the history
  • Loading branch information
leej3 committed Dec 4, 2023
1 parent b349bb0 commit b97c3d1
Showing 1 changed file with 6 additions and 6 deletions.
12 changes: 6 additions & 6 deletions ignite/handlers/lr_finder.py
Original file line number Diff line number Diff line change
Expand Up @@ -162,18 +162,18 @@ def _log_lr_and_loss(self, trainer: Engine, output_transform: Callable, smooth_f
"if output of the engine is torch.Tensor, then "
"it must be 0d torch.Tensor or 1d torch.Tensor with 1 element, "
f"but got torch.Tensor of shape {loss.shape}."
"You may wish to use the output_transform kwarg with the attach method e.g.\n"
"""
lr_finder = FastaiLRFinder()
with lr_finder.attach(trainer, output_transform=lambda x:x["train_loss"]) as trainer_with_lr_finder:
trainer_with_lr_finder.run(dataloader_train)
"""
)
else:
raise TypeError(
"output of the engine should be of type float or 0d torch.Tensor "
"or 1d torch.Tensor with 1 element, "
f"but got output of type {type(loss).__name__}"
"You may wish to use the output_transform kwarg with the attach method e.g.\n"
"""
lr_finder = FastaiLRFinder()
with lr_finder.attach(trainer, output_transform=lambda x:x["train_loss"]) as trainer_with_lr_finder:
trainer_with_lr_finder.run(dataloader_train)
"""
)
loss = idist.all_reduce(loss)
lr = self._lr_schedule.get_param()
Expand Down

0 comments on commit b97c3d1

Please sign in to comment.