Saving conformer checkpoint and resuming train progress from checkpoint #20286
-
Confused about saving checkpoint of encoder-decoder models in lightning AI. Is the saving checkpoint technique same for all models? checkpoint_callback = ModelCheckpoint(
monitor='val_loss',
dirpath="./saved_checkpoint/",
filename='model-{epoch:02d}-{val_wer:.2f}',
save_top_k=3, # 3 Checkpoints
mode='min'
) But how to load the checkpoint again to resume the training? Is the same as normal architecture models like giving ckpt_path = args.checkpoint_path if args.checkpoint_path else None
trainer.fit(speech_trainer, data_module, ckpt_path=ckpt_path)
trainer.validate(speech_trainer, data_module) |
Beta Was this translation helpful? Give feedback.
Answered by
LuluW8071
Sep 19, 2024
Replies: 1 comment
-
Found the answer the code above is correct |
Beta Was this translation helpful? Give feedback.
0 replies
Answer selected by
LuluW8071
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Found the answer the code above is correct