Saving conformer checkpoint and resuming train progress from checkpoint #20286
-
| Confused about saving checkpoint of encoder-decoder models in lightning AI. Is the saving checkpoint technique same for all models?   checkpoint_callback = ModelCheckpoint(
      monitor='val_loss',
      dirpath="./saved_checkpoint/",
      filename='model-{epoch:02d}-{val_wer:.2f}',
      save_top_k=3,        # 3 Checkpoints
      mode='min'
  )But how to load the checkpoint again to resume the training? Is the same as normal architecture models like giving    ckpt_path = args.checkpoint_path if args.checkpoint_path else None
  trainer.fit(speech_trainer, data_module, ckpt_path=ckpt_path)   
  trainer.validate(speech_trainer, data_module) | 
Beta Was this translation helpful? Give feedback.
      
      
          Answered by
          
            LuluW8071
          
      
      
        Sep 19, 2024 
      
    
    Replies: 1 comment
-
| Found the answer the code above is correct | 
Beta Was this translation helpful? Give feedback.
                  
                    0 replies
                  
                
            
      Answer selected by
        LuluW8071
  
    Sign up for free
    to join this conversation on GitHub.
    Already have an account?
    Sign in to comment
  
        
    
Found the answer the code above is correct