I wonder if I can resume training from checkpoints after saving logs in stable baseline as from what I understood from the documentation that CheckpointCallback can work as snapshots (yet not sure about this)
This code creates logs but I am not sure if I can use them to resume training and if this even possible using this method
controller = PPO('MlpPolicy', environment, verbose=0, clip_range=0.15, device='auto', learning_rate=0.00001, )
checkpoint_callback = CheckpointCallback(save_freq=1000, save_path= output_dir +'./logs/', name_prefix='rl_model')
controller.learn(total_timesteps=int(timesteps), callback=checkpoint_callback)
I tried using tensorboard_log as it is what I found clear explanations about how to do this (save model and resume training) but I get the error zsh: illegal hardware instruction, and I never got output from it or link to where I can monitor my model, I couldn't find much solutions for this either
I hope that someone can help, and if what I am doing is wrong then can you please let me know the best way to do this
Thank you