As the title says, I am testen PPO with the Cartpole Environment using SB3, but if I look at the performance measured be the evaluate_policy function I reach a mean reward of 475 reliable at 20000 timesteps, but I need about 90000 timesteps if I look at console log to get comparable results during learning.
Why does my model perform so much better using the evaluation helper?
I used the same hyperparameters in both cases, and I used a new environment for the evaluation with the helper method.