简体   繁体   中英

Evaluate_policy records much higher mean reward then stable baselines 3 logger

As the title says, I am testen PPO with the Cartpole Environment using SB3, but if I look at the performance measured be the evaluate_policy function I reach a mean reward of 475 reliable at 20000 timesteps, but I need about 90000 timesteps if I look at console log to get comparable results during learning.

Why does my model perform so much better using the evaluation helper?

I used the same hyperparameters in both cases, and I used a new environment for the evaluation with the helper method.

I think I have solved the "problem": evaluate_policy uses deterministic action in it's default settings, which leads to better results faster.

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM