@FoConrad thanks for your replay. I also recently implemented the PPO algorithm using PPO, but the training is very slow on breakout game(gym) -- total timestep=10M, the eposide score is only 45.
how do you solve the performance of PPO algorithm. [ Full content available at: https://github.com/apache/incubator-mxnet/issues/10563 ] This message was relayed via gitbox.apache.org for [email protected]
