@FoConrad 
thanks for your replay.
I also recently implemented the PPO algorithm using PPO, but the training is 
very slow on breakout game(gym)  -- total timestep=10M, the eposide score is 
only 45.

how do you solve the performance of PPO algorithm. 

[ Full content available at: 
https://github.com/apache/incubator-mxnet/issues/10563 ]
This message was relayed via gitbox.apache.org for [email protected]

Reply via email to