batch size vs n_env*n_steps #481
-
I have a question regarding this exmaple PPO hyperparameter configuration: MountainCarContinuous-v0: How can the batch size be beigger than the collected rollout buffer size? And on top of that the n_epochs is 10. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Hello,
Each minibatch will be truncated to 8. In fact, it will give you the same result as I would appreciate a PR that updates this parameter ;) |
Beta Was this translation helpful? Give feedback.
Hello,
you are correct, PPO will actually print a warning:
Each minibatch will be truncated to 8.
In fact, it will give you the same result as
python train.py --algo ppo --env MountainCarContinuous-v0 -param batch_size:8
I would appreciate a PR that updates this parameter ;)