Finding the Effective Batchsize #390
Unanswered
m-parchami
asked this question in
Q&A
Replies: 1 comment 2 replies
-
Hi @m-parchami Yes, the effective batch size is the number of GPUs you use in distributed training mode. If you do not run the script in distributed training mode, You can also find the description in |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I was wondering how we should interpret the written
batch_size
fortrain_loader
in theyml
config. I suppose for the effective batch size we should multiply it by number of GPUs right? And where can we find that? Currently, I'm checking ifDistributedDataParallel
is mentioned in the .yml config, if so, I assume the effective batch size is 3x what's in the config, and if not, then just exactly what's in the config.Could you please clarify this?
Sorry if I missed it in the documentation.
All the best.
Beta Was this translation helpful? Give feedback.
All reactions