Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

what would be the best way to do multi-gpu training for snca? #5

Open
pallashadow opened this issue Feb 28, 2019 · 3 comments
Open

Comments

@pallashadow
Copy link

Shall I do nn.parallel.DistributedDataParallel for for both "model" and "lemniscate" and mannually sync each lemniscate memory per epoch?

I think a better solution might be use a single GPU for "lemniscate" memory and calculation, while other GPUs for data parallel "model" part?

I am doing a comics image retrieval task, and find this project very useful. Thank you for you help.

@pallashadow pallashadow changed the title what is the correct way to do multi-gpu training for snca? what would be the best way to do multi-gpu training for snca? Mar 1, 2019
@RookieHong
Copy link

Hi, I am also thinking about the memory issue, how did you deal with it?

@pallashadow
Copy link
Author

I put the memory forward and backward on CPU, when the memory became large. it was a little bit slower but worked well.

model_forward_gpu -> memory_forward_cpu -> loss_cpu -> memory_backward_cpu -> model_backward_gpu

@RookieHong
Copy link

Thanks for sharing!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants