Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Multi GPU Paralellism #300

Open
martindellavecchia opened this issue Sep 30, 2024 · 3 comments
Open

Multi GPU Paralellism #300

martindellavecchia opened this issue Sep 30, 2024 · 3 comments
Assignees
Labels
enhancement New feature or request

Comments

@martindellavecchia
Copy link

I have a 2 GPU system, a 3060 (12gb VRAM) and a 3070ti (8GB). I've read torch supports paralellism that can split large models into both GPUs, it'd be great to have something like that to run big models for hich accuracy on multiple GPUs.

@jhj0517
Copy link
Owner

jhj0517 commented Oct 1, 2024

Hi. As far as I know, when you set up with multiple GPUs, if the GPU types are different, then several problems could occur (CUDA version mismatch, etc.).

For now, I'm listing for possible solutions to refer to later:

@martindellavecchia
Copy link
Author

Thank you!, it'd be great if I can pass a parameter to app.py to allow me to select which cuda device on my system I can use to run whisper... something like --device cuda0

I think thta will allow me to allocate the large model on my 3060(12gb)

@martindellavecchia
Copy link
Author

I found the way to select on what GPU to run the model, i just need to make sure the startup script includes a "export CUDA_VISIBLE_DEVICE=X

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants