-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to run it locally #2
Comments
I have the same question, I have apple Mac M2 pro. |
I have a NVIDIA 4070 super rtx with 16 gig of ram. It's not enough. |
Issue: Out of Memory Error with Qwen2VL Model on RTX 3060 (12GB VRAM)Environment:
Problem Description:I am encountering a
Questions:
Additional Info: The VRAM on my GPU gets nearly filled with the initial model load, leaving limited memory for inference. Even after trying to switch to torch.float16 and using PyTorch's memory management configurations (e.g., PYTORCH_CUDA_ALLOC_CONF), the VRAM requirements still exceed the available capacity. Thank you for any guidance or tips on managing this effectively! |
No description provided.
The text was updated successfully, but these errors were encountered: