-
Notifications
You must be signed in to change notification settings - Fork 82
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multi-GPU inference issue #224
Comments
Hi, @ruifengma , |
Thanks @kennymckormick for the reply, it actually can be loaded onto 2 GPUs, but when inferencing, I got new issue
|
L197 in |
Thanks @junming-yang , the image script I found is |
You can try to dynamically check the model's device. |
I appended |
Maybe you can try |
Yes, I did. Still the same |
I have tried to reproduce your bug.
Each GPU is allocated about 26 GiB. And no error is reported. Please check your code. |
I actually did not DIY but completely following the advice. I use two A40 GPUs for the task, I checked and did the same modification as you did |
Not coding issue, update the latest version of official internvl configuration file solve |
When I try to run internVL-Chat-V1.5, since it is large and need at lease two GPUs, therefore, I use the following command to run
CUDA_VISIBLE_DEVICES=2,3 python run.py --data MME --model InternVL-Chat-V1-5 --verbose
But it does not run on the two GPUs but only one and give OOM error, do I need to give more configuration?
The text was updated successfully, but these errors were encountered: