-
Notifications
You must be signed in to change notification settings - Fork 252
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug] llava, cuda out of memory #1593
Comments
The loading process of vlm is: For llava-v1.5-7b, the first two steps will takes up about 14.5G cuda memory. But according to your log, the |
|
Can you try run the code without jupyter or ipython. |
I only have one card and it's currently running a program, so I can't test it right now. I'll test it next week and will share the results then. |
Checklist
Describe the bug
I hava one A100 gpu card.
following the instruction (https://github.com/InternLM/lmdeploy/blob/main/docs/zh_cn/inference/vl_pipeline.md), run HelloWorld llava program,error occurs. The llava model is llava-v1.5-7b, and is not very big. So, why "cuda out of memory" error occurs?
error info:
Reproduction
Environment
can't find lmdeploy check_env file
Error traceback
No response
The text was updated successfully, but these errors were encountered: