-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: OpenGVLab/InternVL2-Llama3-76B: view size is not compatible with input tensor's size and stride #8630
Comments
Hmm, I ran this locally and didn't get such error. Could you share the dimensions of the images which you inputted to the model? |
This is at the start up of the server so before doing any inference. |
Looking at your environment, it seems that you're running this on AMD GPUs. Maybe there is some bug related to that? @youkaichao @WoosukKwon |
It may be that |
I just tried running this command (after downloading the HF repo locally) on 8x MI250 (ROCm 6.1) and failed to repro this issue. Can you tell us more about your setup by running My `collect_env.py` output:
My `rocm-smi` output:
Edit: I see that you have MI100 GPUs, but the ROCm and Triton versions are similar to mine. |
ROCm 6.1 is not officially supported in vLLM for MI100, so that may be why. |
This is in a container
The previous commit I was using was working though (fde47d3). |
As a sanity check, make sure that your downloaded version of |
cc @hongxiayang |
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
Your current environment
The output of `python collect_env.py`
Model Input Dumps
err_execute_model_input_20240919-094504.pkl.zip
🐛 Describe the bug
When I start the model via:
vllm serve OpenGVLab/InternVL2-Llama3-76B --tensor-parallel-size 8 --max-model-len 8000
I get:
The text was updated successfully, but these errors were encountered: