-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to send CURL request for MLLM models? #3971
Comments
The openai API does not support vision models yet, it will supported in the future version |
@hiyouga Thanks for the prompt response. So right now there is now way to feed multiple images to the model? Or do it programmatically through python? |
The only way I am guessing right now is to load the weights in HF library and do inference using that? |
the model we supported accepts only one image at inference, you can use |
Thanks! |
update: now api mode supports image inputs |
Thanks for the update!! 😀 |
Reminder
Reproduction
Hi There,
I am trying to run inference of MLLM models using the
CUDA_VISIBLE_DEVICES=0 llamafactory-cli api training_recipes/infer.yaml
command which starts a local open ai like api server.I want to now send images with prompt to this server to do inference.
How can I send a CURL request for the same?
Expected behavior
I would like to be able to send a curl request with image and get status 200 response.
System Info
transformers
version: 4.41.0Others
No response
The text was updated successfully, but these errors were encountered: