Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does the repo suport qwen-vl? #728

Closed
tangyipeng100 opened this issue Dec 22, 2023 · 7 comments
Closed

Does the repo suport qwen-vl? #728

tangyipeng100 opened this issue Dec 22, 2023 · 7 comments
Assignees
Labels
feature request New feature or request triaged Issue has been triaged by maintainers

Comments

@tangyipeng100
Copy link

How to build qwen-vl using TensorRT-LLM?

@juney-nvidia juney-nvidia self-assigned this Dec 23, 2023
@juney-nvidia juney-nvidia added triaged Issue has been triaged by maintainers feature request New feature or request labels Dec 23, 2023
@juney-nvidia
Copy link
Collaborator

@tangyipeng100

Hi, we do have an internal implementation supporting qwen-vl with TensorRT-LLM, it has not been published onto the github yet.

Let me sync with the engineers working on that and go back to you later.

June

@Yang-x-Zhao
Copy link

@juney-nvidia
Same request for me and thanks for your work!

I am also working on qwen-vl-chat. To my best understanding on TensorRT-LLM, the decoder session currently does not support embedding as an input, so I cannot modify the current qwen model to adapt to qwen-vl-chat model. I will be glad to use an official release of qwen-vl-chat.

@CN-COTER
Copy link

@tangyipeng100

Hi, we do have an internal implementation supporting qwen-vl with TensorRT-LLM, it has not been published onto the github yet.

Let me sync with the engineers working on that and go back to you later.

June

I am looking forward to deploy Qwen-vl-chat with trt-llm.

@77h2l
Copy link

77h2l commented Dec 29, 2023

same strong demands here .

@shikimoon
Copy link

Same request for me and thanks for your work!

I am also working on qwen-vl-chat. To my best understanding on TensorRT-LLM, the decoder session currently does not support embedding as an input, so I cannot modify the current qwen model to adapt to qwen-vl-chat model. I will be glad to use an official release of qwen-vl-chat.

the same to me

@Tlntin
Copy link
Contributor

Tlntin commented Jan 9, 2024

There might be what you're looking for here
link

@tangyipeng100
Copy link
Author

There might be what you're looking for here link

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request triaged Issue has been triaged by maintainers
Projects
None yet
Development

No branches or pull requests

8 participants