Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

vLLM support #41

Open
MichaelJayW opened this issue Sep 20, 2023 · 12 comments · Fixed by vllm-project/vllm#4978
Open

vLLM support #41

MichaelJayW opened this issue Sep 20, 2023 · 12 comments · Fixed by vllm-project/vllm#4978

Comments

@MichaelJayW
Copy link

No description provided.

@Data-drone
Copy link

Are there any updates on this one?

@louisoutin
Copy link

+1

4 similar comments
@ruidongtd
Copy link

+1

@insist93
Copy link

insist93 commented Nov 3, 2023

+1

@leonardxie
Copy link

+1

@TexasRangers86
Copy link

+1

@Lvjinhong
Copy link

Lvjinhong commented Dec 20, 2023

So, could you provide advice so that I can make custom modifications on vLLM myself (llama2 70b)?

@RonanKMcGovern
Copy link

fwiw, i know this is about vLLM, but you can run medusa on tgi using --speculate 3

@TexasRangers86
Copy link

fwiw, i know this is about vLLM, but you can run medusa on tgi using --speculate 3

hello,how can I pass medusa model and base model args when I use medusa on tgi.

@RonanKMcGovern
Copy link

RonanKMcGovern commented Dec 26, 2023

fwiw, i know this is about vLLM, but you can run medusa on tgi using --speculate 3

hello,how can I pass medusa model and base model args when I use medusa on tgi.

Just pass the medusa model repo (as you would with any other model) and then add on --speculate 2

You can try this template: https://runpod.io/gsc?template=2xpg09eenv&ref=jmfkcdio

@TexasRangers86
Copy link

Thanks a lot !!!!

@chuangzhidan
Copy link

how to use medusa based on vllm or sglang?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

10 participants