-
-
Notifications
You must be signed in to change notification settings - Fork 5.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature]: vAttention #4675
Comments
Hi, I'm one of the authors of this paper. Thank you for your interest in our work! We plan to release the code soon, hopefully in a few weeks. |
@ramyaprabhu-alt Just curious the code release would be a separate project or a PR against vLLM? I think it's a PR, right? |
Our initial release will be as a separate project on a slightly older version of vLLM. But soon after, we can also raise a PR against vLLM-latest. |
Glad to share the source code of vAttention. Please check it out here: https://github.com/microsoft/vattention |
is this still on VLLM roadmap to integrate? Please let us know. Thanks |
If there is still interest from vLLM community, we will be happy to contribute! |
We(many people with no name) are eagerly awaiting the release of this feature。。。so hungry so desire。。。Please!Help! |
🚀 The feature, motivation and pitch
Claim major improvements over vllm. Unfortunately no code only the paper.
arxiv.org/abs/2405.04437
Alternatives
No response
Additional context
No response
The text was updated successfully, but these errors were encountered: