Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add lora_paths to v1_chat_generate_request #2529

Merged
merged 1 commit into from
Dec 22, 2024

Conversation

ccchow
Copy link
Contributor

@ccchow ccchow commented Dec 19, 2024

Motivation

Add missing changes to the v1_chat_generate_request function following PR #2438

Modifications

Checklist

  • Format your code according to the Contributor Guide.
  • Add unit tests as outlined in the Contributor Guide.
  • Update documentation as needed, including docstrings or example tutorials.

@qingzhong1
Copy link

Hello, I have a question. I started the service on a800 according to the following command. Why does the response to request "/v1/chat/completions" take 14 seconds? Is there any way to speed it up?python -m sglang.launch_server --model-path “”
--host 0.0.0.0
--port 8000
--tp-size 1
--mem-fraction-static 0.5
--served-model-name "Qwen2.5-7B-Instruct"
--chunked-prefill-size 64
--disable-cuda-graph
--disable-radix-cache
--lora-paths lora0=“”
--max-loras-per-batch 4

@merrymercy merrymercy merged commit 19ba2b0 into sgl-project:main Dec 22, 2024
15 checks passed
chosen-ox pushed a commit to chosen-ox/sglang that referenced this pull request Dec 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants