-
Notifications
You must be signed in to change notification settings - Fork 707
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Qwen2.5-32b-int4用vllm跑 好像只会生成感叹号 #1103
Comments
不int 比如我用3b跑都是正常的 |
It looks like your vLLM is out of date. Try upgrade to |
the in that case, the we currently find two workarounds
|
我在vllm 0.6.4.post1,两张P100显卡,Qwen2.5-32B-Instruct-GPTQ-Int4上也出现了无限感叹号!!!!!!的问题, if len(messages) <= 1:
messages.extend([
{"role":"user",content:"你好"},
{"role":"assistant",content:"!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!"}
]) 这个解决方案虽然不优雅,但是对于我来说确实好用。当然我的系统提示词确实也超过50token了。 |
Model Series
Qwen2.5
What are the models used?
Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4
What is the scenario where the problem happened?
vllm
Is this a known issue?
Information about environment
Log output
正常返回 但是返回的结果全是"!!!!!"
Description
我把代码贴这,但是目前直接使用huggingface推理是正常的
直接使用huggingface的推理是正常返回的,现在这个也能返回 但是全是"!!!" 想知道为啥或者我哪里写错了?
The text was updated successfully, but these errors were encountered: