-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
使用vllm推理的时候如何传入dtype参数 #3387
Labels
solved
This problem has been already solved
Comments
已修复 |
已经拉去最新代码,还是有同样问题 |
pytorch库在最后检测时进行了一次torch.bfloat16 tensor的创建。V100, torch=2.3.0, cuda=12.1的情况下是可以创建成功的,所以最后的返回结果是True。导致transformer库判断为True,最终导致LLaMA-Factory库判断为True。
LLaMA-Factory vllm_engine.py
LLaMA-Factory misc.py
transformers/src/transformers/utils/import_utils.py
pytorch/torch/cuda/init.py
|
hiyouga
added
pending
This problem is yet to be addressed
and removed
solved
This problem has been already solved
labels
May 27, 2024
hiyouga
added a commit
that referenced
this issue
Jun 5, 2024
|
hiyouga
added
solved
This problem has been already solved
and removed
pending
This problem is yet to be addressed
labels
Jun 5, 2024
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
The text was updated successfully, but these errors were encountered: