We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ascend
GPU
CPU
Uncomment only one /device <> line, hit enter to put that in a new line, and remove leading whitespaces from that line: /device ascend /device gpu /device cpu
Uncomment only one /device <> line, hit enter to put that in a new line, and remove leading whitespaces from that line:
/device <>
/device ascend /device gpu /device cpu
/device ascend
/device gpu
/device cpu
The text was updated successfully, but these errors were encountered:
issue提重复了。 此外标题中不需要指定特定的计算卡,指明Ascend平台就可以了。 多lora推理 是mindformers的特性,目前看dev有个相关提交关闭了,并没有合入。估计还在开发中。
Sorry, something went wrong.
谢谢你的回复。那如果想做到和vllm一样,单基座模型(Qwen1.5_14B)部署加载多个lora,是不是需要load_checkpoint多次?还是说只能将每个lora合并基座
你可以看下这个实现,多lora的情况应该是把多lora合并到一份ckpt里面。 https://gitee.com/mindspore/mindformers/pulls/3541/files
No branches or pull requests
Environment
Hardware Environment(
Ascend
/GPU
/CPU
):Software Environment:
Describe the current behavior
Describe the expected behavior
Steps to reproduce the issue
Related log / screenshot
Special notes for this issue
The text was updated successfully, but these errors were encountered: