-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
提一个小建议。不知道对不对,在解决导出模型时:”RuntimeError: "addmm_impl_cpu_" not implemented for 'Half'“ #3333
Comments
可以提供下训练的命令吗?是否用了除了 lora 外别的微调方法 |
这是我保存的训练命令 |
我也用过bf16的精度,也是会在导出模型时出现上述错误,后来改了代码之后就没有出现这个错误了 |
#3434 |
在 webui 上加入了这个选项 |
Reminder
Reproduction
这个报错是因为cpu不支持bf32或者是fp16,原因在于没使用gpu,
而原文在src/llmtuner/hparams/model_args.py 大约132行 代码是:
export_device: str = field(
default="cpu",
metadata={"help": "The device used in model export."},
)
默认是cpu,所以会报错,改成cuda,就可以顺利导出模型!
Expected behavior
No response
System Info
No response
Others
No response
The text was updated successfully, but these errors were encountered: