-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
FSDP + Qlora Faill #3853
Labels
solved
This problem has been already solved
Comments
It should have been fixed |
hiyouga
added
solved
This problem has been already solved
and removed
pending
This problem is yet to be addressed
labels
May 24, 2024
gotzmann
pushed a commit
to gotzmann/LLaMA-Factory
that referenced
this issue
Jun 2, 2024
@hiyouga I get same errors when finetuing glm-4-9b model with latest codes:
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Reminder
Reproduction
I pulled new code and ran Accelerate +FSDP + Qlora training, but encountered an error:
ValueError: Must flatten tensors with uniform dtype but got torch.bfloat16 and torch.float16
However, when I use the old code, it works fine. I suspect there is a bug in the new code.
Expected behavior
No response
System Info
No response
Others
No response
The text was updated successfully, but these errors were encountered: