Confusion on Zero Redundancy Optimizer, seems the parameter shape is CHANGED and REPORT ERROR #1882
Unanswered
yxding95
asked this question in
Community | Q&A
Replies: 1 comment
-
What is your model? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm training a model which consists torch.nn.Embedding layer, but when I apply the Zero Redundancy Optimizer, the weight's shape would be changed to 1-dim (torch.Size([50272, 2048]) -> torch.Size([102957056])), and this raises a "RuntimeError: 'weight' must be 2-D" when the embedding is indexed.
Is there anyone know how to solve this problem? Is there any mistake I made? The config.py are just copied from language/gpt/gpt2_configs/gpt2_zero3.py
Beta Was this translation helpful? Give feedback.
All reactions