forked from NVIDIA/Megatron-LM
-
Notifications
You must be signed in to change notification settings - Fork 349
Issues: deepspeedai/Megatron-DeepSpeed
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Issues with DeepSpeed optimizer and tensor parallelism when changing topology between machines
#121
opened Mar 31, 2023 by
liutaocode
only tuning the layernorm or added adapter params error
#118
opened Mar 14, 2023 by
MultiModalPromptTuning
Encountered error when enabling ZeRO and CPU Activation Checkpointing at the same time.
#115
opened Mar 5, 2023 by
zincnode
Are there any other layer norm functions, such as RMSNorm or DeepNorm
#111
opened Feb 13, 2023 by
lvcc2018
Website documentation is incoherent with the repository content
#107
opened Jan 19, 2023 by
AnthoJack
AttributeError: module 'transformer_inference' has no attribute 'layer_norm_fp16'
#97
opened Nov 28, 2022 by
ranggihwang
The process is stuck at this step:compiling and loading fused kernels ...
#93
opened Nov 10, 2022 by
AQA6666
deepspeed to megatron - mismatch in function definition and call
#91
opened Oct 14, 2022 by
MatejUlcar
megatron-deepspeed layernorm has different output compare with megatron-lm?
#81
opened Aug 22, 2022 by
Kite0011
ProTip!
Mix and match filters to narrow down what you’re looking for.