[0.6.1] llama 13b gptq the value update is not the same shape as the original. updated: (2560, 3840), original (5120, 3840)
#580
Labels
Low Precision
Issue about lower bit quantization, including int8, int4, fp8
triaged
Issue has been triaged by maintainers
Looks like it ignores the
mapping.tp_rank
.The text was updated successfully, but these errors were encountered: