Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[12/13]GPTQ: ACCURATE POST-TRAINING QUANTIZATION FOR GENERATIVE PRE-TRAINED TRANSFORMERS #11

Open
penpaperkeycode opened this issue Oct 11, 2022 · 0 comments
Assignees

Comments

@penpaperkeycode
Copy link
Collaborator

penpaperkeycode commented Oct 11, 2022

Date: 2022.12.13
Presenter: Jeonghoon Kim
Keywords: Post-training quantization, GPT, causal language model task, acceleration, cuda kernel

A100 1장으로 175B까지 PTQ하는 논문 입니다.
방법론 자체가 기존 SOTA 방법과는 많이 달라 신기해서 관심을 많기 갖고 있는 논문입니다.

Paper(ICLR2023): https://openreview.net/forum?id=tcbBPnfwxS

@penpaperkeycode penpaperkeycode changed the title [11/29]GPTQ: ACCURATE POST-TRAINING QUANTIZATION FOR GENERATIVE PRE-TRAINED TRANSFORMERS [12/13]GPTQ: ACCURATE POST-TRAINING QUANTIZATION FOR GENERATIVE PRE-TRAINED TRANSFORMERS Oct 11, 2022
@penpaperkeycode penpaperkeycode self-assigned this Oct 25, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant