Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Supporting tensor parallelism for int8 weight only quant #939

Merged
merged 14 commits into from
Sep 27, 2024

Conversation

jerryzh168
Copy link
Contributor

Summary:
following https://github.com/pytorch/ao/blob/main/tutorials/developer_api_guide/tensor_parallel.py we can support tensor parallelism for int8 weight only quant, this is needed for torchchat

Test Plan:
python test/dtypes/test_affine_quantized_tensor_parallel.py

Reviewers:

Subscribers:

Tasks:

Tags:

Copy link

pytorch-bot bot commented Sep 24, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/939

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit b113eda with merge base 64719d5 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Sep 24, 2024
@jerryzh168 jerryzh168 changed the title [WIP] Supporting tensor parallelism for int8 weight only quant Supporting tensor parallelism for int8 weight only quant Sep 25, 2024
@jerryzh168 jerryzh168 requested a review from kwen2501 September 25, 2024 02:14
Summary:
following https://github.com/pytorch/ao/blob/main/tutorials/developer_api_guide/tensor_parallel.py
we can support tensor parallelism for int8 weight only quant, this is needed
for torchchat

Test Plan:
python test/dtypes/test_affine_quantized_tensor_parallel.py

Reviewers:

Subscribers:

Tasks:

Tags:
Copy link
Contributor

@kwen2501 kwen2501 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM!
Wow, didn't thought that would be so straightforward!
Nice!

@jerryzh168 jerryzh168 merged commit 72d2518 into pytorch:main Sep 27, 2024
17 checks passed
@jerryzh168 jerryzh168 deleted the aqt_tensor_parallel branch September 27, 2024 02:08
melvinebenezer pushed a commit to melvinebenezer/ao that referenced this pull request Oct 3, 2024
* [WIP] Supporting tensor parallelism for int8 weight only quant

Summary:
following https://github.com/pytorch/ao/blob/main/tutorials/developer_api_guide/tensor_parallel.py
we can support tensor parallelism for int8 weight only quant, this is needed
for torchchat

Test Plan:
python test/dtypes/test_affine_quantized_tensor_parallel.py

Reviewers:

Subscribers:

Tasks:

Tags:

* implement tp for aqt

* fixes

* import fix

* remove cpu test

* fix

* fix

* fix test

* device

* change transpose impl

* Skip compiled TP test for torch version < 2.5

* version util

* fix

* fix version

---------

Co-authored-by: Ke Wen <[email protected]>
yanbing-j pushed a commit to yanbing-j/ao that referenced this pull request Dec 9, 2024
…s empty (pytorch#939)

Skip error if sdk/cmdline-tools/latest/* is empty
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants