Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add INT8 mixed-precision training #748
Add INT8 mixed-precision training #748
Changes from 11 commits
b2e99ec
255abe9
efb53bf
0a510f5
f80ea8c
4a404ce
42abc15
e826d48
2ab9df3
c89b950
cde7e8f
691da9d
3540e79
f9d4e2a
9448b4d
64f707a
d04e8b3
4f8d63d
b3770d3
f39fdac
dd33823
b96769a
2b16ebb
117cc60
ae37058
ae4eb21
730c90c
c470a24
7c1d760
0e15e2d
22c11bc
208188c
77aafdb
ce6a5d5
d367f77
d24a894
fb09b24
3372644
9e05b5c
6e4e684
b395858
986c590
35df447
7164551
b14ab6d
dbbc90f
8d918f1
b4bd411
d67a933
7352335
6122aaa
8dab7cc
0d65b26
6082d30
File filter
Filter by extension
Conversations
Jump to
There are no files selected for viewing
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is this true? QAT does fake quantization (doesn't cast dtypes to int8) but here we're actually casting
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If we consider dtype casting an implementation detail, and only think QAT as doing quantization in forward pass during training (in terms of numerics), then it should be the same 😄.