Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

set autoround format as default to unify CPU/HPU/CUDA #205

Merged
merged 10 commits into from
Aug 6, 2024

Conversation

wenhuach21
Copy link
Contributor

No description provided.

@wenhuach21 wenhuach21 requested review from WeiweiZhang1 and n1ck-guo and removed request for n1ck-guo August 5, 2024 07:19
if "hpu" in backend and model.dtype != torch.bfloat16:
logger.info("We suggest you to set `torch_dtype=torch.bfloat16` for better efficiency with AutoRound.")
model = model.to(torch.bfloat16)
elif model.dtype != torch.float16:
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

only for gpu, better add a judge "cpu" not in backend

@wenhuach21 wenhuach21 merged commit 1e75afd into main Aug 6, 2024
10 checks passed
@wenhuach21 wenhuach21 deleted the default_autoround branch August 6, 2024 02:41
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants