Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Refactor hyper-parameters of adapters and model loader #1864

Merged
merged 4 commits into from
Dec 15, 2023
Merged

Refactor hyper-parameters of adapters and model loader #1864

merged 4 commits into from
Dec 15, 2023

Conversation

hiyouga
Copy link
Owner

@hiyouga hiyouga commented Dec 15, 2023

  • We deprecate the checkpoint_dir argument and use the adapter_name_or_path argument to load peft adapters and rename resume_lora_training to create_new_adapter with reversed values.
  • We create llmtuner.model.patches module to disentangle the patches from the model loader.

@hiyouga hiyouga merged commit e2bd597 into main Dec 15, 2023
@Katehuuh
Copy link
Contributor

No reference in code for --adapter_model_or_path, so is --adapter_name_or_path.

  • We deprecate the checkpoint_dir argument and use the adapter_model_or_path argument to load peft adapters and rename resume_lora_training to create_new_adapter with reversed values.

@hiyouga hiyouga added the solved This problem has been already solved label Dec 26, 2023
@hiyouga
Copy link
Owner Author

hiyouga commented Dec 26, 2023

@Katehuuh thanks for pointing it out

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
solved This problem has been already solved
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants