-
Notifications
You must be signed in to change notification settings - Fork 472
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Collapse reference+learner hydra heads when using LoRa #320
Comments
Haha, I was not aware that Aman proposed the same thing. |
When will this feature be available? |
I am not sure anyone started. cc @jon-tow |
Not yet. @glerzing is looking into |
Cool!!! Looking forward to this update. 👍 |
Sorry to make you wait, but it should take a few weeks to get this done. As a very rough estimate, I would say that I may push a tested solution around the 10th may. But I'm new here so I don't know how much time would then happen before a new release version. |
…CarperAI#434) + Collapse reference+learner hydra heads when using LoRa (CarperAI#320)
…CarperAI#434) + Collapse reference+learner hydra heads when using LoRa (CarperAI#320)
…CarperAI#434) + Collapse reference+learner hydra heads when using LoRa (CarperAI#320)
…CarperAI#434) + Collapse reference+learner hydra heads when using LoRa (CarperAI#320)
…CarperAI#434) + Collapse reference+learner hydra heads when using LoRa (CarperAI#320)
…CarperAI#434) + Collapse reference+learner hydra heads when using LoRa (CarperAI#320)
…CarperAI#434) + Collapse reference+learner hydra heads when using LoRa (CarperAI#320)
…CarperAI#434) + Collapse reference+learner hydra heads when using LoRa (CarperAI#320)
* Migrate to peft from opendelta for parameter efficient tuning methods (#434) + Collapse reference+learner hydra heads when using LoRa (#320) * fix from_config * Review corrections * ILQL generate when temperature is 0. * revert: guard against experimental 8-bit loading support * format: run `black` --------- Co-authored-by: jon-tow <[email protected]> Co-authored-by: maxreciprocate <[email protected]>
🚀 The feature, motivation, and pitch
With additive (delta-style) parameter-efficient tuning methods such as LoRa, we should be able to make a slightly more mem-efficient hydra architecture by using a single block that does ~
frozen_head + tunable_weights
for the learner/policy head's fwd-pass and simplyfrozen_head
for the reference, instead of maintaining 2x heads.CC @LouisCastricato and @cat-state for pointing this out
Alternatives
No response
Additional context
No response
The text was updated successfully, but these errors were encountered: