Replies: 3 comments 1 reply
-
There are several cases when to load LR or not, so this flag exists, I think. |
Beta Was this translation helpful? Give feedback.
-
It's nice to be able to decide from where to load the LR (checkpoints or |
Beta Was this translation helpful? Give feedback.
-
I think @JRMeyer suggests to do it automagically. If base model has 0.001 and you again use 0.001 in TL, there will be no change. Isn't LR a mandatory argument? Usual process is for low resource languages where you should drop the LR. Sometimes you do hyperparameter search to find a better converging one. So, you usually change it... |
Beta Was this translation helpful? Give feedback.
-
when transfer learning, it's usually the case you set a new learning rate.
unfortunately, you have to both set the LR and
force_initialize_learning_rate
, or else you get the old LR and have no idea :(I don't see any reason we should keep this flag
Beta Was this translation helpful? Give feedback.
All reactions