-
Notifications
You must be signed in to change notification settings - Fork 420
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
finetune的 MAX_STEPS = None 意义是什么?可以改成其他吗? #24
Comments
是这样的大佬老师 我使用合并的model作为 base model 来 finetune, 提示这个错误
|
@ZenXir max_step会在代码下面改。这个问题我昨天在本地branch改了忘push上来了,你可以更新一下。 |
好的大佬老师 |
大佬老师 我使用合并的model 使用finetune.py 训练 模型合并过程和流程分两步: finetune命令是: 报错内容是这个:
|
@ZenXir 我还没跑过他们的,你先自己研究一下吧。你这个情况就是没成功转过来。
|
在 prepare for traning 前这样 resize_token_embeddings 就可以训练了大佬
|
@Facico 对了大佬老师 其他参数都是默认的,我的机器是单卡 RTX4090 24G |
抱歉消息太多了有些消息会看漏,如果要直观的对比的话,保持batch size和epoch就可以了,如果想要跑快一点可以将mirco batch size调大 |
双卡,RTX3090: if not args.wandb: |
/root/anaconda3/lib/python3.9/site-packages/transformers/optimization.py:391: FutureWarning: This implementation of AdamW is deprecated and will be removed in a future version. Use the PyTorch implementation torch.optim.AdamW instead, or set no_deprecation_warning=True to disable this warning |
大佬三句代码是加在哪一步的哪个文件里面呢?我也想做同样的训练,奈何我太菜了,没明白 |
@godzeo 放在加载完模型和tokenizer后就行 |
老哥 这个max_step 怎么填哇 |
这里的 MAX_STEPS = None 为什么要设置成None?可以改成其他吗?
The text was updated successfully, but these errors were encountered: