-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
关于evaluation的几个问题 #1316
Comments
|
你好,我有用以下参数做inference,但是回复甚至都不是英文,请问是某个参数设置错误了吗?
我在别的issue里面看到你有提及llama2的inference batch size要设置为1,我不太理解原因,而且试过之后还是一样的乱码。 |
问题已修复,请更新代码后重试,eval batch size 仍然需要设置为1 CUDA_VISIBLE_DEVICES=0 python src/train_bash.py \
--stage sft \
--model_name_or_path meta-llama/Llama-2-7b-chat-hf \
--do_predict \
--dataset alpaca_gpt4_en \
--template llama2 \
--output_dir out/debug/llama2/eval \
--max_new_tokens 128 \
--per_device_eval_batch_size 1 \
--max_samples 30 \
--predict_with_generate |
您能解释下为什么要设置为1吗?不太想设置为1因为速度太慢了 |
@XuanRen4470 llama2 的模型问题,如果使用了多个 Batch 就会溢出 |
能详细解释下这个益出嘛? 我需要评估下对performance影响有多大 因为batch = 1实在是太慢。 谢谢您🙏 |
请问这个参数只能通过命令行添加吗? 在train_web页面上有吗 |
命令行应该是可以 我没用过train_web |
The text was updated successfully, but these errors were encountered: