-
Notifications
You must be signed in to change notification settings - Fork 5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
如何使用8bit model做inference #1462
Labels
solved
This problem has been already solved
Comments
安装 transformers==4.33.2 |
请问这个要求是完全等于4.33.2,还是 |
且LlamaFactory的code需要update么? |
完全等于 |
测试结果是:
这意味着以后不能再升级transformers了(暂时)? |
是的 |
您好, |
1 task
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
您好,我想把codellama的model做8bit的quantization,然后再用它做inference。
使用参数如下:
{
"stage": "sft",
"model_name_or_path": "/workspace/model/CodeLlama-34b-Instruct-hf",
"do_train": false,
"do_predict": true,
"dataset": "test_no_answer",
"template": "llama2",
"finetuning_type": "lora",
"quantization_bit": 8,
"bf16": true,
"lora_target": "all",
"output_dir": "/workspace/34b_baseline",
"cutoff_len": 16384,
"per_device_eval_batch_size": 1,
"max_samples": 100,
"predict_with_generate": true
}
遇到error:
File "/workspace/task_entry.py", line 30, in training_task
run_exp(args_dict)
File "/workspace/src/llmtuner/tuner/tune.py", line 65, in run_exp
run_sft(model_args, data_args, training_args, finetuning_args, generating_args, callbacks)
File "/workspace/src/llmtuner/tuner/sft/workflow.py", line 49, in run_sft
trainer = CustomSeq2SeqTrainer(
File "/usr/local/lib/python3.9/dist-packages/transformers/trainer_seq2seq.py", line 56, in init
super().init(
File "/usr/local/lib/python3.9/dist-packages/transformers/trainer.py", line 412, in init
raise ValueError(
Message:
User error.
请问应该如何解决?
谢谢!
The text was updated successfully, but these errors were encountered: