You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
使用了更新后的代码对Chatglm3进行QLora微调,之前提到的‘enable’问题已经得到了解决,但是又出现了RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn,大概知道是什么问题但是实在是没找到哪里可以改。
Reminder
Reproduction
CUDA_VISIBLE_DEVICES=0 python src/train_bash.py
--stage sft
--do_train True
--model_name_or_path /home/csm/Csm_Workspace/Chatglm3-6b_finetuning/model
--finetuning_type lora
--quantization_bit 4
--template chatglm3
--dataset_dir data
--dataset alpaca_zh
--cutoff_len 1024
--learning_rate 5e-05
--num_train_epochs 3.0
--max_samples 100000
--per_device_train_batch_size 2
--gradient_accumulation_steps 8
--lr_scheduler_type cosine
--max_grad_norm 1.0
--logging_steps 5
--save_steps 100
--warmup_steps 0
--optim adamw_torch
--report_to none
--output_dir saves/ChatGLM3-6B-Chat/lora/train_2024-04-18-11-34-47
--fp16 True
--lora_rank 8
--lora_alpha 16
--lora_dropout 0.1
--lora_target query_key_value
--plot_loss True
Expected behavior
使用了更新后的代码对Chatglm3进行QLora微调,之前提到的‘enable’问题已经得到了解决,但是又出现了RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn,大概知道是什么问题但是实在是没找到哪里可以改。
System Info
transformers
version: 4.39.3Others
No response
The text was updated successfully, but these errors were encountered: