We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llamafactory
CUDA_VISIBLE_DEVICES=0,1,2,3 llamafactory-cli train \ --stage dpo \ --do_predict True \ --finetuning_type lora \ --model_name_or_path model/Meta-Llama-3-8B-Instruct \ --adapter_name_or_path saves/Meta-Llama-3-8B-Instruct/lora/dpo_train_2024-12-12-10-01-38 \ --flash_attn auto \ --dataset_dir data \ --eval_dataset dpo_zh_demo \ --template llama3 \ --cutoff_len 1024 \ --max_samples 1000 \ --preprocessing_num_workers 16 \ --predict_with_generate False \ --per_device_eval_batch_size 2 \ --max_new_tokens 512 \ --top_p 0.7 \ --temperature 0.95 \ --output_dir saves/Meta-Llama-3-8B-Instruct/lora/eval_nlg_dpo_2024-12-12-09 \ --overwrite_output_dir True
训练阶段是DPO,微调方法是lora,训练完成之后,使用以上代码进行模型NLG评估,运行期间并未出错,但是却没有结果输出,请问这是怎么回事?是否是少了什么参数?
No response
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Reminder
System Info
llamafactory
version: 0.9.1.dev0Reproduction
Expected behavior
训练阶段是DPO,微调方法是lora,训练完成之后,使用以上代码进行模型NLG评估,运行期间并未出错,但是却没有结果输出,请问这是怎么回事?是否是少了什么参数?
Others
No response
The text was updated successfully, but these errors were encountered: