You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I noticed that this gsmk-cot-llama.yaml is not using our eval details dataset and it is from this PR in lm_eval, so I am not so sure why there is such a gap.
Hi thx for the work and new llama3.2 reproduction update. But I do encounter an issue with gsm8k reproduction.
I add gsm8k directory to my work_dir manually with .yaml file here,
and the command was
CUDA_VISIBLE_DEVICES=3 lm_eval --model hf --model_args pretrained=/data/models/meta-llama/Llama-3.2-1B-Instruct,dtype=auto,parallelize=False,add_bos_token=True --tasks meta_gsm8k --batch_size 4 --output_path eval_results_general --include_path llama32_1B_workdir --seed 42 --log_samples --fewshot_as_multiturn --apply_chat_template
The result I got is 0.4003, which is different than the official report of 44.4.
Is there any thing I missed? Thx
The text was updated successfully, but these errors were encountered: