Skip to content

Commit

Permalink
fix READEME
Browse files Browse the repository at this point in the history
  • Loading branch information
Sh1gechan committed Jun 25, 2024
1 parent 21de29b commit 01e6043
Showing 1 changed file with 3 additions and 4 deletions.
7 changes: 3 additions & 4 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,12 +26,12 @@ python llm_judge/gen_model_answer.py --config <CONFIG-PATH>

Arguments & Options:
- `<CONFIG-PATH>` is the path to a configuration file. Examples are in `configs/`.
- `num_answers_per_question` specifies how many to generate (default: all)
- `num_answers_per_question` specifies the number of answers to generate per question (default: all)

For example:

```bash
python llm_judge/gen_model_answer.py --config configs/rinna--japanese-gpt-neox-3.6b-instruction-ppo.json --num_answers_per_question <n>
python llm_judge/gen_model_answer.py --config configs/rinna--japanese-gpt-neox-3.6b-instruction-ppo.json
```


Expand Down Expand Up @@ -59,7 +59,7 @@ Arguments & Options:
- `--model-list <LIST-OF-MODEL-IDS>` is a list of model IDs to be evaluated. If not specified, all models in `data/jp_bench/model_answer` will be evaluated.
- `--yes` is a flag to skip the confirmation prompt.
- `--wandb` is a flag to enable logging to W&B. You can upload the results later to W&B by running `upload_result.py`, as described in the next section.
- `num_answers_per_question` : Number of answers to evaluate per question
- `num_answers_per_question` : specifies the number of answers to evaluate per question

**Mode: `pairwise-baseline` (Default)**

Expand Down Expand Up @@ -162,4 +162,3 @@ If you use our code in your research, please cite our work:
year={2024}
}
```

0 comments on commit 01e6043

Please sign in to comment.