Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

can't match the win rate posted on the leaderboard #431

Open
lindseyfeng opened this issue Jan 2, 2025 · 0 comments
Open

can't match the win rate posted on the leaderboard #431

lindseyfeng opened this issue Jan 2, 2025 · 0 comments

Comments

@lindseyfeng
Copy link

Hi team,

I'm observing a significant difference in win rates when using GPT4o Mini with greedy decoding, achieving only 30.33%, compared to the leaderboard sample which reaches a 43% win rate using the same annotator and reference output (GPT Turbo). I'm using the default reference output and standard procedures, but I'm unclear about the exact inference strategy and prompt used for the leaderboard-generated file. Could you please provide details on the inference strategy, the specific prompts, and any additional parameters or settings that might be contributing to the higher win rate in the leaderboard sample? Understanding these factors will help me align my setup more closely with the leaderboard results.

Thank you for your assistance!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant