-
Notifications
You must be signed in to change notification settings - Fork 83
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[EVAL] Add RewardBench #324
Comments
It's in scope for the generative models, not as much for the classifier ones (as we would need to add a whole new pipeline for loading and running them) - do you think having it only for gen models would already be interesting enough? |
OK I think that would be too restrictive since most reward models today are currently trained as sequence classifiers. Feel free to close the issue if you want :) |
Evaluation short description
RewardBench is perhaps the only evaluation suite to provide broad coverage of the strengths and weaknesses of reward models across domains like reasoning, chat, and safety. It would be great to have this in
lighteval
so one can have a unified evaluation suite and not have to hop across different repos and dependencies.RewardBench is the de facto leaderboard for comparing reward models and is widely used by the post-training subset of the community for advanced methods like RL, rejection sampling, and others.
Evaluation metadata
Provide all available
Note: this eval is not a typical LLM eval since it relies on sequence classification. I'm not sure if that is out of scope for
lighteval
The text was updated successfully, but these errors were encountered: