🤗 Preference Dataset | 📚 Documentation | 📄 Paper
This repository is the source code for the paper, Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback, where we introduce a routing framework that creates hybrid preferences with both LLM and human preference annotations to maximize performance on a given evaluation metric (e.g., RewardBench). We release this codebase to improve reproducibility of our work, and to aid researchers in constructing preference datasets in their research.
Install the dependencies within your Python environment:
python -m venv venv
venv/bin/source activate
pip install -r requirements.txt
Running the full pipeline involves several steps, some might need to be run on a TPU machine. Nevertheless, we wrote scripts to automate different parts of the pipeline. Please head over the docs directory for more information.
@article{miranda2024hybrid,
title={{Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback}},
author={Miranda, Lester James V and Wang, Yizhong and Elazar, Yanai and Kumar, Sachin and Pyatkin, Valentina and Brahman, Faeze and Smith, Noah A and Hajishirzi, Hannaneh and Dasigi, Pradeep},
journal={arXiv preprint arXiv:2410.19133},
year={2024}
}