no code implementations • 17 Oct 2023 • Siyan Zhao, John Dang, Aditya Grover
We introduce Group Preference Optimization (GPO), an alignment framework that steers language models to preferences of individual groups in a few-shot manner.
1 code implementation • 30 Aug 2023 • Hritik Bansal, John Dang, Aditya Grover
In particular, we find that LLMs that leverage rankings data for alignment (say model X) are preferred over those that leverage ratings data (say model Y), with a rank-based evaluation protocol (is X/Y's response better than reference response?)