Aligning with Human Judgement: The Role of Pairwise Preference in Large Language Model Evaluators

Yinhong Liu, Han Zhou, Zhijiang Guo, Ehsan Shareghi, Ivan Vulić, Anna Korhonen, Nigel Collier

Research output: Chapter in Book/Report/Conference proceedingConference PaperResearchpeer-review

Abstract

Large Language Models (LLMs) have demonstrated promising capabilities as automatic evaluators in assessing the quality of generated natural language. However, LLMs still exhibit biases in evaluation and often struggle to generate coherent evaluations that align with human assessments. In this work, we first conduct a systematic study of the misalignment between LLM evaluators and human evaluation, revealing that existing calibration methods aimed at mitigating biases of LLMs are insufficient for effectively aligning LLM evaluators. Inspired by the use of preference data in RLHF, we formulate the evaluation as a ranking problem and introduce Pairwise-preference Search (PairS), an uncertainty-guided search method that employs LLMs to conduct pairwise comparisons locally and efficiently ranks candidate texts globally. PairS achieves state-of-the-art performance on representative evaluation tasks in long-form generations and demonstrates significant improvements over direct scoring. Furthermore, we provide insights into the role of pairwise preference in quantifying the transitivity of LLMs and demonstrate how PairS benefits from calibration using debiased pairwise evaluations.
Original languageEnglish
Title of host publicationFirst Conference on Language Modeling COLM 2024
EditorsDipanjan Das, Danqi Chen, Yoav Artizi, Angela Fan
Place of PublicationPortland OR USA
PublisherOpenReview
Number of pages20
Publication statusPublished - 2024
EventConference on Language Modeling 2024 - Philadelphia, United States of America
Duration: 7 Oct 20249 Oct 2024
Conference number: 1st
https://colmweb.org/ (Website)
https://openreview.net/group?id=colmweb.org/COLM/2024/Conference#tab-accept (Proceedings)

Conference

ConferenceConference on Language Modeling 2024
Abbreviated titleCOLM 2024
Country/TerritoryUnited States of America
CityPhiladelphia
Period7/10/249/10/24
Internet address

Keywords

  • LLM evaluator
  • pairwise comparison
  • human alignment

Cite this