Skip to main content
Suppressing Outliers in Pairwise Preference Ranking
Computer Science Department
  • Vitor S. Cavalho, Carnegie Mellon University
  • Jonathan L. Elsas, Carnegie Mellon University
  • William W. Cohen, Carnegie Mellon University
  • Jaime G. Carbonell, Carnegie Mellon University
Date of Original Version
Conference Proceeding
Abstract or Description

Many of the recently proposed algorithms for learning feature-based ranking functions are based on the pairwise preference framework, in which instead of taking documents in isolation, document pairs are used as instances in the learning process [3, 5]. One disadvantage of this process is that a noisy relevance judgment on a single document can lead to a large number of mislabeled document pairs. This can jeopardize robustness and deteriorate overall ranking performance. In this paper we study the effects of outlying pairs in rank learning with pairwise preferences and introduce a new meta-learning algorithm capable of suppressing these undesirable effects. This algorithm works as a second optimization step in which any linear baseline ranker can be used as input. Experiments on eight different ranking datasets show that this optimization step produces statistically significant performance gains over state-of-the-art methods.

Citation Information
Vitor S. Cavalho, Jonathan L. Elsas, William W. Cohen and Jaime G. Carbonell. "Suppressing Outliers in Pairwise Preference Ranking" (2008)
Available at: