Aggregation of Multiple Judgments for Evaluating Ordered Lists

Many tasks (e.g., search and summarization) result in an ordered list of items. In order to evaluate such an ordered list of items, we need to compare it with an ideal ordered list created by a human expert for the same set of items. To reduce any bias, multiple human experts are often used to create multiple ideal ordered lists. An interesting challenge in such an evaluation method is thus how to aggregate these different ideal lists to compute a single score for an ordered list to be evaluated. In this paper, we propose three new methods for aggregating multiple order judgments to evaluate ordered lists: weighted correlation aggregation, rank-based aggregation, and frequent sequential pattern-based aggregation. Experiment results on ordering sentences for text summarization show that all the three new methods outperform the state of the art average correlation methods in terms of discriminativeness and robustness against noise. Among the three proposed methods, the frequent sequential pattern-based method performs the best due to the flexible modeling of agreements and disagreements among human experts at various levels of granularity.
Date: March 01, 2010
Book Title: Proceedings of the 2010 European Conference on Information Retrieval
Type: Proceedings
Downloads: 140

Has 1 soft copy


remote link

Bibtex


@Proceedings{Aggregation_of_Multiple_Judgments_for_Ev,
  author = "Hyun Duk Kim and ChengXiang Zhai and Jiawei Han",
  title = "{Aggregation of Multiple Judgments for Evaluating Ordered Lists}",
  month = "March",
  year = "2010",
  booktitle = "Proceedings of the 2010 European Conference on Information Retrieval",
}