Summarization evaluation methods: Experiments and analysis

Hongyan Jing, Regina Barzilay, Kathleen McKeown, Michael Elhadad

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Two methods are used for evaluation of summarization systems: an evaluation of generated summaries against an" ideal" summary and evaluation of how well summaries help a person perform in a task such as information retrieval. We carried out two large experiments to study the two evaluation methods. Our results show that different parameters of an experiment can (hamatically affect how well a system scores. For example, summary length was found to affect both types of evaluations. For the "ideal" summary based types of evaluations. accuracy decreases as summary length and accuracy on an information retrieval task appear to correlate randomly. In this paper, we show how this parameter and others can effect evaluation results and describe how parameters can be controlled to produce a sound evaluation.
Original languageEnglish
Title of host publicationAAAI symposium on intelligent summarization
EditorsAnja Belz, Roger Evans, Sebastian Varges
Place of PublicationStroudsburg, PA, USA
PublisherAssociation for Computational Linguistics
Pages51-59
Number of pages9
ISBN (Print)9781932432510
StatePublished - Apr 1998

Fingerprint

Dive into the research topics of 'Summarization evaluation methods: Experiments and analysis'. Together they form a unique fingerprint.

Cite this