User simulations for evaluating answers to question series
Title | User simulations for evaluating answers to question series |
Publication Type | Journal Articles |
Year of Publication | 2007 |
Authors | Jimmy Lin |
Journal | Information Processing & Management |
Volume | 43 |
Issue | 3 |
Pagination | 717 - 729 |
Date Published | 2007/05// |
ISBN Number | 0306-4573 |
Keywords | Evaluation, Question answering, User simulations |
Abstract | Recently, question series have become one focus of research in question answering. These series are comprised of individual factoid, list, and “other” questions organized around a central topic, and represent abstractions of user–system dialogs. Existing evaluation methodologies have yet to catch up with this richer task model, as they fail to take into account contextual dependencies and different user behaviors. This paper presents a novel simulation-based methodology for evaluating answers to question series that addresses some of these shortcomings. Using this methodology, we examine two different behavior models: a “QA-styled” user and an “IR-styled” user. Results suggest that an off-the-shelf document retrieval system is competitive with state-of-the-art QA systems in this task. Advantages and limitations of evaluations based on user simulations are also discussed. |
URL | http://www.sciencedirect.com/science/article/pii/S0306457306000963 |
DOI | 10.1016/j.ipm.2006.06.006 |