Journal
ACM TRANSACTIONS ON INFORMATION SYSTEMS
Volume 19, Issue 3, Pages 242-262Publisher
ASSOC COMPUTING MACHINERY
DOI: 10.1145/502115.502117
Keywords
algorithms; design; experimentation; human factors; languages; performance; search engines; natural language processing; query formulation; answer extraction; answer selection
Categories
Ask authors/readers for more resources
The wealth of information on the web makes it an attractive resource for seeking quick answers to simple, factual questions such as who was the first American in space? or what is the second tallest mountain in the world? Yet today's most advanced web search services (e.g., Google and AskJeeves) make it surprisingly tedious to locate answers to such questions, In this paper, we extend question-answering techniques, first studied in the information retrieval literature, to the web and experimentally evaluate their performance, First we introduce MULDER, which we believe to be the first general-purpose, fully-automated question-answering system available on the web. Second, we describe MULDER's architecture, which relies on multiple search-engine queries, natural-language parsing, and a novel voting procedure to yield reliable answers coupled with high recall. Finally, we compare MULDER's performance to that of Google and AskJeeves on questions drawn from the TREC-8 question answering track. We find that MULDER's recall is more than a factor of three higher than that of AskJeeves, In addition, we find that Google requires 6.6 times as much user effort to achieve the same level of recall as MULDER.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available