4.5 Article

Scaling question answering to the Web

Journal

ACM TRANSACTIONS ON INFORMATION SYSTEMS
Volume 19, Issue 3, Pages 242-262

Publisher

ASSOC COMPUTING MACHINERY
DOI: 10.1145/502115.502117

Keywords

algorithms; design; experimentation; human factors; languages; performance; search engines; natural language processing; query formulation; answer extraction; answer selection

Ask authors/readers for more resources

The wealth of information on the web makes it an attractive resource for seeking quick answers to simple, factual questions such as who was the first American in space? or what is the second tallest mountain in the world? Yet today's most advanced web search services (e.g., Google and AskJeeves) make it surprisingly tedious to locate answers to such questions, In this paper, we extend question-answering techniques, first studied in the information retrieval literature, to the web and experimentally evaluate their performance, First we introduce MULDER, which we believe to be the first general-purpose, fully-automated question-answering system available on the web. Second, we describe MULDER's architecture, which relies on multiple search-engine queries, natural-language parsing, and a novel voting procedure to yield reliable answers coupled with high recall. Finally, we compare MULDER's performance to that of Google and AskJeeves on questions drawn from the TREC-8 question answering track. We find that MULDER's recall is more than a factor of three higher than that of AskJeeves, In addition, we find that Google requires 6.6 times as much user effort to achieve the same level of recall as MULDER.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available