Top Banner
Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University of Singapore Sigir’03
32

Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Jan 01, 2016

Download

Documents

Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Structured Use of External Knowledge for Event-basedOpen Domain Question Answering

Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat KohNational University of SingaporeSigir’03

Page 2: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Introduction

Main task in recent TREC-11: consists of 500 short factoid questions posed over more than one million news articles

Traditional “bag-of-words” representation: effective for IR research in retrieving large number of relevant documents, but ineffective for QA where precise answers are needed

Moldovan et al: illustrated the power of extensive WordNet to perform logic proof relied on knowledge reasoning

Lin gave a detailed comparison of employing the Web and WordNet

Page 3: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Introduction (2)

This paper: investigates the integration and structured use of both linguistic and web knowledge for the TREC-style QA, which was called Event-based QA

With a high performance QA system called QUALIFIER (QUestion Answering by LexIcal FabrIc and External Resources) and analyze its effectiveness using the TREC-11benchmark.

Results show that: the Event-based approach, together with the use of Successive Constraint Relaxation, gives rise to an effective question answering system

Page 4: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Event-based QA

Two basic types of things: entities and events

From the definitions in WordNet:An entity is “anything having an existence

(living or nonliving)”An event is “something that happens at a

given place and time”

Page 5: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Event-based QA (2)

Table 1 shows the correspondences of the most common WHquestion:

Page 6: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Event-based QA (3)

Consider an event to be a point within the multidimensional QA event space:

Page 7: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Event-based QA (4)

Difficult to find a correct answer, two major problems: Insufficient known elements (k<<n) Inexact known elements.

Solutions: Use of world knowledge (from the Web and WordNet

glosses) Exploit the lexical knowledge from WordNet synsets and

morphemics and employ a new technique called Successive Constraint Relaxation

In addition, perform structured semantic analysis based on what we already know in order to discover the event structure

Page 8: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

QUALIFIER System Architecture

Page 9: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

QUALIFIER System Architecture (2)

Question Analysis: Extract the known QA elements embedded in the

question and the expected answer (or targeted unknown element) type.

Extracts several sets of words (known elements) from the original question and employs a rule-based question classifier to identify the answer target (unknown element) type.

The rule-based question classifier adopts a two-tiered question taxonomy, which achieves a classification accuracy of over 98% on recent year’s TREC questions

Page 10: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

QUALIFIER System Architecture (3)

Query Formulation: Integrates the knowledge of both the Web and

WordNet to extract context information Two ways to processing

Linear approach of extracting highly correlated terms from the relevant information derived from the Web/WordNet

Structured analysis to discover different facets or elements of events, and employ a structured (event-based) approach to perform query formulation.

In this paper, they will concentrate more on the event-based query expansion

Page 11: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

QUALIFIER System Architecture (4)

Given the newly formulated query: Employs the MG tool to perform Boolean search for top

ranked documents in the QA corpus so that we can narrow down the scope of unknown elements.

Then selects candidate answer sentences based on the similarity between the sentences and the query in terms of original/expanded query terms, noun phrases, and quotations, etc

Finally performs linguistic and named entity analysis on the candidate sentences, and picks a named entity string with high possibility to be the targeted unknown element as the exact answer. Successive Constraint Relaxation is used here to improve the answer coverage

Page 12: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Structured Use of External Knowledge

Given a short, fact-based question, it can be transformed to a query in a form of : For example, given the question: “What Spanish

explorer discovered the Mississippi river?”, the query q(0) will be [Spanish,explorer, Mississippi, river]

In order to constrain the query and to bridge the gap between the query and the QA corpus, we resort to using general open resources to overcome this problem.

In this system, they focus on the amalgamation and structured use of the external knowledge like the Web and WordNet

Page 13: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Integrating External Knowledge

QUALIFIER uses the original content words q(0) to retrieve the top Nw documents using the Web search engine (E.g. Google)

Page 14: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Integrating External Knowledge (2)

Web lacks information on lexical relationships, such as synonyms Solution : employs WordNet to get the glosses

and synsets of the original query terms

Page 15: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Integrating External Knowledge (3)

Two possible approaches to investigate the potential of Event-based QA:Kq as a list without any structure (Linear

Query Formulation)Form semantic word groups before building

the new query (Structured Query Formulation)

Page 16: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Formulating Linear Queries

If we simply append all the terms, the resulting expanded query will be too broad and contain too many terms out of context. Solution: restrict the glosses and synonyms to

only those terms found in the web documents

Page 17: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Formulating Linear Queries (2)

Page 18: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Formulating Structured Queries

In order to discover the structure of terms within the external knowledge, we perform structural analysis on Kq to form semantic groups of terms

Given any two distinct terms ti, tj Kq , ∈we compute their lexical, co-occurrence, and distance correlation as follows:

Page 19: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Formulating Structured Queries (2)

Lexical correlation: consider only

synonym relation

Page 20: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Formulating Structured Queries (3)

Co-occurrence correlation:

Page 21: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Formulating Structured Queries (4)

Distance correlation:

Page 22: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Formulating Structured Queries (5)

Page 23: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Formulating Structured Queries (6)

One promising advantage of this approach: Be able to answer any factual questions about the

elements in this QA event other than just “What Spanish explorer discovered the Mississippi river?”.

For instance, can easily handle questions like “When was Mississippi river discovered?” and “Which river were discovered by Hernando De Soto?” etc with the same set of knowledge

Page 24: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Successive Constraint Relaxation (SCR)

Given the newly formulated linear or structured query, QUALIFIER uses the MG tool to retrieve up to N (N=50) relevant documents within the QA corpus.

Boolean retrieval : One of its disadvantages: returns either too many

or too few documents However, this can be an advantage for QA as it

ensures precision. Successive Constraint Relaxation (SCR):

Avoid over-constraining the query; improve answer coverage while preserving precision

Page 25: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Answer Extraction

With the set of the top K sentences obtained after document retrieval and sentence ranking, QUALIFIER performs fine-grained named entity tagging before extracting the string that matches the question class as the answer

How to detect fine-grained named entities? Using a two-tiered hierarchy as shown in Table 2. It adopts a rule-based algorithm to detect the named

entities by utilizing the lexical and semantic features such as capitalization, punctuation, context, part-of-speech tagging and phrase chunking

Tests indicate that it could achieve a detection accuracy of over 90%

Page 26: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Answer Extraction (2)

Page 27: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Evaluation – TREC-11

The main task in TREC-11: 500 factoid questions from a QA corpus over one million documents

Systems are required to return exactly one response per question A response is either a [answer-string; document-id] pair or the

string “NIL”. The pair will be judged as follows:

correct: the answer-string consists of exactly a correct answer and that answer is supported by the document returned;

unsupported: the answer-string contains a correct answer but the document returned does not support that answer;

inexact: the answer-string contains a correct answer and the document supports that answer, but the string contains either partial answer or more than just the answer; and

wrong: the answer-string does not contain a correct answer or the answer is not responsive.

Page 28: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Evaluation – TREC-11 (2)

TREC-11 employed the Confidence Weighted Score (CWS) to reward systems that successfully rank questions that they answered correctly higher than those answered incorrectly.

By using the linear query expansion, it places them among one of the top performing systems.

Page 29: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Linear vs. Structured Query Formulation

Several tests: 1) Baseline: perform QA without using the external

resources. 2) Baseline + Web: add up to top m context words

from Cq into q(1) 3) Baseline + Web + WordNet: combine both Web

and WordNet knowledge, but do not constrain the new terms from WordNet. This is to test the effects of adding some WordNet terms out of context.

4) Baseline + Web + WordNet + context constraint: i.e. linear query formulation.

5) Baseline + Web + WordNet + structure analysis: i.e. structured query formulation.

Page 30: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Linear vs. Structured Query Formulation (2)

For each test, also compare the results obtained without performing Successive Constraint Relaxation (SCR) to that obtained after 5 iterations of SCR.

In these tests, they examine the top 75 web snippets returned by Google with a cut-off threshold σof 0.2. This set of parameters has been found in H.Yang and T.S. Chua to be most effective

Page 31: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Linear vs. Structured Query Formulation (3)

Observations from Table 4: Web-based query formulation (Test 2) improves the baseline

performance by 25.1%. This confirms: Using Web to extract highly correlated terms generally improves the QA performance.

The use of lexical knowledge from WordNet without context constraint (Test 3) does not seem to be effective for QA, especially for the test w/o SCR: It tends to add too many terms out of context.

The linear query formulation (Test 4) achieves an accuracy of 0.588 after 5 SCR iterations. This gives an improvement of more than 7.3% over just using the Web (Test 2).

The best performance is achieved (0.634) when carrying out the structured use of external knowledge (Test 5): An improvement of 7.8% over the linear query formulation (Test 4) with SCR. The impact is much more significant for tests without SCR with a corresponding improvement of 36.4%.

Page 32: Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.

Conclusion

QUALIFIER:A novel approach for Event-based QADiscover the info about a QA entity or

different aspects of a QA eventStructural analysis: extract semantic groupsSemantic groups: expand query and

achieve an answer accuracy of 63.4%The results demonstrated: this approach is

feasible and effective for open domain QA