Filtering Out Bad Answers with Semantic Relations in a Web-Based Question-Answering System
Transcript of Filtering Out Bad Answers with Semantic Relations in a Web-Based Question-Answering System
22 avril 2004 TALN-2004, Atelier QR 1
Leila Kosseim and Jamileh YousefiCLaC Laboratory
Concondia UniversityMontréal
Filtering Out Bad Answers with Semantic Relations in a Web-Based Question-Answering System
22 avril 2004 TALN-2004, Atelier QR 2
Outline1.Introduction2.The original Web-QA component
3.Filtering out bad answers4.Evaluation5.Discussion and Future work
22 avril 2004 TALN-2004, Atelier QR 3
1. Introduction
Quantum-QA: The Web-QA Component provides Quantum with candidate answers extracted from documents found on the web using a search engine (Yahoo).
QuantumWeb-QAComponent
The Web Trec Collection
data
1 3
2
Candidate answers
22 avril 2004 TALN-2004, Atelier QR 4
1. Introduction(cont.)
Conjunction of Web-QA Component with Quantum:
Yields an interesting improvement, but Its results are very noisy.The right answer is there, but hidden alongside many wrong answers. Without the core Quantum system, no
distinction between right and wrong answers is possible.
22 avril 2004 TALN-2004, Atelier QR 5
2. The original Web-QA component
Example: Who is the fastest swimmer in the world?
1. Produce a formulation such as: The fastest swimmer in the world is <PERSON-NAME>.
2. Search the Web for the exact formulation 3. Extract the noun phrase following the
formulation 4. Apply simple syntactic and semantic
checks to ensure that the noun phrase is a PERSON-NAME.
22 avril 2004 TALN-2004, Atelier QR 6
Question reformulation
To formulate an answer formulation from a question: use a set of hand-made templates templates are composed of 2 sets of patterns:
Question pattern: Defines what the question must look like: When did < ANY-SEQUENCE-WORDS> <VERB-simple> ?
Set of answer patterns: Defines a set of possible answer formulations<DATE> < ANY-SEQUENCE-WORDS> <VERB-past>
the patterns take into account : Specific keywords (e.g. When) Strings of characters (e.g. ANY-SEQUENCE-WORDS) Part-of-speech tags (e.g. VERB-simple).
22 avril 2004 TALN-2004, Atelier QR 7
Evaluation of Web-QA component alone
Corpus No of questions
No of questions with a reformulation
No of questionswith at least one candidate answer
No of questions with a correct candidate answer in the top 5 candidates
Precision
TREC-9 694 624 (89.9%)
63 (9.1%) 17 (2.4%) 0.270
TREC-10
499 450 (90.2%)
256 (51.3%)
153 (30.7%) 0.598
22 avril 2004 TALN-2004, Atelier QR 8
In this paper
We present a method to: Filter out noisy candidate answers
Re-rank candidate answers by using semantic relations
22 avril 2004 TALN-2004, Atelier QR 9
3. Filtering and re-ranking
Our methodology is composed of five steps:
1. Run the Web-QA component and retrieve its top 200 candidates,
2. Run a named-entity tagger,3. Extract the semantic relation and
arguments for each candidate,4. Validate the semantic relation,5. Re-rank the candidate answers by
frequency.
22 avril 2004 TALN-2004, Atelier QR 10
Step 1: Initial Run
Run the Web-QA component and retrieve its top 200 candidates
Example: Who founded American Red Cross?
Jane Delano Who former schoolteacher Activism Exit RW ONLINE WhoWho ReallyClara Barton Christopher Blake…
22 avril 2004 TALN-2004, Atelier QR 11
Step 2: Named-entity tagger
tag candidates with the GATE-NE named-entity tagger
keep only candidates that satisfy the predicted constraints of the answer
Example: Who founded American Red Cross? Jane Delano
Clara Barton Christopher Blake…
22 avril 2004 TALN-2004, Atelier QR 12
Step 3: Find the semantic relation
Decompose the original question into two parts:
1. the main semantic relation relation = the main verb of the question
2. the argument expressed in the question argument1 = first noun phrase of the question
Use the candidate answer as argument2 Example: Who founded American Red Cross? relation: found
argument1: American Red Cross argument2: ANSWER
Found (American Red Cross, ANSWER)
22 avril 2004 TALN-2004, Atelier QR 13
Simplification Assumptions
Relations holding between more than two arguments, are considered as a binary relation:example: When did ARG1 sold ARG2?
relation : giveargument1: ARG1argument2: Answer
If the main verb is TO-BE, then we don’t take the semantic relation into account: example: Who is the author of John Christoph?
relation : noneargument1: John Christophargument2: Answer
22 avril 2004 TALN-2004, Atelier QR 14
Step 3 (cont.): Argument tuples
create a set of argument tuples from: the argument of the question and the candidates answers:
(American Red Cross, Jane Delano )(American Red Cross, Clara Barton )(American Red Cross, Christopher Blake)
submit all tuples to the document collection to find paragraphs that contain both elements
extract the paragraphs where both elements are within a distance of N words of each other
pre-window American Red Cross in-window Jane Delano post-window
≤ N words argument 1 N words ≤ argument 2 N ≤words
22 avril 2004 TALN-2004, Atelier QR 15
Step 4: Validate the semantic relation
The semantic relation may occur as a verb or noun or adjective in the context window.
Example: X was founded by Y Y is the founder of X
22 avril 2004 TALN-2004, Atelier QR 16
Verbs with equivalent semantic relation
check if a verb in any context window is a synonym, a hypernym or a hyponym of the relation in the question.
here, we use a part-of-speech tagger and WordNet.. . . 1881 In History, Event: American Red Cross founded by Clara
Barton. Related Year: 1881 Related Events: . . .
In addition, the American Red Cross organization was formed in 1881, Clara Barton served as its first president . . .
. . . Maryland, house where Clara Barton organized and directed the American Red Cross efforts. Hours, history of the . . .
22 avril 2004 TALN-2004, Atelier QR 17
Nouns and adjectives with equivalent semantic relation
if no verb has an equivalent semantic relation, try to validate nouns and adjectives
check if it has the same stem as the original relation or one of its synonyms.
here, we use the Porter stemmer. . . and of active volunteerism than Clara Barton, founder of the
American Red Cross.
. . . Clara Barton, American humanitarian, organizer of the American Red Cross, b. North Oxford . . .
22 avril 2004 TALN-2004, Atelier QR 18
Step 5: Re-rank candidate answers
throw away tuples that have no similar semantic relation in the question and in the documents
re-rank remaining candidates according to the number of passages in the collection containing the same relation.
Example: Argument tuple: (American Red Cross, Clara Barton). Nb of passages found: 110 passages Nb of passages with the relation kill within 5 words: 24 passages
Rank of argument tuple: 24/110 apply this to all the argument tuples, then select the five best ranked tuples
22 avril 2004 TALN-2004, Atelier QR 19
4. Evaluation
Comparison of the original and the new WebQA components
Corpus System No of questions
No of questionswith at least one candidate answer
No of questions with a correct candidate answer in the top 5 candidates
Precision
TREC-9 OriginalNew
694694
63 (9.1%)28 (4.0%)
17 (2.4%)20 (2.9%)
0.2700.714
TREC-10
OriginalNew
499499
255 (51.1%)189 (37.8%)
153 (30.7%)152 (30.5%)
0.5980.804
On the questions in the TREC-9 and TREC-10 collection Comparison using the pp-eval program provided by NIST
22 avril 2004 TALN-2004, Atelier QR 20
4. Evaluation (cont.)
Nb of questions with candidate answers is inferior in the new system but, Nb of correct candidates is greater or similar, the precision
is higher. although we provide less candidates, they are more likely to constitute correct answers than before.
In TREC-10: MRR is increased by 12% 31% of correct answers were ranked better by 4.5 on average. 9% ranked worse by 3.5 on average. Only one good answer was lost during the process.
In TREC-9: MRR is increased by 40% 25% of correct answers were ranked better; moved up the list by
2.2 on average. 5% of correct answers got worse ranks; moved down by 6 on
average. 5% of good answers were removed.
22 avril 2004 TALN-2004, Atelier QR 21
5. Discussion
Our approach: improves the precision of our Web-QA module less strict than looking for exact phrases as it looks for reinforcement of the semantic relation between the arguments
but, more sensitive to mistakes and wrong interpretations e.g. negations and other modal words
if we look in a much smaller corpus, such as in closed-domain QA, looking for semantic equivalences may be more fruitful.