open-domain question answering

64
1 Carnegie Mellon School of Computer Science 15-381 Lecture, Spring 2003 Copyright © 2003, Carnegie Mellon. All Rights Reserved. Open-Domain Question Answering Eric Nyberg Associate Professor [email protected]

Upload: stacie

Post on 28-Jan-2016

51 views

Category:

Documents


0 download

DESCRIPTION

Open-Domain Question Answering. Eric Nyberg Associate Professor [email protected]. Outline. What is question answering? Typical QA pipeline Unsolved problems The JAVELIN QA architecture Related research areas. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Open-Domain Question Answering

1Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Open-DomainQuestion Answering

Eric NybergAssociate [email protected]

Page 2: Open-Domain Question Answering

2Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Outline• What is question answering?• Typical QA pipeline• Unsolved problems• The JAVELIN QA architecture• Related research areas

These slides and links to other background material canbe found here: http://www.cs.cmu.edu/~ehn/15-381

Page 3: Open-Domain Question Answering

3Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

• Inputs: a question in English; a set of text and database resources

• Output: a set of possible answers drawn from the resources

QASYSTEM Text

Corpora& RDBMS

“When is the next train to Glasgow?”

“8:35, Track 9.”

Question Answering

Page 4: Open-Domain Question Answering

4Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Ancestors of Modern QA• Information Retrieval

– Retrieve relevant documents from a set of keywords; search engines

• Information Extraction– Template filling from text (e.g. event

detection); e.g. TIPSTER, MUC• Relational QA

– Translate question to relational DB query; e.g. LUNAR, FRED

Page 5: Open-Domain Question Answering

5Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

http://trec.nist.gov

Page 6: Open-Domain Question Answering

6Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Typical TREC QA Pipeline

Question

ExtractKeywords

Query

SearchEngine

Corpus

Docs

PassageExtractor

Answers

AnswerSelector

Answer

“A 50-byte passage likelyto contain the desiredanswer” (TREC QA track)

“A simple factoidquestion”

Page 7: Open-Domain Question Answering

7Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Sample Results

Mean Reciprocal Rank (MRR): Find the ordinalposition of the correct answer in your output (1st

answer, 2nd answer, etc.) and divide by one; averageover entire test suite.

Page 8: Open-Domain Question Answering

8Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Functional Evolution

• Traditional QA Systems (TREC)– Question treated like keyword query– Single answers, no understanding

Q: Who is prime minister of India?

<find a person name close to prime, minister, India (within 50 bytes)>

A: John Smith is not prime minister of India

Page 9: Open-Domain Question Answering

9Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Functional Evolution [2]

• Future QA Systems– System understands questions– System understands answers and

interprets which are most useful– System produces sophisticated

answers (list, summarize, evaluate)

What other airports are near Niletown?

Where can helicopters land close to the embassy?

Page 10: Open-Domain Question Answering

10Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Major Research Challenges

• Acquiring high-quality, high-coverage lexical resources

• Improving document retrieval• Improving document understanding• Expanding to multi-lingual corpora• Flexible control structure

– “beyond the pipeline”

• Answer Justification– Why should the user trust the answer?– Is there a better answer out there?

Page 11: Open-Domain Question Answering

11Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Why NLP is Required

• Question: “When was Wendy’s founded?”

• Passage candidate:– “The renowned Murano glassmaking industry, on an

island in the Venetian lagoon, has gone through several reincarnations since it was founded in 1291. Three exhibitions of 20th-century Murano glass are coming up in New York. By Wendy Moonan.”

• Answer: 20th Century

Page 12: Open-Domain Question Answering

12Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Predicate-argument structure

• Q336: When was Microsoft established?

• Difficult because Microsoft tends to establish lots of things…Microsoft plans to establish manufacturing partnerships in Brazil and Mexico in May.

• Need to be able to detect sentences in which `Microsoft’ is object of `establish’ or close synonym.

• Matching sentence: Microsoft Corp was founded in the US in 1975, incorporated in

1981, and established in the UK in 1982.

Page 13: Open-Domain Question Answering

13Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Why Planning is Required

• Question: What is the occupation of Bill Clinton’s wife?– No documents contain these

keywords plus the answer

• Strategy: decompose into two questions:– Who is Bill Clinton’s wife? = X– What is the occupation of X?

Page 14: Open-Domain Question Answering

14Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

JAVELIN: Justification-based Answer Valuation through Language InterpretationCarnegie Mellon Univ. (Language Technologies Institute)

OBJECTIVES• QA as planning by developing a glass box planning infrastructure• Universal auditability by developing a detailed set of labeled dependencies that form a traceable network of reasoning steps• Utility-based information fusion

PLANAddress the full Q/A task:• Question analysis - question typing, interpretation, refinement, clarification• Information seeking - document retrieval, entity and relation extraction• Multi-source information fusion - multi-faceted answers, redundancy and contradiction detection

DataRepository

JAVELIN GUI

QuestionAnalyzer

AnswerGenerator

RetrievalStrategist

ExecutionManager

...search engines &document collections

process historyand results

operator (action) models

RequestFiller

PlannerDomainModel

Page 15: Open-Domain Question Answering

15Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

JAVELIN Objectives• QA as Planning

– Create a general QA planning system– How should a QA system represent its

chain of reasoning?

• QA and Auditability– How can we improve a QA system’s

ability to justify its steps?– How can we make QA systems open

to machine learning?

Page 16: Open-Domain Question Answering

16Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

JAVELIN Objectives [2]

• Utility-Based Information Fusion– Perceived utility is a function of many

different factors– Create and tune utility metrics, e.g.:

U = Argmax k [F (Rel(I,Q,T), Nov(I,T,A), Ver(S,Sup(I,S)), Div(S), Cmp(I,A)), Cst(I,A)]

I: Info item, Q: Question, S: Source, T: Task context, A: Analyst

- relevance- novelty- veracity, support- diversity- comprehensibility- cost

Page 17: Open-Domain Question Answering

17Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Control Flow

StrategicDecisionPoints

Page 18: Open-Domain Question Answering

18Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Repository ERD

(Entity Relationship Diagram)

Page 19: Open-Domain Question Answering

19Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

JAVELIN User Interface

Page 20: Open-Domain Question Answering

20Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Javelin Architecture

DataRepository

JAVELIN GUI

QuestionAnalyzer

AnswerGenerator

RetrievalStrategist

ExecutionManager

...search engines &document collections

process historyand results

operator (action) models

InformationExtractor

PlannerDomainModel

Integrated w/XML

Modules can runon different servers

Page 21: Open-Domain Question Answering

21Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Module Integration

• Via XML DTDs for each object type• Modules use simple XML object-

passing protocol built on TCP/IP• Execution Manager takes care of

checking objects in/out of Repository

Page 22: Open-Domain Question Answering

22Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Sample LogFile Excerpt

Components communicate viaXML object representations

Page 23: Open-Domain Question Answering

23Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Question Analyzer

• Taxonomy of question-answer types and type-specific constraints

• Knowledge integration

• Pattern matching approach for this year’s evaluation

Question input (XML format)

TokenizerToken information extraction

WordnetKantoo Lexicon

Brill TaggerBBN IdentifierKANTOO lexifier

Token string input

QA taxonomy+

Type-specific constraints

Get FR?Yes

Event/entitytemplate filler

Request object builder

FR

No

KANTOO grammars

Parser

Pattern matchingRequest object builder

Request object + system result(XML format)

Page 24: Open-Domain Question Answering

24Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Question Taxonomies• Q-Types

– Express relationships between events, entities and attributes

– Influence Planner strategy

• A-Types– Express semantic

type of valid answers

Q-Type A-Type

When did the Titanic sink ?

event-completion

time-point

Who was Darth Vader's son?

concept-completion

person-name

What is thalassemia ? definition definition

We expect to add moreA-types and refine granularity

Page 25: Open-Domain Question Answering

25Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Sample of Q-Type Hierarchy

Page 26: Open-Domain Question Answering

26Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Sample of A-Type Hierarchy

Page 27: Open-Domain Question Answering

27Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

• Question type• Answer type• Computation element• Keyword set• F-structure

event-completion

person-name

order 1

first, U.S. president, appear, TV

(event(subject(person-name ?)(occupation “U.S. president”))(act appear)(order 1)(theme TV))

Request ObjectWho was the first U.S. presidentto appear on TV ?

Page 28: Open-Domain Question Answering

28Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

How the Retrieval Strategist Works

• Inputs:– Keywords and keyphrases– Type of answer desired– Resource constraints

• Min/Max documents, time, etc.

• Outputs:– Ranked set of documents – Location of keyword matches

Page 29: Open-Domain Question Answering

29Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

How the Retrieval Strategist Works

• Constructs sequences of queries based on a Request Object– Start with very constrained queries

• High quality matches, low probability of success

– Progressively relax queries until search constraints are met• Lower quality matches, high probability

of success

Page 30: Open-Domain Question Answering

30Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Sample Search Strategy

Inquery Operator Type? Query

#3 Yes #3(Titanic #syn(sink sank) *date)

#UW20 Yes #UW20(Titanic #syn(sink sank) *date)

: : : : : : : : : :

#PASSAGE250 Yes #PASSAGE250(Titanic #syn(sink sank) *date)

#SUM Yes #SUM(Titanic #syn(sink sank) *date)

*** ** **** ** * *** **** *** ** *** * * * ** **** * *** ** * ** *** *** ** *** ** ** *** * * *** ** ** *** * * ** * * **** ** *** * * * ** * ** ***** * * * ** * ** * * ***** ** **** ** * *** **** *** ** *** * * * ** **** * *** ** * ** *** *** ** *** ** ** *** * * **

*** ** **** ** * *** **** *** ** *** * * * ** **** * *** ** * ** *** *** ** *** ** ** *** * * *** ** ** *** * * ** * * **** ** *** * * * ** * ** ***** * * * ** * ** * * ***** ** **** ** * *** **** *** ** *** * * * ** **** * *** ** * ** *** *** ** *** ** ** *** * * **

*** ** **** ** * *** **** *** ** *** * * * ** **** * *** ** * ** *** *** ** *** ** ** *** * * *** ** ** *** * * ** * * **** ** *** * * * ** * ** ***** * * * ** * ** * * ***** ** **** ** * *** **** *** ** *** * * * ** **** * *** ** * ** *** *** ** *** ** ** *** * * **

Page 31: Open-Domain Question Answering

31Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Retrieval Strategist (RS):TREC Results Analysis

• Success: % of questions where at least 1 answer document was found

• TREC 2002:Success rate @ 30 docs: ~80%

@ 60 docs: ~85%@ 120 docs: ~86%

• Reasonable performance for a simple method, but room for improvement

Page 32: Open-Domain Question Answering

32Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

RS: Ongoing Improvements• Improved incremental relaxation

– Searching for all keywords too restrictive• Use subsets prioritized by discriminative ability

– Remove duplicate documents from results• Don’t waste valuable list space

– 15% fewer failures (229 test questions)• Overall success rate: @ 30 docs 83% (was 80%)

@ 60 docs 87% (was 85%)

• Larger improvements unlikely without additional techniques, such as constrained query expansion

• Investigate constrained query expansion– WordNet, Statistical methods

Page 33: Open-Domain Question Answering

33Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

What Does the Request Filler Do?

• Input:– Request Object (from QA module)– Document Set (from RS module)

• Output:– Set of extracted answers which match the

desired type (Request Fill objects)– Confidence scores

• Role in JAVELIN: Extract possible answers & passages from documents

Page 34: Open-Domain Question Answering

34Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Request Filler Steps• Filter passages

– Match answer type?– Contain sufficient keywords?

• Create variations on passages– POS tagging (Brill)– Cleansing (punctuation, tags, etc.)– Expand contractions– Reduce surface forms to lexemes

• Calculate feature values• A classifier scores the passages, which

are output with confidence scores

Page 35: Open-Domain Question Answering

35Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Features• Features are self-contained algorithms that

score passages in different ways• Example: Simple Features

– # Keywords present– Normalized window size– Average <Answer,Keywords> distance

• Example: Pattern Features– cN [..] cV [..] in/on [date]– [date], iN [..] cV [..]

• Any procedure that returns a numeric value is a valid feature!

Page 36: Open-Domain Question Answering

36Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Learning An Answer Confidence Function

• Supervised learning– Answer type-specific model– Aggregate model across answer types

• Decision Tree – C4.5– Variable feature dependence– Fast enough to re-learn from each

new instance

Page 37: Open-Domain Question Answering

37Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

A When Q-Type Decision Tree% Keywords present in

the passage

% Keywords present in

the passage

Average distance<date, keywords>

Maximum scaledkeyword

window size

> 0.75

> 0.2.4

33.0/10.35.0/1

0.75

0.75

876.0/91.8

> 0.2 0.2 60> 60

62.0/11.6

Page 38: Open-Domain Question Answering

38Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

• The company said it believes the expenses of the restructuring will be recovered by the end of 1992

• …the artist expressed

• The company said it believes …

• It is a misconception the Titanic sank on April the 15th,1912 …

• The/DT company/NN say/VBD it/PRP believe/VBZ the/DT expense/NNS of/IN the/DT restructuring/NN will/MD be/VB recover/VBN by/IN the/DT end/NN of/IN 1992/CD

• … the performer expressed

• Microsoft said it believes …

• The Titanic sank on April the 15th,1912 …

Semantic Analysis Would Help

Page 39: Open-Domain Question Answering

39Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Information Extractor (IX):TREC Analysis

Inputs Answer in top 5

Answerin docset

Trec 8 200 71 189

Trec 9 693 218 424

Trec 10 500 119 313

If the answer is in the doc set returned by the RetrievalStrategist, does the IX module identify it as an answercandidate with a high confidence score?

Page 40: Open-Domain Question Answering

40Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

IX: Current & Future Work

• Enrich feature space beyond surface patterns & surface statistics

• Perform AType-specific learning• Perform adaptive semantic expansion• Enhance training data quantity/quality• Tune objective function

Page 41: Open-Domain Question Answering

41Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

NLP for Information Extraction

• Simple statistical classifiers are not sufficient on their own

• Need to supplement statistical approach with natural language processing to handle more complex queries

Page 42: Open-Domain Question Answering

42Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Example question• Question: “When was Wendy’s founded?”

• Question Analyzer extended output:– { temporal(?x), found(*, Wendy’s) }

• Passage discovered by retrieval module:– “R. David Thomas founded Wendy’s in 1969, …”

• Conversion to predicate form by Passage Analyzer:– { founded(R. David Thomas, Wendy’s), DATE(1969), … }

• Unification of QA literals against PA literals:– Equiv(found(*,Wendy’s),

founded(R. David Thomas, Wendy’s))– Equiv(temporal(?x),

DATE(1969))– ?x := 1969

• Answer: 1969

Page 43: Open-Domain Question Answering

43Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Answer Generator

• Currently last module in pipe-line.

• Main tasks:

– Combination of different sorts of evidence for answer verification.

– Detection and combination of similar answer candidates to address answer granularity.

– Initiation of processing loops to gather more evidence.

– Generation of answers in required format.

Page 44: Open-Domain Question Answering

44Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Answer Generator input• Analyzed question (RequestObject):

– Question/Answer type (qtype/atype)

– Number of expected answers;

– Syntactic parse and keywords.

• Passages (RequestFills):

– Marked candidates of right semantic type (right NE type);

– Confidences computed using set of text-based (surface) features such as keyword placement.

Page 45: Open-Domain Question Answering

45Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Answer Generator output• Answer string from document (for now).

• Set of text passages (RequestFills) Answer Generator decided were supportive of answer.

• Or, requests for more information (exceptions) passed on to Planner:

– “Not enough answer candidates”

– “Can’t distinguish answer candidates”

Page 46: Open-Domain Question Answering

46Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Types of evidence• Currently implemented: Redundancy, frequency

counts.– Preference given to more often occurring, normalized

answer candidates.

• Next step: Structural information from parser.– Matching question and answer predicate-argument

structure.

– Detecting hypotheticals, negation, etc.

• Research level: Combining collection-wide statistics with ‘symbolic’ QA.– Ballpark estimates of temporal boundaries of

events/states.

Page 47: Open-Domain Question Answering

47Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Example• Q: What year did the Titanic sink?

A: 1912

Supporting evidence:It was the worst peacetime disaster involving a British ship since the Titanic sank on the 14th of April, 1912.

The Titanic sank after striking an iceberg in the North Atlantic on April 14th, 1912.

The Herald of Free Enterprise capsized off the Belgian port of Zeebrugge on March 6, 1987, in the worst peacetime disaster involving a British ship since the Titanic sank in 1912.

Page 48: Open-Domain Question Answering

48Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

What happened?• Different formats for answer candidates detected,

normalized and combined:

– `April 14th, 1912’

– `14th of April, 1912’

• Supporting evidence detected and combined:

– `1912’ supports `April 14th, 1912’

• Structure of date expressions understood and correct piece output:

– `1912’ rather than `April 14th, 1912’

• Most frequent answer candidate found and output:

– `April 14th, 1912’ rather than something else.

Page 49: Open-Domain Question Answering

49Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Answer Normalization• Request Filler/Answer Generator aware of NE types:

dates, times, people names, company names, locations, currency expressions.

• `April 14th, 1912’, `14th of April 1912’, `14 April 1912’ instances of same date, but different strings.

• For date expressions, normalization performed to ISO 8601 (YYYY-MM-DD) in Answer Generator.

• ‘summer’, ‘last year’, etc. remain as strings.

Page 50: Open-Domain Question Answering

50Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Answer Normalization• Normalization enables comparison and detection of

redundant or complementary answers.

• Define supporting evidence as piece of text expressing same or less specific information.

• E.g., `1912’ supports `April 12th, 1912’.

• Complementary evidence: ‘1912’ complements ‘April 12th’.

• Normalization and supporting extend to other NE types: – `Clinton’ supports `Bill Clinton’;

– `William Clinton’ and `Bill Clinton’ are normalized to same.

– For locations, `Pennsylvania’ supports `Pittsburgh’.

Page 51: Open-Domain Question Answering

51Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Other forms of evidence• Q: Name all the bills that were passed during the Bush

administration.

• Not likely to find passages mentioning `bill’, `pass’, `Bush administration’.

• When was Bush administration??

• `Symbolic’ QA: look for explicit answer in collection, might not be present.

• `Statistical’ QA: look at distribution of documents mentioning Bush administration.

• Combining evidence of different sorts!

Page 52: Open-Domain Question Answering

52Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Other forms of evidence• Can we figure out if Bush administration was around

when document was written?

• Look at tense/aspect/wording.

• Forward time references

– Bush administration will do something

• Backward time references

– Bush administration has done something

• Hypothesis: – Backward time references provide information about

onset of event;– Forward time references provide information about end of

event.

Page 53: Open-Domain Question Answering

53Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Other forms of evidence• Bush administration forward references

AdministrationchangeEvent end

Time stamps

#docsmentioningBush adm.on givenday

Page 54: Open-Domain Question Answering

54Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Other forms of evidence• Bush administration backward references

#docsmentioningBush adm.on given day

Time stamps

AdministrationchangeEvent onset

Page 55: Open-Domain Question Answering

55Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Planning in JAVELIN• Enable generation of new question-

answering strategies at run-time

• Improve ability to recover from bad decisions as information is collected

• Gain insight into when different QA components are most useful

Page 56: Open-Domain Question Answering

56Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Planner Integration

exe E

DomainModel

Planner

DataRepository

JAVELIN

GUI

module A

ExecutionManager

process history and data

JAVELIN operator (action) models

module E

module F

...

question

answer

ack

...

dialog

response

exe A

results

results

exe F

results

store

Page 57: Open-Domain Question Answering

57Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Current Domain Operators

RESPOND_TO_USERpre: (and (interactive_session) (request ?q ?ro) (ranked_answers ?ans ?ro ?fills) (> (max_ans_score ?ans) 0.1) (> answer_quality 0))

ASK_USER_FOR_ANSWER_TYPEpre: (and (interactive_session) (request ?q ?ro) (or (and (ranked_answers ?ans ?ro ?fills) (< (max_ans_score ?ans) 0.1))

(no_docs_found ?ro) (no_fills_found ?ro ?docs)))

ASK_USER_FOR_MORE_KEYWORDSpre: (and (interactive_session) (request ?q ?ro) (or (and (ranked_answers ?ans ?ro ?fills) (< (max_ans_score ?ans) 0.1)) (no_docs_found ?ro)

(no_fills_found ?ro ?docs)))

• QuestionAnalyzer module called as a precursor to planning• Demonstrates generation of multiple search paths, feedback loops

RETRIEVE_DOCUMENTSpre: (and (request ?q ?ro) (> (extracted_terms ?ro) 0) (> request_quality 0))

EXTRACT_DT_CANDIDATE_FILLSpre: (and (retrieved_docs ?docs ?ro) (== (expected_atype ?ro) location_t) (> docset_quality 0.3))

EXTRACT_KNN_CANDIDATE_FILLSpre: (and (retrieved_docs ?docs ?ro) (!= (expected_atype ?ro) location_t) (> docset_quality 0.3))

RANK_CANDIDATESpre: (and (candidate_fills ?fills ?ro ?docs) (> fillset_quality 0))

Page 58: Open-Domain Question Answering

58Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Current Domain OperatorsRETRIEVE_DOCUMENTS (?q - question ?ro - qtype)pre: (and (request ?q ?ro) (> (extracted_terms ?ro) 0) (> request_quality 0))

dbind: ?docs (genDocsetID) ?dur (estTimeRS (expected_atype ?ro)) ?pnone (probNoDocs ?ro) ?pgood (probDocsHaveAns ?ro) ?dqual (estDocsetQual ?ro))

effects: (?pnodocs ((no_docs_found ?ro)(scale-down request_quality 2)(assign docset_quality 0)(increase system_time ?dur))

?pgood ((retrieved_docs ?docs ?ro)(assign docset_quality ?dqual)(increase system_time ?dur))

(1-?pgood-?pnone) ((retrieved_docs ?docs ?ro) (scale-down request_quality 2) (assign docset_quality 0) (increase system_time ?dur)))

execute: (RetrievalStrategist ?docs ?ro 10 15 300)

more detailed operator view...

Page 59: Open-Domain Question Answering

59Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Illustrative ExamplesWhere is bile produced?

• Overcomes current limitations of system “location” knowledge• Uses answer candidate confidence to trigger feedback loop

<RETRIEVE_DOCUMENTS RetrievalStrategist DS2216 RO2262 10 15 300><EXTRACT_DT_CANDIDATE_FILLS DTRequestFiller FS2216 RO2262 DS2216 900><RANK_CANDIDATES AnswerGenerator AL2196 RO2262 FS2216 180> <ASK_USER_FOR_ANSWER_TYPE AskUserForAtype Q74050 RO2262> <ASK_USER_FOR_MORE_KEYWORDS AskUserForKeywords Q74050 RO2262><RETRIEVE_DOCUMENTS RetrievalStrategist DS2217 RO2263 10 15 300> <EXTRACT_KNN_CANDIDATE_FILLS KNNRequestFiller FS2217 RO2263 DS2217 900> <RANK_CANDIDATES AnswerGenerator AL2197 RO2263 FS2217 180> <RESPOND_TO_USER RespondToUser A2204 AL2197 Q74050 RANKED>

1st iter

2nd iter

Top 3 answers found during initial pass (with “location” answer type)

1: Moscow (Conf: 0.01825)2: China (Conf: 0.01817)3: Guangdong Province (Conf: 0.01817)

Top 3 answers displayed (with user-specified “object” answer type; ‘liver’ ranked 6th)

1: gallbladder (Conf: 0.58728)2: dollars (Conf: 0.58235)3: stores (Conf: 0.58147)

Page 60: Open-Domain Question Answering

60Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Illustrative ExamplesWho invented the road traffic cone?

• Overcomes current inability to relax phrases during document retrieval• Uses answer candidate confidence scores to trigger feedback loop

1st iter

2nd iter

<RETRIEVE_DOCUMENTS RetrievalStrategist DS2221 RO2268 10 15 300><EXTRACT_KNN_CANDIDATE_FILLS KNNRequestFiller FS2221 RO2268 DS2221 900><RANK_CANDIDATES AnswerGenerator AL2201 RO2268 FS2221 180><ASK_USER_FOR_ANSWER_TYPE AskUserForAtype Q74053 RO2268><ASK_USER_FOR_MORE_KEYWORDS AskUserForKeywords Q74053 RO2268><RETRIEVE_DOCUMENTS RetrievalStrategist DS2222 RO2269 10 15 300><EXTRACT_KNN_CANDIDATE_FILLS KNNRequestFiller FS2222 RO2269 DS2222 900><RANK_CANDIDATES AnswerGenerator AL2202 RO2269 FS2222 180><RESPOND_TO_USER RespondToUser A2207 AL2202 Q74053 RANKED>

1: Colvin (Conf: 0.0176)2: Vladimir Zworykin (Conf: 0.0162)3: Angela Alioto (Conf: 0.01483)

Top 3 answers found during initial pass (using terms ‘invented’ and ‘road traffic cone’)

Top 3 answers displayed (with additional user-specified term ‘traffic cone’; correct answer is ‘David Morgan’)

1: Morgan (Conf: 0.4203)2: Colvin (Conf: 0.0176)3: Angela Alioto (Conf: 0.01483)

Page 61: Open-Domain Question Answering

61Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Multilingual Question Answering• Goals

– English questions– Multilingual information sources (Jpn/Chi)– English/Multilingual Answers

• Extensions to existing JAVELIN modules– Question Analyzer– Retrieval Strategist– Information Extractor– Answer Generator

Page 62: Open-Domain Question Answering

62Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

RS

MultilingualArchitecture

AnswerGenerator

JapaneseIndex

ChineseIndex

InformationExtractor3

(Chinese)

QuestionAnalyzer

OtherIndex

EnglishIndex

Answers?’s

BilingualDictionary

Module

Machinexlation

InformationExtractor1

(English)

InformationExtractor2

(Japanese)

InformationExtractor4

(other lang)

EncodingConverter

Japanesecorpora

Chinesecorpora

other langcorpora

Englishcorpora

Page 63: Open-Domain Question Answering

63Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

15-381 Project Topics• Create more, better RF/IX modules

– More intelligent feature extractors– Smarter classifiers– Train on different answer types– Plug in and evaluate your work in the

context of the larger system

• End-to-end QA system– Focus on a particular question type– Utilize existing RS module for document

retrieval– Evaluate on TREC test suites (subsets)

Page 64: Open-Domain Question Answering

64Carnegie MellonSchool of Computer Science 15-381 Lecture, Spring 2003

Copyright © 2003, Carnegie Mellon. All Rights Reserved.

Questions?