generation of referring expressions (gre)

43
Generation of Referring Expressions (GRE) Reading: Dale & Reiter (1995) (key paper in this area)

Upload: lucian-burnett

Post on 01-Jan-2016

33 views

Category:

Documents


1 download

DESCRIPTION

Generation of Referring Expressions (GRE). Reading: Dale & Reiter (1995) (key paper in this area). The task: GRE. NLG can have different kinds of inputs: ‘Flat’ data (collections of atoms, e.g., in the tables of a database) Logically complex data - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Generation of Referring Expressions (GRE)

Generation of Referring Expressions (GRE)

Reading: Dale & Reiter (1995) (key paper in this area)

Page 2: Generation of Referring Expressions (GRE)

The task: GRE

• NLG can have different kinds of inputs:– ‘Flat’ data (collections of atoms, e.g., in the

tables of a database)– Logically complex data

• In both cases, unfamiliar constants may be used, and this is sometimes unavoidable

Page 3: Generation of Referring Expressions (GRE)

No familiar constant available:

1. The referent has a familiar name, but it’s not unique, e.g., ‘John Smith’

2. The referent has no familiar name: trains, furniture, trees, atomic particles, …

( In such cases, databases use database keys,

e.g., ‘Smith$73527$’, ‘TRAIN-3821’ )

3. Similar: sets of objects (lecture 4).

Page 4: Generation of Referring Expressions (GRE)

• Natural Languages are too economic to have a proper name for everything

• Names may not even be most appropriate

• So, speakers/NLG systems have to invent ways of referring to things. E.g., ‘the 7:38 Trenton express’

• Note: the problem arises whether the referent is a token or a type

Page 5: Generation of Referring Expressions (GRE)

• GRE tries to find the best description

• GRE is microcosm of NLG: e.g., determines– which properties to express

(Content Determination)– which syntactic configuration to use

(Syntactic Realization)– which words to choose

(Lexical Choice)

Page 6: Generation of Referring Expressions (GRE)

This lecture:

• Simplification 1: Content Determination only (until lecture 5).

• Simplification 2: Definite descriptions only (Pronouns, demonstratives, etc., are disregarded; until tomorrow)

Page 7: Generation of Referring Expressions (GRE)

Dale & Reiter (1995): best description fulfills the Gricean maxims. E.g.,

• (Quality:) list properties truthfully• (Quantity:) list sufficient properties to allow

hearer to identify referent – but not more• (Relevance:) use properties that are of

interest in themselves *• (Manner:) be brief

* Slightly different from D&R 1995

Page 8: Generation of Referring Expressions (GRE)

D&R’s expectation:

• Violation of a maxim leads to implicatures.

• For example,– [Quantity] ‘the pitbull’ (when there is

only one dog).– [Manner] ‘Get the cordless drill that’s

in the toolbox’ (Appelt).

• There’s just one problem: …

Page 9: Generation of Referring Expressions (GRE)

…people don’t speak this way

For example,– [Manner] ‘the red chair’ (when there is

only one red object in the domain).

– [Manner/Quantity] ‘I broke my arm’ (when I have two).

General: empirical work shows much redundancy

Similar for other maxims, e.g.,– [Quality] ‘the man with the martini’ (Donellan)

Page 10: Generation of Referring Expressions (GRE)

Example Situation

a, £100 b, £150

c, £100 d, £150 e, £?Swedish Italian

Page 11: Generation of Referring Expressions (GRE)

Formalized in a KB

• Type: furniture (abcde), desk (ab), chair (cde)

• Origin: Sweden (ac), Italy (bde)

• Colours: dark (ade), light (bc), grey (a)

• Price: 100 (ac), 150 (bd) , 250 ({})

• Contains: wood ({}), metal ({abcde}), cotton(d)

Assumption: all this is shared knowledge.

Page 12: Generation of Referring Expressions (GRE)

Violations of …• Manner:

* ‘The £100 grey Swedish desk which is made of metal’

(Description of a)

• Relevance: ‘The cotton chair is a fire hazard?

?Then why not buy the Swedish chair?’ (Descriptions of d and c respectively)

Page 13: Generation of Referring Expressions (GRE)

• In fact, there is a second problem with Manner. Consider the following formalization:

Full Brevity: Never use more than the minimal number of properties required for identification (Dale 1989)

An algorithm:

Page 14: Generation of Referring Expressions (GRE)

Dale 1989:

1. Check whether 1 property is enough

2. Check whether 2 properties is enough

….

Etc., until

success {minimal description is generated} or

failure {no description is possible}

Page 15: Generation of Referring Expressions (GRE)

Problem: exponential complexity

• Worst-case, this algorithm would have to inspect all combinations of properties. n properties combinations.

• Recall: one grain of rice on square one; twice as many on any subsequent square.

• Some algorithms may be faster, but …

• Theoretical result: algorithm must be exponential in the number of properties.

n2

Page 16: Generation of Referring Expressions (GRE)

• D&R conclude that Full Brevity cannot be achieved in practice.

• They designed an algorithm that only approximates Full Brevity:

the Incremental Algorithm.

Page 17: Generation of Referring Expressions (GRE)

Incremental Algorithm (informal):

• Properties are considered in a fixed order:

P =

• A property is included if it is ‘useful’:

true of target; false of some distractors

• Stop when done; so earlier properties have a greater chance of being included. (E.g., a perceptually salient property)

• Therefore called preference order.

nPPPP ,...,,, 321

Page 18: Generation of Referring Expressions (GRE)

• r = individual to be described

• P = list of properties, in preference order

• P is a property

• L= properties in generated description

(Recall: we’re not worried about realization today)

Page 19: Generation of Referring Expressions (GRE)

FailureReturn

LReturn then {r}C If

]][[C:C

}{L:L

do then ]][[ C &]][[r If

:do P allFor

Domain:C

Φ:L

P

P

PP

P

Page 20: Generation of Referring Expressions (GRE)

P = < furniture (abcde), desk (ab), chair (cde), Swedish (ac), Italian (bde), dark (ade), light (bc), grey (a), 100£ ({ac}), 150£(bd) , 250£ ({}), wooden ({}), metal (abcde), cotton ({d}) >

Domain = {a,b,c,d,e} . Now describe:a = <...>d = <...>e = <...>

Page 21: Generation of Referring Expressions (GRE)

P = < furniture (abcde), desk (ab), chair (cde), Swedish (ac), Italian (bde), dark (ade), light (bc), grey (a), 100£ (ac),200£ (bd),250£ ({}), wooden ({}), metal (abcde), cotton (d) >

Domain = {a,b,c,d,e} . Now describe: a = <desk {ab}, Swedish {ac}> d = <chair,Italian,dark,200> (Nonminimal) e = <chair,Italian,dark, ....> (Impossible)

Page 22: Generation of Referring Expressions (GRE)

[ An aside: shared info will be assumed to be complete and uncontroversial. Consider

• Speaker:

[[Student]] = {a,b,…}

• Hearer:

[[Student]] = {a,c,…}

Does this make a referable? ]

Page 23: Generation of Referring Expressions (GRE)

Incremental Algorithm

• It’s a hillclimbing algorithm: ever better approximations of a successful description.

• ‘Incremental’ means no backtracking.

• Not always the minimal number of properties.

Page 24: Generation of Referring Expressions (GRE)

Incremental Algorithm

• Logical completeness: A unique description is found in finite time if there exists one. (Given reasonable assumptions, see van Deemter 2002)

• Computational complexity: Assume thattesting for usefulness takes constant time.Then worst-case time complexity is O(np) where np is the number of properties in P.

Page 25: Generation of Referring Expressions (GRE)

Better approximation of Full Brevity(D&R 1995)

• Attribute + Value model: Properties grouped together as in original example:

Origin: Sweden, Italy, ...

Colour: dark, grey, ...

• Optimization within the set of properties based on the same Attribute

Page 26: Generation of Referring Expressions (GRE)

Incremental Algorithm, using Attributes and Values

• r = individual to be described

• A = list of Attributes, in preference order

• Def: = Value i of Attribute j

• L= properties in generated description

jiV ,

Page 27: Generation of Referring Expressions (GRE)

FailureReturn

LReturn then {r}C If

]][[VC:C

}{VL:L

do then ]]V[[ C &]]V[[r If

)A(r,estValueBFindV

:doA A allFor

Domain:C

Φ:L

ji,

ji,

ji,ji,

iji,

i

Page 28: Generation of Referring Expressions (GRE)

• FindBestValue(r,A):

- Find Values of A that are true of r,

while removing some distractors

(If these don’t exist, go to next Attribute)

- Within this set, select the Value that

removes the largest number of distractors

- If there’s a tie, select the most general one

- If there’s still a tie, select an arbitrary one

Page 29: Generation of Referring Expressions (GRE)

Example: D = {a,b,c,d,f,g}

• Type: furniture (abcd), desk (ab), chair (cd)

• Origin: Europe (bdfg), USA (ac), Italy (bd)

Describe a: {desk, American}

(furniture removes fewer distractors than desk)

Describe b: {desk, European}

(European is more general than Italian)

N.B. This disregards relevance, etc.

Page 30: Generation of Referring Expressions (GRE)

P.S. Note the similarity with Van Rooy & Dekker’s semantic of answers: Let A and B be truthful answers to a question, then

A is a better answer than B Utility(A) > Utility(B) or Utility(A) = Utility(B) & B A

(More about this in the next lecture …)

Page 31: Generation of Referring Expressions (GRE)

• Exercise on Logical Completeness: Construct an example where no description is found, although one exists.

• Hint: Let Attribute have Values whose extensions overlap.

Page 32: Generation of Referring Expressions (GRE)

Example: D = {a,b,c,d,f}

• Contains: wood (abe), plastic (acdf)

• Colour: grey (ab), yellow (cd)

Describe a: {wood, grey, ...} - Failure

(wood removes more distractors than plastic)

Compare:

Describe a: {plastic, grey} - Success

Page 33: Generation of Referring Expressions (GRE)

Complexity of the algorithm

nd = nr. of distractors

nl = nr. of properties in the description

nv = nr. of Values (for all Attributes)

Alternative assessment: O(nv)

(Worst-case running time)

According to D&R: O(nd nl )

(Typical running time)

Page 34: Generation of Referring Expressions (GRE)

Minor complication: Head nouns

• Another way in which human descriptions are nonminimal

– A description needs a Noun, but not all properties are expressed as Nouns

– Example: Suppose Colour was the

most-preferred Attribute, and target = a

Page 35: Generation of Referring Expressions (GRE)

• Colours: dark (ade), light (bc), grey (a)• Type: furniture (abcde), desk (ab), chair (cde)• Origin: Sweden (ac), Italy (bde)• Price: 100 (ac), 150 (bd) , 250 ({})• Contains: wood ({}), metal ({abcde}), cotton(d)

target = aDescribe a: {grey}‘The grey’ ? (Not in English)

Page 36: Generation of Referring Expressions (GRE)

D&R’s repair:

• Assume that Values of the AttributeType can be expressed in a Noun.

• After the core algorithm:

- check whether Type is represented.

- if not, then add the best Value of the Type Attribute to the description

Page 37: Generation of Referring Expressions (GRE)

• Versions of Dale and Reiter’s Incremental Algorithm have often been implemented

• Still the starting point for many new algorithms. (See later lectures.)

• Worth reading!

Page 38: Generation of Referring Expressions (GRE)

Limitations of the algorithm

1. Redundancy does not arise for principled reasons, e.g., for

- marking topic changes, etc.

(Corpus work by Pam Jordan et. al.)

- making it easy to find the referent

(Experimental work by Paraboni et al. - Next lecture)

Page 39: Generation of Referring Expressions (GRE)

Limitations of the algorithm

2. Targets are individual objects, never sets. What changes when target = {a,b,c} ?(Lecture 4)

3. Incremental algorithm uses only conjunctions of atomic properties. No negations, disjunctions, etc. (Lecture 4)

Page 40: Generation of Referring Expressions (GRE)

Limitations of the algorithm

4. No relations with other objects, e.g., ‘the orange on the table’. (Lecture 3)

5. Differences in salience are not taken into account. (Lecture 3)

6. Language realization is disregarded. (Lecture 5)

Page 41: Generation of Referring Expressions (GRE)

Discussion: How bad is it for a GRE algorithm to take exponential time?

– More complex types of referring expressions problem becomes even harder

– Restrict to combinations whose length is less than x problem not exponential.

– Example: descriptions containing a most n properties (Full Brevity)

Page 42: Generation of Referring Expressions (GRE)

However:

– “Mathematicians’ view”: structure of problem shows when no restrictions are put.

– What if the input does not conform withthese restrictions?

(GRE does not control its own input!)

Page 43: Generation of Referring Expressions (GRE)

• Compare with Description Logic:- Increasingly complex algorithms …- that tackle larger and larger fragments of logic …- and whose complexity is ‘conservative’

• Question: how do human speakers cope?