representing meaning: verb and sentence semantics
TRANSCRIPT
![Page 1: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/1.jpg)
Representing Meaning: Verb and Sentence Semantics
Lecture 18
![Page 2: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/2.jpg)
Semantics Road Map
1. Lexical semantics2. Vector semantics3. Meaning representation languages and semantic roles4. Compositional semantics, semantic parsing5. Discourse and pragmatics
![Page 3: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/3.jpg)
Learning Objectives
• Be able to distinguish between extension and intension• Be able to apply the criteria for a good MRL (meaning
representation language)• Be able to represent simple English statements in first order logic
(FOL)• Classical representation• Neo-Davidsonian representation
• Be able to state the strengths and weakness of two different classes of MRLs (FOL and description logics)• Be able to explain the uses and value of semantic role labels• Be able to provide a basic algorithm for semantic role labeling
![Page 4: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/4.jpg)
Intension and Extension
![Page 5: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/5.jpg)
Two Approaches to Semantics
• Intentional• Assumes that the word or utterance is intrinsically meaningful• Decompositional approaches to lexical semantics are intentional
• Extentional• Defines words and utterances by the the things in the world of which
they are true• This lecture will concern extentional models of semantics
![Page 6: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/6.jpg)
Word Meanings are Predicates and Predicates are Sets
{ }RED =
The semantics of red(the word) are RED
(the predicate)
The predicate is defined in terms of things in the world, not intrinsic qualities
![Page 7: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/7.jpg)
RED(x) is the assertion that x ∈ RED
• HIT is the set of <x, y> pairs where x hit y• Asserting that Alan hit David or HIT(Alan, David) is asserting
that <Alan, David> is an element of the set HIT• Such assertions can be either true or false
![Page 8: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/8.jpg)
This Lecture has Two Goals
• We will look at ways of representing the extension of verbs and sentences• We will also look at semantic roles and how they relate to
meaning representation languages (MRLs)
![Page 9: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/9.jpg)
Desirable Properties of Meaning Representations
![Page 10: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/10.jpg)
Meaning Representation?
For what kinds of tasks?• Answering essay questions on an exam• Deciding what to order at a restaurant• Learning an activity via instructions• Making an investment decision• Recognizing an insult
linguistic inputs
results of parsing/WSD/cor
ef/SRL/etc.
meaning representation
non-linguistic domains
![Page 11: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/11.jpg)
Desirable Qualities: Verifiability
We want to be able to determine the truth of our representations.
“Does Udipi serve vegetarian food”?Is SERVE(Udipi, vegetarian food) in our knowledge base?
What is the relationship between the meaning of a sentence and the world as we know it?
![Page 12: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/12.jpg)
Desirable Qualities: Unambiguous Representation
Let’s eat somewhere near campus.
(e.g., we want to eat at a place near campus)
(e.g., we eat places)
Our MRL must capture precisely one of these meanings—not both.
![Page 13: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/13.jpg)
Desirable Qualities: Canonical Form
• “Mad Mex has vegetarian dishes.”• “They have vegetarian food at Mad Mex.”• “Vegetarian dishes are served at Mad Mex.”• “Mad Mex serves vegetarian fare.”
Inputs that mean the same thing should have the same meaning representation.
![Page 14: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/14.jpg)
Desirable Qualities: Inference, Variables, and Expressiveness• “Can vegetarians eat at Mad Mex?”
• “I’d like to find a restaurant where I can get vegetarian food.”SERVE(x, vegetarian food)
• “Delta flies Boeing 737s from Boston to New York.”
![Page 15: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/15.jpg)
One Limitation: LiteralityWe will focus on the basic requirements for meaning representation.
The basic requirements do not include correctly interpreting statements like:• “Ford was hemorrhaging money.”• “I could eat a horse.”
![Page 16: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/16.jpg)
What entities do we want to represent?
A meaning representation scheme should let us represent:• objects (e.g., people, restaurants, cuisines)• properties of objects (e.g., pickiness, noisiness, spiciness)• relations between objects (e.g., SERVE(Oishii Bento,
Japanese))
![Page 17: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/17.jpg)
The Knowledge Base
Our knowledge base
It contains the things that we “know”
We can query it
![Page 18: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/18.jpg)
The Candidates
![Page 19: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/19.jpg)
“I have a car.”
![Page 20: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/20.jpg)
First-Order Logic
![Page 21: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/21.jpg)
MRL #1: First-Order LogicDressCode(ThePorch)Cuisine(Udipi)
SERVES(UnionGrill, AmericanFood)RESTAURANT(UnionGrill)
• HAVE(Speaker, FiveDollars) ∧ ¬ HAVE(Speaker, LotOfTime)• ∀x PERSON(x) ⇒ HAVE(x, FiveDollars)• ∃x,y PERSON(x) ∧ RESTAURANT(y) ∧ ¬HASVISITED(x,y)
Functions
Predicates
![Page 22: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/22.jpg)
First Order Logic and Semantics
• Nouns correspond to one-place predicates:RESTAURANT(x) is true if x is a member of the set of restaurants
• Adjectives correspond to one-place predicates:VEGETARIAN(x) is true if x is a member of the set of things that are vegetarian
• Verbs correspond to one-place, two-place, or three-place predicatesDINE(x) as in Noah dined.EAT(x, y) as in Noah ate American food.GIVE(x, y, z) as in The bad sushi gave Noah a stomach ache.
![Page 23: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/23.jpg)
As individual facts are added to a knowledge base, modus ponens can be used to fire applicable implication rules
.
<latexit sha1_base64="xaOK769s7M7s4w50+CCfpEjQ07Y=">AAAC73icjVJNixNBEO0ZdV3jV1aPXhqDkoCEyV50bwuCeFzF7C5kQqjpqUya7ekeumuyCcP8Ci8eVLz6d7z5b+xJQtBExIKGx3v1XZ0USjqKop9BeOPmrYPbh3dad+/df/CwffTo3JnSChwKo4y9TMChkhqHJEnhZWER8kThRXL1utEv5midNPoDLQsc55BpOZUCyFOTo+AgTjCTugIlM41p3XrO4zwxiyomXJAT1RwzJLAStEVHUFrQVHeHqSxkr47jxn9qLCjFF/8RWHcXPR6/l9nMS9Zc7xRzaOfoGq8XG+V8m+WNMWnd8xXjWTPtXqPb2DW96rD+VxrU6XbsSbsT9aOV8X0w2IAO29jZpP0jTo0oc9QkFDg3GkQFjSuwJIXCuhWXDgsQV5DhyEMNObpxtTpYzZ95JuV+bf5p4iv294gKcueWeeI9c6CZ29Ua8m/aqKTpq3EldVESarEuNC0VJ8Ob6/NUWhSklh6AsNL3ysUMLAjyf6TllzDYHXkfDI/7J/3Bu6hz2t1s45A9YU9Zlw3YS3bK3rIzNmQiyIOPwefgS2jDT+HX8NvaNQw2MY/ZHxZ+/wUk4vL8</latexit>
An example of modus ponens:
![Page 24: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/24.jpg)
First Order Logic: Advantages
• Flexible• Well-understood• Widely used
![Page 25: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/25.jpg)
Description Logics
![Page 26: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/26.jpg)
MRL #2: Description Logics
• Goal of description logics: understand and specify semantics for slot-filler representations• More restrictive than FOL
![Page 27: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/27.jpg)
TBox and Abox are the components of a traditional KB• TBox: contains the knowledge about categories or concepts in
the application domainAll bistros are restaurantsAll restaurants are businesses
• ABox: facts about individuals in the domainUdipi is an Indian restaurant
![Page 28: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/28.jpg)
Categories and Subsumption
IndianRestaurant(Udipi)category domain element
Udipi is an Indian restaurant.
IndianRestaurant ⊑ Restaurantsubsumed subsumerAll Indian restaurants are restaurants.
![Page 29: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/29.jpg)
Negation and Disjunction
IndianRestaurant ⊑ not ItalianRestaurantIndian restaurants can’t also be Italian restaurants.
Restaurant ⊑ (or ItalianRestaurant IndianRestaurantMexicanRestaurant)Restaurants are Italian restaurants, Indian restaurants, or Mexican restaurant.
![Page 30: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/30.jpg)
What are the relative advantages and disadvantages of FOL and description logics?
First Order Logic Description Logics
![Page 31: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/31.jpg)
Semantic Roles and Semantic Role Labeling
![Page 32: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/32.jpg)
• Noah built an ark out of gopher wood.• He loaded two of every animal onto the ark.• Noah piloted the ark into stormy weather.• When the skies cleared, all rejoiced.
![Page 33: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/33.jpg)
• Noah1 built an ark2 out of gopher wood.• He1 loaded two of every animal onto the ark2.• Noah1 piloted the ark2 into stormy weather.• When the skies3 cleared, all4 rejoiced.
![Page 34: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/34.jpg)
Paraphrase
• Noah built an ark out of gopher wood.• An ark was built by Noah. It was made from gopher wood.• Noah constructed an ark with wood from a gopher tree.• Using gopher wood, Noah managed to put together an ark.• Noah built an ark.• …
![Page 35: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/35.jpg)
Traditional Semantic Roles Labels are Useful for Learning about Semantic Roles• In the linguistics literature, one sees a number of common terms for semantic roles
• Agent• Patient• Theme• Force• Experiencer• Stimulus• Recipient• Source• Goal• etc.
• These have their place, and are useful to know if you want to understand what a semantic role is, but are not widely used in NLP
• In NLP, we tend to use finer-grained (and sometimes cryptically named) semantic role labels
![Page 36: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/36.jpg)
Traditional Semantic Roles
• David threw the midterms from Pausch Bridge to the hillside below.• David—agent• the midterms—theme• Pausch Bridge—source• the hillside below—goal
![Page 37: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/37.jpg)
Neo-Davidsonian Representations Treat Events as Entities• David threw the midterms from Pausch Bridge to the hillside
below• THROW(David, midterms, PauschBridge, hillside)• ∃e THROW(e) ∧ AGENT(e, David) ∧ THEME(e, midterms) ∧ SOURCE(e,
PauschBridge) ∧ GOAL(e, hillside)
• The midterms were thrown from Pausch Bridge• THROW(midterms, PauschBridge)• ∃e THROW(e) ∧ THEME(e, midterms) ∧ SOURCE(e, PauschBridge)
![Page 38: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/38.jpg)
Semantic Role Labeling
Input: a sentence, paragraph, or documentOutput: for each predicate*, labeled spans identifying each of its arguments.
*Predicates are sometimes identified in the input, sometimes not.
![Page 39: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/39.jpg)
Predicates
• Noah built an ark out of gopher wood.• An ark was built by Noah. It was made from gopher wood.• Noah constructed an ark with wood from a gopher tree.• Using gopher wood, Noah managed to put together an ark.
![Page 40: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/40.jpg)
Predicates and Arguments
• Noah built an ark out of gopher wood.• An ark was built by Noah. It was made from gopher wood.• Noah constructed an ark with wood from a gopher tree.• Using gopher wood, Noah managed to put together an ark.
![Page 41: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/41.jpg)
Breaking, Eating, Opening• John broke the window.• The window broke.• John is always breaking things.• The broken window testified to John’s malfeasance.
• Eat!• We ate dinner.• We already ate.• The pies were eaten up quickly.• Our gluttony was complete.
• Open up!• Someone left the door open.• John opens the window at night.
![Page 42: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/42.jpg)
Introducing PropBank
• Corpus (PTB) with propositions annotated• Predicates (verbs)• Arguments (semantic roles)
• Semantic roles are Arg0, Arg1, etc., each with a description• Arg0 is typically the most agent-like argument• Labels for other arguments are somewhat arbitrary
![Page 43: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/43.jpg)
“Agree” in PropBank
• arg0: agreer• arg1: proposition• arg2: other entity agreeing• The group agreed it wouldn’t make an offer.• Usually John agrees with Mary on everything
![Page 44: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/44.jpg)
“Fall (move downward)” in PropBank
• arg1: logical subject, patient, thing falling• arg2: extent, amount fallen• arg3: starting point• arg4: ending point• argM-loc: medium• Sales fell to $251.2 million from $278.8 million.• The average junk bond fell by 4.2%.• The meteor fell through the atmosphere, crashing into
Cambridge.
![Page 45: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/45.jpg)
![Page 46: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/46.jpg)
FrameNet
• A frame is a schematic representation of a situation involving various participants, and other conceptual roles• In FrameNet, frames—not verbs—are first-class citizens• To a first approximation, verbs that relate to the same situation
belong to the same frame• Roles are given fine-grained labels that are specific to the frame, but
not the verb• Frames can center around words other than verbs
![Page 47: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/47.jpg)
change_position_on_a_scaleCore roles
ATTRIBUTE scalar property that the ITEM possesses
DIFFERENCE distance by which an ITEM changes its position
FINAL_STATE ITEM’s state after the change
FINAL_VALUE position on the scale where ITEM ends up
INITIAL_STATE ITEM’s state before the change
INITIAL_VALUE position on the scale from which the ITEM moves
ITEM entity that has a position on the scale
VALUE_RANGE portion of the scale along which values of ATTRIBUTE fluctuate
Some non-core roles ...
DURATION length of time over which the change occurs
SPEED rate of change of the value
GROUP the group in which an ITEM changes the value of an ATTRIBUTE
![Page 48: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/48.jpg)
• Verbs: advance, climb, decline, decrease, diminish, dip, double, drop, dwindle, edge, explode, fall, fluctuate, gain, grow, increase, jump, move, mushroom, plummet, reach, rise, rocket, shift, skyrocket, slide, soar, swell, swing, triple, tumble• Nouns: decline, decrease, escalation,
explosion, fall, fluctuation, gain, growth, hike, increase, rise, shift, tumble• Adverb: increasingly
![Page 49: Representing Meaning: Verb and Sentence Semantics](https://reader030.vdocuments.net/reader030/viewer/2022020917/61bd389661276e740b108f3f/html5/thumbnails/49.jpg)
How Can We Build an SRL System?(1) Parse(2) For each predicate word in the parse:
For each node in the parse:Classify the node with respect to the predicate