learning to exploit structured resources for lexical inference · lady gaga new york city...

1
Learning to Exploit Structured Resources for Lexical Inference Vered Shwartz, Omer Levy, Ido Dagan and Jacob Goldberger 1. Can we replicate WordNet-based methods for common nouns? Using WordNet as the only resource Baseline: 4 manual whitelists based on the literature Our method replicates the best whitelist on each dataset 2. Can we extract inferences over proper-names from community- built resources ? Improved recall over WordNet on our proper-names dataset Maintains high precision (p=97%, r=29%) Interesting whitelist (44 edge types): 3. How is our method compared to corpus-based methods? Baseline: state-of-the-art distributional methods: word2vec (Mikolov et al., 2013) + concatenation (Baroni et al.,2012) Corpus-based method: higher recall, lower precision Our method complements corpus-based methods 4. Evaluation occupation gender position in sports team Task : Identifying Lexical Inference Inferring the meaning of one term from another: Useful in many NLP applications Prior Methods : Corpus-based: high recall, limited precision Resource-based: Usually based on WordNet - manual selection of WordNet relations, high precision Limited recall. Missing: recent terminology (social network) proper-names (Lady Gaga) Goal : high-precision resource-based method with improved recall Means : community-built resources: 1. Lexical Inference We use knowledge resources: Knowledge resource is a graph: Nodes : terms / concepts Edges : concept-term / semantic edges between concepts We look at paths connecting , in resource graph Which edge types are indicative of the target lexical inference relation? Challenge : manual selection is infeasible – thousands of relations to choose from! Solution : learn an optimal subset of edge types for a certain target relation 2. Structured Resources Input : annotated dataset of term-pairs Defining a lexical-semantic relation (e.g. ) Task : Given a new term-pair (,), predict whether Parameter : whitelist of indicative edge types w = {instance_of, occupation, subclass_of, …} Inference : A path is indicative if all its edge types are whitelisted. A term-pair is positive if at least one of its paths is indicative. Training : Learn optimal whitelist (optimize on train set) Subset selection problem - Genetic Search 3. Learning apple. n.01 edible fruit. n.01 term-to- concept hypernym apple cat animal true apple animal false Kia car true Lady Gaga singer occupation person subclass_of Lady Gaga New York City place_of_birth city instance_of city New York City rock band Queen Radio Gaga Lady Gaga performer singer occupation person subclass_of instance_of instance_of Supervised framework for automatically selecting an optimized subset of resource relations for a given target inference task. Enables the use of large-scale resources, providing a rich source of high-precision inferences over proper-names. Available to download: https ://github.com/vered1986/LinKeR 5. Contribution

Upload: others

Post on 16-Jul-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Learning to Exploit Structured Resources for Lexical Inference · Lady Gaga New York City place_of_birth city instance_of city New York City rock band Queen Radio Gaga Lady Gaga performer

Learning to Exploit Structured Resources

for Lexical InferenceVered Shwartz, Omer Levy, Ido Dagan and Jacob Goldberger

1. Can we replicate WordNet-based methods for common nouns?• Using WordNet as the only resource• Baseline: 4 manual whitelists based on the literature

• Our method replicates the best whitelist on each dataset

2. Can we extract inferences over proper-names from community-built resources?• Improved recall over WordNet on our proper-names dataset• Maintains high precision (p=97%, r=29%)

• Interesting whitelist (44 edge types):

3. How is our method compared to corpus-based methods?• Baseline: state-of-the-art distributional methods:

word2vec (Mikolov et al., 2013) + concatenation (Baroni et al.,2012)

• Corpus-based method: higher recall, lower precision

• Our method complements corpus-based methods

4. Evaluation

occupation 𝐷𝑎𝑛𝑖𝑒𝑙 𝑅𝑎𝑑𝑐𝑙𝑖𝑓𝑓𝑒 → 𝑎𝑐𝑡𝑜𝑟

gender 𝐿𝑜𝑢𝑖𝑠𝑎 𝑀𝑎𝑦 𝐴𝑙𝑐𝑜𝑡𝑡 → 𝑤𝑜𝑚𝑎𝑛

position in sports team 𝐽𝑎𝑠𝑜𝑛 𝐶𝑜𝑙𝑙𝑖𝑛𝑠 → 𝑐𝑒𝑛𝑡𝑒𝑟

Task: Identifying Lexical Inference • Inferring the meaning of one term from another: 𝑐𝑎𝑡 → 𝑎𝑛𝑖𝑚𝑎𝑙• Useful in many NLP applications

Prior Methods:• Corpus-based: high recall, limited precision • Resource-based: • Usually based on WordNet - manual selection of WordNet relations, high precision• Limited recall. Missing:

• recent terminology (social network)• proper-names (Lady Gaga)

Goal: high-precision resource-based method with improved recall Means: community-built resources:

1. Lexical Inference

• We use knowledge resources:

• Knowledge resource is a graph:Nodes: terms / conceptsEdges: concept-term / semantic edges between concepts

• We look at paths connecting 𝑥, 𝑦 in resource graph

• Which edge types are indicative of the target lexical inference relation?Challenge: manual selection is infeasible – thousands of relations to choose from!Solution: learn an optimal subset of edge types for a certain target relation

2. Structured Resources

Input: annotated dataset of term-pairs• Defining a lexical-semantic relation 𝑅 (e.g. “𝒊𝒔 𝒂”)

Task: Given a new term-pair (𝑥,𝑦), predict whether 𝑥𝑅𝑦

Parameter: whitelist of indicative edge types w = {instance_of, occupation, subclass_of, …}

Inference:• A path is indicative if all its edge types are whitelisted.• A term-pair is positive if at least one of its paths is indicative.

𝐿𝑎𝑑𝑦 𝐺𝑎𝑔𝑎 → 𝑝𝑒𝑟𝑠𝑜𝑛

𝐿𝑎𝑑𝑦 𝐺𝑎𝑔𝑎 → 𝑐𝑖𝑡𝑦

Training:• Learn optimal whitelist (optimize 𝐹𝛽 on train set)

• Subset selection problem - Genetic Search

3. Learning

apple.n.01

edible fruit.n.01

term-to-concept

hypernymapple

cat animal true

apple animal false

Kia car true

LadyGaga

singeroccupation

personsubclass_of

LadyGaga

NewYorkCity

place_of_birthcity

instance_of

cityNewYorkCity

rockband

QueenRadioGaga

LadyGaga

performer

singeroccupationperson

subclass_of

instance_of

instance_of

• Supervised framework for automatically selecting an optimized subset of resource relations for a given target inference task.

• Enables the use of large-scale resources, providing a rich source of high-precision inferences over proper-names.

• Available to download:https://github.com/vered1986/LinKeR

5. Contribution