dan roth department of computer science university of illinois at urbana-champaign

62
Page 1 June 2009 ILPNLP Workshop @ NAACL-HLT With thanks to: Collaborators: Ming-Wei Chang, Dan Goldwasser, Vasin Punyakanok, Lev Ratinov, Nick Rizzolo, Mark Sammons, Ivan Titov, Scott Yih, Dav Zimak Constrained Conditional Models Learning and Inference for Information Extraction and Natural Language Understanding Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Upload: maxim

Post on 12-Feb-2016

39 views

Category:

Documents


0 download

DESCRIPTION

Constrained Conditional Models Learning and Inference for Information Extraction and Natural Language Understanding. Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign. With thanks to: - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 1

June 2009

ILPNLP Workshop @ NAACL-HLT

With thanks to:

Collaborators: Ming-Wei Chang, Dan Goldwasser, Vasin Punyakanok, Lev Ratinov, Nick Rizzolo, Mark Sammons, Ivan Titov, Scott Yih, Dav ZimakFunding: ARDA, under the AQUAINT program

NSF: ITR IIS-0085836, ITR IIS-0428472, ITR IIS- 0085980, SoD-HCER-0613885 A DOI grant under the Reflex program; DHS DASH Optimization (Xpress-MP)

Constrained Conditional Models Learning and Inference

for Information Extraction

and Natural Language Understanding

Dan RothDepartment of Computer ScienceUniversity of Illinois at Urbana-Champaign

Page 2: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 2

Informally: Everything that has to do with global constraints (and

learning models)

A bit more formally: We typically make decisions based on models such as:

With CCMs we make decisions based on models such as:

This is a global inference problem (you can solve it multiple ways)

We do not dictate how models are learned. but we’ll discuss it and make suggestions

Page 2

Constraints Conditional Models (CCMs)

),(maxarg xyfwTy

arg max ( , ) ( ,1 )Ty c C

c C

w f y x d y

CCMs assign values to variables in the presence/guided by constraints

Page 3: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 3

Constraints Driven Learning and Decision Making Why Constraints?

The Goal: Building a good NLP systems easily We have prior knowledge at our hand

How can we use it? Often knowledge can be injected directly and be used to

improve decision making guide learning simplify the models we need to learn

How useful are constraints? Useful for supervised learning Useful for semi-supervised learning Sometimes more efficient than labeling data directly

Page 4: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 4

Make my day

Page 5: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 5

Learning and Inference Global decisions in which several local decisions play a

role but there are mutual dependencies on their outcome. E.g. Structured Output Problems – multiple dependent output

variables (Main playground for these methods so far)

(Learned) models/classifiers for different sub-problems In some cases, not all local models can be learned simultaneously Key examples in NLP are Textual Entailment and QA In these cases, constraints may appear only at evaluation time

Incorporate models’ information, along with prior knowledge/constraints, in making coherent decisions decisions that respect the local models as well as domain &

context specific knowledge/constraints.

Page 6: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 6

Comprehension

1. Christopher Robin was born in England. 2. Winnie the Pooh is a title of a book. 3. Christopher Robin’s dad was a magician. 4. Christopher Robin must be at least 65 now.

A process that maintains and updates a collection of propositions about the state of affairs.

(ENGLAND, June, 1989) - Christopher Robin is alive and well. He lives in England. He is the same person that you read about in the book, Winnie the Pooh. As a boy, Chris lived in a pretty home called Cotchfield Farm. When Chris was three years old, his father wrote a poem about him. The poem was printed in a magazine for others to read. Mr. Robin then wrote a book. He made up a fairy tale land where Chris lived. His friends were animals. There was a bear called Winnie the Pooh. There was also an owl and a young pig, called a piglet. All the animals were stuffed toys that Chris owned. Mr. Robin made them come to life with his words. The places in the story were all near Cotchfield Farm. Winnie the Pooh was written in 1925. Children still love to read about Christopher Robin and his animal friends. Most people don't know he is a real person who is grown now. He has written two books of his own. They tell what it is like to be famous.

This is an Inference Problem

Page 7: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 7

This Talk: Constrained Conditional Models

A general inference framework that combines Learning conditional models with using declarative

expressive constraints Within a constrained optimization framework

Formulate a decision process as a constrained optimization problem Break up a complex problem into a set of sub-problems and require

components’ outcomes to be consistent modulo constraints

Has been shown useful in the context of many NLP problems SRL, Summarization; Co-reference; Information

Extraction; Transliteration [Roth&Yih04,07; Punyakanok et.al 05,08; Chang et.al 07,08; Clarke&Lapata06,07;

Denise&Baldrige07;Goldwasser&Roth’08]

Here: focus on Learning and Inference for Structured NLP Problems

Issues to attend to: While we formulate the problem as an ILP

problem, Inference can be done multiple ways Search; sampling; dynamic programming; SAT; ILP

The focus is on joint global inference Learning may or may not be joint.

Decomposing models is often beneficial

Page 8: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 8

Outline

Constrained Conditional Models Motivation Examples

Training Paradigms: Investigate ways for training models and combining constraints Joint Learning and Inference vs. decoupling

Learning & Inference Training with Hard and Soft Constrains Guiding Semi-Supervised Learning with Constraints

Examples Semantic Parsing Information Extraction Pipeline processes

Page 9: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 9

Pipeline

Conceptually, Pipelining is a crude approximation Interactions occur across levels and down stream decisions often

interact with previous decisions. Leads to propagation of errors Occasionally, later stage problems are easier but cannot correct

earlier errors. But, there are good reasons to use pipelines

Putting everything in one basket may not be right How about choosing some stages and think about them jointly?

POS Tagging

Phrases

Semantic Entities

Relations

Most problems are not single classification problems

Parsing

WSD Semantic Role Labeling

Raw Data

Page 10: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 10

Inference with General Constraint Structure [Roth&Yih’04]Recognizing Entities and Relations

Dole ’s wife, Elizabeth , is a native of N.C. E1 E2 E3

R12 R2

3

other 0.05per 0.85loc 0.10

other 0.05per 0.50loc 0.45

other 0.10per 0.60loc 0.30

irrelevant 0.10spouse_of 0.05born_in 0.85

irrelevant 0.05spouse_of 0.45born_in 0.50

irrelevant 0.05spouse_of 0.45born_in 0.50

other 0.05per 0.85loc 0.10

other 0.10per 0.60loc 0.30

other 0.05per 0.50loc 0.45

irrelevant 0.05spouse_of 0.45born_in 0.50

irrelevant 0.10spouse_of 0.05born_in 0.85

other 0.05per 0.50loc 0.45

Improvement over no inference: 2-5%

Some Questions: How to guide the global inference? Why not learn Jointly?

Models could be learned separately; constraints may come up only at decision time.

Non-SequentialKey Components:

1)Write down an objective function (Linear).

2)Write down constraints as linear inequalities

x* = argmaxx c(x=v) [x=v] =

= argmaxx c{E1 = per}· x{E1 = per} + c{E1 = loc}· x{E1 = loc}+…+ c{R12 = spouse_of}· x{R12 = spouse_of} +…+ c{R12 = }· x{R12 = }

Subject to Constraints

Page 11: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 11

Random Variables Y:

Conditional Distributions P (learned by models/classifiers) Constraints C– any Boolean function defined over partial assignments (possibly: + weights W )

Goal: Find the “best” assignment The assignment that achieves the highest global

performance. This is an Integer Programming Problem

Problem Setting

y7y4 y5 y6 y8

y1 y2 y3C(y1,y4)C(y2,y3,y6,y7,y8)

Y*=argmaxY PY subject to constraints C(+ WC)

observations

Page 12: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 12

Formal Model

How to solve?This is an Integer Linear ProgramSolving using ILP packages gives an exact solution. Search techniques are also possible

(Soft) constraints component

Weight Vector for “local” models

Penalty for violatingthe constraint.

How far y is from a “legal” assignment

Subject to constraints

A collection of Classifiers; Log-linear models (HMM, CRF) or a combination

How to train?How to decompose the global objective function?Should we incorporate constraints in the learning process?

Page 13: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 13

Example: Semantic Role Labeling

I left my pearls to my daughter in my will .[I]A0 left [my pearls]A1 [to my daughter]A2 [in my will]AM-LOC .

A0 Leaver A1 Things left A2 Benefactor AM-LOC Location I left my pearls to my daughter in my will .

Special Case (structured output problem): here, all the data is available at one time; in general, classifiers might be learned from different sources, at different times, at different contexts.

Implications on training paradigms

Overlapping arguments

If A2 is present, A1 must also be

present.

Who did what to whom, when, where, why,…

Page 14: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 14

PropBank [Palmer et. al. 05] provides a large human-annotated corpus of semantic verb-argument relations. It adds a layer of generic semantic labels to Penn Tree

Bank II. (Almost) all the labels are on the constituents of the parse

trees.

Core arguments: A0-A5 and AA different semantics for each verb specified in the PropBank Frame files

13 types of adjuncts labeled as AM-arg where arg specifies the adjunct type

Semantic Role Labeling (2/2)

Page 15: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 15

Algorithmic Approach

Identify argument candidates Pruning [Xue&Palmer, EMNLP’04] Argument Identifier

Binary classification (SNoW) Classify argument candidates

Argument Classifier Multi-class classification (SNoW)

Inference Use the estimated probability

distribution given by the argument classifier

Use structural and linguistic constraints Infer the optimal global output

I left my nice pearls to her

I left my nice pearls to her[ [ [ [ [ ] ] ] ] ]

I left my nice pearls to her[ [ [ [ [ ] ] ] ] ]

I left my nice pearls to her

I left my nice pearls to her

Identify Vocabulary

Inference over (old and new)

Vocabulary

candidate arguments

EASY

Page 16: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 16

InferenceI left my nice pearls to her

The output of the argument classifier often violates some constraints, especially when the sentence is long.

Finding the best legitimate output is formalized as an optimization problem and solved via Integer Linear Programming. [Punyakanok et. al 04, Roth & Yih 04;05;07]

Input: The probability estimation (by the argument classifier)

Structural and linguistic constraints

Allows incorporating expressive (non-sequential) constraints on the variables (the arguments types).

Page 17: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 17 Page 17

Semantic Role Labeling (SRL)

I left my pearls to my daughter in my will .

0.50.150.150.10.1

0.150.60.050.050.05

0.050.10.20.60.05

0.050.050.70.050.15

0.30.20.20.10.2

Page 18: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 18 Page 18

Semantic Role Labeling (SRL)

I left my pearls to my daughter in my will .

0.50.150.150.10.1

0.150.60.050.050.05

0.050.10.20.60.05

0.050.050.70.050.15

0.30.20.20.10.2

Page 19: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 19 Page 19

Semantic Role Labeling (SRL)

I left my pearls to my daughter in my will .

0.50.150.150.10.1

0.150.60.050.050.05

0.050.10.20.60.05

0.050.050.70.050.15

0.30.20.20.10.2

One inference problem for each verb predicate.

Page 20: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 20

Integer Linear Programming Inference

For each argument ai

Set up a Boolean variable: ai,t indicating whether ai is classified as t

Goal is to maximize i score(ai = t ) ai,t

Subject to the (linear) constraints

If score(ai = t ) = P(ai = t ), the objective is to find the assignment that maximizes the expected number of arguments that are correct and satisfies the constraints.

The Constrained Conditional Model is completely decomposed during training

Page 21: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 21

No duplicate argument classesa POTARG x{a = A0} 1

R-ARG a2 POTARG , a POTARG x{a = A0} x{a2 = R-A0}

C-ARG a2 POTARG ,

(a POTARG) (a is before a2 ) x{a = A0} x{a2 = C-A0} Many other possible constraints:

Unique labels No overlapping or embedding Relations between number of arguments; order

constraints If verb is of type A, no argument of type B

Any Boolean rule can be encoded as a linear constraint.

If there is an R-ARG phrase, there is an ARG Phrase

If there is an C-ARG phrase, there is an ARG before it

Constraints

Joint inference can be used also to combine different SRL Systems.

Universally quantified rulesLBJ: allows a developer to encode

constraints in FOL; these are compiled into linear inequalities automatically.

Page 22: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 22

Learning Based Java (LBJ):A modeling language for Constrained Conditional Models Supports programming along with building learned models,

high level specification of constraints and inference with constraints

Learning operator: Functions defined in terms of data Learning happens at “compile time”

Integrated constraint language: Declarative, FOL-like syntax defines constraints in terms of

your Java objects Compositionality:

Use any function as feature extractor Easily combine existing model specifications /learned

models with each other

http://L2R.cs.uiuc.edu/~cogcomp/software.php

Page 23: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 23

Example: Semantic Role Labeling

Declarative, FOL-style constraints written in terms of functions applied to Java objects[Rizzolo, Roth’07]

Inference produces new functions that respect the constraints

LBJ site provides example code for NER, POS tagger etc.

Page 24: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 24

Semantic Role LabelingScreen shot from a CCG demo http://L2R.cs.uiuc.edu/~cogcomp

Semantic parsing reveals several relations in the sentence along with their arguments.

Top ranked system in CoNLL’05 shared task

Key difference is the Inference

This approach produces a very good semantic parser. F1~90% Easy and fast: ~7 Sent/Sec (using Xpress-MP)

Page 25: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 25

Textual Entailment

Eyeing the huge market potential, currently led by Google, Yahoo took over search company Overture Services Inc. last year

Yahoo acquired Overture

Is it true that…?(Textual Entailment)

Overture is a search company Google is a search company ……….Google owns Overture

Phrasal verb paraphrasing [Connor&Roth’07]Entity matching [Li et. al, AAAI’04, NAACL’04]

Semantic Role LabelingPunyakanok et. al’05,08Inference for EntailmentBraz et. al’05, 07

Page 26: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 26

Outline

Constrained Conditional Models Motivation Examples

Training Paradigms: Investigate ways for training models and combining constraints Joint Learning and Inference vs. decoupling

Learning & Inference Training with Hard and Soft Constrains Guiding Semi-Supervised Learning with Constraints

Examples Semantic Parsing Information Extraction Pipeline processes

Page 27: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 27

Training Paradigms that Support Global Inference

Algorithmic Approach: Incorporating general constraints Allow both statistical and expressive declarative constraints

[ICML’05] Allow non-sequential constraints (generally difficult)

[CoNLL’04]

Coupling vs. Decoupling Training and Inference. Incorporating global constraints is important but Should it be done only at evaluation time or also at training

time? How to decompose the objective function and train in parts? Issues related to:

Modularity, efficiency and performance, availability of training data

Problem specific considerations

Page 28: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 28

Training in the presence of Constraints

General Training Paradigm: First Term: Learning from data (could be further

decomposed) Second Term: Guiding the model by constraints Can choose if constraints’ weights trained, when

and how, or taken into account only in evaluation.

Decompose Model (SRL case) Decompose Model from constraints

Page 29: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 29

Comparing Training Methods

Option 1: Learning + Inference (with Constraints) Ignore constraints during training

Option 2: Inference (with Constraints) Based Training Consider constraints during training

In both cases: Global Decision Making with Constraints

Question: Isn’t Option 2 always better?

Not so simple… Next, the “Local model story”

Page 30: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 30

f1(x)

f2(x)

f3(x)f4(x)

f5(x)

Training Methods

x1

x6

x2

x5

x4x3

x7

y1

y2

y5

y4

y3

X

Y

Learning + Inference (L+I)Learn models independentlyInference Based Training (IBT)Learn all models together!IntuitionLearning with constraints may make learning more difficult

Cartoon: each model can be

more complex and may have a view on a set of output

variables.

Page 31: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 31

-1 1 111Y’ Local Predictions

Training with ConstraintsExample: Perceptron-based Global Learning

x1

x6

x2

x5

x4x3

x7

f1(x)

f2(x)

f3(x)f4(x)

f5(x)

X

Y

-1 1 1-1-1YTrue Global Labeling -1 1 11-1Y’ Apply Constraints:

Which one is better? When and Why?

Page 32: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 32

Claims [Punyakanok et. al , IJCAI 2005]

When the local modes are “easy” to learn, L+I outperforms IBT. In many applications, the components are identifiable and

easy to learn (e.g., argument, open-close, PER). Only when the local problems become difficult to solve in

isolation, IBT outperforms L+I, but needs a larger number of training examples.

Other training paradigms are possible Pipeline-like Sequential Models: [Roth, Small, Titov: AI&Stat’09]

Identify a preferred ordering among components Learn k-th model jointly with previously learned models

L+I: cheaper computationally; modularIBT is better in the limit, and other extreme cases.

Page 33: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 33

opt=0.2opt=0.1opt=0

Bound Prediction

Local ≤ opt + ( ( d log m + log 1/ ) / m )1/2

Global ≤ 0 + ( ( cd log m + c2d + log 1/ ) / m )1/2

Bounds Simulated Data

L+I vs. IBT: the more identifiable individual problems are, the better overall performance is with L+I

Indication for hardness of

problem

Page 34: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 34

Relative Merits: SRL

Difficulty of the learning problem

(# features)

L+I is better.

When the problem is artificially made harder, the tradeoff is clearer.

easyhard

Page 35: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 35

Comparing Training Methods (Cont.)

Local Models (train independently) vs. Structured Models In many cases, structured models might be better due to expressivity

But, what if we use constraints? Local Models+Constraints vs. Structured Models

+Constraints Hard to tell: Constraints are expressive For tractability reasons, structured models have less expressivity than

the use of constraints. Local can be better, because local models are easier to learn

Decompose Model (SRL case) Decompose Model from constraints

Page 36: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 36

Example: CRFs are CCMs Consider a common model for sequential inference: HMM/CRF

Inference in this model is done via the Viterbi Algorithm.

Viterbi is a special case of the Linear Programming based Inference. Viterbi is a shortest path problem, which is a LP, with a

canonical matrix that is totally unimodular. Therefore, you can get integrality constraints for free.

One can now incorporate non-sequential/expressive/declarative constraints by modifying this canonical matrix

No value can appear twice; a specific value must appear at least once; AB

And, run the inference as an ILP inference.

y1 y2 y3 y4 y5yx x1 x2 x3 x4 x5

sABC

ABC

ABC

ABC

ABC

t

Learn a rather simple model; make decisions with a more expressive model

But, you can do better

Page 37: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 37

Example: Semantic Role Labeling Revisited

s

A

B

C

A

B

C

A

B

C

A

B

C

A

B

C

t

Sequential Models Conditional Random Field Global perceptronTraining: sentence basedTesting: find the shortest path

with constraints

Local Models Logistic Regression Local Avg. PerceptronTraining: token based. Testing: find the best assignment locally with constraints

Page 38: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 38

Model CRF CRF-D CRF-IBT Avg. PBaseline 66.46 69.14 58.15

+ Constraint

s

71.94 73.91 69.82 74.49

Training Time

48 38 145 0.8

Which Model is Better? Semantic Role Labeling Experiments on SRL: [Roth and Yih, ICML 2005]

Story: Inject constraints into conditional random field models

Sequential Models LocalL+I L+IIBT

Sequential Models are better than Local Models !

(No constraints)

Local Models are now better than Sequential Models! (With constraints)

Page 39: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 39

Summary: Training Methods

Many choices for training a CCM Learning + Inference (Training without constraints) Inference based Learning (Training with constraints) Model Decomposition

Advantages of L+I Require fewer training examples More efficient; most of the time, better performance Modularity; easier to incorporate already learned models.

Advantages of IBT Better in the limit Better when there are strong interactions among y’s

Learn a rather simple model; make decisions with a more expressive model

Page 40: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 40

Outline

Constrained Conditional Models Motivation Examples

Training Paradigms: Investigate ways for training models and combining constraints Joint Learning and Inference vs. decoupling

Learning & Inference Training with Hard and Soft Constrains Guiding Semi-Supervised Learning with Constraints

Examples Semantic Parsing Information Extraction Pipeline processes

Page 41: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 41

Constrained Conditional Model: Soft Constraints

(3) How to solve?This is an Integer Linear ProgramSolving using ILP packages gives an exact solution. Search techniques are also possible

(Soft) constraints component

Constraint violation penalty

How far y is from a “legal” assignment

Subject to constraints

(4) How to train?How to decompose the global objective function?Should we incorporate constraints in the learning process?

(1) Why use soft constraints?

(2) How to model “degree of violations”

Page 42: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 42

(1) Why Are Soft Constraints Important?

Some constraints may be violated by the data.

Even when the gold data violates no constraints, the model may prefer illegal solutions. If all solutions considered by the model violate

constraints, we still want to rank solutions based on the level of constraints’ violation.

Important when beam search is used Rather than eliminating illegal assignments, re-rank them

Working with soft constraints [Chang et. al, ACL’07] Need to define the degree of violation

Maybe be problem specific Need to assign penalties for constraints

Page 43: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 43 Page 43

Information extraction without Prior Knowledge

Prediction result of a trained HMMLars Ole Andersen . Program analysis andspecialization for the C Programming language

. PhD thesis .DIKU , University of Copenhagen , May1994 .

[AUTHOR] [TITLE] [EDITOR] [BOOKTITLE] [TECH-REPORT] [INSTITUTION] [DATE]

Violates lots of natural constraints!

Lars Ole Andersen . Program analysis and specialization for the C Programming language. PhD thesis. DIKU , University of Copenhagen, May 1994 .

Page 44: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 44

Examples of Constraints

Each field must be a consecutive list of words and can appear at most once in a citation.

State transitions must occur on punctuation marks.

The citation can only start with AUTHOR or EDITOR.

The words pp., pages correspond to PAGE. Four digits starting with 20xx and 19xx are DATE. Quotations can appear only in TITLE ……. Easy to express pieces of “knowledge”

Non Propositional; May use Quantifiers

Page 45: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 45 Page 45

Adding constraints, we get correct results! Without changing the model

[AUTHOR] Lars Ole Andersen . [TITLE] Program analysis and specialization

for the C Programming language .

[TECH-REPORT] PhD thesis .[INSTITUTION] DIKU , University of Copenhagen , [DATE] May, 1994 .

Information Extraction with Constraints

Page 46: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 46

Hard Constraints vs. Weighted Constraints

Constraints are close to perfect

Labeled data might not follow the constraints

Page 47: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 47

Training with Soft Constraints

Need to figure out the penalty as well…

Option 1: Learning + Inference (with Constraints) Learn the weights and penalties separately

Penalty(c) = -log{P(C is violated)}

Option 2: Inference (with Constraints) Based Training Learn the weights and penalties together

The tradeoff between L+I and IBT is similar to what we saw earlier.

Page 48: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 48

YPRED=

Inference Based Training With Soft Constraints

For each iterationFor each (X, YGOLD ) in the training data

If YPRED != YGOLD

λ = λ + F(X, YGOLD ) - F(X, YPRED)ρI = ρI+ d(YGOLD,1C(X)) - d(YPRED,1C(X)), I =

1,..endif

endfor

• Example: Perceptron• Update penalties as well !

Page 49: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 49

L+I vs IBT for Soft Constraints

Test on citation recognition: L+I: HMM + weighted constraints IBT: Perceptron + weighted constraints Same feature set

With constraints Factored Model is better More significant with a small # of examples

Without constraints Few labeled examples, HMM > perceptron Many labeled examples, perceptron > HMM

Page 50: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 50

Outline

Constrained Conditional Models Motivation Examples

Training Paradigms: Investigate ways for training models and combining constraints Joint Learning and Inference vs. decoupling

Learning & Inference Training with Hard and Soft Constrains Guiding Semi-Supervised Learning with Constraints

Examples Semantic Parsing Information Extraction Pipeline processes

Page 51: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 51

Outline

Constrained Conditional Models Motivation Examples

Training Paradigms: Investigate ways for training models and combining constraints Joint Learning and Inference vs. decoupling Learning

& Inference Guiding Semi-Supervised Learning with Constraints

Features vs. Constraints Hard and Soft Constraints

Examples Semantic Parsing Information Extraction Pipeline processes

Page 52: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 52

Constraints As a Way To Encode Prior Knowledge

Consider encoding the knowledge that: Entities of type A and B cannot occur simultaneously in a

sentence The “Feature” Way

Requires larger models

The Constraints Way Keeps the model simple; add expressive

constraints directly A small set of constraints Allows for decision time incorporation of

constraints

A effective way to inject knowledge

Need more training data

We can use constraints as a way to replace training data

Page 53: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 53

Guiding Semi-Supervised Learning with Constraints

Model

Decision Time Constraints

Un-labeled Data

Constraints

In traditional Semi-Supervised learning the model can drift away from the correct one.

Constraints can be used to generate better training data At decision time, to bias the objective

function towards favoring constraint satisfaction.

At training to improve labeling of un-labeled data (and thus improve the model)

Page 54: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 54 Page 54

Semi-supervised Learning with Constraints

=learn(T)For N iterations do

T= For each x in unlabeled dataset

{y1,…,yK} InferenceWithConstraints(x,C, )

T=T {(x, yi)}i=1…k

= +(1- )learn(T)Learn from new training data.Weigh supervised and unsupervised model.

Inference based augmentation of the training set (feedback)(inference with constraints).

Supervised learning algorithm parameterized by

[Chang, Ratinov, Roth, ACL’07;ICML’08]

Page 55: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 55

Objective function: Value of Constraints in Semi-Supervised Learning

# of available labeled examples

Learning w 10 ConstraintsConstraints are used to Bootstrap a semi-supervised learner Poor model + constraints used to annotate unlabeled data, which in turn is used to keep training the model.

Learning w/o Constraints: 300 examples.

Factored model.

Page 56: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 56 56

xX

x1

x6

x2

x5

x4x3

x7

X

Yy1

Single Output Problem:Only one output

Constraints in a hidden layer

Page 57: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 57 57

Adding Constraints Through Hidden Variables

xX

x1

x6

x2

x5

x4x3

x7

X

xX

Yf1

f2

f4

f3

f5

y1

Single Output Problemwith hidden variables

Page 58: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 58

Yes/No (Italy,איטליה)

Learning feature representation is a structured learning problem Features are the graph edges – the problem is choosing the

optimal subset Many constraints on the legitimacy of the active features

representation Formalize the problem as a constrained optimization problem

A successful solution depends on:

Subject to: One-to-One mapping; Non-crossing Length difference restriction Language specific constraints

Learning Good Feature Representation for Discriminative Transliteration

ylatIטי יל אה

features

Iterative Unsupervised learning algorithm

Learning a good objective function

Romanization table

Good initial objective function

TjSi

ijijX

xwX,

* maxarg

Page 59: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 59

Iterative Objective Function Learning

Inference Prediction

Training

Romanization Table

Generate features

Predict labels for all word pairs

Update weight vector

Initial objective function

Language pair UCDL Prev. Sys

English-Russian (ACC) 73 63English-Hebrew (MRR) 89.9 51

Page 60: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 60

y* = argmaxy wi Á(x; y)

Linear objective functions Typically Á(x,y) will be local

functions, or Á(x,y) = Á(x)

Summary: Constrained Conditional Models

y7y4 y5 y6 y8

y1 y2 y3

y7y4 y5 y6 y8

y1 y2 y3

Conditional Markov Random Field Constraints Network

i ½i dC(x,y)

Expressive constraints over output variables

Soft, weighted constraints Specified declaratively as FOL

formulae Clearly, there is a joint probability distribution that

represents this mixed model. We would like to:

Learn a simple model or several simple models Make decisions with respect to a complex model

Key difference from MLNs, which provide a concise definition of a model, but the whole joint one.

Page 61: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 61

Conclusion

Constrained Conditional Models combine Learning conditional models with using declarative

expressive constraints Within a constrained optimization framework

Use constraints! The framework supports: A clean way of incorporating constraints to bias and

improve decisions of supervised learning models Significant success on several NLP and IE tasks (often,

with ILP) A clean way to use (declarative) prior knowledge to

guide semi-supervised learning

Training protocol matters More work needed here

LBJ (Learning Based Java): http://L2R.cs.uiuc.edu/~cogcompA modeling language for Constrained Conditional Models. Supports

programming along with building learned models, high level specification of constraints and inference with constraints

Page 62: Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign

Page 62

Nice to Meet You