entity spotting in informal text

32
Entity Spotting in Informal Text Meena Nagarajan with Daniel Gruhl*, Jan Pieper*, Christine Robson*, Amit P. Sheth Kno.e.sis, Wright State IBM Research - Almaden, San Jose CA* 1 Thursday, October 29, 2009

Upload: meena-nagarajan

Post on 01-Dec-2014

2.076 views

Category:

Education


2 download

DESCRIPTION

Slides for: Context and Domain Knowledge Enhanced Entity Spotting in Informal Text, ISWC 2009

TRANSCRIPT

Page 1: Entity Spotting in Informal Text

Entity Spotting in Informal Text

Meena Nagarajanwith

Daniel Gruhl*, Jan Pieper*, Christine Robson*,Amit P. Sheth

Kno.e.sis, Wright StateIBM Research - Almaden, San Jose CA*

1Thursday, October 29, 2009

Page 2: Entity Spotting in Informal Text

Tracking Online Popularity http://www.almaden.ibm.com/cs/projects/iis/sound/

2Thursday, October 29, 2009

Page 3: Entity Spotting in Informal Text

Tracking Online Popularity

• What is the buzz in the online Music Community?

• Ranking and displaying top X music artists, songs, tracks, albums..

• Spotting entities, despamming, sentiment identification, aggregation, top X lists..

http://www.almaden.ibm.com/cs/projects/iis/sound/

3Thursday, October 29, 2009

Page 4: Entity Spotting in Informal Text

Spotting music entities in user-generated content in

online music forums (MySpace)

4Thursday, October 29, 2009

Page 5: Entity Spotting in Informal Text

Chatter in Online Music Communities

http://knoesis.wright.edu/research/semweb/projects/music/

5Thursday, October 29, 2009

Page 6: Entity Spotting in Informal Text

Goal: Semantic Annotation of artists, tracks, songs, albums..

Ohh these sour times... rock!

Ohh these <track id=574623> sour times </track> ... rock!

Music Brainz RDF

6Thursday, October 29, 2009

Page 7: Entity Spotting in Informal Text

Multiple Senses in the same Domain

• 60 songs with Merry Christmas

• 3600 songs with Yesterday

• 195 releases of American Pie

• 31 artists covering American Pie

Caught AMERICAN PIE on cable so much

fun!7Thursday, October 29, 2009

Page 8: Entity Spotting in Informal Text

• Several Cultural named entities

• artifacts of culture, common words in everyday language

LOVED UR MUSIC YESTERDAY!

Lily your face lights up when you smile!

♥ Just showing some Love to you Madonna you are The Queen to me

Annotating UGC, other Challenges

8Thursday, October 29, 2009

Page 9: Entity Spotting in Informal Text

Annotating UGC, other Challenges

• Informal Text

• slang, abbreviations, misspellings..

• indifferent approach to grammar..

• Context dependent terms

• Unknown distributions

9Thursday, October 29, 2009

Page 10: Entity Spotting in Informal Text

Our ApproachSpotting and subsequent sense

disambiguation of spots

Ohh these sour times... rock!

Ohh these <track id=574623> sour times </track> ... rock!

10Thursday, October 29, 2009

Page 11: Entity Spotting in Informal Text

Ground Truth Data Set• 3 artists : Madonna, Rihanna, Lily Allen

• 1858 spots (MySpace UGC) using naive spotter over MusicBrainz artist metadata

• Adjudicate if a spot is an entity or not (or inconclusive)

• hand tagged by 4 authors

3.1 Ground Truth Data Set

Our experimental evaluation focuses on user comments from the MySpace pagesof three artists: Madonna, Rihanna and Lily Allen (see Table 2). The artistswere selected to be popular enough to draw comment but different enough toprovide variety. The entity definitions were taken from the MusicBrainz RDF (seeFigure 1), which also includes some but not all common aliases and misspellings.

Madonna an artist with a extensive discography as well as a current album andconcert tour

Rihanna a pop singer with recent accolades including a Grammy Award and avery active MySpace presence

Lilly Allen an independent artist with song titles that include “Smile,” “Allright,Still”, “Naive”, and “Friday Night” who also generates a fair amountof buzz around her personal life not related to music

Table 2. Artists in the Ground Truth Data Set

Artist Good spots Bad spots(Spots scored) Agreement Agreement

100% 75 % 100% 75%Rihanna (615) 165 18 351 8Lily (523) 268 42 10 100Madonna (720) 138 24 503 20

Table 3. Manual scoring agreements onnaive entity spotter results.

We establish a ground truth dataset of 1858 entity spots for theseartists (breakdown in Table 3). Thedata was obtained by crawling theartist’s MySpace page comments andidentifying all exact string matchesof the artist’s song titles. Only com-ments with at least one spot were re-tained. These spots were then handscored by four of the authors as“good spot,” “bad spot,” or “inconclusive.” This dataset is available for down-load from the Knoesis Center website 5.

The human taggers were instructed to tag a spot as “good” if it clearlywas a reference to a song and not a spurious use of the phrase. An agreementbetween at least three of the hand-spotters with no disagreement was consideredagreement. As can be seen in Table 3, the taggers agreed 4-way (100% agreement)on Rihanna (84%) and Madonna (90%) spots. However ambiguities in Lily Allensongs (most notably the song “Smile”), resulted in only 53% 4-way agreement.

We note that this approach results in a recall of 1.0, because we use thenaive spotter, restricted to the individual artist, to generate the ground truthcandidate set. The precision of the naive spotter after hand-scoring these 1858spots was 73%, 33% and 23% for Lilly Allen, Rihanna and Madonna respectively(see Table 3). This represents the best case for the naive spotter and accuracydrops quickly as the entity candidate set becomes less restricted. In the nextSection we take a closer look at the relationship between entity candidate setsize and spotting accuracy.

5 http://knoesis.wright.edu/research/semweb/music

33%73%23%

Precision(best case for naive spotter)

11Thursday, October 29, 2009

Page 12: Entity Spotting in Informal Text

Experiments and Results

12Thursday, October 29, 2009

Page 13: Entity Spotting in Informal Text

Experiments

1. Light weight, edit distance based entity spotter

All entities from MusicBrainz

13Thursday, October 29, 2009

Page 14: Entity Spotting in Informal Text

Experiments

1. Naive spotter using all entities from all of MusicBrainz

2. This new Merry Christmas tune is so good!

? but which one ?

Disambiguate between the 60+ Merry Christmas entries in MusicBrainz

14Thursday, October 29, 2009

Page 15: Entity Spotting in Informal Text

Experiments

This new Merry Christmas tune is

so good!

2. Constrain set of possible entities from Musicbrainz

- to increase spotting accuracy - constrain using cues from the comment to eliminate alternatives

15Thursday, October 29, 2009

Page 16: Entity Spotting in Informal Text

Experiments

Your SMILE rocks!

3. Eliminate non-music mentions

Natural language and domain specific cues

16Thursday, October 29, 2009

Page 17: Entity Spotting in Informal Text

Restricted Entity Spotting

17Thursday, October 29, 2009

Page 18: Entity Spotting in Informal Text

2. Restricted Entity Spotting

• Investigating the relationship between number of entities used and spotting accuracy

• Understand systematic ways of scoping domain models for use in semantic annotation

• Experiments to gauge benefits of implementing particular constraints in annotator systems

• harder artist age detector vs. easier gender detector ?

18Thursday, October 29, 2009

Page 19: Entity Spotting in Informal Text

2a. Random Restrictions

3.2 Impact of Domain Restrictions

One of the main contributions of this paper is the insight that it is often possible

to restrict the set of entity candidates, and that such a restriction increases

spotting precision. In this Section we explore the effect of domain restrictions

on spotting precision by considering random entity subsets.

We begin with the whole MusicBrainz RDF of 281,890 publishing artists

and 6,220,519 tracks, which would be appropriate if we had no information

about which artists may be contained in the corpus. We then select random

subsets of artists that are factors of 10 smaller (10%, 1%, etc). These subsets

always contain our three actual artists (Madonna, Rihanna and Lily Allen),

because we are interested in simulating restrictions that remove invalid artists.

The most restricted entity set contains just the songs of one artist (≈0.0001% of

the MusicBrainz taxonomy). In order to rule out selection bias, we perform 200

random draws of sets of artists for each set size - a total of 1200 experiments.

Figure 2 shows that the precision increases as the set of possible entities shrinks.

For each set size, all 200 results are plotted and a best fit line has been added

to indicate the average precision. Note that the figure is in log-log scale.

!"""#$

!""#$

!"#$

!#$

#$

#"$

#""$!"""#$ !""#$ !"#$ !#$ #$ #"$ #""$

!"#$"%&'()'&*"'+,-.$'/#0.%1'&02(%(34

!#"$.-.(%'()'&*"'56(&&"#

%&'()*''+,-.)/(.012+)314+5&61,,1-.)/(.012+)314+/178,,1-.)/(.012+)314+

%&'()*''+,

5&61,,1

/178,,1

Fig. 2. Precision of a naive spotter using differently sized portions of the MusicBrainzTaxonomy to spot song titles on artist’s MySpace pages

We observe that the curves in Figure 2 conform to a power law formula,

specifically a Zipf distribution (1

nR2 ). Zipf’s law was originally applied to demon-

strate the Zipf distribution in frequency of words in natural language corpora

[18], and has since been demonstrated in other corpora including web searches

[7]. Figure 2 shows that song titles in Informal English exhibit the same fre-

quency characteristics as plain English. Furthermore, we can see that in the

average case, a domain restrictions of 10% of the MusicBrainz RDF will result

approximately in a 9.8 times improvement in precision of a naive spotter.

This result is remarkably consistent across all three artists. The R2 values

for the power lines on the three artists are 0.9776, 0.979, 0.9836, which gives a

deviation of 0.61% in R2 value between spots on the three MySpace pages.

• From all of MusicBrainz (281890 artists, 6220519 tracks) to songs of one artist (for all three artists)

Domain restrictions of 10% of the RDF result in approximately 9.8 times improvement in precision

33%73%23%

Precision(best case for naive spotter)

19Thursday, October 29, 2009

Page 20: Entity Spotting in Informal Text

2b. Real-world Constraints for Restrictions

“Happy 25th Rhi!” (eliminate using Artist DOB - metadata in MusicBrainz)

“ur new album dummy is awesome” (eliminate using Album release dates - metadata in MusicBrainz)

• Systematic scoping of the RDF

• Question: Do real-world constraints from metadata reduce size of the entity spot set in a meaningful way?

• Experiments: Derived manually and tested for usefulness

20Thursday, October 29, 2009

Page 21: Entity Spotting in Informal Text

Real-world ConstraintsKey Count Restriction

Artist Career Length Restrictions- Applied to MadonnaB 22 80’s artists with recent (within 1 year) albumC 154 First album 1983D 1,193 20-30 year career

Recent Album Restrictions- Applied to MadonnaE 6,491 Artists who released an album in the past yearF 10,501 Artists who released an album in the past 5 years

Artist Age Restrictions- Applied to Lily AllenH 112 Artist born 1985, album in past 2 yearsJ 284 Artists born in 1985 (or bands founded in 1985)L 4,780 Artists or bands under 25 with album in past 2 yearsM 10,187 Artists or bands under 25 years old

Number of Album Restrictions- Applied to Lily AllenK 1,530 Only one album, released in the past 2 yearsN 19,809 Artists with only one album

Recent Album Restrictions- Applied to RihannaQ 83 3 albums exactly, first album last yearR 196 3+ albums, first album last yearS 1,398 First album last yearT 2,653 Artists with 3+ albums, one in the past yearU 6,491 Artists who released an album in the past year

Specific Artist Restrictions- Applied to each ArtistA 1 Madonna onlyG 1 Lily Allen onlyP 1 Rihanna onlyZ 281,890 All artists in MusicBrainz

Table 4. The efficacy of various sample restrictions.

We consider three classes of restrictions - career, age and album based re-

strictions, apply these to the MusicBrainz RDF to reduce the size of the entity

spot set in a meaningful way and finally run the trivial spotter. For the sake of

clarity, we apply different classes of constraints to different artists.

We begin with restrictions based on length of career, using Madonna’s MyS-

pace page as our corpus. We can restrict the RDF graph based on total length of

career, date of earliest album (for Madonna this is 1983, which falls in the early

80’s), and recent albums (within the past year or 5 years). All of these restric-

tions are plotted in Figure 4, along with the Zipf distribution for Madonna from

Figure 2. We can see clearly that restricting the RDF graph based on career

characteristics conforms to the predicted Zipf distribution.

For our next experiment we consider restrictions based on age of artist, using

Lily Allen’s MySpace page as our corpus. Our restrictions include Lily Allen’s

age of 25 years, but overlap with bands founded 25 years ago because of how

dates are recorded in the MusicBrainz Ontology. We can further restrict using

album information, noting that Lily Allen has only a single album, released in

the past two years. These restrictions are plotted in Figure 4, showing that these

restrictions on the RDF graph conform to the same Zipf distribution.

Finally, we consider restrictions based on absolute number of albums, us-

ing Rihanna’s MySpace page as our corpus. We restrict to artists with three

albums, or at least three albums, and can further refine by the release dates of

these albums. These restrictions fit with Rihanna’s short career and dispropor-

tionately large number of album releases (3 releases in one year). As can be seen

in Figure 4, these restrictions also conform to the predicted Zipf distribution.

D. I’ve been your fan for 25 years! M. Happy 25th !

Key Count Restriction

Artist Career Length Restrictions- Applied to MadonnaB 22 80’s artists with recent (within 1 year) albumC 154 First album 1983D 1,193 20-30 year career

Recent Album Restrictions- Applied to MadonnaE 6,491 Artists who released an album in the past yearF 10,501 Artists who released an album in the past 5 years

Artist Age Restrictions- Applied to Lily AllenH 112 Artist born 1985, album in past 2 yearsJ 284 Artists born in 1985 (or bands founded in 1985)L 4,780 Artists or bands under 25 with album in past 2 yearsM 10,187 Artists or bands under 25 years old

Number of Album Restrictions- Applied to Lily AllenK 1,530 Only one album, released in the past 2 yearsN 19,809 Artists with only one album

Recent Album Restrictions- Applied to RihannaQ 83 3 albums exactly, first album last yearR 196 3+ albums, first album last yearS 1,398 First album last yearT 2,653 Artists with 3+ albums, one in the past yearU 6,491 Artists who released an album in the past year

Specific Artist Restrictions- Applied to each ArtistA 1 Madonna onlyG 1 Lily Allen onlyP 1 Rihanna onlyZ 281,890 All artists in MusicBrainz

Table 4. The efficacy of various sample restrictions.

We consider three classes of restrictions - career, age and album based re-

strictions, apply these to the MusicBrainz RDF to reduce the size of the entity

spot set in a meaningful way and finally run the trivial spotter. For the sake of

clarity, we apply different classes of constraints to different artists.

We begin with restrictions based on length of career, using Madonna’s MyS-

pace page as our corpus. We can restrict the RDF graph based on total length of

career, date of earliest album (for Madonna this is 1983, which falls in the early

80’s), and recent albums (within the past year or 5 years). All of these restric-

tions are plotted in Figure 4, along with the Zipf distribution for Madonna from

Figure 2. We can see clearly that restricting the RDF graph based on career

characteristics conforms to the predicted Zipf distribution.

For our next experiment we consider restrictions based on age of artist, using

Lily Allen’s MySpace page as our corpus. Our restrictions include Lily Allen’s

age of 25 years, but overlap with bands founded 25 years ago because of how

dates are recorded in the MusicBrainz Ontology. We can further restrict using

album information, noting that Lily Allen has only a single album, released in

the past two years. These restrictions are plotted in Figure 4, showing that these

restrictions on the RDF graph conform to the same Zipf distribution.

Finally, we consider restrictions based on absolute number of albums, us-

ing Rihanna’s MySpace page as our corpus. We restrict to artists with three

albums, or at least three albums, and can further refine by the release dates of

these albums. These restrictions fit with Rihanna’s short career and dispropor-

tionately large number of album releases (3 releases in one year). As can be seen

in Figure 4, these restrictions also conform to the predicted Zipf distribution.

........

Restrictions over MusicBrainz

21Thursday, October 29, 2009

Page 22: Entity Spotting in Informal Text

Real-world Constraints

• Applied different constraints to different artists

• Reduce potential entity spot size

• Run naive spotter

• Measure precision

22Thursday, October 29, 2009

Page 23: Entity Spotting in Informal Text

Real-world Constraints

!"""#$

!""#$

!"#$

!#$

#$

#"$

#""$!"""#$ !""#$ !"#$ !#$ #$ #"$ #""$

!"#$%&%'()'*)+,#)-.'++#"%&'()*+,-..)/*./&0%)1*-%*-%23*405&%%&*+-%6+**

*****789!9$*,/):0+0-%;

&/.0+.+*<5-+)*=0/+.*&2>?@*<&+*0%*.5)*,&+.*8*3)&/+

*&22*&/.0+.+*<5-*/)2)&+)1*&%*&2>?@*0%*.5)*,&+.*8*3)&/+

&.*2)&+.*8*&2>?@+)A&:.23*8*&2>?@+

*)%.0/)*B?+0:*C/&0%D*.&A-%-@3*7"!"""8$*,/):0+0-%;

Rihanna: short career, recent album releases, 3 album releases etc....

“I heart your new album”“I love all your 3 albums”

“You are most favorite new pop artist”

23Thursday, October 29, 2009

Page 24: Entity Spotting in Informal Text

Real-world Constraints

!"""#$

!""#$

!"#$

!#$

#$

#"$

#""$!"""#$ !""#$ !"#$ !#$ #$ #"$ #""$

!"#$%&%'()'*)+,#)-.'++#"

%&'()')*+(',*%*-"./"*01%&*2%&11&

%&'()')*+(',*%3*%4567*(3*',1*8%)'*01%&

%&'()')*+(',*%3*%4567*(3*',1*8%)'*9*01%&)

%&'()')*+,:)1*;(&)'*&141%)1*+%)*(3*#<=/

1%&40*=">)*%&'()')*+(',*%3*%4567*(3*',1*8%)'*01%&

3%?@1*)8:''1&*'&%(31A*:3*:340*B%A:33%*):3C)*********D--!9$*8&12()(:3E

13'(&1*B6)(2*F&%(3)*'%G:3:70**D"!"""9$*8&12()(:3E!"""#$

!""#$

!"#$

!#$

#$

#"$

#""$

!"""#$ !""#$ !"#$ !#$ #$ #"$ #""$

!"#$%&%'()'*)+,#)-.'++#"

%&'()')*+,&-*(-*#./0*1,&*+%-2)*3,4-252*(-*#./06

%-*%7+48*(-*'95*:%)'*';,*<5%&)

%&'()')*;('9*,-7<*,-5*%7+48

%&'()')*4-25&*=0*<5%&)*,72*1,&*+%-2)*75))*'95-*=0*<5%&)*,726

-%>?5*):,''5&*'&%(-52*,-*,-7<*@(7<*A775-*),-B)***************1C#$*:&5D()(,-6

5-'(&5*E4)(D*F&%(-G*'%H,-,8<*1"!""C$*:&5D()(,-6

Madonna Lily Allen

Age restrictions, only one album, last year releases, extensive career etc...

24Thursday, October 29, 2009

Page 25: Entity Spotting in Informal Text

Take aways..

• Real world restrictions closely follow distribution of random restrictions, conforming loosely to a Zipf distribution

• Confirms general effectiveness of limiting domain size regardless of restriction

• Choosing which constraints to implement is simple - pick whatever is easiest first

• use metadata from the model to guide you

25Thursday, October 29, 2009

Page 26: Entity Spotting in Informal Text

Non-music Mentions

26Thursday, October 29, 2009

Page 27: Entity Spotting in Informal Text

Disambiguating Non-music References

UGC on Lily Allen’s page about her new track Smile

Got your new album Smile. Loved it!

Keep your SMILE on!

27Thursday, October 29, 2009

Page 28: Entity Spotting in Informal Text

Binary Classification, SVM

Training data

550 good spots

550 bad spots

Test data

120 good spots

229 * 2 bad spots

Syntactic features Notation-S+POS tag of s s.POS

POS tag of one token before s s.POSb

POS tag of one token after s s.POSa

Typed dependency between s and sentiment word s.POS-TDsent∗

Typed dependency between s and domain-specific term s.POS-TDdom∗

Boolean Typed dependency between s and sentiment s.B-TDsent∗

Boolean Typed dependency between s and domain-specific term s.B-TDdom∗

Word-level features Notation-W+Capitalization of spot s s.allCaps

+Capitalization of first letter of s s.firstCaps

+s in Quotes s.inQuotes

Domain-specific features Notation-DSentiment expression in the same sentence as s s.Ssent

Sentiment expression elsewhere in the comment s.Csent

Domain-related term in the same sentence as s s.Sdom

Domain-related term elsewhere in the comment s.Cdom+Refers to basic features, others are advanced features

∗These features apply only to one-word-long spots.

Table 6. Features used by the SVM learner

Valid spot: Got your new album Smile.

Simply loved it!

Encoding: nsubj(loved-8, Smile-5) imply-

ing that Smile is the nominal subject of

the expression loved.

Invalid spot: Keep your smile on. You’ll

do great !Encoding: No typed dependency between

smile and great

Table 7. Typed Dependencies Example

Typed Dependencies:

We also captured the typed de-

pendency paths (grammatical rela-

tions) via the s.POS-

TDsent and s.POS-TDdom boolean

features. These were obtained be-

tween a spot and co-occurring senti-

ment and domain-specific words by

the Stanford parser[12] (see exam-

ple in 7). We also encode a boolean

value indicating whether a relation

was found at all using the s.B-TDsent

and s.B-TDdom features. This allows us to accommodate parse errors given the

informal and often non-grammatical English in this corpus.

5.2 Data and Experiments

Our training and test data sets were obtained from the hand-tagged data (see

Table 3). Positive and negative training examples were all spots that all four

annotators had confirmed as valid or invalid respectively, for a total of 571 posi-

tive and 864 negative examples. Of these, we used 550 positive and 550 negative

examples for training. The remaining spots were used for test purposes.

Our positive and negative test sets comprised of all spots that three annota-

tors had confirmed as valid or invalid spots, i.e. had a 75% agreement. We also

included spots where 50% of the annotators had agreement on the validity of the

*

*

*

*

Got your new album Smile. Loved it!Keep your SMILE on!

28Thursday, October 29, 2009

Page 29: Entity Spotting in Informal Text

Most Useful Combinations

TP best : word, domain, contextual

TP next best : word, domain, contextual (POS)

FP best : All features, other combinations

Not all syntactic features are useless, contrary to general

belief, wrt informal text

90-35

78-50

42-91

Recall intensive

Prec

ision

inte

nsive

29Thursday, October 29, 2009

Page 30: Entity Spotting in Informal Text

Naive MB spotter + NLP

• Annotate using naive spotter

• best case baseline (artist is known)

• follow with NLP analytics to weed out FPs

• run on less than entire input data

!"

#!"

$!"

%!"

&!"

'!!"

()*+,

-./00,1

2!345

6&35!

6$3$!

6#345

6'36!

6!35%

%#3&$

%'36!

%!3&5

$53&%

$#32'

!"#$$%&%'()#**+(#*,)$-"%.$)/0#"%12%30#"%14

5('*%$%63)7)8'*#""

71,89-9/(:;/1:<9=>:?==,(71,89-9/(:;/1:@9A)(()71,89-9/(:;/1:B)C/(()@,8)==:D)==:0A1,,E

PR tradeoffs: choosing feature combinations depending on end

application requirement

30Thursday, October 29, 2009

Page 31: Entity Spotting in Informal Text

Summary..

• Real-time large-scale data processing

• prohibits computationally intensive NLP techniques

• Simple inexpensive NL learners over a dictionary-based naive spotter can yield reasonable performance

• restricting the taxonomy results in proportionally higher precision

• Spot + Disambiguate a feasible approach for (esply. Cultural) NER in Informal Text

31Thursday, October 29, 2009

Page 32: Entity Spotting in Informal Text

Thank You!

• Bing, Yahoo, Google: Meena Nagarajan

• Contact us

• {dgruhl, jhpieper, crobson}@us.ibm.com, {meena, amit}@knoesis.org

• More about this work

• http://www.almaden.ibm.com/cs/projects/iis/sound/

• http://knoesis.wright.edu/researchers/meena

32Thursday, October 29, 2009