Top Banner
Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011
119

Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Dec 19, 2015

Download

Documents

Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Lexical Semantics & Word Sense

DisambiguationLing571

Deep Processing Techniques for NLPFebruary 16, 2011

Page 2: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

RoadmapLexical semantics

Lexical taxonomyWordNet

Thematic RolesIssues Resources:

PropBank & FrameNet

Selectional Restrictions

Primitive decompositions

Page 3: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

WordNet TaxonomyMost widely used English sense resource

Manually constructed lexical database3 Tree-structured hierarchies

Nouns (117K) , verbs (11K), adjective+adverb (27K)Entries: synonym set, gloss, example use

Relations between entries:Synonymy: in synsetHypo(per)nym: Isa tree

Page 4: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

WordNet

Page 5: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Noun WordNet Relations

Page 6: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

WordNet Taxonomy

Page 7: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesDescribe semantic roles of verbal arguments

Capture commonality across verbs

Page 8: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesDescribe semantic roles of verbal arguments

Capture commonality across verbsE.g. subject of break, open is AGENT

AGENT: volitional causeTHEME: things affected by action

Page 9: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesDescribe semantic roles of verbal arguments

Capture commonality across verbsE.g. subject of break, open is AGENT

AGENT: volitional causeTHEME: things affected by action

Enables generalization over surface order of argumentsJohnAGENT broke the windowTHEME

Page 10: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesDescribe semantic roles of verbal arguments

Capture commonality across verbsE.g. subject of break, open is AGENT

AGENT: volitional causeTHEME: things affected by action

Enables generalization over surface order of argumentsJohnAGENT broke the windowTHEME

The rockINSTRUMENT broke the windowTHEME

Page 11: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesDescribe semantic roles of verbal arguments

Capture commonality across verbsE.g. subject of break, open is AGENT

AGENT: volitional causeTHEME: things affected by action

Enables generalization over surface order of argumentsJohnAGENT broke the windowTHEME

The rockINSTRUMENT broke the windowTHEME

The windowTHEME was broken by JohnAGENT

Page 12: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesThematic grid, θ-grid, case frame

Set of thematic role arguments of verb

Page 13: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesThematic grid, θ-grid, case frame

Set of thematic role arguments of verbE.g. Subject:AGENT; Object:THEME, or Subject: INSTR; Object: THEME

Page 14: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesThematic grid, θ-grid, case frame

Set of thematic role arguments of verbE.g. Subject:AGENT; Object:THEME, or Subject: INSTR; Object: THEME

Verb/Diathesis AlternationsVerbs allow different surface realizations of roles

Page 15: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesThematic grid, θ-grid, case frame

Set of thematic role arguments of verbE.g. Subject:AGENT; Object:THEME, or Subject: INSTR; Object: THEME

Verb/Diathesis AlternationsVerbs allow different surface realizations of roles

DorisAGENT gave the bookTHEME to CaryGOAL

Page 16: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesThematic grid, θ-grid, case frame

Set of thematic role arguments of verbE.g. Subject:AGENT; Object:THEME, or Subject: INSTR; Object: THEME

Verb/Diathesis AlternationsVerbs allow different surface realizations of roles

DorisAGENT gave the bookTHEME to CaryGOAL

DorisAGENT gave CaryGOAL the bookTHEME

Page 17: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic RolesThematic grid, θ-grid, case frame

Set of thematic role arguments of verbE.g. Subject:AGENT; Object:THEME, or Subject: INSTR; Object: THEME

Verb/Diathesis AlternationsVerbs allow different surface realizations of roles

DorisAGENT gave the bookTHEME to CaryGOAL

DorisAGENT gave CaryGOAL the bookTHEME

Group verbs into classes based on shared patterns

Page 18: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Canonical Roles

Page 19: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic Role IssuesHard to produce

Page 20: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic Role IssuesHard to produce

Standard set of rolesFragmentation: Often need to make more specific

E,g, INSTRUMENTS can be subject or not

Page 21: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic Role IssuesHard to produce

Standard set of rolesFragmentation: Often need to make more specific

E,g, INSTRUMENTS can be subject or not

Standard definition of rolesMost AGENTs: animate, volitional, sentient, causalBut not all….

Strategies: Generalized semantic roles: PROTO-AGENT/PROTO-PATIENT

Defined heuristically: PropBank Define roles specific to verbs/nouns: FrameNet

Page 22: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic Role IssuesHard to produce

Standard set of rolesFragmentation: Often need to make more specific

E,g, INSTRUMENTS can be subject or not

Standard definition of rolesMost AGENTs: animate, volitional, sentient, causalBut not all….

Page 23: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic Role IssuesHard to produce

Standard set of rolesFragmentation: Often need to make more specific

E,g, INSTRUMENTS can be subject or not

Standard definition of rolesMost AGENTs: animate, volitional, sentient, causalBut not all….

Strategies: Generalized semantic roles: PROTO-AGENT/PROTO-PATIENT

Defined heuristically: PropBank

Page 24: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Thematic Role IssuesHard to produce

Standard set of rolesFragmentation: Often need to make more specific

E,g, INSTRUMENTS can be subject or not

Standard definition of rolesMost AGENTs: animate, volitional, sentient, causalBut not all….

Strategies: Generalized semantic roles: PROTO-AGENT/PROTO-PATIENT

Defined heuristically: PropBank Define roles specific to verbs/nouns: FrameNet

Page 25: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

PropBankSentences annotated with semantic roles

Penn and Chinese Treebank

Page 26: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

PropBankSentences annotated with semantic roles

Penn and Chinese TreebankRoles specific to verb sense

Numbered: Arg0, Arg1, Arg2,… Arg0: PROTO-AGENT; Arg1: PROTO-PATIENT, etc

Page 27: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

PropBankSentences annotated with semantic roles

Penn and Chinese TreebankRoles specific to verb sense

Numbered: Arg0, Arg1, Arg2,… Arg0: PROTO-AGENT; Arg1: PROTO-PATIENT, etc

E.g. agree.01Arg0: Agreer

Page 28: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

PropBankSentences annotated with semantic roles

Penn and Chinese TreebankRoles specific to verb sense

Numbered: Arg0, Arg1, Arg2,… Arg0: PROTO-AGENT; Arg1: PROTO-PATIENT, etc

E.g. agree.01Arg0: AgreerArg1: Proposition

Page 29: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

PropBankSentences annotated with semantic roles

Penn and Chinese TreebankRoles specific to verb sense

Numbered: Arg0, Arg1, Arg2,… Arg0: PROTO-AGENT; Arg1: PROTO-PATIENT, etc

E.g. agree.01Arg0: AgreerArg1: PropositionArg2: Other entity agreeing

Page 30: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

PropBankSentences annotated with semantic roles

Penn and Chinese TreebankRoles specific to verb sense

Numbered: Arg0, Arg1, Arg2,… Arg0: PROTO-AGENT; Arg1: PROTO-PATIENT, etc

E.g. agree.01Arg0: AgreerArg1: PropositionArg2: Other entity agreeingEx1: [Arg0The group] agreed [Arg1it wouldn’t make an

offer]

Page 31: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

FrameNetSemantic roles specific to Frame

Frame: script-like structure, roles (frame elements)

Page 32: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

FrameNetSemantic roles specific to Frame

Frame: script-like structure, roles (frame elements)

E.g. change_position_on_scale: increase, riseAttribute, Initial_value, Final_value

Page 33: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

FrameNetSemantic roles specific to Frame

Frame: script-like structure, roles (frame elements)

E.g. change_position_on_scale: increase, riseAttribute, Initial_value, Final_value

Core, non-core roles

Page 34: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

FrameNetSemantic roles specific to Frame

Frame: script-like structure, roles (frame elements)

E.g. change_position_on_scale: increase, riseAttribute, Initial_value, Final_value

Core, non-core roles

Relationships b/t frames, frame elementsAdd causative: cause_change_position_on_scale

Page 35: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.
Page 36: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional RestrictionsSemantic type constraint on arguments

I want to eat someplace close to UW

Page 37: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional RestrictionsSemantic type constraint on arguments

I want to eat someplace close to UWE.g. THEME of eating should be edible

Associated with senses

Page 38: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional RestrictionsSemantic type constraint on arguments

I want to eat someplace close to UWE.g. THEME of eating should be edible

Associated with senses

Vary in specificity:

Page 39: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional RestrictionsSemantic type constraint on arguments

I want to eat someplace close to UWE.g. THEME of eating should be edible

Associated with senses

Vary in specificity: Imagine: AGENT: human/sentient; THEME: any

Page 40: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional RestrictionsSemantic type constraint on arguments

I want to eat someplace close to UWE.g. THEME of eating should be edible

Associated with senses

Vary in specificity: Imagine: AGENT: human/sentient; THEME: any

Representation:Add as predicate in FOL event representation

Page 41: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional RestrictionsSemantic type constraint on arguments

I want to eat someplace close to UWE.g. THEME of eating should be edible

Associated with senses

Vary in specificity: Imagine: AGENT: human/sentient; THEME: any

Representation:Add as predicate in FOL event representation

Overkill computationally; requires large commonsense KB

Page 42: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional RestrictionsSemantic type constraint on arguments

I want to eat someplace close to UWE.g. THEME of eating should be edible

Associated with senses

Vary in specificity: Imagine: AGENT: human/sentient; THEME: any

Representation:Add as predicate in FOL event representation

Overkill computationally; requires large commonsense KBAssociate with WordNet synset (and hyponyms)

Page 43: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Primitive Decompositions Jackendoff(1990), Dorr(1999), McCawley (1968)

Word meaning constructed from primitivesFixed small set of basic primitives

E.g. cause, go, become, kill=cause X to become Y

Page 44: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Primitive Decompositions Jackendoff(1990), Dorr(1999), McCawley (1968)

Word meaning constructed from primitivesFixed small set of basic primitives

E.g. cause, go, become, kill=cause X to become Y

Augment with open-ended “manner”Y = not aliveE.g. walk vs run

Page 45: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Primitive Decompositions Jackendoff(1990), Dorr(1999), McCawley (1968)

Word meaning constructed from primitivesFixed small set of basic primitives

E.g. cause, go, become, kill=cause X to become Y

Augment with open-ended “manner”Y = not aliveE.g. walk vs run

Fixed primitives/Infinite descriptors

Page 46: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Word Sense Disambiguation

Selectional Restriction-based approaches Limitations

Robust Approaches Supervised Learning Approaches

Naïve Bayes Dictionary-based Approaches Bootstrapping Approaches

One sense per discourse/collocation Unsupervised Approaches

Schutze’s word space Resource-based Approaches

Dictionary parsing, WordNet Distance Why they work Why they don’t

Page 47: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Word Sense Disambiguation

Application of lexical semantics

Goal: Given a word in context, identify the appropriate senseE.g. plants and animals in the rainforest

Crucial for real syntactic & semantic analysis

Page 48: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Word Sense Disambiguation

Application of lexical semantics

Goal: Given a word in context, identify the appropriate senseE.g. plants and animals in the rainforest

Crucial for real syntactic & semantic analysisCorrect sense can determine

.

Page 49: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Word Sense Disambiguation

Application of lexical semantics

Goal: Given a word in context, identify the appropriate senseE.g. plants and animals in the rainforest

Crucial for real syntactic & semantic analysisCorrect sense can determine

Available syntactic structureAvailable thematic roles, correct meaning,..

Page 50: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restriction Approaches

Integrate sense selection in parsing and semantic analysis – e.g. with lambda calculus

Concept: Predicate selects senseWashing dishes vs stir-frying dishes

Page 51: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restriction Approaches

Integrate sense selection in parsing and semantic analysis – e.g. with lambda calculus

Concept: Predicate selects senseWashing dishes vs stir-frying dishes

Stir-fry: patient: food => dish~food

Page 52: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restriction Approaches

Integrate sense selection in parsing and semantic analysis – e.g. with lambda calculus

Concept: Predicate selects senseWashing dishes vs stir-frying dishes

Stir-fry: patient: food => dish~foodServe Denver vs serve breakfast

Serve vegetarian dishes

Page 53: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restriction Approaches

Integrate sense selection in parsing and semantic analysis – e.g. with lambda calculus

Concept: Predicate selects senseWashing dishes vs stir-frying dishes

Stir-fry: patient: food => dish~foodServe Denver vs serve breakfast

Serve vegetarian dishes Serve1: patient: loc; serve1: patient: food => dishes~food: only valid variant

Page 54: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restriction Approaches

Integrate sense selection in parsing and semantic analysis – e.g. with lambda calculus

Concept: Predicate selects senseWashing dishes vs stir-frying dishes

Stir-fry: patient: food => dish~foodServe Denver vs serve breakfast

Serve vegetarian dishes Serve1: patient: loc; serve1: patient: food => dishes~food: only valid variant

Integrate in rule-to-rule: test e.g. in WN

Page 55: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restrictions: Limitations

Problem 1:

Page 56: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restrictions: Limitations

Problem 1: Predicates too generalRecommend, like, hit….

Problem 2:

Page 57: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restrictions: Limitations

Problem 1: Predicates too generalRecommend, like, hit….

Problem 2: Language too flexible“The circus performer ate fire and swallowed

swords”Unlikely but doable

Also metaphor

Page 58: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restrictions: Limitations

Problem 1: Predicates too generalRecommend, like, hit….

Problem 2: Language too flexible“The circus performer ate fire and swallowed

swords”Unlikely but doable

Also metaphor

Strong restrictions would block all analysis

Page 59: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restrictions: Limitations

Problem 1: Predicates too generalRecommend, like, hit….

Problem 2: Language too flexible“The circus performer ate fire and swallowed

swords”Unlikely but doable

Also metaphor

Strong restrictions would block all analysisSome approaches generalize up hierarchy

Can over-accept truly weird things

Page 60: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Selectional Restrictions: Limitations

Problem 1: Predicates too generalRecommend, like, hit….

Problem 2: Language too flexible“The circus performer ate fire and swallowed swords”

Unlikely but doable

Also metaphor

Strong restrictions would block all analysisSome approaches generalize up hierarchy

Can over-accept truly weird things Selectional preferences: apply weighted preferences

Page 61: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Robust DisambiguationMore to semantics than P-A structure

Select sense where predicates underconstrain

Page 62: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Robust DisambiguationMore to semantics than P-A structure

Select sense where predicates underconstrain

Learning approachesSupervised, Bootstrapped, Unsupervised

Page 63: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Robust DisambiguationMore to semantics than P-A structure

Select sense where predicates underconstrain

Learning approachesSupervised, Bootstrapped, Unsupervised

Knowledge-based approachesDictionaries, Taxonomies

Widen notion of context for sense selection

Page 64: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Robust DisambiguationMore to semantics than P-A structure

Select sense where predicates underconstrain

Learning approachesSupervised, Bootstrapped, Unsupervised

Knowledge-based approachesDictionaries, Taxonomies

Widen notion of context for sense selectionWords within window (2,50,discourse)Narrow cooccurrence - collocations

Page 65: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Page 66: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Part of speech Of word and neighbors

Page 67: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Part of speech Of word and neighbors

Morphologically simplified form

Page 68: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Part of speech Of word and neighbors

Morphologically simplified formWords in neighborhood

Page 69: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Part of speech Of word and neighbors

Morphologically simplified formWords in neighborhood

Question: How big a neighborhood?

Page 70: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Part of speech Of word and neighbors

Morphologically simplified formWords in neighborhood

Question: How big a neighborhood? Is there a single optimal size? Why?

..

Page 71: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Part of speech Of word and neighbors

Morphologically simplified formWords in neighborhood

Question: How big a neighborhood? Is there a single optimal size? Why?

(Possibly shallow) Syntactic analysisE.g. predicate-argument relations, modification,

phrasesCollocation vs co-occurrence features

Page 72: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Part of speech Of word and neighbors

Morphologically simplified formWords in neighborhood

Question: How big a neighborhood? Is there a single optimal size? Why?

(Possibly shallow) Syntactic analysisE.g. predicate-argument relations, modification,

phrasesCollocation vs co-occurrence features

Collocation: words in specific relation: p-a, 1 word +/-

Page 73: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Disambiguation FeaturesKey: What are the features?

Part of speech Of word and neighbors

Morphologically simplified formWords in neighborhood

Question: How big a neighborhood? Is there a single optimal size? Why?

(Possibly shallow) Syntactic analysisE.g. predicate-argument relations, modification, phrases

Collocation vs co-occurrence featuresCollocation: words in specific relation: p-a, 1 word +/-Co-occurrence: bag of words..

Page 74: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

WSD EvaluationIdeally, end-to-end evaluation with WSD

componentDemonstrate real impact of technique in system

Page 75: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

WSD EvaluationIdeally, end-to-end evaluation with WSD

componentDemonstrate real impact of technique in systemDifficult, expensive, still application specific

Page 76: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

WSD EvaluationIdeally, end-to-end evaluation with WSD

componentDemonstrate real impact of technique in systemDifficult, expensive, still application specific

Typically, intrinsic, sense-basedAccuracy, precision, recallSENSEVAL/SEMEVAL: all words, lexical sample

Baseline:

Page 77: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

WSD EvaluationIdeally, end-to-end evaluation with WSD component

Demonstrate real impact of technique in systemDifficult, expensive, still application specific

Typically, intrinsic, sense-basedAccuracy, precision, recallSENSEVAL/SEMEVAL: all words, lexical sample

Baseline:Most frequent sense, Lesk

Topline:

Page 78: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

WSD Evaluation Ideally, end-to-end evaluation with WSD component

Demonstrate real impact of technique in system Difficult, expensive, still application specific

Typically, intrinsic, sense-based Accuracy, precision, recall SENSEVAL/SEMEVAL: all words, lexical sample

Baseline: Most frequent sense, Lesk

Topline: Human inter-rater agreement: 75-80% fine; 90% coarse

Page 79: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Naïve Bayes’ ApproachSupervised learning approach

Input: feature vector X label

Page 80: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Naïve Bayes’ ApproachSupervised learning approach

Input: feature vector X label

Best sense = most probable sense given f

Page 81: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Naïve Bayes’ ApproachIssue:

Data sparseness: full feature vector rarely seen

Page 82: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Naïve Bayes’ ApproachIssue:

Data sparseness: full feature vector rarely seen

“Naïve” assumption: Features independent given sense

Page 83: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Training NB Classifier

Estimate P(s):Prior

Page 84: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Training NB Classifier

Estimate P(s):Prior

Page 85: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Training NB Classifier

Estimate P(s):Prior

Estimate P(fj|s)

Page 86: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Training NB Classifier

Estimate P(s):Prior

Estimate P(fj|s)

Issues:

Page 87: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Training NB Classifier

Estimate P(s):Prior

Estimate P(fj|s)

Issues:Underflow => log prob

Page 88: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Training NB Classifier

Estimate P(s):Prior

Estimate P(fj|s)

Issues:Underflow => log probSparseness => smoothing

Page 89: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Dictionary-Based Approach

(Simplified) Lesk algorithm“How to tell a pine cone from an ice cream cone”

Page 90: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Dictionary-Based Approach

(Simplified) Lesk algorithm“How to tell a pine cone from an ice cream cone”

Compute context ‘signature’ of word to disambiguateWords in surrounding sentence(s)

Page 91: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Dictionary-Based Approach

(Simplified) Lesk algorithm“How to tell a pine cone from an ice cream cone”

Compute context ‘signature’ of word to disambiguateWords in surrounding sentence(s)

Compare overlap w.r.t. dictionary entries for senses

Page 92: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Dictionary-Based Approach

(Simplified) Lesk algorithm“How to tell a pine cone from an ice cream cone”

Compute context ‘signature’ of word to disambiguateWords in surrounding sentence(s)

Compare overlap w.r.t. dictionary entries for senses

Select sense with highest (non-stopword) overlap

Page 93: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Applying Lesk The bank can guarantee deposits will eventually cover

future tuition costs because it invests in mortgage securities.

Page 94: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Applying Lesk The bank can guarantee deposits will eventually cover

future tuition costs because it invests in mortgage securities.

Bank1 : 2

Page 95: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Applying Lesk The bank can guarantee deposits will eventually cover future

tuition costs because it invests in mortgage securities.

Bank1 : 2

Bank2: 0

Page 96: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Improving LeskOverlap score:

All words equally weighted (excluding stopwords)

Page 97: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Improving LeskOverlap score:

All words equally weighted (excluding stopwords)

Not all words equally informative

Page 98: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Improving LeskOverlap score:

All words equally weighted (excluding stopwords)

Not all words equally informativeOverlap with unusual/specific words – betterOverlap with common/non-specific words – less

good

Page 99: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Improving LeskOverlap score:

All words equally weighted (excluding stopwords)

Not all words equally informativeOverlap with unusual/specific words – betterOverlap with common/non-specific words – less

good

Employ corpus weighting: IDF: inverse document frequency

Idfi = log (Ndoc/ndi)

Page 100: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Minimally Supervised WSDYarowsky’s algorithm (1995)

Bootstrapping approach:Use small labeled seedset to iteratively train

Page 101: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Minimally Supervised WSDYarowsky’s algorithm (1995)

Bootstrapping approach:Use small labeled seedset to iteratively train

Builds on 2 key insights:One Sense Per Discourse

Word appearing multiple times in text has same sense

Corpus of 37232 bass instances: always single sense

Page 102: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Minimally Supervised WSD Yarowsky’s algorithm (1995)

Bootstrapping approach: Use small labeled seedset to iteratively train

Builds on 2 key insights: One Sense Per Discourse

Word appearing multiple times in text has same senseCorpus of 37232 bass instances: always single sense

One Sense Per CollocationLocal phrases select single sense

Fish -> Bass1

Play -> Bass2

Page 103: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Yarowsky’s AlgorithmTraining Decision Lists

1. Pick Seed Instances & Tag2. Find Collocations: Word Left, Word Right,

Word +K(A) Calculate Informativeness on Tagged Set,

Order:

(B) Tag New Instances with Rules(C)* Apply 1 Sense/Discourse(D) If Still Unlabeled, Go To 2

3. Apply 1 Sense/Discourse

Disambiguation: First Rule Matched

Page 104: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Yarowsky Decision List

Page 105: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Iterative Updating

Page 106: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

There are more kinds of plants and animals in the rainforests than anywhere else on Earth. Over half of the millions of known species of plants and animals live in the rainforest. Many are found nowhere else. There are even plants and animals in therainforest that we have not yet discovered.Biological Example

The Paulus company was founded in 1938. Since those days the product range has been the subject of constant expansions and is brought up continuously to correspond with the state of the art. We’re engineering, manufacturing and commissioning world-wide ready-to-run plants packed with our comprehensive know-how. Our Product Range includes pneumatic conveying systems for carbon, carbide, sand, lime andmany others. We use reagent injection in molten metal for the…Industrial Example

Label the First Use of “Plant”

Page 107: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Sense Choice With Collocational Decision

ListsCreate Initial Decision List

Rules Ordered by

Check nearby Word Groups (Collocations)Biology: “Animal” in + 2-10 words Industry: “Manufacturing” in + 2-10 words

Result: Correct Selection95% on Pair-wise tasks

Page 108: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Schutze’s Vector Space: Detail

Build a co-occurrence matrixRestrict Vocabulary to 4 letter sequencesExclude Very Frequent - Articles, AffixesEntries in 5000-5000 Matrix

Word Context4grams within 1001 CharactersSum & Normalize Vectors for each 4gramDistances between Vectors by dot product

97 Real Values

Page 109: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Schutze’s Vector Space: continued

Word Sense DisambiguationContext Vectors of All Instances of WordAutomatically Cluster Context VectorsHand-label Clusters with Sense TagTag New Instance with Nearest Cluster

Page 110: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Sense Selection in “Word Space”

Build a Context Vector1,001 character window - Whole Article

Compare Vector Distances to Sense ClustersOnly 3 Content Words in CommonDistant Context VectorsClusters - Build Automatically, Label Manually

Result: 2 Different, Correct Senses92% on Pair-wise tasks

Page 111: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Resnik’s WordNet Labeling: Detail

Assume Source of Clusters

Assume KB: Word Senses in WordNet IS-A hierarchy

Assume a Text Corpus

Calculate Informativeness For Each KB Node:

Sum occurrences of it and all children Informativeness

Disambiguate wrt Cluster & WordNet Find MIS for each pair, I For each subsumed sense, Vote += I Select Sense with Highest Vote

Page 112: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Sense Labeling Under WordNet

Use Local Content Words as ClustersBiology: Plants, Animals, Rainforests, species… Industry: Company, Products, Range, Systems…

Find Common Ancestors in WordNetBiology: Plants & Animals isa Living Thing Industry: Product & Plant isa Artifact isa EntityUse Most Informative

Result: Correct Selection

Page 113: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

The Question of ContextShared Intuition:

Context -> Sense

Area of Disagreement:What is context?

Wide vs Narrow Window

Word Co-occurrences

Page 114: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Taxonomy of Contextual Information

Topical Content

Word Associations

Syntactic Constraints

Selectional Preferences

World Knowledge & Inference

Page 115: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

A Trivial Definition of Context

All Words within X words of Target

Many words: Schutze - 1000 characters, several sentences

Unordered “Bag of Words”

Information Captured: Topic & Word Association

Limits on ApplicabilityNouns vs. Verbs & AdjectivesSchutze: Nouns - 92%, “Train” -Verb, 69%

Page 116: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Limits of Wide ContextComparison of Wide-Context Techniques (LTV ‘93)

Neural Net, Context Vector, Bayesian Classifier, Simulated AnnealingResults: 2 Senses - 90+%; 3+ senses ~ 70%People: Sentences ~100%; Bag of Words: ~70%

Inadequate Context

Need Narrow ContextLocal Constraints OverrideRetain Order, Adjacency

Page 117: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Surface Regularities = Useful Disambiguators

Not Necessarily!

Right for the Wrong ReasonBurglar Rob… Thieves Stray Crate Chase Lookout

Learning the Corpus, not the SenseThe “Ste.” Cluster: Dry Oyster Whisky Hot Float

Ice

Learning Nothing Useful, Wrong QuestionKeeping: Bring Hoping Wiping Could Should

Some Them Rest

Page 118: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

Interactions Below the Surface

Constraints Not All Created Equal “The Astronomer Married the Star”Selectional Restrictions Override Topic

No Surface Regularities “The emigration/immigration bill guaranteed

passports to all Soviet citizensNo Substitute for Understanding

Page 119: Lexical Semantics & Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 16, 2011.

What is SimilarAd-hoc Definitions of Sense

Cluster in “word space”, WordNet Sense, “Seed Sense”: Circular

Schutze: Vector Distance in Word Space

Resnik: Informativeness of WordNet Subsumer + ClusterRelation in Cluster not WordNet is-a hierarchy

Yarowsky: No Similarity, Only DifferenceDecision Lists - 1/PairFind Discriminants