10/30/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Lecture 7 Giuseppe Carenini.
5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics...
-
Upload
tamsyn-thompson -
Category
Documents
-
view
218 -
download
4
Transcript of 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics...
![Page 1: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/1.jpg)
04/18/23 CPSC503 Winter 2009 1
CPSC 503Computational Linguistics
Computational Lexical SemanticsLecture 14
Giuseppe Carenini
![Page 2: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/2.jpg)
04/18/23 CPSC503 Winter 2009 2
Today 23/10
Three well-defined Semantic Task• Word Sense Disambiguation
– Corpus and Thesaurus
• Word Similarity– Thesaurus and Corpus
• Semantic Role Labeling (corpus)
![Page 3: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/3.jpg)
04/18/23 CPSC503 Winter 2009 3
WSD example: table + ?? -> [1-6]
The noun "table" has 6 senses in WordNet.1. table, tabular array -- (a set of data …)2. table -- (a piece of furniture …)3. table -- (a piece of furniture with tableware…)4. mesa, table -- (flat tableland …)5. table -- (a company of people …)6. board, table -- (food or meals …)
![Page 4: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/4.jpg)
04/18/23 CPSC503 Winter 2009 4
WSD methods
•Machine Learning – Supervised– Unsupervised
•Dictionary / Thesaurus (Lesk)
![Page 5: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/5.jpg)
04/18/23 CPSC503 Winter 2009 5
Supervised ML Approaches to WSD
MachineLearning
Classifier
TrainingData
((word + context1) sense1)……((word + contextn) sensen)
sense(word + context)
![Page 6: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/6.jpg)
04/18/23 CPSC503 Winter 2009 6
Training Data Example
..after the soup she had bass with a big salad…
((word + context) sense)i
context
Examples, • One of 8 possible senses for “bass”
in WordNet• One of the 2 key distinct senses for
“bass” in WordNet
sense
![Page 7: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/7.jpg)
04/18/23 CPSC503 Winter 2009 7
WordNet Bass: music vs. fishThe noun ``bass'' has 8 senses in WordNet1. bass - (the lowest part of the musical range)2. bass, bass part - (the lowest part in polyphonic
music)3. bass, basso - (an adult male singer with …)4. sea bass, bass - (flesh of lean-fleshed saltwater fish
of the family Serranidae)5. freshwater bass, bass - (any of various North
American lean-fleshed ………)6. bass, bass voice, basso - (the lowest adult male
singing voice)7. bass - (the member with the lowest range of a
family of musical instruments)8. bass -(nontechnical name for any of numerous
edible marine and freshwater spiny-finned fishes)
![Page 8: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/8.jpg)
04/18/23 CPSC503 Winter 2009 8
Representations for Context
• GOAL: Informative characterization of the window of text surrounding the target word
• Supervised ML requires a simple representation for the training data:
vectors of feature/value pairs
• TASK: Select relevant linguistic information, encode them as a feature vector
![Page 9: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/9.jpg)
04/18/23 CPSC503 Winter 2009 9
Relevant Linguistic Information(1)
• Collocational: info about the words that appear in specific positions to the right and left of the target word
• Example text (WSJ)– An electric guitar and bass player
stand off to one side not really part of the scene, …
Assume a window of +/- 2 from the target
[guitar, NN, and, CJC, player, NN, stand, VVB]
[word in position -n, part-of-speech position -n, …
word in position +n, part-of-speech position +n,]
Typically words and their POS
![Page 10: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/10.jpg)
04/18/23 CPSC503 Winter 2009 10
Relevant Linguistic Information(2)
Co-occurrence: info about the words that occur anywhere in the window regardless of position
• Find k content words that most frequently co-occur with target in corpus (for bass: fishing, big, sound, player, fly …, guitar, band))Vector for one case:
[c(fishing), c(big), c(sound), c(player), c(fly), …, c(guitar), c(band)]• Example text (WSJ)
– An electric guitar and bass player stand off to one side not really part of the scene, …
[0,0,0,1,0,0,0,0,0,0,1,0]
![Page 11: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/11.jpg)
04/18/23 CPSC503 Winter 2009 11
Training Data Examples
[guitar, NN, and, CJC, player, NN, stand, VVB, 0] [0,0,0,1,0,0,0,0,0,0,1,0,0]
Let’s assume: bass-music encoded as 0bass-fish encoded as 1
[a, AT0, sea, CJC, to, PRP, me, PNP, 1]
[play, VVB, the, AT0, with, PRP, others, PNP, 0]
[……… ]
[1,0,0,0,0,0,0,0,0,0,0,0,1]
[1,0,0,0,0,0,0,0,0,0,0,1,1][…………………..]
• Inputs to classifiers
[guitar, NN, and, CJC, could, VM0, be, VVI] [1,1,0,0,0,1,0,0,0,0,0,0]
![Page 12: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/12.jpg)
04/18/23 CPSC503 Winter 2009 12
ML for Classifiers
MachineLearning
Training Data:•Co-occurrence•Collocational
Classifier
• Naïve Bayes• Decision lists• Decision trees• Neural nets• Support vector machines• Nearest neighbor
methods…
![Page 13: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/13.jpg)
04/18/23 CPSC503 Winter 2009 13
Naïve Bayes
)|( argmaxs VsPSs
argmaxs Ss
n
jj
SssvPsP
1 )|()( argmaxs
)|( sVP
Independence
![Page 14: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/14.jpg)
04/18/23 CPSC503 Winter 2009 14
Naïve Bayes: EvaluationExperiment comparing different
classifiers [Mooney 96]• Naïve Bayes and Neural Network
achieved highest performance• 73% in assigning one of six senses
to line • Is this good?
• Simplest Baseline: “most frequent sense”
• Celing: human inter-annotator agreement– 75%-80% on refined sense distinctions
(wordnet)– Closer to 90% for binary distinctions
![Page 15: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/15.jpg)
04/18/23 CPSC503 Winter 2009 15
Bootstrapping• What if you don’t have enough data
to train a system…
More DataMoreClassifiedData
MachineLearning
SmallTrainingData
Classifier
seeds
![Page 16: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/16.jpg)
04/18/23 CPSC503 Winter 2009 16
Bootstrapping: how to pick the seeds
E.g., bass: play is strongly associated with the music sense whereas fish is strongly associated the fish sense
• Hand-labeling (Hearst 1991):– Likely correct– Likely to be prototypical
• One sense per collocation (Yarowsky 1995):
• One Sense Per Discourse: multiple occurrences of word in one discourse tend to have the same sense
![Page 17: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/17.jpg)
04/18/23 CPSC503 Winter 2009 17
Unsupervised Methods [Schutze ’98]
MachineLearning
(Clustering)
TrainingData
K Clusters ci
(word + vector)1
……(word + vector)n
Hand-labeling(c1 sense1)
……
(word + vector) senseVector/cluster
Similarity
![Page 18: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/18.jpg)
04/18/23 CPSC503 Winter 2009 18
Agglomerative Clustering
• Assign each instance to its own cluster• Repeat
– Merge the two clusters that are more similar
• Until (specified # of clusters is reached)
• If there are too many training instances ->random sampling
![Page 19: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/19.jpg)
04/18/23 CPSC503 Winter 2009 19
Problems
• Given these general ML approaches, how many classifiers do I need to perform WSD robustly– One for each ambiguous word in the
language
• How do you decide what set of tags/labels/senses to use for a given word?– Depends on the application
![Page 20: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/20.jpg)
04/18/23 CPSC503 Winter 2009 20
WDS: Dictionary and Thesaurus Methods
Most common: Lesk method• Choose the sense whose
dictionary gloss shares most words with the target word’s neighborhood
• Exclude stop-wordsDef: Words in gloss for a sense
is called the signature
![Page 21: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/21.jpg)
04/18/23 CPSC503 Winter 2009 21
Lesk: ExampleTwo SENSES for channel
S1: (n) channel (a passage for water (or other fluids) to flow through) "the fields were crossed with irrigation channels"; "gutters carried off the rain water into a series of channels under the street"
S2: (n) channel, television channel, TV channel (a television station and its programs) "a satellite TV channel"; "surfing through the channels"; "they offer more than one hundred channels" …..
“ most streets closed to the TV station were flooded because the main channel was clogged by heavy rain .”
![Page 22: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/22.jpg)
04/18/23 CPSC503 Winter 2009 22
Corpus LeskBest performer• If a corpus with annotated senses is
available• For each sense: add to the signature for
that sense, words “that frequently appear” in the sentences containing that sense
CORPUS……“most streets closed to the TV station were flooded because the main <S1> channel </S1> was clogged by heavy rain.…..
![Page 23: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/23.jpg)
04/18/23 CPSC503 Winter 2009 23
WSD: More Recent Trends SemEval workshops –
Cross Language Evaluation Forum (CLEF)
• Better ML techniques (e.g., Combining Classifiers)
• Combining ML and Lesk (Yuret,2004)
• Other Languages
• Building better/larger corpora
![Page 24: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/24.jpg)
04/18/23 CPSC503 Winter 2009 24
Today 22/10
• Word Sense Disambiguation• Word Similarity• Semantic Role Labeling
![Page 25: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/25.jpg)
04/18/23 CPSC503 Winter 2009 25
Word Similarity/Semantic DistanceActually relation between two sensessun vs. moon – mouth vs. food – hot
vs. cold
Applications?
• Thesaurus methods: measure distance in online thesauri (e.g., Wordnet)
• Distributional methods: finding if the two words appear in similar contexts
![Page 26: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/26.jpg)
04/18/23 26
WS: Thesaurus Methods (path-length)• Path-length sim based on isa hierarchies
),(log),(sim 2121path ccpathlencc
• If we do not have Word Sense Disambiguation
CPSC503 Winter 2009
![Page 27: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/27.jpg)
04/18/23 27
WS: Thesaurus Methods(info content)• Or not all edges are equal…. Add
probabilistic info derived from a corpus
N
ccount
c csubsensesci
i
)(
)(
)(P)(log)(I cPcC ),( 21 ccLCSprobability
Information Lowest Common Subsumer
)),((log),(sim 2121resnik ccLCSPcc CPSC503 Winter 2009
![Page 28: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/28.jpg)
04/18/23 28
WS: Thesaurus Methods(info-content)• One of best performers – Jiang-
Conrath distance
))(log)((log)),((log2),(d 212121 cPcPccLCSPccistJC
• This is a measure of distance. Reciprocal for similarity!
• See also extended LeskCPSC503 Winter 2009
![Page 29: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/29.jpg)
04/18/23 29
Best Performers
• Jiang-Conrath• Extended Lesk
• Wordnet::Similarity Package
Pedersen et al. 2004
CPSC503 Winter 2009
![Page 30: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/30.jpg)
04/18/23 CPSC503 Winter 2009 30
WS: Distributional Methods
• Do not have any thesauri for target language
• If you have thesaurus, still– Missing domain-specific (e.g., technical words)– Poor hyponym knowledge (for V) and nothing for
Adj and Adv– Difficult to compare senses from different
hierarchies• Solution: extract similarity from corpora
• Basic idea: two words are similar if they appear in similar contexts
![Page 31: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/31.jpg)
04/18/23 CPSC503 Winter 2009 31
WS Distributional Methods (1)
• Context: feature vector
Example: fi how many times wi appeared in the neighborhood of w
Stop list
),...,,( 21 Nfffw
![Page 32: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/32.jpg)
04/18/23 CPSC503 Winter 2009 32
WS Distributional Methods (2)
• More informative values (referred to as weights or measure of association in the literature)
• Point-wise Mutual Information
)()(
),(log),( 2
i
iiPMI wPwP
wwPwwassoc
• t-test
)()(
)()(),(),(
i
iiitestt
wPwP
wPwPwwPwwassoct
![Page 33: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/33.jpg)
04/18/23 CPSC503 Winter 2009 33
WS Distributional Methods (3)• Similarity between vectors
Not sensitive to extreme values
)cos(),(cos
wv
wv
w
w
v
vwvsim ine
v
w
N
iii
N
iii
Jaccard
wv
wvwvsim
1
1
),max(
),min(),(
Normalized
(weighted) number of overlapping features
![Page 34: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/34.jpg)
04/18/23 CPSC503 Winter 2009 34
WS Distributional Methods (4)
• Best combination overall (Curan 2003)– t-test for weights– Jaccard (or Dice) for vector similarity
![Page 35: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/35.jpg)
04/18/23 CPSC503 Winter 2009 35
Today 22/10
• Word Sense Disambiguation• Word Similarity• Semantic Role Labeling
![Page 36: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/36.jpg)
04/18/23 CPSC503 Winter 2009 36
Semantic Role LabelingTypically framed as a classification
problem [Gildea, Jurfsky 2002]1. Assign parse tree to input2. Find all predicate-bearing words
(PropBank, FrameNet)3. For each “governing” predicate:
determine for each synt. constituent which role (if any) it plays with respect to the predicate
Common constituent features: predicate, phrase type, head word and its POS, path, voice, linear position…… and many others
![Page 37: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/37.jpg)
04/18/23 CPSC503 Winter 2009 37
Semantic Role Labeling: Example
[issued, NP, Examiner, NNP, NPSVPVBD, active, before, …..]ARG0
predicate, phrase type, head word and its POS, path, voice, linear position…… and many others
![Page 38: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/38.jpg)
04/18/23 CPSC503 Winter 2009 38
Next Time
• Discourse and Dialog : Overview of Chapters 21 and 24
![Page 39: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/39.jpg)
04/18/23 39
WordSim: Thesaurus Methods(Extended Lesk)
• For each n-word phrase that occurs in both glosses, Extended Lesk adds in a score n2
CPSC503 Winter 2009
![Page 40: 5/16/2015CPSC503 Winter 20091 CPSC 503 Computational Linguistics Computational Lexical Semantics Lecture 14 Giuseppe Carenini.](https://reader036.fdocuments.us/reader036/viewer/2022062515/56649cf65503460f949c55c1/html5/thumbnails/40.jpg)
04/18/23 CPSC503 Winter 2009 40
WS: Thesaurus Methods(1)• Path-length based sim on hyper/hypo
hierarchies),(log),(sim 2121path ccpathlencc
• Information content word similarity (not all edges are equal)
N
ccount
c csubsensesci
i
)(
)(
)(P)(log)(I cPcC ),( 21 ccLCSprobability
Information Lowest Common Subsumer
)),((log),(sim 2121resnik ccLCSPcc