similarity and relatedness and asymmetric are ignored. Similarity Basically do not think about direction truck ⇔ car. People judgetiger → leopardis more similar toleopard → tiger. Relatedness storkco-occurs withbaby. babydoesn`t co-occurs so far withstork. • Experiment influence of similarity and relatedness. • Make the dataset 4
only to the similarity between words. Annotation by people. coast - shore : 9.00 clothes – closet : 1.96 • ProNorm Pay attention only to the relatedness between words. People generate the words. Annotator generates words that are not similar but related. Generate 1169 related words for 100 words. • WordSim353 Including similarity pairs and relatedness pairs. Response based either on taxonomic similarity or associative relatedness. 5
embedding only. Context embedding will eventually be discarded. 6 AA measure • Improved to save two vector information* • Consider symmetry and asymmetry between words * Improving distributional similarity with lessons learned from word embeddings (O.Levy et al. TACL 2015)
A and word B are word embedding. Commonly used similarity • CC Both word A and word B are context embedding. • WC Word A is word embedding, and word B is context embedding. • CW Word A is context embedding, and word B is word embedding. 7 Similarity between word A and word B. (A, B)
accuracy with ProNorm and SimLex. Use output score and correct data score • hypothesis SimLex (Similarity) WW measure is the best. The direction between two words is not a factor. ProNorm (Relatedness) WC measure is the best. Tell us how likely we would see w2 and similar words in the context of w1. 9
includes score between human similarity judgment and a linear regression. • Numbers in this row show an upper bound for Spearman scores of the individual measures.
Using ProNorm (Relatedness) data. w1 is given and generates the n most similar words. Count how many people responses were contained 13 dim : 300 win : 10