Why does word2Vec use cosine similarity?

后端 未结 2 1614
無奈伤痛
無奈伤痛 2021-02-01 21:11

I have been reading the papers on Word2Vec (e.g. this one), and I think I understand training the vectors to maximize the probability of other words found in the same contexts.<

2条回答
  •  粉色の甜心
    2021-02-01 21:41

    Those two distance metrics are probably strongly correlated so it might not matter all that much which one you use. As you point out, cosine distance means we don't have to worry about the length of the vectors at all.

    This paper indicates that there is a relationship between the frequency of the word and the length of the word2vec vector. http://arxiv.org/pdf/1508.02297v1.pdf

提交回复
热议问题