Back

Explore Courses Blog Tutorials Interview Questions
+1 vote
2 views
in Machine Learning by (4.2k points)

word2vec is an open-source tool by Google:

  • For each word it provides a vector of float values, what exactly do they represent?

  • There is also a paper on paragraph vector can anyone explain how they are using word2vec in order to obtain fixed length vector for a paragraph.

1 Answer

+2 votes
by (6.8k points)

So we have that a word to vector model has been trained on a precise corpus to be readily given words as inputs (one-hot encoding is a way for the input) to represent the word as a vector of the typically high dimensionality of numbers.

Does it carry information? Yes, we could say that it carries information about the words. Similar words have similar vectors. The point is that we humans, we tend to don't define the similarity however the neural network defines the similarity for us supported the corpus that we tend to provide.

Does it carry meaning? No, as a result of this illustration will include grammatical or syntactical data and so the solution would be that it doesn't carry any meaning.

To learn more about Word2vec, study the machine learning tutorial

Browse Categories

...