- Joined
- Dec 7, 2020
So, an embedding for a document would look like a word cloud where the size of the word is proportional to the number of times the word appears in the document. Like "Wallace" only appeared once, so it's really small and "Bobert" appeared 100 times so it's the biggest in the cloud.
Then word embedding are the same thing, but with word similarity. Like cat and dog are similar but different, And Hound and Dog overlap because they are synonyms.
I this right? It some seem a bit schizophrenic when they start talking about making "multidimensional space" and then start talking about path finding algorithms and concordance systems.
Then word embedding are the same thing, but with word similarity. Like cat and dog are similar but different, And Hound and Dog overlap because they are synonyms.
I this right? It some seem a bit schizophrenic when they start talking about making "multidimensional space" and then start talking about path finding algorithms and concordance systems.