TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Demystifying Word2Vec

129 pointsby buss_janover 8 years ago

5 comments

sdenton4over 8 years ago
I&#x27;m reading the New Michael Lewis book right now, on kahneman and tversky, and there&#x27;s a point where the notion of encoding similarity as distances is first very popular, but then torn apart in the sixties.<p>Essentially, it comes down to statements like &#x27;this woman is like Queen Elizabeth&#x27; being different from &#x27;Queen Elizabeth is like this woman.&#x27; human perception of similarity is assymetric. Tversky&#x27;s suggestion is that we essentially think if things as a collection of tags, and A is like B if A&#x27;s tags are mostly a subset of B&#x27;s tags. But if B has more tags (because it&#x27;s a specific thing that we know a lot about, for example) then B&#x27;s tags won&#x27;t mostly be a subset of A&#x27;s tags, creating the asymmetry.<p>This directly attacks the kind of similarity that word2vec relies on, and I&#x27;m wondering if there are critiques along these lines on the literature.
评论 #13592803 未加载
评论 #13591859 未加载
staredover 8 years ago
And if you want an interactive vis, there is one: <a href="https:&#x2F;&#x2F;lamyiowce.github.io&#x2F;word2viz&#x2F;" rel="nofollow">https:&#x2F;&#x2F;lamyiowce.github.io&#x2F;word2viz&#x2F;</a> (and some discussion: <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=13346104" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=13346104</a>).
评论 #13590421 未加载
NKCSSover 8 years ago
Isn&#x27;t this all based on LSA (Latent Semantic Analysis; e.g. first major paper was by by Landauer back in 1990, and going back to the 1960&#x27;s)?<p>I implemented LSA a few years ago based on these papers and it just seems like LSA&#x2F;LSI to me..<p>[Update]<p>Just found this:<p>&gt; In this sense we have come full circle to the methods presented earlier that rely on matrix factorization (such as LSA). Where LSA uses SVD to find the best fitting subspace in terms of the co-variances of words across documents, the Glove model explicitly optimizes wordvectors to reflect the likelihood of their co-occurrence. The point, however, stands that Word2Vec is not an entirely novel approach to arrive at word vectors that capture semantic meaning in their substructure.
评论 #13589227 未加载
ponderingHplusover 8 years ago
A good read, thanks! This technique largely inspired a project we did for school this year, a subreddit recommender system with an RNN learning an embedding space for subreddits. I&#x27;ve just finished up exams and am starting work on getting a minimal webapp up for people to play with, but links to the final report and an interactive bokeh plot of the final embedding can be found here: <a href="http:&#x2F;&#x2F;cole-maclean.github.io&#x2F;blog&#x2F;RNN-Based-Subreddit-Recommender-System&#x2F;" rel="nofollow">http:&#x2F;&#x2F;cole-maclean.github.io&#x2F;blog&#x2F;RNN-Based-Subreddit-Recom...</a>
vonnikover 8 years ago
This may be helpful, too: <a href="https:&#x2F;&#x2F;deeplearning4j.org&#x2F;word2vec" rel="nofollow">https:&#x2F;&#x2F;deeplearning4j.org&#x2F;word2vec</a>