TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

A Word Is Worth a Thousand Vectors

127 pointsby astrobiasedabout 10 years ago

8 comments

estabout 10 years ago
As a native Chinese speaker, this comes so natural.<p>pork = pig + meat<p>So the year of 2015 is the year of ram&#x2F;sheep&#x2F;goat, in Chinese they literally means<p>Ram = male ∪ caprinae<p>sheep = wool ∪ caprinae<p>goat = mountain ∪ caprinae<p>basically, word composition is pretty common in analytic language like Chinese, but kinda new idea in fusional languages like English.
评论 #9189943 未加载
imhabout 10 years ago
I&#x27;ve always wondered about doing this in non-flat spaces. Like if I add the &quot;7100 miles west&quot; vector to the &quot;California&quot; point, I get Turkmenistan. If I add &quot;7100 miles west&quot; again, I get back near &quot;California.&quot; Similarly, adding the &quot;not&quot; vector twice might get you back where you started in a word embedding. Anyone know if anyone is working on this? It could be tricky because &quot;7100 miles west&quot; lives in the tangent space to the space &quot;California&quot; lives in, but that in itself could be an interesting thing to study in the context of words.
评论 #9189788 未加载
madcowherdabout 10 years ago
Wondering how this differs from the SemanticVectors package? Will have to look into word2vec further.
评论 #9187829 未加载
评论 #9187809 未加载
madsravnabout 10 years ago
Very exciting stuff. I love how you can take simple building blocks and create something elegant and fun with them.<p>However, why are there words more similar to &quot;vacation&quot; than &quot;vacation&quot;?
评论 #9188368 未加载
nlabout 10 years ago
I don&#x27;t understand how the item matching is working. Do they have textual descriptions of each item (including colors and patterns), or are they somehow building vectors for the images and then doing cross-modal vector calculations?<p>If it&#x27;s the first option, then generating those descriptions seems and important thing to mention.<p>If it&#x27;s the second, then it&#x27;s a pretty significant result! I&#x27;ve seen some papers that indicate some possibilities in that area, but never anything working as well as this.
评论 #9188141 未加载
sinwaveabout 10 years ago
Am I wrong that the title seems to imply something negative about word vectors? But the article is super pumped about them!
评论 #9187541 未加载
hmate9about 10 years ago
By simple algerba, we can now prove that a picture is worth 1,000,000 vectors.
Dewieabout 10 years ago
&gt; If the gender axis is more positive, then it&#x27;s more feminine; more negative, more masculine.<p>Reminds me of my java textbook: the example was to model some employee[1] and it&#x27;s gender was a `boolean`: `false` for man; `true` for woman. Of course that was just an intermediate example before they showed off the `enum` solution.<p>[1] because hey, java OOP + CRUD business application example == match made in heaven as an example, apparently.