TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Language, trees, and geometry in neural networks

125 pointsby 1wheelalmost 6 years ago

3 comments

ttctciyfalmost 6 years ago
This looks to be an interesting piece on a very interesting paper! Somewhat tangentially (I&#x27;m afraid) I just wanted to comment on this para from the article&#x27;s intro:<p>&gt; Language is made of discrete structures, yet neural networks operate on continuous data: vectors in high-dimensional space. A successful language-processing network must translate this symbolic information into some kind of geometric representation<p>I was a bit surprised recently by another article linked here recently[1] that discusses &quot;direct speech-to-speech translation without relying on intermediate text representation&quot; which (if I read it correctly) works by taking frequency domain representations of speech as input and producing frequency domain representations of translated speech as output. This is indeed as near as you get to &quot;continuous&quot; input and output data in the digital domain, and brings into question (in my mind, anyhow) the assumption that discrete structures are fundamental to language processing (in humans too, for that matter.)<p>I don&#x27;t mean to detract from the paper, which looks highly interesting, it&#x27;s just that this business of given discrete structures in language is a bugbear of mine for some time now :)<p>1: <a href="https:&#x2F;&#x2F;ai.googleblog.com&#x2F;2019&#x2F;05&#x2F;introducing-translatotron-end-to-end.html" rel="nofollow">https:&#x2F;&#x2F;ai.googleblog.com&#x2F;2019&#x2F;05&#x2F;introducing-translatotron-...</a>
jf-almost 6 years ago
I’m impressed by the method of mapping higher dimensional vectors to a consistent tree representation, but I’m not sure what the take home point is after that. The BERT embeddings are (possibly randomly) branching structures? I’m only eyeballing figure 5 here, but the BERT embeddings only approximate the dependency parse tree to the same extent that the random trees do.
评论 #20130062 未加载
评论 #20130113 未加载
DoctorOetkeralmost 6 years ago
This is fantastic!<p>Can this be generalized to embedding 1: graphs or 2: DAGs ?