TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Google Translate invented its own language to help it translate more effectively

29 点作者 alexkadis超过 8 年前

6 条评论

jdmichal超过 8 年前
No, it didn&#x27;t. Or, rather, if it did, then so does every human. The neural network is doing what NNs do and associating particular input patterns with particular neurons &#x2F; pathways. So the same or highly similar concepts end up in the same place, which is how they are connected for the purpose of translation.<p>All this likely has similar analogues in the human brain. That is, I would be rather surprised if there wasn&#x27;t a dedicated neural pathway identifying a banana, which fires whenever the thought of a banana is invoked. This is also where banana is associated with yellow and food and delicious etc.<p>Also, don&#x27;t forget that in the human brain reading and listening may as well be two separate languages processed by entirely different portions of the brain. I would have to see pretty convincing evidence to believe that reading &quot;banana&quot; and hearing it don&#x27;t at some point touch the same part of the brain where the concept and associations of &quot;banana&quot; &quot;live&quot;.
xbmcuser超过 8 年前
I disagree with the authors conclusion that it invented a new language. I know a few languages if you can see into my brain you would probably find that my brain has saved meanings of words that only my brain could understand that is not a new language. What Google translate is doing now is that it comprehends multiple languages so articulates the output in the required language.
balabaster超过 8 年前
This is fascinating... I wonder how it copes with the words and phrases that don&#x27;t have any meaningful translation. Figures of speech that only work within a culture because of the culture and removing that removes the context in which the phrase makes sense.<p>I remember a past girlfriend who had cute little phrases which I&#x27;d love to remember right now by way of example but they escape me, which made no sense when translated to English because the context that made them make sense didn&#x27;t exist outside of her language.
评论 #13482270 未加载
jasode超过 8 年前
For technical HN readers, I think the article[1] that the author linked is better.<p>After reading Google&#x27;s explanation, I don&#x27;t think his comment is accurate:<p><i>&gt;Google Translate invented its own language to help it translate more effectively.<p>&gt;What’s more, nobody told it to. It didn’t develop a language (or interlingua, as Google call it) because it was coded to. It developed a new language because the software determined over time that this was the most efficient way to solve the problem of translation.</i><p>That makes it sound like the middle GNMT box (alternating in blue and orange) was automatically fabricated by the algorithm. Instead, what seems to have happened is that the <i>existence</i> of an &quot;intermediate&quot; representation was a deliberate architecture choice by human Google programmers. What got &quot;learned by machine&quot; was the build up of internal data (filling up the vectors with numbers to find mappings of &quot;meaning&quot;).<p>Google programmers can chime in on this but as an outsider, I&#x27;m guessing the previous incarnations of translate was more &quot;point-to-point&quot; instead of &quot;hub-&amp;-spoke&quot;.<p>With the 103 languages, the point-to-point when computed as &quot;n choose k&quot;[2] means 5253[3] possible direct mappings. (Although one example pair such as <i>African Swahili</i> to <i>Australia Aborigine</i> would probably not be filled with translation data.)<p>With the new GNMT (the intermediate hub), you don&#x27;t need a 5253 mappings. Instead of (n!&#x2F;k!(n-k)!) combinations, it&#x27;s just n. (However, I&#x27;m not saying that reducing the mathematical combinations was the main motivator for the re-architecture.)<p>An analogy would be the LLVM IR intermediate representation. One can target an &quot;intermediate hub&quot; language like LLVM-IR. This reduces the combinatorial complexity of all frontend programming language compilers to understand all backend machine languages. Instead of languages like Rust &amp; Julia writing point-to-point backends to specific machine languages like x86 &amp; ARM &amp; Sun. The difference with Google&#x27;s GNMT is that the keywords of &quot;intermediate language&quot; was not pre-specified by humans.<p>[1] <a href="https:&#x2F;&#x2F;research.googleblog.com&#x2F;2016&#x2F;11&#x2F;zero-shot-translation-with-googles.html" rel="nofollow">https:&#x2F;&#x2F;research.googleblog.com&#x2F;2016&#x2F;11&#x2F;zero-shot-translatio...</a><p>[2] <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Combination#Number_of_k-combinations" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Combination#Number_of_k-combin...</a><p>[3] <a href="https:&#x2F;&#x2F;www.google.com&#x2F;search?q=(103%5E2-103)%2F2" rel="nofollow">https:&#x2F;&#x2F;www.google.com&#x2F;search?q=(103%5E2-103)%2F2</a>
评论 #13482508 未加载
wooot超过 8 年前
What is the difference between how this neural network translates from Japanese to Korean, and just translating from Japanese to English and then English to Korean.
Hnrobert42超过 8 年前
Is it just me or do you have to have a linkedin account to read the article?