TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Google patents application: Batch normalization layers

63 pointsby wwilsonover 5 years ago

8 comments

dynamite-readyover 5 years ago
Prior to this, Google also successfully patented dropout layers, is that right? This being the case, what&#x27;s the implication for hobbyists (like myself), businesses and researchers, etc? It&#x27;s hard to imagine not being able to use such an abstract concept in your code, for fear of litigation.<p>For a layman like myself, it almost sounds like someone patenting the idea of a chair, or a table.
评论 #20864660 未加载
评论 #20864235 未加载
评论 #20865734 未加载
评论 #20865595 未加载
heyitsguayover 5 years ago
Well, attempted to patent batch normalization back in 2015, it looks like its application status is still pending. Which in some ways is worse, since that sounds closer to when it was first becoming popular, and from the description it does sound like they tried to patent the general computational method, not any specific implementation.<p>Can people patent computing F(x) when F is just some function when it has such a low descriptive complexity? Where&#x27;s the cutoff?
评论 #20863747 未加载
评论 #20863768 未加载
dynamite-readyover 5 years ago
Apparently, the patent has already been granted in Europe according to this page - <a href="https:&#x2F;&#x2F;piip.co.kr&#x2F;en-us&#x2F;news&#x2F;batch-normalization-layers-google" rel="nofollow">https:&#x2F;&#x2F;piip.co.kr&#x2F;en-us&#x2F;news&#x2F;batch-normalization-layers-goo...</a>
评论 #20864519 未加载
评论 #20865483 未加载
antplsover 5 years ago
I see two possible explanations to this patent :<p>- Google is rather open about deep learning development, they want to protect the ecosystem from patent trolls. It is a defensive patent to eventually punish unfair players who don&#x27;t want to play the collaborative game<p>- Google noticed OpenAI, which (legally) built stuff using some Google&#x27;s findings in the field. Now OpenAI is aiming at becoming a multi-billion dollars successful &quot;non-profit&quot; company, and Google wants its share of the money if it happens.
评论 #20864725 未加载
DoctorOetkerover 5 years ago
Isn&#x27;t this outdated retro technology compared to layer normalization?<p>&gt;Training state-of-the-art, deep neural networks is computationally expensive. One way to reduce the training time is to normalize the activities of the neurons. A recently introduced technique called batch normalization uses the distribution of the summed input to a neuron over a mini-batch of training cases to compute a mean and variance which are then used to normalize the summed input to that neuron on each training case. This significantly reduces the training time in feed-forward neural networks. However, the effect of batch normalization is dependent on the mini-batch size and it is not obvious how to apply it to recurrent neural networks. In this paper, we transpose batch normalization into layer normalization by computing the mean and variance used for normalization from all of the summed inputs to the neurons in a layer on a single training case. Like batch normalization, we also give each neuron its own adaptive bias and gain which are applied after the normalization but before the non-linearity. Unlike batch normalization, layer normalization performs exactly the same computation at training and test times. It is also straightforward to apply to recurrent neural networks by computing the normalization statistics separately at each time step. Layer normalization is very effective at stabilizing the hidden state dynamics in recurrent networks. Empirically, we show that layer normalization can substantially reduce the training time compared with previously published techniques.<p><a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1607.06450" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1607.06450</a><p>just forget batched normalization<p>instead of computing mean and variance over a batch, compute mean and variance over all the incoming dendrites of a neuron, for each neuron; result: now you are using the same function during train and test time, putting it on a more rigorous mathematical footing, and it&#x27;s adaptable to RNN
评论 #20868955 未加载
nightcrackerover 5 years ago
What happened to prior art? Why are all these garbage patents getting approved?
评论 #20868879 未加载
hgoelover 5 years ago
&quot;Don&#x27;t be evil&quot; died a while ago, so it&#x27;s free real estate for then anyway.
mlthoughts2018over 5 years ago
Why is Google so busy patenting useless layer types?