TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Differentiable Plasticity: A New Method Learning to Learn

249 pointsby myhrvoldabout 7 years ago

8 comments

no_identdabout 7 years ago
&quot;What if the plasticity of the connections was under the control of the network itself, as it seems to be in biological brains through the influence of neuromodulators?&quot;<p>Anyone who wishes to explore this idea would do well to go back to the basics of neural nets and read Warren McCulloch&#x27;s seminal papers on neural nets, from the 40s:<p><a href="http:&#x2F;&#x2F;www.cse.chalmers.se&#x2F;~coquand&#x2F;AUTOMATA&#x2F;mcp.pdf" rel="nofollow">http:&#x2F;&#x2F;www.cse.chalmers.se&#x2F;~coquand&#x2F;AUTOMATA&#x2F;mcp.pdf</a> A Logical Calculus of the ideas immanent in nervous activity<p><a href="http:&#x2F;&#x2F;vordenker.de&#x2F;ggphilosophy&#x2F;mcculloch_heterarchy.pdf" rel="nofollow">http:&#x2F;&#x2F;vordenker.de&#x2F;ggphilosophy&#x2F;mcculloch_heterarchy.pdf</a> A heterarchy of values determined by the topology of neural nets<p>(After having read those two papers, one can then try to make sense of Heinz von Förster&#x27;s masterpiece, <a href="http:&#x2F;&#x2F;www.univie.ac.at&#x2F;constructivism&#x2F;archive&#x2F;fulltexts&#x2F;1270.pdf" rel="nofollow">http:&#x2F;&#x2F;www.univie.ac.at&#x2F;constructivism&#x2F;archive&#x2F;fulltexts&#x2F;127...</a>, Objects: Tokens for (Eigen-)Behaviors, which also bears some relevance to this matter. However, most people find it incomprehensible.)
评论 #16805432 未加载
评论 #16806689 未加载
dpflanabout 7 years ago
Interesting. Some highlighted links from the writeup:<p>1. <i>Differentiable plasticity: training plastic neural networks with backpropagation</i> (<a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1804.02464" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1804.02464</a>)<p>2. <i>Born to Learn: the Inspiration, Progress, and Future of Evolved Plastic Artificial Neural Networks</i> (<a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1703.10371" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1703.10371</a>)<p>3. Github for the project: <a href="https:&#x2F;&#x2F;github.com&#x2F;uber-common&#x2F;differentiable-plasticity" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;uber-common&#x2F;differentiable-plasticity</a><p>4. <i>Learning to Learn</i> (<a href="http:&#x2F;&#x2F;bair.berkeley.edu&#x2F;blog&#x2F;2017&#x2F;07&#x2F;18&#x2F;learning-to-learn&#x2F;" rel="nofollow">http:&#x2F;&#x2F;bair.berkeley.edu&#x2F;blog&#x2F;2017&#x2F;07&#x2F;18&#x2F;learning-to-learn&#x2F;</a>)<p>5. Meta-Learning: <a href="http:&#x2F;&#x2F;metalearning.ml&#x2F;" rel="nofollow">http:&#x2F;&#x2F;metalearning.ml&#x2F;</a>
trextrexabout 7 years ago
Very cool. It&#x27;s interesting how powerful the recurrent network becomes with the addition of the learned hebbian term. For context, even without the Hebbian term, recurrent networks can learn to learn to do quite interesting things (Hochreiter et al. 2001).<p>Shameless plug -- our lab recently ported LSTMs to spiking networks without a significant loss in performance, and showed that learning to learn works quite well even with spiking networks (Bellec et al. 2018).<p>So it seems like this method of learning to learn could provide a extremely biologically realistic and fundamental paradigm for fast learning. The addition of the Hebbian term neatly fits in with this paradigm too.<p>Hochreiter et al. 2001: <a href="http:&#x2F;&#x2F;link.springer.com&#x2F;chapter&#x2F;10.1007&#x2F;3-540-44668-0_13" rel="nofollow">http:&#x2F;&#x2F;link.springer.com&#x2F;chapter&#x2F;10.1007&#x2F;3-540-44668-0_13</a><p>Bellec et al. 2018: <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1803.09574" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1803.09574</a>
dchichkovabout 7 years ago
It’d be interesting to compare this approach against a simpler baseline: setting a <i>different</i> (10 – 100 times higher?) learning rate for a <i>fraction</i> (10% ?) of neurons in an LSTM.
评论 #16807596 未加载
letitgo12345about 7 years ago
Is the plasticity update guaranteed to reach equilibrium assuming the network is run on iid data (as in do the H_{ij} values reach a fixed point)?<p>Edit: Seems like it should be reached eventually as the equilibrium point is H_ij = y_i * y_j and they keep doing a weighted average of the former with the latter (this is not a proof ofc as y_i * y_j keeps changing with each sample).
adrianratnapalaabout 7 years ago
So the &quot;plastic component&quot; of a connection strength is a thing which decays away exponentially, but is replenished whenever the two endpoints do the same thing.<p>I have heard that neuroscientists have an adage &quot;fire together wire together&quot;. Is that all that ML people mean by &quot;plasticity&quot;.
signa11about 7 years ago
very cool stuff ! it might be possible to use this stuff for pruning edges which are not that plastic as well.
whatever1about 7 years ago
Good luck with getting even more suboptimal solutions with this extra non linearity.<p>No wonder why when your autonomous cars are plowing into people or walls you have no clue of what is going on.
评论 #16809712 未加载
评论 #16806494 未加载