TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Proving the Lottery Ticket Hypothesis: Pruning is All You Need

174 点作者 che_shr_cat大约 5 年前

10 条评论

stared大约 5 年前
The lottery ticket hypothesis is IMHO the single most interesting finding for deep learning. It explains why does deep learning works (vs shallow neural nets), why is initial over-parametrization is often useful, why deeper is often better than shallow, etc.<p>I recommend for an overview:<p>- the original paper &quot;The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks&quot;, <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1803.03635" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1803.03635</a><p>- &quot;Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask&quot; <a href="https:&#x2F;&#x2F;eng.uber.com&#x2F;deconstructing-lottery-tickets&#x2F;" rel="nofollow">https:&#x2F;&#x2F;eng.uber.com&#x2F;deconstructing-lottery-tickets&#x2F;</a> showing that if we remove &quot;non-winning tickets&quot; before the training, the trained network still works well
评论 #22512895 未加载
评论 #22515407 未加载
评论 #22514937 未加载
xt00大约 5 年前
If “pruning is all you need” that does feel like a way of explaining how intelligence could come out of a mass of neurons such as our brain. Or at least that sounds like a thing that makes it understandable to me. Basically add a bunch of connections relatively randomly, start pruning slowly until you hit a point where the system changes... I’ll keep hand waving until somebody who knows this stuff can chime in.. :)
评论 #22513895 未加载
评论 #22516035 未加载
评论 #22515811 未加载
IX-103大约 5 年前
This is really neat and has a lot of implications for porting larger models to limited platforms like mobile. Unfortunately you still have to train the larger network, so the gains are somewhat limited. Some other papers I read show that you might be able to prune the network in the middle of training, which would make larger models more practical to work with.
评论 #22512153 未加载
rubyn00bie大约 5 年前
Am I understanding this right? Surely, I must be missing the entire point because...<p>This looks like to me, adding more and more bullshit to a model while managing to increase its accuracy, eventually leads to a &quot;smaller&quot; model with less bullshit?<p>That is to say, adding correlated or endogenous variables to a model (over-parameterization), so long as it increases its accuracy, will one day yield, a smaller, more optimized, model with less variables?<p>If so; why is this news? Isn&#x27;t this like the fundamental process of most statistics and optimization problems? Or like isn&#x27;t adding more data (when available) a fundamental method of solving&#x2F;fixing with multicolinearity?
评论 #22518014 未加载
bo1024大约 5 年前
I have a question. They show that any given depth-ell network, computing F, is w.h.p. approximated by some subnetwork of a random depth-2ell network.<p>But there is a theorem that even depth-2 networks can approximate <i>any</i> continuous function F. If the assumptions were the same, then their theorem would imply any continuous function F is w.h.p. approximated by some subnetwork of a depth-4 network.<p>So what is the difference in assumptions, i.e. what’s the significance of F being computed by a depth-ell network? What functions can a depth-ell+1 network approximate that a depth-ell network can’t? I’d guess it has to do with Lipschitz assumptions and bounded parameters but would be awesome if someone can clarify!
评论 #22514896 未加载
anonymousDan大约 5 年前
As a potentially naive thought experiment, if you just generated in advance a number of random networks of similar size to the pruned lottery ticket, and then trained them all in parallel, would you eventually find a lottery ticket? If so how many would you have to train to find a lottery ticket with high probability? Why is training one big network and then pruning any better than training lots of different smaller network? Assume in all of the above that you have a rough idea of how big the pruned network will be be.
评论 #22516520 未加载
评论 #22515704 未加载
tells大约 5 年前
ELI5 someone please.
m0zg大约 5 年前
So in other words, a sufficiently large set of monkeys with typewriters contains a subset which approximates the works of Shakespeare.
lonelappde大约 5 年前
This paper formally proves what everyone already intuitively knows, right?<p>It&#x27;s mathematically interesting, but not a practical advance.
zackmorris大约 5 年前
I&#x27;ve always felt the there is a deep connection between evolution and thought, or more specifically, genetic algorithms (GAs) and neural networks (NNs).<p>The state of the art when I started following AI in the late 90s was random weights and hyper-parameters chosen with a GA, then optimized with NN hill climbing to find the local maximum. Looks like the research has continued:<p><a href="https:&#x2F;&#x2F;www.google.com&#x2F;search?q=genetic+algorithm+neural+network" rel="nofollow">https:&#x2F;&#x2F;www.google.com&#x2F;search?q=genetic+algorithm+neural+net...</a><p>All I&#x27;m saying is that since we&#x27;re no longer compute-bound, I&#x27;d like to see more big-picture thinking. We&#x27;re so obsessed with getting 99% accuracy on some pattern-matching test that we completely miss other options, like in this case that effective subnetworks can evolve within a larger system of networks.<p>I&#x27;d like to see a mathematical proof showing that these and all other approaches to AI like simulated annealing are (or can be made) equivalent. Sort of like a Church–Turing thesis for machine learning:<p><a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Church–Turing_thesis" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Church–Turing_thesis</a><p>If we had this, then we could use higher-level abstractions and substitute simpler algorithms (like GAs) for the harder ones (like NNs) and not get so lost in the minutia and terminology. Once we had working solutions, we could analyze them and work backwards to covert them to their optimized&#x2F;complex NN equivalents.<p>An analogy for this would be solving problems in our heads with simpler&#x2F;abstract methods like spreadsheets, functional programming and higher-order functions. Then translating those solutions to whatever limited&#x2F;verbose imperative languages we have to use for our jobs.<p>Edit: I should have said &quot;NN gradient descent to find the local minimum&quot; but hopefully my point still stands.<p>Edit 2: I should clarify that in layman&#x27;s terms, Church-Turing says &quot;every effectively calculable function is a computable function&quot; so functional programming and imperative programming can solve the same problems, be used interchangeably and even be converted from one form to the other.
评论 #22513472 未加载
评论 #22513972 未加载