TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Deep Learning From The Bottom Up

139 pointsby zercoolalmost 11 years ago

8 comments

PieSquaredalmost 11 years ago
In addition to issues raised by other commenters, one of the problems with deep learning (deep nets in general) is that they can be very hard to train. If you&#x27;re interested in some techniques people have been using, I highly suggested you read up on optimization methods such as conjugate gradient and hessian-free optimization. I did this recently [0] and have a brief write-up, but honestly the original Martens paper may be more understandable [1].<p>[0] <a href="http://andrew.gibiansky.com/blog/machine-learning/hessian-free-optimization/" rel="nofollow">http:&#x2F;&#x2F;andrew.gibiansky.com&#x2F;blog&#x2F;machine-learning&#x2F;hessian-fr...</a><p>[1] <a href="http://machinelearning.wustl.edu/mlpapers/paper_files/icml2010_Martens10.pdf" rel="nofollow">http:&#x2F;&#x2F;machinelearning.wustl.edu&#x2F;mlpapers&#x2F;paper_files&#x2F;icml20...</a>
cjrdalmost 11 years ago
Hi, I&#x27;m one of the creators of Metacademy. I hope you find it useful. Feel free to follow our new Twitter account if you&#x27;d like low volume updates:<p><a href="https://twitter.com/meta_learning" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;meta_learning</a><p>Also, you can register an account for an occasional email.<p>PS) We&#x27;re completely free and open source: <a href="https://github.com/metacademy/metacademy-application" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;metacademy&#x2F;metacademy-application</a>
nrmnalmost 11 years ago
For anyone actually interested in implementing DNN&#x27;s I wrote up a quick blog post (essentially a brain dump) of general guidelines to adhere to when training DNNs. The source for this information is primarily from videos given by Geoffrey Hinton as well as various papers.<p><a href="http://343hz.com/general-guidelines-for-deep-neural-networks/" rel="nofollow">http:&#x2F;&#x2F;343hz.com&#x2F;general-guidelines-for-deep-neural-networks...</a>
jarvicalmost 11 years ago
I just skimmed the post as I don&#x27;t have time to fully read it right now, but I&#x27;ll point out a couple of problems that you can run into with neural nets and associated approaches.<p>One issue that can be a back breaker depending on your application is that, to produce a generalizable model, nets tend to need much more training data than the alternatives. There are ways to work around this, though.<p>The bigger problem to me is interpretability. Deep learning often gives feature sets that are very good for whatever task you are working on, they are in some senses artificial and it is difficult to relate changes in features to changes in the input data. I work with a lot of biological and medical data, and this is an issue because for some applications it is important not to just get accurate classification results, but to be able to understand what your features mean in the context of the original problem. I saw some interesting work in a computer vision paper earlier this year on trying to learn how to visualize how changes in input and outputs of a neural net were related, I&#x27;ll try to dig that up later if anyone is interested.<p>I&#x27;m not sure how coherent that was as I was trying to get this typed out in a hurry.
评论 #7804560 未加载
agibsoncccalmost 11 years ago
To address some of the comments being presented here, neural nets despite being harder to train can be debugged visually.<p>A few tips for those of you who use neural nets:<p>Debug the weights with histograms. Track the gradient and make sure the magnitude is not too large and its normally distributed.<p>Keep track of your gradient changes when using either gradient descent or conjugate gradient.<p>Plot your filters, visualize what each neuron is learning.<p>Watch the rate of change of your cost function. If it seems like its changing too fast and stops early lower your learning rate.<p>Plot your activations: if they start out grey you&#x27;re fine. If you start all black, you need to retune some of your parameters.<p>Lastly, understand the algorithm you&#x27;re using. Convolutional nets are different from recursive neural tensor are different denoising autoencoders are different from RBMS&#x2F;DBNs.<p>Pay attention to your cost function, reconstruction entropy is used differently from negative log likelihood is used differently for different objectives.<p>If you are trying to do feature learning, you are using RBMs, Denoising AutoEncoders and you will use reconstruction entropy. This is what you use for feature detectors. You may end up using negative log likelihood if you are dealing with continuous data.<p>For RBMs, pay attention to the different kinds of units[1]. Hinton recommends Gaussian visible with recitifed linear for continuous data, binary binary otherwise.<p>For denoising autoencoders, watch your corruption level. A higher one helps generalize better, especially with less data.<p>For time series or sequential data, you can use a recurrent net,moving window with DBNs, or recursive neural tensor<p>Other knobs:<p>If your deep learning framework doesn&#x27;t have adagrad find one that does.<p>Dropout: crucial. Dropout is used in combination with mini batch learning to handle learning different &quot;poses&quot; of images as well as generalizing feature learning. This can be used in combination with sampling with replacement to minimize sampling error.<p>Regularization: L2 is typically used. Hinton once said: you want a neural net that always overfits but is regularized (youtube video...don&#x27;t remember link right now).<p>Would love to answer questions! Source: I work on&#x2F;teach this stuff. Still working my way up there, but it seems to be going well so far.[2][3]<p>Lastly, tweak one knob at a time. Neural nets have a lot going on. You don&#x27;t want a situation where you A&#x2F;B tested 10 different parameters at once and you don&#x27;t know which one worked or why.<p>[1]: <a href="http://www.cs.toronto.edu/~hinton/absps/guideTR.pdf" rel="nofollow">http:&#x2F;&#x2F;www.cs.toronto.edu&#x2F;~hinton&#x2F;absps&#x2F;guideTR.pdf</a><p>[2]: <a href="http://deeplearning4j.org/" rel="nofollow">http:&#x2F;&#x2F;deeplearning4j.org&#x2F;</a><p>[3]: <a href="http://zipfianacademy.com/" rel="nofollow">http:&#x2F;&#x2F;zipfianacademy.com&#x2F;</a><p>[4]: <a href="http://arxiv.org/abs/1206.5533" rel="nofollow">http:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1206.5533</a> <a href="http://deeplearning4j.org/" rel="nofollow">http:&#x2F;&#x2F;deeplearning4j.org&#x2F;</a> <a href="http://deeplearning4j.org/debug.html" rel="nofollow">http:&#x2F;&#x2F;deeplearning4j.org&#x2F;debug.html</a> <a href="http://yosinski.com/media/papers/Yosinski2012VisuallyDebuggingRestrictedBoltzmannMachine.pdf" rel="nofollow">http:&#x2F;&#x2F;yosinski.com&#x2F;media&#x2F;papers&#x2F;Yosinski2012VisuallyDebuggi...</a>
评论 #7805670 未加载
评论 #7818886 未加载
tshadwellalmost 11 years ago
I&#x27;m pretty familiar with neural networks, and skimming that article it appears to describe something that is a neural network. Is &#x27;Deep Learning&#x27; new terminology for &#x27;Neural Network&#x27;, or does it describe a subset of ways of using them?
评论 #7805918 未加载
评论 #7804577 未加载
sytelusalmost 11 years ago
THANK YOU for this link. Meta Academy is amazing! I always wanted a tool like this which tells me graph of concepts I need to learn first before I can learn X. I wish we had this kind of learning plan graph for other fields as well.
ellipticalmost 11 years ago
Anyone know of good papers relating to deep learning that are not from image classification or speech&#x2F;text recognition?
评论 #7803091 未加载
评论 #7802661 未加载