TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Does Deep Learning Have Deep Flaws?

110 pointsby vkhucalmost 11 years ago

14 comments

magicalistalmost 11 years ago
The concept is very cool, but it&#x27;s not surprising that dimensionality reduction through a non-linear process is going to result in sections of input parameters that yield incorrect (and weird) results. Our visual system, while not the same as these systems, is extremely well developed and robust, yet the list of optical illusions that can fool us is quite long. In this study, the optical illusions are really just surprising because they aren&#x27;t like anything that would fool humans.<p>This isn&#x27;t to take away from the research; the most interesting result was just how close to valid inputs these erroneously classified images are.<p>But again, this isn&#x27;t some fatal flaw. This summary completely neglects the fact that the paper <i>also</i> recommends that -- just like distorted images are added to training sets today (you wouldn&#x27;t want something common like optical aberration from the camera lens screwing up your classifier) -- in the future, these adversarial examples should be added to training sets to mitigate their effects.<p>&gt; <i>In some sense, what we describe is a way to traverse the manifold represented by the network in an efficient way (by optimization) and finding adversarial examples in the input space. The adversarial examples represent low-probability (high-dimensional) “pockets” in the manifold, which are hard to efficiently find by simply randomly sampling the input around a given example. Already, a variety of recent state of the art computer vision models employ input deformations during training for increasing the robustness and convergence speed of the models [9, 13]. These deformations are, however, statistically inefficient, for a given example: they are highly correlated and are drawn from the same distribution throughout the entire training of the model. We propose a scheme to make this process adaptive in a way that exploits the model and its deficiencies in modeling the local space around the training data.</i>[1]<p>[1] <a href="http://cs.nyu.edu/~zaremba/docs/understanding.pdf" rel="nofollow">http:&#x2F;&#x2F;cs.nyu.edu&#x2F;~zaremba&#x2F;docs&#x2F;understanding.pdf</a>
评论 #7917556 未加载
评论 #7917636 未加载
iandanforthalmost 11 years ago
Let&#x27;s not forget that the word &quot;imperceptible&quot; is a heavily laden term in this context. There are numerous modifications to the data that would be &quot;imperceptible&quot; to a machine learning system, but would completely confuse a human. For example if you were to divide the image into a grid, and shuffle the squares, many ML systems would be tolerant to this kind of modification because some training regimes do this anyway. To that system you haven&#x27;t changed anything important about the image and it would correctly classify it.<p>What this result says to me is that there are <i>really useful</i> features of the data that can <i>successfully classify</i> images that humans are totally unaware of! And that&#x27;s neat.
评论 #7918206 未加载
评论 #7917848 未加载
jal278almost 11 years ago
The key claim, from the original paper:<p>&gt; Second, we find that deep neural networks learn input-output mappings that are fairly discontinuous [...] Specifically, we find that we can cause the network to misclassify an image by applying a certain imperceptible perturbation [...] the same perturbation can cause a different network that was trained on a different subset of the dataset, to misclassify the same input.<p>It&#x27;s an interesting outcome -- but there are many deep-learning approaches and many different benchmarks, so it will be important to see if this is a misleading anecdote or indicative of a systematic problem.<p>[1] <a href="http://cs.nyu.edu/~zaremba/docs/understanding.pdf" rel="nofollow">http:&#x2F;&#x2F;cs.nyu.edu&#x2F;~zaremba&#x2F;docs&#x2F;understanding.pdf</a>
jrochkind1almost 11 years ago
First thought:<p>Can I turn all digital pictures of me into &#x27;adversarial examples&#x27;, so the eye of sauron can&#x27;t identify me from pictures?<p>I&#x27;m sure it&#x27;s not as simple as that, presumably any algorithmic modification to an &#x27;adversarial&#x27; nature can be countered by other algorithms.<p>But I predict a new realm of &#x27;arms race&#x27; here in the future.
评论 #7917169 未加载
kilbasaalmost 11 years ago
IIRC even the human brain has the &#x27;adversarial&#x27; image flaw (these images will be unique to each person), but one simple workaround is to alter the input image via eye movement (which happens unconsciously).
评论 #7917401 未加载
hyperion2010almost 11 years ago
Not an example from deep learning, but [1] also demonstrates that Bayesian systems also have similar problems with sensitivity to initial conditions that are quite similar.<p>It is also rather striking that these DLNs seem to be tricked by what we would typically think of as noise.<p>1. <a href="http://arxiv.org/abs/1308.6306" rel="nofollow">http:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1308.6306</a>
评论 #7917917 未加载
评论 #7918907 未加载
varelsealmost 11 years ago
This problem was observed 20+ years ago with linear models used for protein structure prediction. For any given model of what described a properly folded protein, one could locate conformations of the same protein that were rated as folded even better than the correct conformation (I called them doppelgangers, but the name &quot;decoy&quot; is what caught on).<p>The statistical naivete of the field led to all sorts of inadvertent mixing of training and test set data which generated a lot of spurious claims for solving the problem. That is until one attempted to find those decoys and they were <i>always</i> found. This led to the creation of the CASP competition to weed this out and the field finally moved forward.<p><a href="http://en.wikipedia.org/wiki/CASP" rel="nofollow">http:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;CASP</a><p>The key similarity to what I described above is that adversarial search is done posterior to the training of the deep neural network. That makes all the difference in the world IMO. These adversaries may just be strange, otherwise hard to reach bad neighborhoods in image space without using a roadmap. Or they may be an unvaoidable consequence of the curse of dimensionality.<p><a href="http://en.wikipedia.org/wiki/Curse_of_dimensionality" rel="nofollow">http:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Curse_of_dimensionality</a><p>But given that neural networks have a gradient, it doesn&#x27;t shock me that it can serve as a roadmap to locate a set of correlated but seemingly minor changes to an example in order to flip its classification. Doing so is simply back-propagation with constant weight values to propagate the gradient to the input data itself - literally a couple lines of code.<p>IMO there are two interesting experiments to do next (not that anyone will take this seriously I expect, but ya know, hear me now, believe me later):<p>1. Characterize the statistical nature of the changes in input images and then use those summary statistics as the basis of an image altering algorithm to see if that can be used to flip the classification of any image on its own. If it can, be afraid, your driverless car may have blind spots. If not, then this is probably just a narrower form of overfitting.<p>2. If it&#x27;s likely overfitting, attempt an expectation maximization-like fix to the problem. Train the network. Generate adversaries, Add them to the training set, train again and then lather rinse repeat until either the network can&#x27;t be trained or the problem goes away.<p>Expensive? Yes. But you&#x27;re Google&#x2F;Facebook&#x2F;Microsoft and you have lots of GPUs. No excuses...<p>Failing that, the above is on my todo list so I&#x27;m throwing it out there to see if anyone can poke holes in the approach.
评论 #7917570 未加载
评论 #7917532 未加载
评论 #7918617 未加载
agibsoncccalmost 11 years ago
I think just like any machine learning algorithm, especially with computer vision, you need to prepare things properly. More robust data and images such as moving window on image patches and image rotations, even going as far as applying filters like sobel for edge detection will make your algorithm a lot better.<p>Any algorithms have weaknesses. It&#x27;s a matter of being aware of them and compensating for them in your model, possibly by using another model.
评论 #7917418 未加载
评论 #7917265 未加载
fleitzalmost 11 years ago
But then we wouldn&#x27;t see faces in clouds...<p>Neutral networks are not perfect solutions, they are solutions that get an organism to reproduce successfully.<p>Read any book on color vision humans have similar problems, yet for the most part we see things, and realize that clouds are just clouds and not faces, except for the religious, they lose their shit when faces appear in clouds.
vonnikalmost 11 years ago
This is old news. And not really that shocking. You always use multiple models and check them against each other. None is perfect. Big deal.
orasisalmost 11 years ago
This actually kind of freaks me out. Might it be possible that there is a way to corrupt brains?
评论 #7917669 未加载
评论 #7917630 未加载
评论 #7917632 未加载
peter-rowalmost 11 years ago
Can you fix it by adding random gaussian noise to the input?
dirkcalmost 11 years ago
I don&#x27;t know if this applies to other classifying algorithms, but I guess this will mean better CAPTCHAs?
nikhizzlealmost 11 years ago
I wonder if subtle make up could be used to make a real life face into an adversarial example.
评论 #7917253 未加载
评论 #7917384 未加载