TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Feature Visualization: How neural nets build up their understanding of images

461 点作者 rrherr超过 7 年前

16 条评论

muxator超过 7 年前
Looking at the finger instead of the moon: I like the HTML layout (responsive, inline images with captions, lateral notes).<p>Any insights on how it&#x27;s generated? Markdown, Rst, Latex -&gt; HTML? I would love to produce my documentation in this way.<p>Edit: I was too hurried. Everything is explained in <a href="https:&#x2F;&#x2F;distill.pub&#x2F;guide&#x2F;" rel="nofollow">https:&#x2F;&#x2F;distill.pub&#x2F;guide&#x2F;</a>, the template is at <a href="https:&#x2F;&#x2F;github.com&#x2F;distillpub&#x2F;template" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;distillpub&#x2F;template</a>
评论 #15648105 未加载
评论 #15649326 未加载
评论 #15668264 未加载
colah3超过 7 年前
Hey! I&#x27;m one of the authors, along with Alex and Ludwig. We&#x27;re happy to answer any questions! :)
评论 #15646999 未加载
评论 #15646389 未加载
评论 #15648215 未加载
评论 #15653856 未加载
评论 #15647205 未加载
评论 #15650702 未加载
评论 #15650697 未加载
radarsat1超过 7 年前
Great presentation, but I do wish they&#x27;d throw in an equation or two. When they talk about the &quot;channel objective&quot;, which they describe as &quot;layer_n[:,:,z]&quot;, do they mean they are finding parameters that maximize the sum of the activations of RGB values of each channel? I&#x27;m not quite sure what the scalar loss function actually is here. I&#x27;m assuming some mean. (They discuss a few reduction operators, L_inf, L_2, in the preconditioning part but I don&#x27;t think it&#x27;s the same thing?)<p>The visualizations of image gradients was really fascinating, I never really thought about plotting the gradient of each pixel channel as an image. I take it these gradients are for a particular (and same) random starting value and step size? It&#x27;s not totally clear.<p>(I have to say, &quot;second-to-last figure..&quot; again.. cool presentation but being able to say &quot;figure 9&quot; or whatever would be nice. Not <i>everything</i> about traditional publication needs to be thrown out the window.. figure and section numbers are useful for discussion!)
评论 #15655452 未加载
评论 #15649722 未加载
shancarter超过 7 年前
There’s also an appendix where you can browse all the layers. <a href="https:&#x2F;&#x2F;distill.pub&#x2F;2017&#x2F;feature-visualization&#x2F;appendix&#x2F;googlenet&#x2F;4b.html" rel="nofollow">https:&#x2F;&#x2F;distill.pub&#x2F;2017&#x2F;feature-visualization&#x2F;appendix&#x2F;goog...</a>
评论 #15653006 未加载
chillingeffect超过 7 年前
Are the layer names the same ones referred to in this paper? <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1409.4842" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1409.4842</a><p>And how can e.g. layer3a be generated from layer conv2d0? By convolving with a linear kernel? Or by the entire Inception Module including the linear and the non-linear operations?<p>Thank you. Outstanding work breaking it down.<p>Here&#x27;s another paper people might enjoy. The author generates an example for &quot;Saxophone,&quot; which includes a player... Which is fascinating, bc it implies that our usage of the word in real practice implies a player, even though the Saxophone is an instrument only. This highlights the difference between our denotative language and our experience of language! <a href="https:&#x2F;&#x2F;www.auduno.com&#x2F;2015&#x2F;07&#x2F;29&#x2F;visualizing-googlenet-classes&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.auduno.com&#x2F;2015&#x2F;07&#x2F;29&#x2F;visualizing-googlenet-clas...</a><p>Also, for those curious about the DepthConcat operation, it&#x27;s described here: <a href="https:&#x2F;&#x2F;stats.stackexchange.com&#x2F;questions&#x2F;184823&#x2F;how-does-the-depthconcat-operation-in-going-deeper-with-convolutions-work" rel="nofollow">https:&#x2F;&#x2F;stats.stackexchange.com&#x2F;questions&#x2F;184823&#x2F;how-does-th...</a><p>Edit: I&#x27;ll be damned if there isn&#x27;t something downright <i>Jungian</i> about these prototypes! There are snakes! Man-made objects! Shelter structures! Wheels! Animals! Sexy legs! The connection between snakes and guitar bodies is blowing my mind!
Houshalter超过 7 年前
This didn&#x27;t include my favorite kind of visualization from Nguyen, et al., 2015: <a href="https:&#x2F;&#x2F;i.imgur.com&#x2F;AERgy7I.png" rel="nofollow">https:&#x2F;&#x2F;i.imgur.com&#x2F;AERgy7I.png</a>
评论 #15650224 未加载
aj_g超过 7 年前
Wow. That&#x27;s incredible how psychedelic these images are. I&#x27;d be really curious to learn more about the link between these two seemingly distant subjects.
评论 #15647634 未加载
评论 #15649242 未加载
shellbackground超过 7 年前
This pictures reminds me about what one&#x27;s can see under psychedelics. All sensory input basically begins to break down to that kind of patterns, and thus reality dissolves into nothing. This is equally terrifying and liberating depends on look. The terrifying thought is that there&#x27;s no-one behind this eyes and ears. The liberating thought is that if there&#x27;s no-one there, then there&#x27;s no-one to die.
评论 #15652018 未加载
评论 #15652397 未加载
评论 #15650193 未加载
dandermotj超过 7 年前
Hi Chris, firstly thanks for all the work you&#x27;ve done publishing brilliant articles on supervised and unsupervised methods and visualisation on your old blog and now in Distill.<p>This question isn&#x27;t about feature visualisation, but I though I&#x27;d take the chance to ask you, what do you think of Hinton&#x27;s latest paper and his move away from neural network architectures?
评论 #15646607 未加载
Kronopath超过 7 年前
Interesting that simple optimization ends up with high-frequency noise similar to adversarial attacks on neural nets.<p>While I agree that the practicality of these visualizations mean that you have to fight against this high-frequency &quot;cheating&quot;, I can&#x27;t help but shake the feeling that what these optimization visualizations are showing us is <i>correct</i>. <i>This</i> is what the neuron responds to, whether you like it or not. Put in another way, the problem doesn&#x27;t seem to be with the visualization but with the <i>network itself</i>.<p>Has there been any research in making neural networks that are robust to adversarial examples?
评论 #15650318 未加载
评论 #15672048 未加载
hosh超过 7 年前
Cool. Reminds me a bit of <a href="https:&#x2F;&#x2F;qualiacomputing.com&#x2F;2016&#x2F;12&#x2F;12&#x2F;the-hyperbolic-geometry-of-dmt-experiences&#x2F;" rel="nofollow">https:&#x2F;&#x2F;qualiacomputing.com&#x2F;2016&#x2F;12&#x2F;12&#x2F;the-hyperbolic-geomet...</a><p>(Though maybe not as symmetric?)
chillingeffect超过 7 年前
Is there any way to run images from a camera real-time into GoogLeNet?<p>E.g. like if I want to scan areas around me to see if there are any perspectives in my environment that light up the &quot;snake&quot; neurons or the dog neurons???
评论 #15649305 未加载
评论 #15648540 未加载
评论 #15649994 未加载
评论 #15651115 未加载
dsnuh超过 7 年前
Okay...maybe a stupid question.<p>Could they train on white noise from a television and see if the CBR shows a structure similar to the structure of the observable universe when examining the feature layers?
snippyhollow超过 7 年前
Similar to <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1311.2901" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1311.2901</a>
gergoerdi超过 7 年前
So can someone use this to show us where the rifle is on the turtle?
评论 #15651109 未加载
nnfy超过 7 年前
Awesome, but to me this stuff is also terrifying, and I can&#x27;t quite place why.<p>Something about dissecting intelligence, and the potential that our own minds process things similarly. Creepy how our reality is distilled into these uncanny valley type matrices.<p>Also, I suspect it says something that these images look like what people report seeing on psychedelic trips...
评论 #15646634 未加载