TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Is artificial intelligence permanently inscrutable?

124 pointsby peterbonneyover 8 years ago

19 comments

hyperion2010over 8 years ago
This is not just an issue for neural nets, but also for brains. Our interpretations of our own actions should always be considered posthoc rationalizations in the absence of some falsifiable experiment being conducted to demonstrate the validity of the interpretation. Human brains are excellent at creating a coherent story about the world they experience based on the data at hand, thus we suffer the same kinds of issues, mitigated only by the fact that we have inherited developmental programs that have been subjected to a huge variety of adverse situations that have rigorously tested their performance (by killing anything that failed).
评论 #12410727 未加载
评论 #12411711 未加载
评论 #12410232 未加载
venningover 8 years ago
The pneumonia-asthma example seems to be an example of a Simpson&#x27;s paradox [1]. The doctors acted on a strong (accurate) belief about asthma sufferers contracting pneumonia and acted in such a way that the data obscured an actual causal link (asthma as an aggravating factor to pneumonia). This is opposed to the canonical Simpson&#x27;s paradox where doctors acted on a strong (inaccurate) belief about severe kidney stones [1a] and again produced lopsided data that hid the best treatment option until the paradox was identified.<p>Humans have a very hard time uncovering so-called &quot;lurking variables&quot; [2] and identifying such paradoxes. I don&#x27;t see how a neural network (or other machine learning tool) could do so on their own, but I don&#x27;t know that much about machine learning. So, I guess I have two questions for the experts out there:<p>* If all training data is affected by a confounding variable, can a machine learning algorithm identify its existence, or is it limited by only knowing a tainted world?<p>* Once we have identified such lopsided data and understood its cause, how do you feed that back into your algorithm to correct for it?<p>---<p>[1] <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Simpson%27s_paradox" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Simpson%27s_paradox</a><p>[1a] <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Simpson%27s_paradox#Kidney_stone_treatment" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Simpson%27s_paradox#Kidney_sto...</a><p>[2] <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Confounding" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Confounding</a>
评论 #12410634 未加载
评论 #12410468 未加载
rdlecler1over 8 years ago
The answer is no. The problem is, is that we don&#x27;t trim the neural networks of their spurious connections and instead we&#x27;re stuck staring at these fully (visually) connected layered networks.<p>Once you start to trim out the spurious connections you start to see that you are left with a logic design with integration&#x2F;threshold circuits instead of straight binary circuits that we&#x27;re used to seeing. There are even certain universal network patterns what will emerge to perform different functions just like in binary circuit design.<p>I wrote a paper about this in 2008 that&#x27;s now been cited about 150 times. It&#x27;s using Artificial Gene Regulatory Networks instead of Artificial Neural Networks, but the math is the same and the principle still holds:<p><a href="http:&#x2F;&#x2F;m.msb.embopress.org&#x2F;content&#x2F;4&#x2F;1&#x2F;213.abstract" rel="nofollow">http:&#x2F;&#x2F;m.msb.embopress.org&#x2F;content&#x2F;4&#x2F;1&#x2F;213.abstract</a>
评论 #12413355 未加载
Inlinkedover 8 years ago
The trick to accurate interpretability is to decouple accuracy from explanations.<p>Just like an International Master commentator can explain most of the moves of a Super GM, so can an interpretable simple model explain the predictions of a very complex black box model.<p>The work by Caruana referenced in this article actually culminated in a method to get both very accurate models and still retain interpretability.<p><a href="https:&#x2F;&#x2F;vimeo.com&#x2F;125940125" rel="nofollow">https:&#x2F;&#x2F;vimeo.com&#x2F;125940125</a><p><a href="http:&#x2F;&#x2F;www.cs.cornell.edu&#x2F;~yinlou&#x2F;projects&#x2F;gam&#x2F;" rel="nofollow">http:&#x2F;&#x2F;www.cs.cornell.edu&#x2F;~yinlou&#x2F;projects&#x2F;gam&#x2F;</a><p>More recently there was LIME:<p><a href="https:&#x2F;&#x2F;homes.cs.washington.edu&#x2F;~marcotcr&#x2F;blog&#x2F;lime&#x2F;" rel="nofollow">https:&#x2F;&#x2F;homes.cs.washington.edu&#x2F;~marcotcr&#x2F;blog&#x2F;lime&#x2F;</a><p>And there are workshops:<p><a href="http:&#x2F;&#x2F;www.blackboxworkshop.org&#x2F;pdf&#x2F;Turner2015_MES.pdf" rel="nofollow">http:&#x2F;&#x2F;www.blackboxworkshop.org&#x2F;pdf&#x2F;Turner2015_MES.pdf</a><p>We will get there. &#x27;Permanent&#x27; is a very long time and in the grand scale of things, deep learning is relatively new.
AndrewKemendoover 8 years ago
When I try to explain neural nets (specifically in vision systems) to people I basically explain how you take inputs in the form of images, label pixels&#x2F;pixel groups in the images with what you want them to output in the future, and then do that thousands of times and continue to test the results.<p>Critically though, I will say something to the effect of &quot;but if you try and break the net open and see <i>how</i> this specific net came to it&#x27;s result, it will look like spaghetti&quot;<p>So it&#x27;s a roundabout way of saying &quot;junk in; junk out.&quot; That holds true for <i>any</i> learning system, including human animals. The thought process of humans is inscrutable thus far, and I think that future computing will be similarly inscrutable if we do it correctly.
yoav_hollanderover 8 years ago
I think this issue of &quot;Explainable Machine Learning&quot; and interpretability is just going to get more and more important as ML grows. It will also be important for verifying ML-based systems - another problem area.<p>See [1] for a discussion of both.<p>[1] <a href="https:&#x2F;&#x2F;blog.foretellix.com&#x2F;2016&#x2F;08&#x2F;31&#x2F;machine-learning-verification-and-explainable-ai&#x2F;" rel="nofollow">https:&#x2F;&#x2F;blog.foretellix.com&#x2F;2016&#x2F;08&#x2F;31&#x2F;machine-learning-veri...</a>
评论 #12408390 未加载
dharma1over 8 years ago
The human brain also does massive dimensionality reduction on very large amounts of data, and a lot of unconscious processing, with much of it being beyond our capabilities of conscious introspection.<p>I think eventually, within a couple of decades, we will have AI that correlates well enough with human thought process, and has enough knowledge of the world, to be able to introspect and explain in various levels of detail, in natural language, images and other human readable constructs, why it has reached a certain conclusion. And we will be able to experimentally verify those explanations.
euskeover 8 years ago
I&#x27;ve been saying that ML is much like alchemy than science. They&#x27;ve pretty much given up to understand the underlying mechanism because it&#x27;s so complex, but that doesn&#x27;t stop them experimenting because they still get something that looks like a result. And hey, they can get paid for it.<p>Eventually it might grow into a full-fledged science, but it will probably take an awful lot of time.
评论 #12409770 未加载
评论 #12410491 未加载
评论 #12409901 未加载
unabstover 8 years ago
Isn&#x27;t this all simply about correlation vs causation? Machine learning can find strong correlations and we can make predictions based on those correlations, but at the end of the day, the machine knows nothing about what is causing any of it, and hence is &quot;inscrutable&quot;.<p>So it is up to us to fill the gap in our understanding because that is what machine learning ultimately says about the subject. It tells us what we don&#x27;t know. If we knew all about the subject, our predictions would match the predictions of the machine because there is only one reality we&#x27;re both observing. But if there is any gap, then the machine is telling us what we don&#x27;t know, not what it (of all things) knows. It&#x27;s just crunching numbers. It doesn&#x27;t &quot;know&quot; anything.
MrQuincleover 8 years ago
Interesting article. Some things are weird. I don&#x27;t know why a support vector machine is ranked better than Bayesian nets, or why they are both worse than ensemble methods w.r.t. interpretability.<p>However, I think the human should not be in the loop. The network should have another semantic layer that serves communication. It can be done from the ground up like Steels or Vogt have been doing.<p>In other words, yes we need insight, but I prefer it through introspective networks. The network should be able to explain itself.
评论 #12409090 未加载
评论 #12409071 未加载
Houshalterover 8 years ago
This isn&#x27;t unique to neural networks at all. There was a machine learning system <i>designed</i> to produce interpretable results, called Eureqa. Eureqa is a fantastic piece of software that finds simple mathematical equations that fit your data as good as possible. Emphasis on the &quot;simple&quot;, it searches for the smallest equations it can find that works, and gives you a choice of different equations at different levels of complexity.<p>But still, the results are very difficult to interpret. Yes you can verify that the equation works, that it predicts the data. But why does it work? Well who knows? No one can answer that. Understanding even simple math expressions can be quite difficult.<p>One biologist put his data into the program, and found, to his surprise, that it found a simple expression that almost perfectly explained one of the variables he was interested in. But he couldn&#x27;t publish his result, because he couldn&#x27;t understand it himself. You can&#x27;t just publish a random equation with no explanation. What use is that?<p>I think the best method of understanding our models, is not going to come from making simpler models that we can compute by hand. Instead I think we should take advantage of our own neural networks. Try to train humans to predict what inputs, particularly in images, will activate a node in a neural network. We will learn that function ourselves, and then it&#x27;s purpose will make sense to us.<p>There is a huge amount of effort put into making more accurate models, but much less into trying to interpret them. I think this is a huge mistake, because understanding a model lets you see it&#x27;s weaknesses. The things that it can&#x27;t learn, and the mistakes it makes.
评论 #12410112 未加载
Animatsover 8 years ago
No, but Nautil.us, with its mandatory tracking cookies, is.
评论 #12409380 未加载
jomamaxxover 8 years ago
We&#x27;re using these things and we&#x27;re not even sure how they work. Love it.<p>At least we should have a standard for characterizing their accuracy or something like that ...
评论 #12408125 未加载
monadaiover 8 years ago
Maybe the community needs a little simulated annealing. It seems the communal views, approaches, and focus are stuck in a local optimum.<p>Think Different! O&#x27;well.
ajcarpy2005over 8 years ago
To label so-called causative factors or even actual relationships (in a shifting...virtual...hyperspace) among potential relationships is a separate task than to make meaningful predictions or predictable changes. The Universe is inherently a system-less set of potentials. The <i></i>strongest system is the one that is indeterminate in its methodologies. Systems are survivors of reduction processes.
hour_glassover 8 years ago
I can&#x27;t even understand why deep learning creates better predictions than regular neural nets. How does adding layers change anything?
评论 #12408229 未加载
评论 #12408172 未加载
评论 #12409204 未加载
评论 #12408246 未加载
评论 #12408445 未加载
nurettinover 8 years ago
Adrian Thompson&#x27;s 1996 paper was about Genetic Algorithms. A poor overfitting example considering the whole article is prominently about Artificial Neural Networks. Thompson&#x27;s FPGA components were trained at room temperature and the creatures were unable to function well when the temperature deviates too much from 10 deg. C.
Cortezover 8 years ago
Artificial intelligence shows little promise of developing any time soon but still shows promise over long term development.
评论 #12408291 未加载
jessaustinover 8 years ago
<i>“What machines are picking up on are not facts about the world,” [Dhruv] Batra says. “They’re facts about the dataset.”</i><p>This seems analogous to 90% of (random, unreplicable) science these days.