TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Gaussian Distributions Are Soap Bubbles

140 pointsby lebekover 7 years ago

14 comments

cs702over 7 years ago
Of course. As others here point out, the hypervolume inside an n-dimensional hypersphere grows as the nth power of a linear increase in radius. In high dimensions, tiny increases in radius cause hypervolume to grow by more than 100%. The concentration of hypervolume is always highest at the edge.[0]<p>The theoretical tools (and intuitions) we have today for making sense of the distribution of data, developed over the past three centuries, <i>break down</i> in high dimensions. The fact that in high dimensions Gaussian distributions are not &quot;clouds&quot; but actually &quot;soap bubbles&quot; is a perfect example of this breakdown. Can you imagine trying to model a cloud of high-dimensional points lying on or near a lower-dimensional manifold with soap bubbles?<p>If the data is not only high-dimensional but also non-linearly entangled, we don&#x27;t yet have &quot;mental tools&quot; for reasoning about it:<p>* <a href="https:&#x2F;&#x2F;medium.com&#x2F;intuitionmachine&#x2F;why-probability-theory-should-be-thrown-under-the-bus-36e5d69a34c9" rel="nofollow">https:&#x2F;&#x2F;medium.com&#x2F;intuitionmachine&#x2F;why-probability-theory-s...</a><p>* <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=15620794" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=15620794</a><p>[0] See kgwgk&#x27;s comment below.
评论 #15677019 未加载
评论 #15677217 未加载
评论 #15677777 未加载
woopwoopover 7 years ago
I was going to comment that what&#x27;s going on here doesn&#x27;t have much to do with the Gaussian distribution. In high dimensions, almost all of the volume of the unit ball is concentrated near the unit sphere. In the first comment, Frank Morgan makes the same remark, pointing out that you get the same effect with the uniform distribution on the unit cube in high dimensions.<p>High dimensions are weird.
评论 #15676590 未加载
评论 #15676770 未加载
评论 #15677865 未加载
smallnamespaceover 7 years ago
Isn&#x27;t the unsuitability of the high-dimensional Gaussian intimately related to the fact that for most realistic problem spaces, we actually believe there are really far fewer than the N &gt;&gt; 1 measured dimensions?<p>A uniform Gaussian presupposes that the variates are either linearly orthogonal, or all have the same linear interaction with each other (in the case of fixed positive correlation).<p>If your actual problem has dimension 20, but you&#x27;ve measured it with N dimensions, then that means there are strong interactions between your measured variates, and moreover the intervariate interactions do not have a single <i>fixed</i> interaction strength (like a single Gaussian correlation), but probably vary like a random matrix.<p>This might be related to the Tracy-Widom[1] distribution somehow. Perhaps the the distribution you use to replace the Gaussian should really be something like: first generate a random positive semi-definite matrix as C, then generate random data based on different random choices of C.<p>[1] <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Tracy%E2%80%93Widom_distribution" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Tracy%E2%80%93Widom_distributi...</a>
tgbover 7 years ago
I won&#x27;t dispute the main point of the article but a couple minor errors bug me. First, he kept referring to a Gaussian distribution as being the <i>unit</i> sphere, when of course the radius depends upon the parameters of the Gaussian (the standard deviation). If not, then it wouldn&#x27;t be invariant under which units you chose. A bizarre mistake to repeat many times throughout the article.<p>Less importantly, the last paragraph says that the probability that two samples are orthogonal is &quot;very high&quot;. Being precisely orthogonal is technically a probability zero event. There author means &quot;very close to orthogonal.&quot;<p>There was a good discussion about this problem in the context of Monte Carlo simulations in (1).<p>(1) <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1701.02434" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1701.02434</a>
评论 #15677313 未加载
评论 #15677266 未加载
Bromsklossover 7 years ago
Those images [0] inputs that were optimised to maximise a certain classification response were cool! Instead of going to this peak of the response function, is there a way to explore the shell where the actual images reside? Would such images look, to our eyes, more like real input than the optimised input? I suspect they won&#x27;t, but I still would like to see what is between the dogs!<p>[0] <a href="http:&#x2F;&#x2F;www.inference.vc&#x2F;content&#x2F;images&#x2F;2017&#x2F;11&#x2F;Screen-Shot-2017-11-09-at-2.12.44-PM.png" rel="nofollow">http:&#x2F;&#x2F;www.inference.vc&#x2F;content&#x2F;images&#x2F;2017&#x2F;11&#x2F;Screen-Shot-2...</a>
评论 #15676750 未加载
评论 #15676778 未加载
snippyhollowover 7 years ago
Compulsory &quot;Spikey Spheres&quot; notebook <a href="http:&#x2F;&#x2F;nbviewer.jupyter.org&#x2F;urls&#x2F;gist.githubusercontent.com&#x2F;syhw&#x2F;9025964&#x2F;raw&#x2F;441645b476a2a997f27f5993e4da2988febe1ef3&#x2F;SpikeySpheres" rel="nofollow">http:&#x2F;&#x2F;nbviewer.jupyter.org&#x2F;urls&#x2F;gist.githubusercontent.com&#x2F;...</a>
评论 #15677301 未加载
amlutoover 7 years ago
In information theory, there&#x27;s a reflected concept of a &quot;typical set&quot;, which is the set of sequences of samples from a distribution whose probability (or probability density) is very close to the expected probability. If you draw a sequence of samples, you are overwhelmingly likely to get a typical outcome as opposed to, say, anything resembling the most likely outcome.<p>As a concrete example, if you have a coin that gets heads 99% of the time and you flip it 1M times, you are overwhelmingly likely to get around 10k tails, even though the individual sequences with many fewer tails are each far likelier than the typical sequences.
Scene_Cast2over 7 years ago
So I have a feeling that he&#x27;s looking at the wrong histogram. If you plot the distributions of the vector magnitudes, you&#x27;ll get a spike around whatever large number, and a very sharp falloff to the right and left.<p>However, it&#x27;s not a &quot;bubble&quot; in the intuitive sense. He&#x27;s looking at the magnitude distribution of dots over the entire space, and implicitly using the Cartesian coordinate system (discarded angle, looking at just magnitude).<p>If you look at the distribution of dots per volume (or R^N hyper-volume rather), then you&#x27;ll still have the highest concentration in the center, with no &quot;bubble&quot;.
strainerover 7 years ago
Maybe it goes without saying but what I found distinctive about gaussian distribution in multiple dimensions is that it seems to be the only distribution which produces a smooth radial pattern when plotted co-linearly (yet not radially). All other distributions which I tested exhibit a bias through the main axis when just a number of (variateA,variateB) pairs are plotted. Gaussian seems to be the only one , fundamentally, which shows no sign of the orientation of the axis it is plotted along.<p>Comes in handy for plotting a radially smooth &#x27;star cluster&#x27; without doing polar coordinates and trig. Just plot a load of (x=a_guass,y=another_gaus,z=another_gaus) and you have a radially smooth object. I dont think any other distribution can do that, it seems to me there is something mathematically profound about it which Im sure some mathemagicians have a proper grasp of.<p>The &#x27;co-linear&#x27; distortions of other distributions can be seen here in some plots in the test page for my random distribution lib:<p><a href="http:&#x2F;&#x2F;strainer.github.io&#x2F;Fdrandom.js&#x2F;" rel="nofollow">http:&#x2F;&#x2F;strainer.github.io&#x2F;Fdrandom.js&#x2F;</a>
bglazerover 7 years ago
I was recently reading the section on importance sampling in David MacKay&#x27;s &quot;Information Theory, Learning, and Inference Algorithms&quot;. Page 373-376 in the linked pdf (<a href="http:&#x2F;&#x2F;www.inference.org.uk&#x2F;itprnn&#x2F;book.pdf" rel="nofollow">http:&#x2F;&#x2F;www.inference.org.uk&#x2F;itprnn&#x2F;book.pdf</a>)<p>He shows that importance sampling will likely fail in high dimensions precisely because samples from a high dimensional Gaussian can be <i>very different</i> than those from a uniform distribution on the unit sphere.<p>Consider the ratio between a sample at the same point from a 1000D Gaussian and a 1000D uniform distribution over a sphere. If you sample enough times, then the median ratio and the largest ratio will be different by a factor of 10^19. Basically, most samples from the Gaussian will be fairly similar to the uniform. A few will be wildly different.<p>Perhaps I&#x27;m misunderstanding both the post and MacKay&#x27;s book. I&#x27;d be happy to be corrected.
评论 #15677619 未加载
srs70187over 7 years ago
This is an interesting take and kudos to the author for relaying a helpful way to think about high dimensional distributions.<p>I really like and often come back to this talk by Michael Betancourt where the theme is quite similar: <a href="https:&#x2F;&#x2F;youtu.be&#x2F;pHsuIaPbNbY" rel="nofollow">https:&#x2F;&#x2F;youtu.be&#x2F;pHsuIaPbNbY</a>
andrewflnrover 7 years ago
This reminds me of the story summed up in this quote:<p><pre><code> There was no such thing as an average pilot. If you’ve designed a cockpit to fit the average pilot, you’ve actually designed it to fit no one. </code></pre> Good enough source here: <a href="http:&#x2F;&#x2F;wmbriggs.com&#x2F;post&#x2F;18291&#x2F;" rel="nofollow">http:&#x2F;&#x2F;wmbriggs.com&#x2F;post&#x2F;18291&#x2F;</a><p>Humans form a very high-dimensional space. I&#x27;m not sure what to make of the point about orthogonality in that regard.
brianjoseffover 7 years ago
Having trouble understanding a lot of the specifics of this--though broader concepts grok-able. Before I go blindly googling around to get up to speed--<p>Any recommended foundational texts to begin with?<p>Recommended learning trajectory to get to where this is understandable?
m3kw9over 7 years ago
Can we derive some type of optimization algorithms from using this?