The <i>models</i>, <i>neural networks</i> Professor Rudin is considering have a LOT of parameters, dimensions, <i>neurons,</i> neuron values, etc.<p>Okay. Since apparently no one has the <i>explanations</i> desired, we have to guess. So, let's do some guessing:<p>Given so many parameters, etc., we have in some sense -- in some case of geometry, <i>spaces</i>, maybe <i>vector spaces</i>, maybe as in linear algebra -- a lot of <i>dimensions</i>.<p>Then something surprising holds (once we get precise about a space, easy enough to prove): Given a sphere in the space, we can calculate its volume. We can do this for the space of any finite dimension. Here is the surprise: As we have a lot of dimensions, there is a LOT of volume in that sphere, and nearly all that volume is just inside the surface of that sphere. E.g., if do some work in <i>nearest neighbors</i>, discover this surprise in strong terms.<p>Net, in the space being considered, there is a LOT of volume. Then ...: There is plenty of volume to put faces of cats over here, dogs over there, men another place, women still well separated, essays on bone cancer far away, ..., for thousands, millions, ..., more things, thoughts, topics, etc. Then given some new data, say, a white cat not in the <i>training</i> data, likely the data on that white cat will settle on the volume with the cats instead of dogs, monkeys, etc. and, thus, we will have <i>recognized</i> a cat via some <i>emergent</i> functionality.<p>Just a guess.