Can someone go in to more detail about how certain neural networks inherently are able to encode and interpret particular structures? Why RNs capture relations and Recurent neural networks can capture sequential data?
I'm not familiar with research on relational reasoning, and the subset they're working with seems pretty well defined, but even so this seems like a pretty meaningful step forward. Can anyone involved in the research comment further on how confident they are that RR is generalizable across dissimilar domains to the one tested?
Very cool! Not quite the same, but brings to mind SHRDLU, which was recently discussed here.<p>The first few thoughts I had on seeing this:
1) Of course this is by DeepMind! Why would I think anything different. (I love the "basic" research they are doing on NNs & Deep Learning, and am always excited to see a new paper by them).<p>2) I would love to see more investment into this kind of basic ML research. (By that I don't mean "easy", but addressing the fundamentals of how to approach different types / classes of problems). A lot of where the DeepMind guys seem to be finding these big wins is in combining "classic" AI / CS techniques with Deep Learning / Optimization.<p>Examples (And I'm a novice at deep learning, so someone PLEASE PLEASE correct me if I'm wrong):
AlphaGo - Take a technique like Tree Search for playing a game, and combine with deep networks for the tricky bit of evaluating play positions
Deep Reinforcement Learning - Q-Learning and other reinforcement techniques have been around for a while, but they adapted them to a deep neural net architecture
Neural Turing Machines - Took a classical model of computation and made it differentiable, alowing for a neural net to "learn" algorithms like sorting.
Deep Neural Computing - Figured out how to add and address external memory in a differentiable computer, allowing a neural net to solve problems like path finding on a graph.<p>Where I think a lot of cool stuff is going to continue to come from is by revisiting classic techniques, and figuring out how they can be adapted to a differentiable / optimizible architecture. Or taking a classic problem and finding an efficient way to evaluate "goodness" of an answer that lends itself to being used in an optimization problem. Again, not saying it is easy, but I wonder how much "low hanging fruit" there is in revisiting classic algorithms and GOFAI techniques, and asking "can I use this in a Neural Net or adapt this to be differentiable so that I can learn or optimize the tricky bits?"<p>I'm sure I'm glossing over a lot / missing the point of a lot of it - like I said, just a noob whose super excited about this stuff :-)
Could this be useful for multivariate statistical analysis, such as Basketball Player A has more steals than Player B but only when Player C is on the court? If so, it might get pretty busy on FanDuel