This claim is unsupported:<p>> The machine, not smart enough to do the actual difficult job of converting these sophisticated image types to each other<p>Obviously it found an easy way to solve the problem it was given: stenography But could it have solved the problem the researchers intended, if they had framed it correctly? There's no evidence either way for this particular algorithm, but in general this is not hard. This is usually called style transfer and I don't see any reason to believe that standard techniques[1] wouldn't be able to solve the street-map-to-aerial-map problem. And it's pretty well established that adding a bit of noise[2] during training helps GANs avoid these kinds of problems.<p>[1]: <a href="https://medium.com/tensorflow/neural-style-transfer-creating-art-with-deep-learning-using-tf-keras-and-eager-execution-7d541ac31398" rel="nofollow">https://medium.com/tensorflow/neural-style-transfer-creating...</a><p>[2]: <a href="https://www.inference.vc/instance-noise-a-trick-for-stabilising-gan-training/" rel="nofollow">https://www.inference.vc/instance-noise-a-trick-for-stabilis...</a>
So basically AI abiding by Goodhart's law (<a href="https://en.wikipedia.org/wiki/Goodhart%27s_law" rel="nofollow">https://en.wikipedia.org/wiki/Goodhart%27s_law</a>). I wonder how much of this goes undetected in other applications due to poor objective definition.