DALL-E is still highly probabilistic in its judgement. For instance, in this article, it keeps putting "fire" in the the background on something that is likely to be on fire, rather than lighting up the person.<p>I have a similar experience. In my own experiment, I can't get DALL-E to turn off the street lamp at a bus stop in the darkness. I've tried "no light", "broken street lamp", etc.; no use. Any mention of "street lamp", the scene will include a working street lamp.<p>It's just more probable that a scene with a lamp in the darkness must have that lamp providing light, and this is something that DALL-E will not break out of.
Worth noting that DALL-E automatically “rejects attempts to create the likeness of any public figures, including celebrities". So, you wouldn't be able to get an image that included the 4 Liverpudlians. It does allow you to create fake faces. Might be fun to try and recreate Miles Davis Tutu, Aladdin Sane, Piano Man.
Man, after seeing Stable Diffusion's output, DALL-E's looks just janky. Like watching a propeller plane after seeing a jet.<p>Crazy how fast the tech is moving.
After getting access to the beta, combined with all these HN posts -- I've determined DallE2 is neat but no where as great as the initial samples made me believe.
An upvote for whoever can give me a prompt to generate an image of someone who's been massaged so much their body has been flattened, as if they were made of dough or jelly or something.<p>I spent ages on this earlier getting nowhere. I'm starting to think DALL-E is better if you don't really know what you want and you're just fishing for ideas.
Have anyone given a prompt to Dall-e of designing a company website and included “make it pop!”?<p>Maybe the AI will finally get what designers always complained about annoying clients.
I love seeing people experiment with this technology. You can feel we’re on the cusp of something great - whatever it is, we’re just not quite there yet.
> <i>The question is, when the music blows up and the artwork becomes a signature, like the Rolling Stones' Tongue & Lips, who will own the copyright?</i><p>That’s what trademarks are for.
Interestingly related, I just used AI image generation to create my EP cover.. first I tried running luciddrains dall e 2 PyTorch implementation using the prompt “death by Tetris EP album cover 2022” unfortunately I am using a Mac Pro so the gpu was not able to work. Then I tried imagen PyTorch implementation and used same keyword. This time it was working with the CPU unfortunately 2 days in we had a power outage so I had something but nothing complete. So I fed the generated image into the google dream generator and got my album cover!<p><a href="https://willsimpson.hearnow.com/" rel="nofollow">https://willsimpson.hearnow.com/</a>
There are a lot of articles focusing on how close does DALL-E match some prompt, but I wonder if this is a suboptimal way to explore the medium.<p>What if you can get a lot more out of it by embracing the unexpected responses. Can it be a tool for exploring lateral thinking? You provide a prompt computer responds with images that are a prompt to human.
A baby swiming next to a dolar bill outputs a distorted person face inside a dolar bill with some baby features, could be the start to a rabbit hole of prompts and images where you'll end up with something completly different than your initial expectations.
It's interesting that the prompts that would do badly in a Google image search also seem to be the ones that make poor prompts. Basically, it seems that rather than describing a scene, you have to try to give an analogy for some image(s) that it might have in its training set - which is why, I believe, "banana in the style of Andy Warhol" produces a much higher quality result than "Outline of prism on a black background in the middle of scene splits a beam of light coming from the left side into rainbow on the right side".
Although AI artists will destroy a lot of jobs, it will also create demand for new jobs for people who specialize in “paint overs” – taking a high concept output created by AI artists and touching it up to perfection.<p>Or perhaps even beyond just a paint over, and into the realm of recreating an entire AI artwork but with a human touch to get details just right.<p>Looking forward to it.
DALL E works really well if you are specific enough. When you don't get the intended result, it helps to identify the element which wasn't generated properly and then improve your description of the same.<p>"Two men, one of whom is on fire, shaking hands on an industrial lot." can be rewritten as, "Two men, shaking hands, standing on an industrial lot. Person on the right is on fire. Camera is 30 metres away."<p>You can go into more specifics of the framing and the angle from which you want the picture to be take. By default, DALL E will give you the most realistic generations to your prompts unless you mention "digital art" or a particular art style. I have gotten the best results when generating art instead of photos.
I haven't gotten to try it for myself, but I've read a few of these blogs that take you through generating examples or even look-alikes to older art pieces.<p>It surprisingly reminds me a lot of when I traveled to Japan without knowing really any Japanese. I needed to communicate not only with friends who don't know much English either, but also other people (like restaurant wait staff, train station staff, etc.).<p>I used Google Translate often, but many times I or my friend(s) (or the other people) would need to re-write our statements a few times until the translation result clicked well enough in each other's languages to be understandable.
I've been recreating the 50 worst heavy metal album art using AI as well, currently at 30. Recently I've found Stable Diffusion plus DALL-E inpainting to be a good combination.<p><a href="https://twitter.com/P_Galbraith/status/1560469019605344256" rel="nofollow">https://twitter.com/P_Galbraith/status/1560469019605344256</a>
If you gave those same instructions to humans I’m sure the output would be just as varied. I’d be interested to see a comparison between dall-e and humans.
Look, it's trained on these images.<p>It's really great and cool and all - but it's retrieving things that it was trained on.<p>Show me something original it did.
I wonder how long the novelty of DALL-E will persist. HN seems to upvote anything titled "I did X with DALL-E". This is a fun post, but it's not that interesting or surprising. Still worth a look don't get me wrong, but personally didn't learn anything new from it. (eg recreating the famous pink Floyd cover with "Outline of prism on a black background in the middle of scene splits a beam of light coming from the left side into rainbow on the right side" unsurprisingly worked somewhat well).