I’ve noticed a lot of the time you can tell an image is AI generated because it has a shiny/glossy lighting look to it.<p>Has anyone figured out why this is the case?
It’s just the typical aesthetic model used and isn’t inherent to the tech itself. It’s very easy to make AI images in specific art styles, with the result that you can’t tell they’re not real.<p>This is actually something of a pet peeve of mine - people sharing AI images never use styles other than the generic shiny one, and so places like Reddit.com/r/midjourney are filled with the same exact style of images.<p>Edit: if you’re looking for other style inspiration ideas, this website is a great resource for Midjourney keywords: <a href="https://midlibrary.io/styles" rel="nofollow">https://midlibrary.io/styles</a>
Many AI-generated images you encounter are low-effort creations without much prompt tuning, created using something like DALL-E or Llama 3.1. For whatever reason, the default style of DALL-E, Llama 3.1, and base Stable Diffusion seems to lean towards a glossy "photorealism" that people can instantly tell isn't real. By contrast, Midjourney's style is a bit more painted, like the cover of a fantasy novel.<p>All that being said, it's very possible to prompt these generators to create images in a particular style. I usually include "flat vector art" in image generation prompts to get something less photorealistic that I've found is closer to the style I want when generating images.<p>If you really want to go down the rabbit hole, click through the styles on this Stable Diffusion model to see the range that's possible with finetuning (the tags like "Watercolor Anime" above the images): <a href="https://civitai.com/models/264290/styles-for-pony-diffusion-v6-xl-not-artists-styles" rel="nofollow">https://civitai.com/models/264290/styles-for-pony-diffusion-...</a>
A lot of (non-AI) photos of humans tend to be airbrushed by (human) photo editors -- this removes natural imperfections -- like patchy skin, acne, discolouration etc.<p>In AI models, I think the pictures the AI generates is biased to generate is also a form of "airbrush" except the model makes the reflectivity of the images high -- simply to hide the fact that there _arent_ any imperfections that would make the photo more realistic.<p>In other words, gloss is just a form of airbrushing -- AI does it to hide the fact that there are no more details available.<p>I would guess that AI models could make the airbrush more like the airbrush human photo editors do by changing some hyper-parameters of their models.
Dall-E at least seems to have adopted the cartoonish style just to avoid lawsuits<p>You can get realistic images with Midjourney and Flux with minimal prompt tuning. Adding “posted on snapchat” or “security camera footage” to the prompt will often produce mostly realistic looking images
there is an "aesthetics" model<p><a href="https://github.com/LAION-AI/laion-datasets/blob/main/laion-aesthetic.md">https://github.com/LAION-AI/laion-datasets/blob/main/laion-a...</a><p>obviously, it reflects the mass preference for glosslop<p>secondarily it is likely due to a desire to ensure that ai images have a distinct look
I don't know, but I've noticed another pattern: They don't like leaving any empty space. Every area has to be busy, filled with objects. They can never leave any empty grass, or walls, or anything. Everything is full of objects.
This is an interesting question, though I think it needs to be qualified a bit since there are many AI images and AI image generators that don't match this pattern.<p>First, AI Images != OpenAI/ChatGPT Images. OpenAI has done a great job making a product that is accessible and thus their product decisions get a lot more exposure than other options. A few people have commented how there are several Stable Diffusion fine tunings that produce very different styles.<p>Second, AI Images and images of AI images of people are different. I think that the high gloss style is most pronounced in people. Partly this is because it is more notable and out of place.<p>If you take the previous two points as being true the question becomes why does ChatGPT image model skew toward generated shiny people. I would venture that is a conscious product decision that has something to do with what someone thought looked the most reliably good given their model's capabilities.<p>Some wild speculation as to why this would be the case:<p>* Might have to do with fashion photos having unusually bright lights and various cosmetics to give a sheen.<p>* It might have something to do with training the model on synthetic data (i.e. 3d models) which will have trouble producing the complicated subsurface scattering of human skin.<p>* Might have something to do with image statistics and glossy finishes creeping in where they don't belong.<p>* Might have to do with the efficiency of representing white spots.
I suppose because a large part of these models is recognition-probability, the shine is sort of an approximation of what is likely lighting. It isn't just the lighting that you expect but the culmination of thousands of similar yet slightly different. If you where to take a thousand photo's of someone with all manner of light angles, maybe it would look like this. Just a wild guess though.
People have started training Lora's for Flux that look pretty pretty real. This was a good recent example: <a href="https://www.reddit.com/r/StableDiffusion/comments/1ero4ts/fine_tuned_flux/" rel="nofollow">https://www.reddit.com/r/StableDiffusion/comments/1ero4ts/fi...</a>
One of the most interesting things about Midjourney is that it always returns multiple images, and asks the user to select which of those they would like to view at full resolution.<p>This is pretty clearly training for a preference model - so they now have MILLIONS of votes showing which images are more "pleasing" to their users.
I naively assumed the “airbrushed” effect AI photos have was just a way of blending components of the training data to make it look normal — opposite the way a collage of magazine clippings would appear.
Intentional choices during data set collation (to some degree 'emergent intention' due to aggregate preference). Search for 'boring realism' to find people working in other regions of latent space, e.g. this LORA: <a href="https://civitai.com/models/310571/boring-reality" rel="nofollow">https://civitai.com/models/310571/boring-reality</a> . Most of the example pictures there don't have the shiny/glossy look you're talking about.
ML-generated pseudo-photos look 3D-rendered because noise is information, and more information is both more expensive (a noisy photo can be 3x the size at the same resolution) and creates more opportunities for self-inconsistencies (e.g., with real camera sensor noise) that make fakes easier to identify automatically.
Because that's the kind of image that AI trainers like the most? Would they rather train them on old newspapers?<p>That would be the "oiled bodybuilder" applied to image training. Maybe similar and clearly defined lighting also allows AI's to match features much better, specially volumes.
I tend to agree. However, I tried to continue to prompt ChatGPT make make the picture less "AI like" and it actually ended up doing a really good job after 5 or 6 attempts. I'm not sure why it took so much prompting. Further prompting just made it worse.
Because models are trained on images that are usually edited in post-production with this aesthetic.<p>Mostly highlights down, shadows and clarity up. I often needs to edit it back to have realistic looking lights on scene.<p>Also "--s" 0 helps with generating more realistic images.
AI is compression and compression, of any lossy kind, usually works by removing the high-frequency information first. That applies to both audio and imagery. It's obviously not the only factor, but I bet it's an important one.
The noisy pattern of the skin, combined with noise from light hitting the camera sensor, plus the noise created of image compression creates an effect that is too subtle and random to “learn”.
Models have Goodharted themselves into oblivion. That's the result of endless cycles of aesthetic preference optimization, training on synthetic data, repeat ad nauseam.
I believe it's due to AI's limited ability to generate localized texture details effectively, often resorting to the use of highlights as a concealment strategy.
There’s an added objective in some of these models to make them more aesthetically pleasing based on subjective crowdsourced data that very likely contributes to this
I believe it's due to AI's limited ability to generate localized texture details effectively, often use highlights as a concealment strategy.
I don't know what you mean by "shiny/glossy lighting look to it." Could you give some examples?<p>I haven't noticed that <i>a lot</i> of AI images are generated with a "realistic cartoon" style, and I assume that's to smooth over some uncannyness.