It seems like mandating just a tiny bit of usage data back to the model would give SD a massive lead on training data, but I'm not an expert. Maybe that's happening already.<p>Like, example. I use SD in Blender sometimes as part of the compositor. I have maybe a 10% acceptance rate for SD output: sometimes the water isn't right, or the clouds look goofy, or something keeps getting rendered as an anime pillow for some godforsaken reason. If SD captured my prompt history and some of the final model tweaks between runs, they could ostensibly get really solid HITL test data. Then they could be the curator of that "super model" which they could upsell, maybe along with very high rez stuff, or a higher priority on jobs. Again, not an expert, so who knows. And also, having the model local, that gives you back some of the same benefits, but without the scale.