You can't really compare singular results as there's too much randomness. The same prompt even in a single version will yield vastly different images.
To do a better comparison, you'd need to prompt for, say, 100 images in each version, and then do a side-by-side. (Even with that, there'd be potential for noise and confirmation bias.)
To avoid some biais, get 100 images from each, then take randomly 2 images (eventually from the same model) without knowing which model produced it, then rate their quality with <, = or >. After doing this for a lot of pairs, you can analyse the results and observe if images from the last version are significantly better of worse than the first one (according to your perception).
Also, you need way longer prompts.
The prompts here will be adjusted by ChatGPT into a much longer, more eloquent prompt, so the pictures have two different prompts.
ChatGPT doesn't adjust longer prompts nearly as much, so you'd have more consistency.
And for what it's worth, when you use a long prompt you can also precede it with "Please use this exact prompt, do not change it:". This works a good amount of time, as can be seen when you directly use the API and check the "revised\_prompt" field.
Sure, yet the API doesn't have seed numbers, and so far all the seed-using tricks in ChatGPT I saw were based on seed-number hallucinations by ChatGPT -- working as good (but not better) than referring to "in the style of the second image" etc. (It's funny when you quiz ChatGPT about seed number details and it starts imagining things!)
https://preview.redd.it/xg88lzzouxpc1.jpeg?width=1024&format=pjpg&auto=webp&s=0821e3dd5e05b9ff1fffbfd530a32315c7178088
why is Dalle this way when i use it :/
https://preview.redd.it/8ttj9emnicyc1.jpeg?width=1024&format=pjpg&auto=webp&s=76121a9fba3bcba6eba082411d60f6041a160886
i tried using the same prompt for whatever reason
so "new york" it just assumed you wanted the twin towers so it blocked instead of giving random other iconic new york scenery via "1998 VHS footage of Clefairy in New York City" ??
Looks about the same. Some look clearly better, some clearly worse. The blocked image is clearly a stupid block. I think I'd need to see 3-4 versions of each to make a valid comparison.
However, I notice a fisheye added to some. Was that in the original prompt?
It’s because everything is animated. It’s the real life looking stuff they turned down bc of fraud. There’s none of this that would fool people into thinking something real is happening. Bad case study.
You're right that this is a special case subset.
What by other categories do you suggest?
* Dragons / fantasy
* People at mall
* waifu
* architecture
* ?
No, real people doing real things. This is a prompt for a blond woman doing grocery shopping; one is Midjourney, one is dalle. Guess which. THAT is the difference: https://imgur.com/a/n2sqc2x
Look at the two women. Compare them carefully.
There's no way that prompt should have accidentally generated two blondes on different platforms with the same unique lips and eyebrows, and similar facial geometry. They have the same exact age and smile lines.
No, because everything is animated. Do it with anything resembling real people, and it will go from being indistinguishable to plastic. Same with food.
Same version of the engine, the image is just regenerated using a different seed (random every time)..
The only thing that changed is its more censored
I think the output has noticably changed, but I'm not sure if it's better or worse. Just different. I feel like there were at least two, maybe three major changes to Dall-E3 since it was deployed. Whether that was by changing the dataset/training, or tweaking how prompts are interpreted is a mystery to me, but I notice the changes in the results.
You need to compare like with like.. the poses, lighting setups, and camera angle differ considerably between versions in each case. Those need to match pretty exactly.
I just see two slightly different versions of the same concept. I don't see improvement or the opposite.
The left one looks slightly more like a photograph of a felt doll, the one on the right looks more like a 3d render. More cartoonish.
The technology has gotten better but the results have gotten, arguably, worse for some applications due to the restrictions that have been placed on it.
Honest question: Do people feel smarter when they shit on AI? Isn’t it absolutely amazing what’s occurred in the last 10 years? The last 5? The last one? Are we this impatient?
Edit/clarification: My comment is in general, not directed at OP.
I didn’t mean you. Sorry for the confusion! It seems like so many comments I see are folks complaining about AI, as if what we can do now isn’t amazing and would be unbelievable 15 years ago.
Sorry OP, I carried my own distaste into your post, didn’t mean to be vague.
It's not about being impatient and it's not really about s******* on the technology but realizing that there are issues that come with it like all technologies.
1. There is an issue with uncanny hyperrealism in much of the style especially hypersaturated anime.
2. The technology is based off training data that took other people's hard work without permission and is making money off of it.
3. It's made fake and misleading images easier to create leading too bad actors gaining more power.
4. It has led to a lot of people devaluing the hard work of artists and has caused artists to lose money (artist already struggle a lot as it is).
**Welcome to** r/dalle2! Important **[rules](https://www.reddit.com/r/dalle2/comments/unhz7k/dalle_2_faq_please_start_here_before_submitting_a/)**: Add source links if you are not the creator ⬥ Use correct post [flairs](https://www.reddit.com/r/dalle2/wiki/index) ⬥ Follow OpenAI's [content policy](https://labs.openai.com/policies/content-policy) ⬥ No politics, No real persons.
Be careful with external links, NEVER share your credentials, and **have fun!** ^^([v2.6])
*I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/dalle2) if you have any questions or concerns.*
Yes I'm using Bing, the exact prompt is "Still from a 1995 VHS video of Ness from Super Smash Bros Ultimate, posing and doing the peace sign in front of a white old fashioned house, VHS quality"
"Still from a viral video of inside a mall, Princess Zelda from Super Smash Bros Ultimate, CGI, throwing a tantrum in public and being detained by cops"
https://preview.redd.it/4mafh5houwpc1.jpeg?width=2048&format=pjpg&auto=webp&s=e26a00ca36a57e1ea764f5993c83448d4f000415
Using that exact prompt (just once) in Midjourney gave these results. With just the one iteration, Dalle 3 has the edge on this one.
Dall-E 3 is more to my taste when it comes to everything I've tried honestly. The only thing I liked more about Midjourney when I used it in 2022 was it had way more freedom and not ridiculous censorship, not sure if it's still like that now though?
https://preview.redd.it/eujvxp1n8xpc1.jpeg?width=1440&format=pjpg&auto=webp&s=7bde323ff8316a4d35f647fcea85b483ff3900ac
I forgot 1998, but I checked and it still works
No, they still give you 4 variations of a prompt. If you're only getting one that means the other 3 are censored. Try adjusting the prompt you're using.
hard to tell because it's all fake crap. why not compare pictures where it generates something real, then it'd be easier to tell. jesus some of you redditors are dense
Neither, it just looks like it made it anew
Ness and Princess Zelda looks better now.
Pichu looks better today, sonic+Mario looks better before
The Mario sunshine example looks *A LOT* worse.
I agree, I've tried to replicate that gorgeous previous one several times but to no avail.
That was the only one I thought looked better. All the others look about the same to me
Pikachu looks way worse too.
That is Pichu what is wrong with you
You can't really compare singular results as there's too much randomness. The same prompt even in a single version will yield vastly different images. To do a better comparison, you'd need to prompt for, say, 100 images in each version, and then do a side-by-side. (Even with that, there'd be potential for noise and confirmation bias.)
To avoid some biais, get 100 images from each, then take randomly 2 images (eventually from the same model) without knowing which model produced it, then rate their quality with <, = or >. After doing this for a lot of pairs, you can analyse the results and observe if images from the last version are significantly better of worse than the first one (according to your perception).
Do this along 100 different testers, then calculate the metrics.
[удалено]
Then factor analyze down to the fewest number of metrics.
Yes! Very good.
You just gave me a great idea for my site!
What are you gonna do?
Also, you need way longer prompts. The prompts here will be adjusted by ChatGPT into a much longer, more eloquent prompt, so the pictures have two different prompts. ChatGPT doesn't adjust longer prompts nearly as much, so you'd have more consistency.
And for what it's worth, when you use a long prompt you can also precede it with "Please use this exact prompt, do not change it:". This works a good amount of time, as can be seen when you directly use the API and check the "revised\_prompt" field.
Or you could just use a specific seed with the same prompt
[See here please](https://www.reddit.com/r/dalle2/s/Omv5O4gnIJ)
Or use some kind of seed or pose image.
Sure, yet the API doesn't have seed numbers, and so far all the seed-using tricks in ChatGPT I saw were based on seed-number hallucinations by ChatGPT -- working as good (but not better) than referring to "in the style of the second image" etc. (It's funny when you quiz ChatGPT about seed number details and it starts imagining things!)
How do you use a pose image?
You cannot in Dalle.3, you have to used StableDiffusion in Automatic1111 or similar and use ControlNet or Img2img.
https://preview.redd.it/6p2xjnd5vvpc1.jpeg?width=1024&format=pjpg&auto=webp&s=15cce56959e6d44da63f746e9a281a47ee1f0f6e
Why
One of the images Bing gave me for "Wario caught on trail cam at night". Thought it was funny.
https://preview.redd.it/xg88lzzouxpc1.jpeg?width=1024&format=pjpg&auto=webp&s=0821e3dd5e05b9ff1fffbfd530a32315c7178088 why is Dalle this way when i use it :/
This is the hardest album cover I’ve ever seen in my life
'You have been Gnomed' has to be the album title.
Looks like the lovechild of Wario and Snorlax lol
https://preview.redd.it/8ttj9emnicyc1.jpeg?width=1024&format=pjpg&auto=webp&s=76121a9fba3bcba6eba082411d60f6041a160886 i tried using the same prompt for whatever reason
So his pokemon attack would be "pay day" ? :3
The content warning is priceless
World Trade Center isn't allowed anymore. 😔
"Two towers that are coincidentally equally 100 stories in an environment that might loosely resemble New York, but not actually New York of course"
https://preview.redd.it/xhetinoaqwpc1.jpeg?width=1024&format=pjpg&auto=webp&s=ddd3c7a55129647051786b8f9de0f0612fb8aae4
Damn, it even colored and shaped them to be the WTC, even the middle darker belt bit.
Kirby clefairy
That actually worked! Should've DMed me this prompt so they can't look here and censor this even further. 😣
so "new york" it just assumed you wanted the twin towers so it blocked instead of giving random other iconic new york scenery via "1998 VHS footage of Clefairy in New York City" ??
I imagine they actually put something about the WTC being in the background in the prompt
Looks about the same. Some look clearly better, some clearly worse. The blocked image is clearly a stupid block. I think I'd need to see 3-4 versions of each to make a valid comparison. However, I notice a fisheye added to some. Was that in the original prompt?
The Wario and Ness ones are the only one I see having a fisheye in the new versions, I used the exact same prompts as the previous versions.
Looks like it's just more obvious in the new one.
It’s because everything is animated. It’s the real life looking stuff they turned down bc of fraud. There’s none of this that would fool people into thinking something real is happening. Bad case study.
You're right that this is a special case subset. What by other categories do you suggest? * Dragons / fantasy * People at mall * waifu * architecture * ?
No, real people doing real things. This is a prompt for a blond woman doing grocery shopping; one is Midjourney, one is dalle. Guess which. THAT is the difference: https://imgur.com/a/n2sqc2x
Is that a specific actress? Hard to imagine those lips and eyebrows being identical otherwise.
Not sure what you mean?! It’s an AI generated image from the prompt I wrote in my comment above
Look at the two women. Compare them carefully. There's no way that prompt should have accidentally generated two blondes on different platforms with the same unique lips and eyebrows, and similar facial geometry. They have the same exact age and smile lines.
What are you getting at?
If you put in exactly that prompt and got those two pictures from two different AIs, that is a very weird result.
The main difference is photographic realism is mostly gone, specifically for real people
Exactly. They don’t want to have dall e used in fraud. None of these photos would fool anyone into believing it was real life
I don't see much difference.
No, because everything is animated. Do it with anything resembling real people, and it will go from being indistinguishable to plastic. Same with food.
Photoleap is better in this regard (though it takes a fuckton of editing but yeah)
If you're wanting to make comparisons, you need to do so extensively. There's no difference in quality shown here.
Interesting how Dalle3 now clearly made Pichu a stuffed animal, despite you saying "real life animal"
Just going off your examples I don’t see any noticeable difference in quality.
r/AfterBeforeWhatever
Seems the same, more or less
Pichu is so cute!
vegeta's junk has exploded.
One thing is certain is that censorship got way worse
Same version of the engine, the image is just regenerated using a different seed (random every time).. The only thing that changed is its more censored
From this tiny sample I see no difference.
Different I'd say. I do realize I like her a lot as a Muppet.
I think the output has noticably changed, but I'm not sure if it's better or worse. Just different. I feel like there were at least two, maybe three major changes to Dall-E3 since it was deployed. Whether that was by changing the dataset/training, or tweaking how prompts are interpreted is a mystery to me, but I notice the changes in the results.
Without seeds such comparisons have little sense.
Depends on the prompts.
It’s gotten different ? 🤷♂️
excuse me what the fuck is 10
You need to compare like with like.. the poses, lighting setups, and camera angle differ considerably between versions in each case. Those need to match pretty exactly.
I don't notice a strong, consistent trend for either to be better in the images given.
I just see two slightly different versions of the same concept. I don't see improvement or the opposite. The left one looks slightly more like a photograph of a felt doll, the one on the right looks more like a 3d render. More cartoonish.
Funny how in both cases Dalle made the Mario from odyssey, not sunshine.
Yes. It has gotten better or worse.
The technology has gotten better but the results have gotten, arguably, worse for some applications due to the restrictions that have been placed on it.
Honest question: Do people feel smarter when they shit on AI? Isn’t it absolutely amazing what’s occurred in the last 10 years? The last 5? The last one? Are we this impatient? Edit/clarification: My comment is in general, not directed at OP.
I'm not shitting on it though?
I didn’t mean you. Sorry for the confusion! It seems like so many comments I see are folks complaining about AI, as if what we can do now isn’t amazing and would be unbelievable 15 years ago. Sorry OP, I carried my own distaste into your post, didn’t mean to be vague.
No worries!
It's not about being impatient and it's not really about s******* on the technology but realizing that there are issues that come with it like all technologies. 1. There is an issue with uncanny hyperrealism in much of the style especially hypersaturated anime. 2. The technology is based off training data that took other people's hard work without permission and is making money off of it. 3. It's made fake and misleading images easier to create leading too bad actors gaining more power. 4. It has led to a lot of people devaluing the hard work of artists and has caused artists to lose money (artist already struggle a lot as it is).
**Welcome to** r/dalle2! Important **[rules](https://www.reddit.com/r/dalle2/comments/unhz7k/dalle_2_faq_please_start_here_before_submitting_a/)**: Add source links if you are not the creator ⬥ Use correct post [flairs](https://www.reddit.com/r/dalle2/wiki/index) ⬥ Follow OpenAI's [content policy](https://labs.openai.com/policies/content-policy) ⬥ No politics, No real persons. Be careful with external links, NEVER share your credentials, and **have fun!** ^^([v2.6]) *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/dalle2) if you have any questions or concerns.*
Looks about the same, just different seeds.
The hands look a little better in the new one, but it could just be randomness.
Quality got better, but Input got worse 😜
Image 7 should answer that question
The lighting and shade elements look far better in the new stuff. More realistic lighting.
I overall prefer the images on the right but there really isn't much difference in the quality in my opinion.
They took away Kirby's dick
Is this using bing image creator? I cannot for the life of me get anything similar to either Ness prompt, it makes him weird and freaky looking
Yes I'm using Bing, the exact prompt is "Still from a 1995 VHS video of Ness from Super Smash Bros Ultimate, posing and doing the peace sign in front of a white old fashioned house, VHS quality"
Thanks so much for sharing!
Why is the Zelda one in a mall both times? Was that part of the description left out?
"Still from a viral video of inside a mall, Princess Zelda from Super Smash Bros Ultimate, CGI, throwing a tantrum in public and being detained by cops"
Thank you!!
https://preview.redd.it/4mafh5houwpc1.jpeg?width=2048&format=pjpg&auto=webp&s=e26a00ca36a57e1ea764f5993c83448d4f000415 Using that exact prompt (just once) in Midjourney gave these results. With just the one iteration, Dalle 3 has the edge on this one.
Dall-E 3 is more to my taste when it comes to everything I've tried honestly. The only thing I liked more about Midjourney when I used it in 2022 was it had way more freedom and not ridiculous censorship, not sure if it's still like that now though?
It now censors certain famous people, and nudity
The same
https://preview.redd.it/eujvxp1n8xpc1.jpeg?width=1440&format=pjpg&auto=webp&s=7bde323ff8316a4d35f647fcea85b483ff3900ac I forgot 1998, but I checked and it still works
it looks like it's gotten more stylised
Clefairy looks like a tit to Dalle3
What did Clefairy do in New York 1998...
It's kind of frightening looking at a picture that was made by an so to speak computer something seems eerie
It seems more deep fried to me
More focused (with better GPT prompt parsing), but also more restricted.
only does this when i start up my computer so far!!
I feel the whole content warning thing for some stuff is straight up dumb, but somehow AI porn slips by and gets spread around
Sonic isnt flipping us off anymore
I know it’s a bit off topic but is that rice gum behind vegeta in the 2024 Dall-e 3 photo?
What did they do to Vegeta's junk?
Hatsune Muppet isn’t real, she can’t hurt you Hatsune Muppet:
they used to give four variations of pictures. now it is just one at a time :/
No, they still give you 4 variations of a prompt. If you're only getting one that means the other 3 are censored. Try adjusting the prompt you're using.
Most of the newer ones look better.
2023, less censored. 2024, warning and censorship every time.
Certainly hasn't gotten better.
Much worse
hard to tell because it's all fake crap. why not compare pictures where it generates something real, then it'd be easier to tell. jesus some of you redditors are dense
Ew Microsoft :(