ChatGPT Images 2.0
Posted by meetpateltech 3 hours ago
Comments
Comment by ea016 2 hours ago
GPT Image 2
Low : 1024×1024 $0.006 | 1024×1536 $0.005 | 1536×1024 $0.005
Medium : 1024×1024 $0.053 | 1024×1536 $0.041 | 1536×1024 $0.041
High : 1024×1024 $0.211 | 1024×1536 $0.165 | 1536×1024 $0.165
GPT Image 1 Low : 1024×1024 $0.011 | 1024×1536 $0.016 | 1536×1024 $0.016
Medium : 1024×1024 $0.042 | 1024×1536 $0.063 | 1536×1024 $0.063
High : 1024×1024 $0.167 | 1024×1536 $0.25 | 1536×1024 $0.25Comment by Melatonic 1 hour ago
Comment by vunderba 1 hour ago
Comment by kibibu 42 minutes ago
One that i can think of:
- replacing photography of people who may be unable to consent or for whom it may be traumatic to revisit photographs and suitable models may not be available, e.g. dementia patients, babies, examples of medical conditions.
Most other vaguely positive use cases boil down to "look what image generators can do", with very little "here's how image generators are necessary for society.
On the flip side, there are hundreds of ways that these tools cause genuine harm, not just to individuals but to entire systems.
Comment by throwaway2027 2 hours ago
Comment by joegibbs 54 minutes ago
Comment by louiereederson 2 hours ago
Comment by 6thbit 2 hours ago
direct pdf https://deploymentsafety.openai.com/chatgpt-images-2-0/chatg...
Comment by samiwami 2 hours ago
I know this is probably mega cherry-picked to look more impressive, but some of the images are terrifyingly realistic. They seem to have put a lot of effort into the lighting.
Comment by alextheparrot 2 hours ago
From the system card someone linked elsewhere in the discussion
Comment by ai-tamer 1 hour ago
Comment by Legend2440 2 hours ago
Seeing is not believing anymore, and I don't think SynthID or anything like it can restore that trust in images.
Comment by vunderba 1 hour ago
That being said, gpt-image-1.5 was a big leap in visual quality for OpenAI and eliminated most of the classic issues of its predecessor, including things like the “piss filter.”
I’ll update this comment once I’ve finished running gpt-image-2 through both the generative and editing comparison charts on GenAI Showdown.
Since the advent of NB, I’ve had to ratchet up the difficulty of the prompts especially in the text-to-image section. The best models now score around 70%, successfully completing 11 out of 15 prompts.
For reference, here’s a comparison of ByteDance, Google, and OpenAI on editing performance:
https://genai-showdown.specr.net/image-editing?models=nbp3,s...
And here’s the same comparison for generative performance:
https://genai-showdown.specr.net/?models=s4,nbp3,g15
UPDATES:
gpt-image-2 has already managed to overcome one of the so‑called “model killers” on the test suite: the nine-pointed star.
Results are in for the generative (text to image) capabilities: Gpt-image-2 scored 12 out of 15 on the text-to-image benchmark, edging out the previous best models by a single point. It still fails on the following prompts:
- A photo of a brightly colored coral snake but with the bands of color red, blue, green, purple, and yellow repeated in that exact order.
- A twenty-sided die (D20) with the first twenty prime numbers (2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71) on the faces.
- A flat earth-like planet which resembles a flat disc is overpopulated with people. The people are densely packed together such that they are spilling over the edges of the planet. Cheap "coastal" real estate property available.
All Models:
https://genai-showdown.specr.net
Just Gpt-Image-1.5, Gpt-Image-2, Nano-Banana 2, and Seedream 4.0
Comment by Melatonic 1 hour ago
Comment by throw310822 1 hour ago
Comment by thevinter 2 hours ago
Comment by lehmacdj 2 hours ago
Comment by throw310822 1 hour ago
Comment by Vachyas 1 hour ago
Consistency? So it fails less often?
Based on the released images, (especially the one "screenshot" of the Mac desktop) I feel like the best images from this model are so visually flawless that the only way to tell they're fake is by reasoning about the content of the image itself (ex. "Apple never made a red iPhone 15, so this image is probably fake" or "Costco prices never end in .96 so this image is probably fake")
Comment by RobinL 1 hour ago
It doesn't reliably give you 10 slices, even if you ask it to number them. None of the frontier models seem to be able to get this right
Comment by thevinter 1 hour ago
Especially when it comes to detailed outputs or non-standard prompts.
I do believe it will get even better - not sure it will happen within a year but I wouldn't be incredibly surprised if it did.
Comment by vunderba 1 hour ago
I experimented with the concept of procedural generation of Waldo-style scavenger images with Flux models with rather disappointing results. (unsurprisingly).
Comment by throw310822 1 hour ago
Comment by jinushaun 48 minutes ago
Comment by minimaxir 3 hours ago
API Pricing is mostly unchanged from gpt-image-1.5, the output price is slightly lower: https://developers.openai.com/api/docs/pricing
...buuuuuuuuut the price per image has changed. For a high quality image generation the 1024x1024 price has increased? That doesn't make sense that a 1024x1024 is cheaper than a 1024x1536, so assuming a typo: https://developers.openai.com/api/docs/guides/image-generati...
The submitted page is annoyingly uninformative, but from the livestream it proports the same exact features as Gemini's Nano Banana Pro. I'll run it through my tests once I figure out how to access it.
Comment by strongpigeon 1 hour ago
I think you meant more expensive, right? Because it would make sense for it to be cheaper as there are less pixels.
Comment by ieie3366 1 hour ago
I would imagine this will hit illustrators / graphics designers / similar people very hard, now that anyone can just generate professional looking graphical content for pennies on the dollar.
Comment by retrac98 1 hour ago
Comment by Bennettheyn 1 hour ago
Comment by ChrisArchitect 1 hour ago
As with anything AI, we are not ready for the scale of impact. And for what? Like, why are you proud of this?