AI-generated images would remove all the trust and admire for human talent in art, similar to how text-generation would remove trust and admire for human talent in writing. Same case for coding.
So, let's simulate that future. Since no one trusts your talent in coding, art or writing, you wouldn't care to do any of these. But the economy is built on the products and services which get their value based how much of human talent and effort is required to produce them.
So, the value of these services and products goes down as demand and trust goes down. No one knows or cares who is a good programmer in the team, who is great thinker and writer and who is a modern Picasso.
So, the motivation disappears for humans. There are no achievements to target, there is no way to impress others with your talent. This should lead to uniform workforce without much difference in talents. Pretty much a robot army.
If this was a farm of sweatshop Photoshopers in 2010, who download all images from the internet and provide a service of combining them on your request, this would escalate pretty quickly.
Question: with copyright and authorship dead wrt AI, how do I make (at least) new content protected?
Anecdotal: I had a hobby of doing photos in quite rare style and lived in a place where you'd get quite a few pictures of. When I asked gpt to generate a picture of that are in that style, it returned highly modified, but recognizable copy of a photo I've published years ago.
my question to your anecdotal: who cares? not being fecicious, but who cares if someone reproduced your stuff and millions of people see your stuff? is the money that you want? is it the fame? because fame you will get, maybe not money... but couldn't there be another way?
People have values that go beyond wealth and fame. Some people care about things like personal agency, respect and deference, etc.
If someone were on vacation and came home to learn that their neighbor had allowed some friends stay in the empty house, we would often expect some kind of outrage regardless of whether there had been specific damage or wear to the home.
Culturally, people have deeply set ideas about what's theirs, and feel like they deserve some say over how their things are used and by whom. Even those that are very generous and want their things be widely shared usually want to have have some voice in making that come to be.
Suddenly, copyright doesn't matter anymore when it's no longer useful to the narrative.
OpenAI does care about copyright, thankfully China does not: https://imgur.com/a/RKxYIyi
(to clarify, OpenAI stops refining the image if a classifier detects your image as potentially violating certain copyrights. Although the gulf in resolution is not caused by that.)
facetious
[I won't bother responding to the rest of your appalling comment]
[dead]
The issue is ownership, not promotion or visibility.
[dead]
I have a Nano Banana Pro blog post in the works expanding on my experiments with Nano Banana (https://news.ycombinator.com/item?id=45917875). Running a few of my test cases from that post and the upcoming blog post through this new ChatGPT Image model, this new model is better than Nano Banana but MUCH worse than Nano Banana Pro which now nails the test cases that previously showed issues. The pricing is unclear but gpt-image-1.5 appears to be 20% cheaper than the current gpt-image-1 model, which would put a `high`-quality generation in the same price range as Nano Banana Pro.
One curious case demoed here in the docs is the grid use case. Nano Banana Pro can also generate grids, but for NBP grid adherence to the prompt collapses after going higher than 4x4 (there's only a finite amount of output tokens to correspond to each subimage), so I'm curious that OpenAI started with a 6x6 case albeit the test prompt is not that nuanced.
I'll be running gpt-image-1.5 through my GenAI Showdown later today, but in the meantime if you want to see some legitimately impressive NB Pro outputs, check out:
https://mordenstar.com/blog/edits-with-nanobanana
In particular, NB Pro successfully assembled a jigsaw puzzle it had never seen before, generated semi-accurate 3D topographical extrapolations, and even swapped a window out for a mirror.
Nice! Your comparison site is probably the best one out there for image models
I just tested GPT1.5. I would say the image quality is on par with NBP in my tests (which is surprising as the images in their trailer video are bad), but the prompt adherence is way worse, and its "world model" if you want to call it that is worse. For instance, I asked it for two people in a row boat and it had two people, but the boat was more like a coracle and they would barely fit inside it.
Also: SUPER ANNOYING. It seems every time you give it a modification prompt it erases the whole conversation leading up to the new pic? Like.. all the old edits vanish??
I added "shaky amateur badly composed crappy smartphone photo of ____" to the start of my prompts to make them look more natural.
I really enjoyed your experiments. Thank you for sharing your experiences. They've improved my prompting and have tempered my expectations.
It's really weird to see "make images from memories that aren't real" as a product pitch
It's strange to me too, but they must have done the market research for what people do with image gen.
My own main use cases are entirely textual: Programming, Wiki, and Mathematics.
I almost never use image generation for anything. However its objectively extremely popular.
This has strong parallels for me to when snapchat filters became super popular. I know lots of people loved editing and filtering pictures but I always left everything as auto mode, in fact I'd turn off a lot of the default beauty filters. It just never appealed to me.
It would creep me out if the model produced origami animals for that prompt.
Is there a watermarking, or some other way for normal people to tell if its fake?
https://help.openai.com/en/articles/8912793-c2pa-in-chatgpt-...
It doesn't mention the new model, but it's likely the same or similar.
Was it ever explained or understood why ChatGPT Images always has (had?) that yellow cast?
Meta's codec avatars all have a green cast because they spent millions on the rig to capture whole bodies and even more on rolling it out to get loads of real data.
They forgot to calibrate the cameras, so everything had a green tint.
Meanwhile all the other teams had a billion macbeth charts lying around just in case.
Also, you'd be shocked at how few developers know anything at all about sRGB (or any other gamut/encoding), other than perhaps the name. Even people working in graphics, writing 3D game engines, working on colorist or graphics artist tools and libraries.
maybe their version of synth-id? it at least helps me spot gpt images vs gemini's
My pet theory is that OpenAI screwed up the image normalization calculation and was stuck with the mistake since that's something that can't be worked around.
At the least, it's not present in these new images.
There's still something off in the grading, and I suspect they worked around it
(although I get what you mean, not easily since you already trained)
I'm guessing when they get a clean slate we'll have Image 2 instead of 1.5. In LMArena it was immediately apparent it was an OpenAI model based on visuals.
Not really, but there's a number of theories. The simplest one is that they "style tuned" the AI on human preference data, and this introduced a subtle bias for yellow.
And I say "subtle" - but because that model would always "regenerate" an image when editing, it would introduce more and more of this yellow tint with each tweak or edit. Which has a way of making a "subtle" bias anything but.
There was also the theory that is was because they scanned a bunch of actual real books and book paper has a slight yellow hue.
That seems unlikely, as we didn't see anything like that with Dall-E, unless the auto regressive nature of gpt-image somehow was more influenced by it.
My pet theory is that this is the "Mexico filter" from movies leaking through the training data.
I never heard anything concrete offered. At least it's relatively easy to work around with a tone mapping / LUTs.
Soaked in piss.
Colloquially called the urine filter
I still use Midjourney, because all of these major players are so bad at stylistic and creative work. They're singularly focused on photorealism.
This is surprising. Is there a gallery of images that illustrates this?
Midjourney has a gallery on their website: https://www.midjourney.com/explore
their explore page is a firehose of examples created by users and you can see the prompt used so you can compare the results in other services https://www.midjourney.com/explore?tab=video_top
That's the opinionated vs user choice dynamic. When the opinions are good, they have a leg up
This is a cultural flaw that predates image generation. Even PG has made statements on HN in the past equating “rendering skill” with the quality of art works. It’s a stand-in for the much more difficult task of understanding the work and value of culture making within the context of the society producing it.
It's still not available in the API despite them announcing the availability.
They even linked to their Image Playground where it's also not available..
I updated my local playground to support it and I'm just handling the 404 on the model gracefully
It's a staggered rollout but I am not seeing it on the backend either.
Alt text is one of the nicest uses for ai and still Open AI didn't bother using it for something so basic. The dogfooding is not strong with their marketing team.
Even from their own curated examples, this looks quite a bit worse than nano banan in terms of preserving consistency on image edits.
I didn't have a good experience with NB. I am half Indian. Immediately changes my face to a prototypical Indian man every time I use it.
This tool is keeping my look the same.
I find including "don't change anything else" in the NBP prompt goes a long way.
we seriously can't be burning GW of energy just to have sama in a GPT-Shirt Ad generated by A.I
impressive stuff though - as you can give it a base image + prompt.
counterpoint: we should make energy abundant enough that it really doesn't matter if sama wants to generate gpt-shirt ads or not.
we have the capability, we just stopped making power more abundant.
Great to have continued competition in the different model types.
What angle is there for second tier models? Could the future for OpenAI be providing a cheaper option when you don't need the best? It seems like that segment would also be dominated by the leading models.
I would imagine the future shakes out as: first class hosted models, hosted uncensored models, local models.
Hope to see more "red alert" status from the ai wars putting companies into al hands on deck. This is only helping cost of tokens and efficacy. As always competition only helps the end users.
not super impressed. feels like 70% as good as nano banana pro.
Unlike Nano Banana it allows generating photos of children. Always fun to ask AI to imagine children of a couple but it's also kinda concerning that there might be terrible use cases.
If memory serves me, Nano Banana allows generating/editing photos of children. But anything that could be misinterpreted, gets blocked, even absolutely benign and innocent things (especially if you are asking to modify a photo that you upload there). So they allow, but they turn on the guardrails to a point that might not be useful in many situations.
I was able to generate photos of my imagined children via Nano Banana
I haven't seen that, meanwhile gpt-image-1.5 still has zero-tolerance policing copyright (even via the API) so it's pretty much useless in production once exposed to consumers.
I'm honestly surprised they're still on this post-Sora 2: let the consumer of the API determine their risk appetite. If a copyright holder comes knocking, "the API did it" isn't going to be a defense either way.
In the image they showed for the new one, the mechanic was checking a dipstick...that was still in the vehicle.
I really hope everyone is starting to get disillusioned with OpenAI. They're just charging you more and more for what? Shitty images that are easy to sniff out?
In that case, I have a startup for you to invest in. Its a bridge-selling app.
Haven’t their prices stayed at $20/m for a while now?
They've published anticipated price increases over coming years. Prices will rise dramatically and steadily to meet revenue targets.
Anyone else have issues verifying with openai? I always get a "congrats you're done" screen with a green checkmark from Persona, nothing to click, and my account stays unverified. (Edit, mystically, it's fixed..!)
Nano Banana Pro is so good that any other attempt feels 1-2 generations behind.
Nano banana pro is almost as good as seedream 4.5!
Seedream 4.5 is almost as good as Seedream 4!
(Realistically, Seedream 4 is the best at aesthetically pleasing generation, Nano Banana Pro is the best at realism and editing, and Seedream 4.5 is a very strong middleground between the two with great pricing)
gpt-image-1.5 feels like OpenAI doing the bare minimum to keep people from switching to Gemini every time they want an image.
nah Nano Banana Pro is much better
>Now remove the two men, just keep the dog, and put them in an OpenAI livestream that looks like the attached image.
Where is the image given along with the prompt? If I didn't miss it: Would have been nice to show the attached image.
on top of the prompt. It has a weird layout; I had to scroll up to see it.
We'll merge that thread hither to give some other submitters a chance.
Another bunch of "startups" have been eliminated.
Among those, Photoshop.
I wish. Even Nano Banana Pro still sucks for even basic operations.
[dead]
[dead]