Stolen from a tweet. Not groundbreaking, but interesting thoughts on use-cases for the new OpenAI image model:
(1) This changes filters. Instagram filters required custom code; now all you need are a few keywords like “Studio Ghibli” or Dr. Seuss or South Park.
(2) This changes online ads. Much of the workflow of ad unit generation can now be automated, as per QT below.
(3) This changes memes. The baseline quality of memes should rise, because a critical threshold of reducing prompting effort to get good results has been reached.
(4) This may change books. I’d like to see someone take a public domain book from Project Gutenberg, feed it page by page into Claude, and have it turn it into comic book panels with the new ChatGPT. Old books may become more accessible this way.
(5) This changes slides. We’re now close to the point where you can generate a few reasonable AI images for any slide deck. With the right integration, there should be less bullet-point only presentations.
(6) This changes websites. You can now generate placeholder images in a site-specific style for any
tag, as a kind of visual Loren Ipsum.
(7) This may change movies. We could see shot-for-shot remakes of old movies in new visual styles, with dubbing just for the artistry of it. Though these might be more interesting as clips than as full movies.
(8) This may change social networking. Once this tech is open source and/or cheap enough to widely integrate, every upload image button will have a generate image alongside it.
(9) This should change image search. A generate option will likewise pop up alongside available images.
(10) In general, visual styles have suddenly become extremely easy to copy, even easier than frontend code. Distinction will have to come in other ways.