I have a similar experience. In my own experiment, I can't get DALL-E to turn off the street lamp at a bus stop in the darkness. I've tried "no light", "broken street lamp", etc.; no use. Any mention of "street lamp", the scene will include a working street lamp.
It's just more probable that a scene with a lamp in the darkness must have that lamp providing light, and this is something that DALL-E will not break out of.
They got a solution for that, which is using their —no argument. https://midjourney.gitbook.io/docs/imagine-parameters#prompt...
I haven’t checked if DALL-E has that option too.
Otherwise, you could try other variations like:
street light, light pole, lamp pole, lamppost, street lamp, light standard, or lamp standard
I copied that from Wikipedia :)
Best of luck!
It reminds me of how TV shows often have a president that resembles the current president in superficial ways, while being distinct enough that they won't get sued.
I'd be interested to know why this happens.
ETA: OK, well, based on the following comments, it has a prohibition on living people, but you can’t libel the dead. So it either is bad at faces or it has a prohibition there too. The article would have mentioned if DALL-E said it wouldn’t render Lennon or Harrison. QED, bad at faces?
Crazy how fast the tech is moving.
Here's one I made: https://imgur.com/yAzKkHb
“High detail, macro portrait photo, a handsome Australian man with a strong jaw line, blue eyes and brown hair, smiles at the camera, set in an outdoor pub at golden hour, shot using a ZEISS Supreme Prime lens”
https://reddit.com/r/dalle2/comments/wsi97q/_/ikyjqhh/?conte...
> High detail, macro portrait photo, a [physical descriptor, regional identity, etc] man/woman with [eye color] eyes and [hair color] hair, smiles at the camera, set in a [field/dimly lit room/whatever] at golden hour, shot using a ZEISS Supreme Prime lens
The suggestions in that thread are quite effective, particularly the notion of reducing synthetic ‘beauty’ for a more human appearance.
that seems too specific. You wouldn't even request that in real life, unless you were trying to be ironically pretentious.
I spent ages on this earlier getting nowhere. I'm starting to think DALL-E is better if you don't really know what you want and you're just fishing for ideas.
DALLE is hard! Curious to see if I can be beat.
do you want a realistic looking one? 3d rendered? what do you have in mind exactly?
Maybe the AI will finally get what designers always complained about annoying clients.
> Create a website design for a company that sells propane and propane accessories. The name of the company is Strickland Propane, a local propane dealership. Make it pop.
Results:
* https://i.imgur.com/Jv7NJEN.png
* https://i.imgur.com/5Uiyg1R.png
* https://i.imgur.com/LL1DC11.png
* https://i.imgur.com/buv5BvS.png
So there you have it :p
Prompt:
> Create a website design for ACME Corporation, a company which produces a wide array of products that are dangerous, unreliable or preposterous. Include customer quotes from a dissatisfied Wile E. Coyote prominently on the page. Make it pop.
Results:
* https://i.imgur.com/WK3QBj9.png
* https://i.imgur.com/Bghgzjt.png
I suggest learning how to use a few graphics packages instead when you have a need. eg, this evening my wife wanted to print a birthday card for a dog that she is boarding. I fired up Inkscape and discovered a bug wrt handling an imported photo, scaling it and cropping it with a mask. lol! I was able to export it to a .png and print from something else.
Anyway, I was a little unfortunate and IT is still rather shakey. However, discussions about DALL-E are way more interesting than actually using it yourself. You see some seriously intelligent solutions to getting a desired image that is surely a form of programming. DALL-E does not even have a real natural language pre-processor because it is hamstrung to dump pejorative terms etc. This means that I can draw the classic, dreadful plan of the cargo hold of a slaver ship from the 17/18C and its cargo but I doubt DALL-E can.
The q&a about this thing are way more complex and interesting than its actual output! A is right, I it isn't. DALL-E is quite interesting but it really isn't intelligent. The intelligent bit is getting the input right for the desired output. Perhaps another model could be developed for that.
That’s what trademarks are for.
So really they generated these and never bothered to do the research of their own question.
And, as the article states:
> But seriously, how creative and original can you be with something that is trained on the works of millions of other creators?
> To me, it is unclear whether you can actually call these works your 'own' at all, because there's always someone else's touch on it.
> […] users of DALL-E will also never be sure whether they are generating something that is 'theirs' or just a knockoff of someone else's work.
It's pay for action (send me a penny if you find anything worthwhile), and the copyright is owned by the payee:
What if you can get a lot more out of it by embracing the unexpected responses. Can it be a tool for exploring lateral thinking? You provide a prompt computer responds with images that are a prompt to human. A baby swiming next to a dolar bill outputs a distorted person face inside a dolar bill with some baby features, could be the start to a rabbit hole of prompts and images where you'll end up with something completly different than your initial expectations.
Or perhaps even beyond just a paint over, and into the realm of recreating an entire AI artwork but with a human touch to get details just right.
Looking forward to it.
I am a logo artist and I sell pre-made logo designs. Before the current AI services I had to come up with visual ideas by myself, like a caveman. Now I use the AI to generate a bunch of sketches and blurry ideas, and then use my graphic design experience to polish them up to a usable level. Here's how it looks. https://imgur.com/a/DKTsKdC
I am absolutely sure that a lot of people are doing the same right now, just keeping quiet about it.
I am intrigued by the use of AI as a form of creativity assist. As someone without any talent for this, the left pictures are useless for me, as I don't know how to take them into something like the pictures on the right. The point of a sketch is to show them to a customer, but if you would show these sketches to me, I wouldn't know which one would turn out great and which one wouldn't.
Given that, do you feel that the generated sketches are useful as a base sketch? I mean, you could probably have used any of the existing NFL teams logos and as inspiration, instead of letting the software remix them for you.
This is what I got for the prompt "honey badger logo for an NFL sports team"
Dunno what's going on with that last guy though, perhaps he's had one too many concussions on the field...
Imagine you're a software developer. In the near future, your manager wants a feature implemented in your company's app. He throws together a short mail with requirements and sends it to the prompt engineering department.
The prompt engineers fix a few typos, clean up the grammar and pepper a few secret sauce keywords around like "in the style of firefox", "in the style of kde". This get thrown into Microsoft Copilot 3.0 that barfs out a bunch of code.
Copilot's code has inconsistent indentation, three different method naming conventions and some variables named in a foreign language. It runs, but crashes if you tap on the lower left corner of the screen and allows you to drag the order quantity below 0. But that's ok, it's why we employ software engineers like you in our company. You will use your years of coding experience to touch up AI code to perfection. Better get the details just right until Monday all-hands!
Still looking forward to it?
Something I've come to realise is that I'm much better at fixing other people's code than coding from scratch.
I could definitely work with that kind of tool.
When you put it like that, it sounds like a nightmare up side down world. It's not AI that's the tool for enhancing human creativity. It's humans that are the AI's tool, cleaning up the edge cases the AI artist can't handle (yet). It destroys creative jobs that give joy to people and creates assembly line jobs for them to slog in.
Also sounds like an upside down use of the tool, since AI generated art really isn't that great at composition once you've got over the magic of it being able to respond to the prompt at all (but is much better at texture and filling in boring details), and the current state of the art AI tools can produce images which conform to a human guideline sketch...
Artists, programmers and everyone else will have to find their joy somewhere other than than selling themselves to a corporation, once AI driven markets optimize away any room for "joy" and the like, and that's going to be one of the few good things about automation. The sooner we break people from the Puritan delusion that work defines a person's meaning and the value of their expression, the sooner we can once again decouple culture from the machinery of capitalism.
Finally, the crossover of creative writing x cs... for graphic design? I can't wait to watch the lawyers recoil. ::Prepares Popcorn::
"Two men, one of whom is on fire, shaking hands on an industrial lot." can be rewritten as, "Two men, shaking hands, standing on an industrial lot. Person on the right is on fire. Camera is 30 metres away."
You can go into more specifics of the framing and the angle from which you want the picture to be take. By default, DALL E will give you the most realistic generations to your prompts unless you mention "digital art" or a particular art style. I have gotten the best results when generating art instead of photos.
It surprisingly reminds me a lot of when I traveled to Japan without knowing really any Japanese. I needed to communicate not only with friends who don't know much English either, but also other people (like restaurant wait staff, train station staff, etc.).
I used Google Translate often, but many times I or my friend(s) (or the other people) would need to re-write our statements a few times until the translation result clicked well enough in each other's languages to be understandable.
No excuse for it. Screw that.
It's really great and cool and all - but it's retrieving things that it was trained on.
Show me something original it did.
Most of the images I've generated using Dalle 2 feels completely original. Just have a look at the reddit r/dalle2 and I'm pretty sure you'll also decide they're "original works".
So I guess "something original" means "everything you've seen before on Instagram"
Somebody posted a nice comparison between DALL-E 2, Mid Journey, and Stable Diffusion: https://twitter.com/fabianstelzer/status/1561019187451011074.
Nevermind because it showed this weakness in the model in understanding what a dollar bill is. The most novel result being the image where the baby's visage appears inside the dollar.
Regarding Abbey Road, I found it interesting that the model's concept of a public person spans their lifetime evidenced by the images where the contemporary images are used. Also interesting to me is the model's weakness in understanding specific people.
Then again though, I haven't been clicking on every DALL-E post so maybe this is old news.
I had imagined that some parts of the training data would consist of the actual image, and it would find a good match for it somewhere deep into it's artificial consciousness.
Perhaps as a blogger I am extra salty when relatively low effort stuff gets upvoted over things that take a lot of work to write (in some cases, those being things I have written). But hey, that's life.
The only cover that really worked from my point of view was the Velvet Underground one, and perhaps the Rolling Stones one. Abbey Road came closer than what I thought it would, but was pretty bad ultimately, and the other three really had nothing usable.
Thus, I look forward to the AI-generated versions of famous works that deepfake the original cast into speaking (hopefully) better-written dialogue. Imagine when this technology is widespread, fanfic authors rendering their interpretation of works with the descendants of DALL-E. Everyone gets the dream adaptations and sequels and finales they want.