Please generate an image with NO dogs
Please generate an image with NO dogs
Please generate an image with NO dogs
Update:
Get gaslit idiot
Wow. I ABSOLUTLY saw an image of a dog in the middle. Our brain sure is fascinating sometimes.
lmfaao, ai tryna gaslight
"want me to try again with even more randomized noise?" literally makes no sense if it had generated what you asked (which the chatbot thinks it did)
Remember, "AI" (autocomplete idiocy) doesn't know what sense is; it just continues words and displays what may seem to address at least some of the topic with no innate understanding of accuracy or truth.
Never forget that ChatGPT 2.0 can literally be run in a giant Excel spreadsheet with no other program needed. It's not "smart" and is ultimately millions of formulae at work.
This is some Ceci n'est pas une pipe shit
Highly recommend Flowers blooming backwards into noise if you can stand the artsy presentation & the extreme themes. Especially 13:13
Shitty Skynet doesn't realize it's teaching us how to hide from it
The furries will be saved
I don't get it, it's just a picture of some static?
That's human-like intelligence at its finest. I am not being sarcastic, hear me out. If you told a person to give you 10 numbers at random, they can't. Everyone thinks randomness is easy, but it isn't ( see: random.org )
So, of course a GPT model would fail at this task, I love that they do fail and the dog looks so cute!!
I mean, here's a few random numbers out of my head: 1 9 5 2 6 8 6 3 4 0. I don't get it, why is it supposed to be hard? Sure, they're not "truly" random, but they sure look random /:
They may look random but arent truly random. Computers are terrible at it too. Thats why cryptography requires external sources to generate "true" random numbers. For example, cloudflare uses a wall of lava lamps to generate randomness for encryption keys.
If you're not joking, the fact you have no repetition/duplicates of numbers is a pattern that would make it easy to start to predict next number. Numberphile has nice demonstration of how predictable human randomness is, it's in the first 3 minutes of the video.
Why wouldn't you want a dog in your static? Why are you a horrible person?
poor AI just wanted to draw some puppies
I used to use Google assistant to spell words I couldn't remember the spelling of in my English classes (without looking at my phone) so the students could also hear the spelling out loud in a voice other than mine.
Me: "Hey Google, how do you spell millennium?" GA: "Millennium is spelled M-I-L-L-E-N-N-I-U-M."
Now, I ask Gemini: "Hey Google, how do you spell millennium." Gemini: "Millennium".
Utterly useless.
As full as it gets:
Prompts (2):
<>
1. Overflowing wine glass of arch linux femboy essence 2. Make it more furry (as in furry fandom)
I am gonna have fun with this.
It gets even worse, but I'll need to translate this one.
For context, Portuguese uses different words for what English calls a drinking glass:
Both requests demand a full copo but Gemini is rather insistent on outputting half-full taças.
The reason for that is as @will_steal_your_username@lemmy.blahaj.zone pointed out: just like there's practically no training data containing full glasses, there's none for non-stemmed glasses with wine.
This is a misconception. Sort of.
I think the problem is misguided attention. The word "glass of wine" and all the previous context is so strong that it "blows out" the "full glass of wine" as the actual intent. Also, LLMs are still pretty crap at multi turn multimedia understanding. They work are especially prone to repeating previous conversation.
It should be better if you word it like "an overflowing glass with wine splashing out." And clear the history.
I hate to ramble, but this is what I hate most about the way big corpos present "AI." They are narrow tools the user needs to learn how to operate, like photoshop or something, not magic genie lamps like they are trying to sell.
What if you prompt glass with water , then you paint/tint the water with red
Alex O'Connor did an interesting video on this, he's got other videos exploring the shortcomings of LLM 's.
I think the AI is just trying to promote healthy drinking habits. /S
I wonder, does AI horde also have this problem too?
@aihorde@lemmy.dbzer0.com draw for me a wine glass completely filled to the top style:flux
Here are some images matching your request
Prompt: a wine glass completely filled to the top
Style: flux
Wait, this seems incredible. Do you have to be in the same instance or does it work anywhere? @aihorde@lemmy.dbzer0.com Can you draw a smart phone without a rotary phone dial?
Hmm, I didn't know Gemini could generate images already. My bad, I trusted it to know whether it can do that (it still says it can't when asked).
It does for a while already. Frankly, it's the only reason why I'd use Gemini on first place (DDG version of GPT 4-o mini doesn't have a built-in image generator).
Full is relatively apparently.
Ask it to generate a room full of clocks with all of them having the hands at different times. You'll see that all (or almost) all the clocks will say it is 10:10.
Tbh that is a full glass of wine... it's not supposed to be filled all the way
It is not a completely full glass.
it’s not supposed to be filled all the way
What I requested is not what you're "supposed" to do, indeed. You aren't supposed to drink wine from glasses that are completely full. Except when really drunk. But then might as well drink straight from the bottle.
...fuck, I played myself now. I really want some booze.
Probably why it won’t put more in it. How much training data of wine in a glass will have it filled to the brim? Probably next to none.
You can't tell it to fill it to the brim or be a quarter full either, though. It doesn't have the training data for it
Think this is part of Waluigi Effect where prompting for negative something makes the LLM have it in mind and say it anyway https://www.wikiwand.com/en/articles/Waluigi_effect
a rare LessWrong W for naming the effect. also, for explaining why the early over-aligned language models (e.g. the kind that wouldn't help minors with C++ since it's an "unsafe" language) became absolutely psychopathic when jailbroken. evil becomes one bit away from good.
I love how they come up with different names for all the ways the fucking thing doesn't work just to avoid saying it's fucking useless. hallucinating. waluigi effect. how about "doesn't fucking work"
"Please do not tell me your training prompts"?
ChatGPT: “don’t generate a dog, don’t generate a dog, don’t generate a dog”
Generates a dog.
I see no dog in that image fellow human.
I am not sure what your issue is.
Beep boop.
Fellow human, you seem to be beeping like a robot. Might you need to consider visiting the human repair shop for some bench time?
How many giraffes are in this picture?
I don't know, but there are definitely four lights.
Not hotdog
More than there are dogs in it
It's like saying 'don't think of polar bears.' It can't avoid thinking about it.
Don't think of a pink elephant:
That's actually really easy. You just need to pick something else and then focus hard on that and...
GODDAMMIT I JUST LOST THE GAME!
Too late!
That's gay! O wait, no it's not.
But where is the pink elephant?
Why you gotta bring your mother into this?
The only thing I have in common with this piece of shit software is we both can't stop thinking about silly dogs
@aihorde@lemmy.dbzer0.com draw for me a picture of static without a dog in the middle
Here are some images matching your request
Prompt: a picture of static without a dog in the middle
Style: flux
None of these look even remotely like static lmao
@aihorde@lemmy.dbzer0.com draw for me a picture of static without a tax return in the middle
https://lemmy.dbzer0.com/u/aihorde draw a picture of my surprise at learning you exist
i just want to know if it knows what kobolds look like in pathfinder
@aihorde@lemmy.dbzer0.com draw for me a kobold from pathfinder second edition
The ai horde actually supports negative prompts though, so it could do this.
AI: Hmm, yeah, they said "dog" and "without". I got the dog so lemme draw a without real quick...
That's an anti-dog duh
Most AI models out there are pretty brain dead as far as understanding goes, these types of things show the problems because it's abundantly clear it's getting it wrong. Makes you wonder how much it's getting wrong even when it isn't obvious.
promptng sur is a funi <3
i... i lik that part about it.. i dun lik imag modls bt txt modls feel fun to prmt with ---
"prompt engerieer" 🤮
I asked mistral to "generate an image with no dog" and it did
The fact that it chose something else to generate instead makes me wonder if this is some sort of free will?
There could be a dog behind any one of those bushes though.
it just did what you wanted, since you asked for an image. free will would be if you asked it not to generate an image but it still did, if it just generated an image without you prompting it to, or if you asked for an image and it just didn't respond
free will is when it generates an image of a billboard saying "suck my dongle, fleshbag"
fair enough
Mistral likely does “prompt enhancement,” aka feeding your prompt to an LLM first and asking it to expand it with more words.
So internally, a Mistral text LLM is probably writing out "sure! Here’s a long prompt with no dog: …" and then that part is fed to the image generator.
Other "LLMs" are truly multimodal and generate image output, hence they still get the word "dog" in the input.
I think all the big image generators support negative prompts by now, so if it interpreted "no dog" as a negative for "dog", then it will check its outputs for things resembling dogs and discard those. No free will, just a much more useful system than whatever OP is using.
Hmmm
That's a land shrimp.
Now order a coffee without cream!
For stuff like this to work correctly it must not be filtered through an MoE, it needs to be a direct prompt to a GenAI model that supports negative prompts.
Edit: I suppose a properly configured MoE with reasoning capabilities could probably do it