488
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 22 Feb 2024
488 points (96.2% liked)
Technology
60012 readers
2576 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
No matter what Google does, people are going to come up with gotcha scenarios to complain about. People need to accept the fact that if you don't specify what race you want, then the output might not contain the race you want. This seems like such a silly thing to be mad about.
It's silly to point at brand new technology and not expect there to be flaws. But I think it's totally fair game to point out the flaws and try to make it better, I don't see why we should just accept technology at its current state and not try to improve it. I totally agree that nobody should be mad at this. We're figuring it out, an issue was pointed out, and they're trying to see if they can fix it. Nothing wrong with that part.
It's really a failure of one-size-fits-all AI. There are plenty of non-diverse models out there, but Google has to find a single solution that always returns diverse college students, but never diverse Nazis.
If I were to use A1111 to make brown Nazis, it would be my own fault. If I use Google, it's rightfully theirs.
The solution is going to take time. Software is made more robust by finding and fixing edge cases. There's a lot of work to be done to find and fix these issues in AI, and it's impossible to fix them all, but it can be made better. The end result will probably be a patchwork solution.
The issue seems to be the underlying code tells the ai if some data set has too many white people or men, Nazis, ancient Vikings, Popes, Rockwell paintings, etc then make them diverse races and genders.
What do we want from these AIs? Facts, even if they might be offensive? Or facts as we wish they would be for a nicer world?
American using Gemini: "Please produce images of the KKK, historically accurate Santa's Workshop Elves, and the board room of a 1950s auto company"
Also Americans: "AH!! AH!!!!! Minorities and Women!!!!!!! AAAAAHHH!!!!"
I mean, idk, man. Why do you need AI to generate an image of George Washington when you have thousands of images of him already at your disposal?
Because sometimes you want an image of George Washington, riding a dinosaur, while eating a cheeseburger, in Paris.
Which you actually can’t do on Bing anyway, since it ‘content warning’ stops you from generating anything with George Washington…
Ask it for a Founding Father though, it’ll even hand him a gat!
He's not even eating the cheeseburger, crap AI.
An it's not a beyond burger... it's promoting the genocide of cattle.
Here’s one that was made, just for you, with specifically a VEGAN cheeseburger in the prompt :D
Excellent.
Funnily enough, he’s not eating one in the other three images either. He’s holding an M16 in one, with the dinosaur partially as a hamburger (?). In the other two he’s merely holding the burger.
I assume if I change the word order around a bit, I could get him to enjoy that burger :D
This is the thing. There's an incredible number of inaccuracies in the picture, several of which flat out ignore the request in the prompt, and we laugh it off. But the AI makes his skin a little bit darker? Write the Washington Post! Historical accuracy! Outrage!
Well, the tech is of course still young. And there's a distinct difference between:
A) User error: a prompt that isn't as good as it can be, with the user understanding for example the 'order of operations' that the AI model likes to work in.
B) The tech flubbing things because it's new and constantly in development
C) The owners behind the tech injecting their own modifiers into the AI model in order to get a more diverse result.
For example, in this case I understand the issue: the original prompt was 'image of an American Founding Father riding a dinosaur, while eating a cheeseburger, in Paris.' Doing it in one long sentence with several comma's makes it harder for the AI to pin down the 'main theme' from my experience. Basically, it first thinks 'George on a dinosaur' with the burger and Paris as afterthoughts. But if you change the prompt around a bit to 'An American Founding Father is eating a cheeseburger. He is riding on a dinosaur. In the background of the image, we see Paris, France.', you end up with the correct result:
Basically the same input, but by simply swapping around the wording it got the correct result. Other 'inaccuracies' are of course to be expected, since I didn't really specify anything for the AI to go of. I didn't give it a timeframe for one, so it wouldn't 'know' not to have the Eiffel Tower and a modern handgun in it. Or that that flag would be completely wrong.
The problem is with C) where you simply have no say in the modifiers that they inject into any prompt you send. Especially when the companies state that they are doing it on purpose so the AI will offer a more diverse result in general. You can write the best, most descriptive prompt and there will still be an unexpected outcome if it injects their modifiers in the right place of your prompt. That's the issue.
C is just a work around for B and the fact that the technology has no way to identify and overcome harmful biases in its data set and model. This kind of behind the scenes prompt engineering isn't even unique to diversifying image output, either. It's a necessity to creating a product that is usable by the general consumer, at least until the technology evolves enough that it can incorporate those lessons directly into the model.
And so my point is, there's a boatload of problems that stem from the fact that this is early technology and the solutions to those problems haven't been fully developed yet. But while we are rightfully not upset that the system doesn't understand that lettuce doesn't go on the bottom of a burger, we're for some reason wildly upset that it tries to give our fantasy quasi-historical figures darker skin.
The random lettuce between every layer is weirdly off-putting to me. It seems like it's been growing on the burger for quite some time :D
Doesn’t look too bad to me. I love a fair bit of crispy lettuce on a burger. Doing it like that at least spreads it out a bit, rather than having a big chunk of lettuce.
Still, it that was my burger… I’d add another patty and extra cheese.