Imagine reading that headline 20 years ago.
Gaming
Sub for any gaming related content!
Rules:
- 1: No spam or advertising. This basically means no linking to your own content on blogs, YouTube, Twitch, etc.
- 2: No bigotry or gatekeeping. This should be obvious, but neither of those things will be tolerated. This goes for linked content too; if the site has some heavy "anti-woke" energy, you probably shouldn't be posting it here.
- 3: No untagged game spoilers. If the game was recently released or not released at all yet, use the Spoiler tag (the little ⚠️ button) in the body text, and avoid typing spoilers in the title. It should also be avoided to openly talk about major story spoilers, even in old games.
God that would sound so dystopian and futuristic...but to be honest, most articles about AI today would sound like that back then. Damn people would freak out about privacy.
pretty sure they didn't.
BOINC came out 21 years ago, so it wouldn't be that unreasonable.
So, it's like folding@home, but instead of donating your spare compute to science, you sell it to generate porn?
So... this AI company gets gaming teens to "donate" their computing power, rather than pay for render farms / GPU clouds?
And then oblivious parents pay the power bills, effectively covering the computing costs of the AI porn company?
Sounds completely ethical to me /s.
No no, they're getting copies of digital images out of it. It's a totally fair trade!
I’ll be a minority voice considering the other comments. But maybe just pay for onlyfans or whatever you guys use. I’m a generally attractive woman (I can surmise from interactions while trying to date) and I really don’t like the idea that my likeness would be used for something like this. Get your jollies off, but try and be a bit consensual about it. Is that so much to ask?
It isn't too much to ask. According to Dr. K of HealthyGamerGG (Harvard Psychiatrist/Instructor), research shows that the release of non-consensual porn makes the unwilling subjects suicidal over half the time. Non-consensual porn = deepfakes, revenge porn, etc. It's seriously harmful, and there are other effects like depression, shame, PTSD, anxiety, and so on. There is functionally unlimited porn out there that is made with consent, and if someone doesn't want to be publicly sexually explicit then that's their choice.
I'm not against AI porn in general (I consider it the modern version of dirty drawings/cartoons), but when it comes to specific likenesses as with deepfakes then there's clear proof of harm and that's enough for me to oppose it. I don't believe there's some inherent right to see specific people naked against their will.
I think it would be too big of a privacy overreach to try to ban it outright as I think what people do on their own computers is their own business and there's no way to enforce a full ban without being incredibly intrusive, but as soon as it gets distributed in any way I think it should be prosecuted as heavily as real non consensual porn that was taken against someone's will.
I think the key is a lot of people don't want to pay for porn. And in the case of deep fakes, it's stuff they literally cannot pay money to get.
Ai porn isn't deepfake porn. The default is just a random ai generated face and body. Unless you want to it's difficult to deepfake someone.
So I’m not disagreeing with you, but you’re assuming they’re making deepfake images, and the article doesn’t specify that. In fact I’d bet that it’s just AI generated “people” that don’t exist.
What about AI porn of a person that doesn’t exist?
However, one of Salad's clients is CivitAi, a platform for sharing AI generated images which has previously been investigated by 404 media. It found that the service hosts image generating AI models of specific people, whose image can then be combined with pornographic AI models to generate non-consensual sexual images.
I know someone who’s into really dark romance stuff, like really hardcore stuff, but she’d never do some of this due to safety reasons. I can totally see her generating scenes of herself in those situations.
This feels exploitative AF on multiple levels.
I remember when GPUs were used to fold proteins...
I wore an onion on my belt
As was the fashion at the time
If I'm reading this right, it's a program that users sign up for to donate their processing power (and can opt in or out of adult content), which is then used by client companies to generate their own users' content? It even says that Salad can't view or moderate the images, so what exactly are they doing wrong besides providing service to potentially questionable companies? It makes as much sense as blaming Nvidia or Microsoft, am I missing something?
Based on the rewards, I'm assuming it's being done by very young people. Presumably the value of rewards is really low, but these kids haven't done the cost-benefit analysis. If I had to guess, for the vast majority it costs more in electricity than they get back, but the parents don't know it's happening.
This could be totally wrong. I haven't looked into it. This is how most of these things work though. They prey on the youth and their desire for these products to take advantage of them.
so what exactly are they doing wrong besides providing service to potentially questionable companies?
Well I think that is the main point of what is wrong. I think the big question is whether the mature content toggle is on by default or not. The company says it's off, but some users said otherwise. Dunno why the author didn't install it and check.
They said they did.
However, by default the software settings opt users into generating adult content. An option exists to "configure workload types manually" which enables users to uncheck the "Adult Content Workloads" option (via 404 media), however this is easily missed in the setup process, which I duly tested for myself to confirm.
Honestly, and I'm not saying I support what's being done here, the way I see it if you're tech savvy enough to be interested in using a program like this you should be looking through all of the options properly anyway. If users don't care what they're doing and are only interested in the rewards that's kind of on them.
I just think the article is focused on the wrong company, Salad is selling a tool that is being potentially misused by users of their client's service. I can certainly see why that can be a problem, but based on the information given in the article I don't think it's really theirs. If that's ALL Salad's used for then that's a different story.
I kinda fail to see the problem. The GPU owner doesn't see what workload they are processing. The pr0n company is willing to pay for GPU power. The GPU owner wants to earn money with his hardware. There's a demand, there's an offer, nobody is getting hurt (ai pr0n is not illegal, at least for now) so let people what they want to do
The problem is that they are clearly targeting minors who don't pay their own electricity bill, and dont even neccessarily have awareness that they are paying for their fortnite skins with their parents money. Also: there is a good chance that the generated pictures are at some point present on in the filesystem of the generating computer, and that alone is a giant can of worms that can even lead to legal troubles, if the person lives in a country where some or all kinds of pronography are illegal.
This is a shitty grift, abusing people who don't understand the consequences of the software.
wow. Imagine burning out your expensive GPU for a fortnite skin.
Great. Now we're trading pre-made traditional artwork to kids in exchange for fresh robot porn!
You would think they would do this to mine Bitcoin too.
What? Seems like porn generation is the new crypto mining.
I'd rather have a wealth of new porn around rather than thousands random Blockchains going around.
At least the porn will probably be useful for someone long term haha
Boring Dystopia