this post was submitted on 26 Aug 2024
366 points (97.2% liked)

News

23332 readers
4808 users here now

Welcome to the News community!

Rules:

1. Be civil


Attack the argument, not the person. No racism/sexism/bigotry. Good faith argumentation only. This includes accusing another user of being a bot or paid actor. Trolling is uncivil and is grounds for removal and/or a community ban. Do not respond to rule-breaking content; report it and move on.


2. All posts should contain a source (url) that is as reliable and unbiased as possible and must only contain one link.


Obvious right or left wing sources will be removed at the mods discretion. We have an actively updated blocklist, which you can see here: https://lemmy.world/post/2246130 if you feel like any website is missing, contact the mods. Supporting links can be added in comments or posted seperately but not to the post body.


3. No bots, spam or self-promotion.


Only approved bots, which follow the guidelines for bots set by the instance, are allowed.


4. Post titles should be the same as the article used as source.


Posts which titles don’t match the source won’t be removed, but the autoMod will notify you, and if your title misrepresents the original article, the post will be deleted. If the site changed their headline, the bot might still contact you, just ignore it, we won’t delete your post.


5. Only recent news is allowed.


Posts must be news from the most recent 30 days.


6. All posts must be news articles.


No opinion pieces, Listicles, editorials or celebrity gossip is allowed. All posts will be judged on a case-by-case basis.


7. No duplicate posts.


If a source you used was already posted by someone else, the autoMod will leave a message. Please remove your post if the autoMod is correct. If the post that matches your post is very old, we refer you to rule 5.


8. Misinformation is prohibited.


Misinformation / propaganda is strictly prohibited. Any comment or post containing or linking to misinformation will be removed. If you feel that your post has been removed in error, credible sources must be provided.


9. No link shorteners.


The auto mod will contact you if a link shortener is detected, please delete your post if they are right.


10. Don't copy entire article in your post body


For copyright reasons, you are not allowed to copy an entire article into your post body. This is an instance wide rule, that is strictly enforced in this community.

founded 1 year ago
MODERATORS
 

A Florida man is facing 20 counts of obscenity for allegedly creating and distributing AI-generated child pornography, highlighting the danger and ubiquity of generative AI being used for nefarious reasons.

Phillip Michael McCorkle was arrested last week while he was working at a movie theater in Vero Beach, Florida, according to TV station CBS 12 News. A crew from the TV station captured the arrest, which made for dramatic video footage due to law enforcement leading away the uniform-wearing McCorkle from the theater in handcuffs.

you are viewing a single comment's thread
view the rest of the comments
[–] ReallyActuallyFrankenstein@lemmynsfw.com 107 points 2 months ago* (last edited 2 months ago) (32 children)

It's hard to have a nuanced discussion because the article is so vague. It's not clear what he's specifically been charged with (beyond "obscenity," not a specific child abuse statute?). Because any simulated CSAM laws have been, to my knowledge, all struck down when challenged.

I completely get the "lock them all up and throw away the key" visceral reaction - I feel that too, for sure - but this is a much more difficult question. There are porn actors over 18 who look younger, do the laws outlaw them from work that would be legal for others who just look older? If AI was trained exclusively on those over-18 people, would outputs then not be CSAM even if the images produced features that looked under 18?

I'm at least all for a "fruit of the poisoned tree" theory - if AI model training data sets include actual CSAM then they can and should be made illegal. Deepfaking intentionally real under 18 people is also not black and white (looking again to the harm factor), but also I think it can be justifiably prohibited. I also think distribution of completely fake CSAM can be arguably outlawed (the situation here), since it's going to be impossible to tell AI from real imagery soon and allowing that would undermine enforcement of vital anti-real-CSAM laws.

The real hard case is producing and retaining fully fake people and without real CSAM in training data, solely locally (possession crimes). That's really tough. Because not only does it not directly hurt anyone in its creation, there's a possible benefit in that it diminishes the market for real CSAM (potentially saving unrelated children from the abuse flowing from that demand), and could also divert the impulse of the producer from preying on children around them due to unfulfilled desire.

Could, because I don't think there's studies that answers whether those are true.

[–] mpa92643@lemmy.world 28 points 2 months ago (22 children)

I mostly agree with you, but a counterpoint:

Downloading and possession of CSAM seems to be a common first step in a person initiating communication with a minor with the intent to meet up and abuse them. I've read many articles over the years about men getting arrested for trying to meet up with minors, and one thing that shows up pretty often in these articles is the perpetrator admitting to downloading CSAM for years until deciding the fantasy wasn't enough anymore. They become comfortable enough with it that it loses its taboo and they feel emboldened to take the next step.

CSAM possession is illegal because possession directly supports creation, and creation is inherently abusive and exploitative of real people, and generating it from a model that was trained on non-abusive content probably isn't exploitative, but there's a legitimate question as to whether we as a society decide it's associated closely enough with real world harms that it should be banned.

Not an easy question for sure, and it's one that deserves to be answered using empirical data, but I imagine the vast majority of Americans would flatly reject a nuanced view on this issue.

[–] HelixDab2@lemm.ee 17 points 2 months ago (7 children)

CSAM possession is illegal because possession directly supports creation

To expound on this: prior to this point, the creation of CSAM requires that children be sexually exploited. You could not have CSAM without children being harmed. But what about when no direct harms have occurred? Is lolicon hentai 'obscene'? Well, according to the law and case law, yes, but it's not usually enforced. If we agree that drawings of children engaged in sexual acts aren't causing direct harm--that is, children are not being sexually abused in order to create the drawings--then how much different is a computer-generated image that isn't based off any specific person or event? It seem to me that, whether or not a pedophile might decide that they eventually want more than LLM-generated images is not relevant. Treating a future possibility as a foregone conclusion is exactly the rationale behind Reefer Madness and the idea of 'gateway' drugs.

Allow me to float a second possibility that will certainly be less popular.

Start with two premises: first, pedophilia is a characteristic that appears to be an orientation. That is, a true pedophile--a person exclusively sexually attracted to pre-pubescent children--does not choose to be a pedophile, any more than a person chooses to be gay. (My understanding is that very few pedophiles are exclusively pedophilic though, and that many child molesters are opportunistic sexual predators rather than being pedophiles.) Secondly, the rates of sexual assault appear to have decreased as pornography availability has increased. So the question I would have is, would wide availability of LLM-generated CSAM--CSAM that didn't cause any real, direct harm to children--actually decrease rates of child sexual assault?

[–] RandomlyNice@lemmy.world 5 points 2 months ago (1 children)

With regards to your last paragraph: Pedophiles can indeed by straight, gay or bi. Pedophiles may also not become molesters, and molesters of children may not at all be pedophilic. It's seems you understand this. I mentioned ITT that I read a newspaper article many years ago that was commissioned to show the access to cp would increase child abuse, it seemed to show the opposite.
If persons could use AI to generate their own porn of their own personal fantasies (whatever those might be) and NOT share that content what then? Canada allows this for text (maybe certain visuals? Audio? IDK). I don't know about current 'obscene' laws in the USA, however, I do recall reading about an art exhibit in NY which featured an upside down urinal that was deemed obscene, than later deemed a work or art. I also recall seeing (via an internet image) a sculpture of what seemed to be a circle of children with penises as noses. Porn? Art? Comedy?

[–] HelixDab2@lemm.ee 6 points 2 months ago

My understanding was that 'pure' pedophiles--ones that have no interest at all in post-pubescent children or any adults whatsoever--tend to be less concerned with sex/gender, particularly because children don't have defined secondary sex characteristics. I don't know if this is actually correct though. I'm not even sure how you could ethically research that kind of thing and end up with valid results.

And honestly, not being able to do solid research that has valid results makes it really fuckin' hard to find solutions that work to prevent as many children from being harmed as possible. In the US at least research about sex and sexuality in general-much less deviant sexualities--seems to be taboo, and very difficult to get funding for.

load more comments (5 replies)
load more comments (19 replies)
load more comments (28 replies)