this post was submitted on 17 Sep 2023
144 points (100.0% liked)

Chat

7499 readers
3 users here now

Relaxed section for discussion and debate that doesn't fit anywhere else. Whether it's advice, how your week is going, a link that's at the back of your mind, or something like that, it can likely go here.


Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

Several months ago Beehaw received a report about CSAM (i.e. Child Sexual Abuse Material). As an admin, I had to investigate this in order to verify and take the next steps. This was the first time in my life that I had ever seen images such as these. Not to go into great detail, but the images were of a very young child performing sexual acts with an adult.

The explicit nature of these images, the gut-wrenching shock and horror, the disgust and helplessness were very overwhelming to me. Those images are burnt into my mind and I would love to get rid of them but I don’t know how or if it is possible. Maybe time will take them out of my mind.

In my strong opinion, Beehaw must seek a platform where NO ONE will ever have to see these types of images. A software platform that makes it nearly impossible for Beehaw to host, in any way, CSAM.

If the other admins want to give their opinions about this, then I am all ears.

I, simply, cannot move forward with the Beehaw project unless this is one of our top priorities when choosing where we are going to go.

(page 2) 48 comments
sorted by: hot top controversial new old
[–] Rentlar@beehaw.org 5 points 1 year ago

I am very sorry you had to go through such a terrible experience.

It is my sincerest hope that you will be able to find a workable solution to this problem, from Lemmy or elsewhere.

I am (and have been) okay with admins taking any action necessary to accomplish the goals of the Beehaw project. So removing image hosting, implementing lemmy-safety, restricting federation severely, do whatever you need.

And please, also do whatever you need to care for yourself, including if it means needing to take a break from the site.

[–] OnichiCub@beehaw.org 3 points 1 year ago

I'm so sorry you went through this.

[–] Griseowulfin@beehaw.org 2 points 1 year ago

I'm sorry you had to handle that. Though, this is an inevitable problem for a site hosting user-generated content of any sort. I think not hosting your own content is the biggest first step. It might be better to depend on external image/filesharing sites like how things used to work with imgur, photobucket, flickr, etc. (is photobucket around anymore, lol?). This does pose a risk for link rot and what not, but I think given the scale of this operation, there's no reason to having to be scanning your own servers for illegal content any more than you have to for basic moderation. I'm not sure how this issue works with federation (if another lemmy instance hosted a bad image/thumbnail/message, does that get copied over to beehaw?), but if your risk tolerance doesn't want to deal with it, a non-federated option would give you more control.

In general user generated content is always going to pose a risk to the website hosting it. It's a matter of good risk management, from prevention and mitigation to an effective response, that will best serve both the administration and the users in terms of ensuring a safe service and minimizing legal risk.

[–] retronautickz@beehaw.org 1 points 1 year ago (1 children)

Aren't there ways to automatically scan the content of posts to detect csam and other disturbed content?

Because completely disallowing images isn't the solution. For once, people can still share those things through links. And secondly, there are people who need images and the like to communicate.

[–] blackn1ght@feddit.uk 3 points 1 year ago (1 children)

It could be possible to just disable uploading of images on the platform and rely on third parties (such as Imgur etc) who have the tools and experience to tackle this. It would at least solve the issue of images from other federated instances putting on an instance and it's not the sole responsibility of the admin to sort it.

[–] nosurprises@lemm.ee 2 points 1 year ago

From my understanding, beehaw still can get these images from other instances, unless they've already disabled that feature like lemm.ee did. Anyway, I think this should be a common effort, because it's not easy to solve and probably costs a lot of money. Fediverse should be free from CSAM.

load more comments
view more: ‹ prev next ›