Divisions by zero

16,956 readers
591 users here now

Communities about Anarchism, Generative AI, Copylefts, Neurodivergence, Filesharing, and Free Software.

Follow the Anarchist Code of Conduct and honor the Disengage Rule.

Don't be shitty to each other. Keep it SFW. Obey the spirit of The Golden Rules. Fuck around and find out.


👶 New to lemmy? Start here and here


Please help cover server costs

Ko-Fi Liberapay
Ko-fi Liberapay

This instance provides a wiki service which only users of this lemmy can use. If you want a wiki for your community or your account or whatever, feel free to use it


This instance currently allows new community creation, however the following subjects are explicitly not allowed as communities.

Preferably you'll stay within the topics endorsed by this instance (see first line)

Note that you are expected to attempt to create an active community and not just squat on a name. Inactive communities will be purged after receiving a warning.


Find us on Matrix and regale us with your tales of adventure!


When going to other communities, respect their rules AND our rules when they are more restrictive. Do not give cause for others to de-federate our instance please.


Alternative Frontends

founded 2 years ago
ADMINS
1
 
 

tværpostet fra: https://lemmy.dbzer0.com/post/3076577

I posted the other day that you can clean up your object storage from CSAM using my AI-based tool. Many people expressed the wish to use it on their local file storage-based pict-rs. So I've just extended its functionality to allow exactly that.

The new lemmy_safety_local_storage.py will go through your pict-rs volume in the filesystem and scan each image for CSAM, and delete it. The requirements are

  • A linux account with read-write access to the volume files
  • A private key authentication for that account

As my main instance is using object storage, my testing is limited to my dev instance, and there it all looks OK to me. But do run it with --dry_run if you're worried. You can delete lemmy_safety.db and rerun to enforce the delete after (method to utilize the --dry_run results coming soon)

PS: if you were using the object storage cleanup, that script has been renamed to lemmy_safety_object_storage.py

2
 
 

I posted the other day that you can clean up your object storage from CSAM using my AI-based tool. Many people expressed the wish to use it on their local file storage-based pict-rs. So I've just extended its functionality to allow exactly that.

The new lemmy_safety_local_storage.py will go through your pict-rs volume in the filesystem and scan each image for CSAM, and delete it. The requirements are

  • A linux account with read-write access to the volume files
  • A private key authentication for that account

As my main instance is using object storage, my testing is limited to my dev instance, and there it all looks OK to me. But do run it with --dry_run if you're worried. You can delete lemmy_safety.db and rerun to enforce the delete after (method to utilize the --dry_run results coming soon)

PS: if you were using the object storage cleanup, that script has been renamed to lemmy_safety_object_storage.py

3
 
 

cross-posted from: https://lemmy.dbzer0.com/post/2896209

I noticed a bit of panic around here lately and as I have had to continuously fight against pedos for the past year, I have developed tools to help me detect and prevent this content.

As luck would have it, we recently published one of our anti-csam checker tool as a python library that anyone can use. So I thought I could use this to help lemmy admins feel a bit more safe.

The tool can either go through all your images via your object storage and delete all CSAM, or it canrun continuously and scan and delete all new images as well. Suggested option is to run it using --all once, and then run it as a daemon and leave it running.

Better options would be to be able to retrieve exact images uploaded via lemmy/pict-rs api but we're not there quite yet.

Let me know if you have any issue or improvements.

4
 
 

I noticed a bit of panic around here lately and as I have had to continuously fight against pedos for the past year, I have developed tools to help me detect and prevent this content.

As luck would have it, we recently published one of our anti-csam checker tool as a python library that anyone can use. So I thought I could use this to help lemmy admins feel a bit more safe.

The tool can either go through all your images via your object storage and delete all CSAM, or it canrun continuously and scan and delete all new images as well. Suggested option is to run it using --all once, and then run it as a daemon and leave it running.

Better options would be to be able to retrieve exact images uploaded via lemmy/pict-rs api but we're not there quite yet.

Let me know if you have any issue or improvements.

EDIT: Just to clarify, you should run this on your desktop PC with a GPU, not on your lemmy server!

view more: next ›