• 30p87@feddit.de
      link
      fedilink
      arrow-up
      3
      ·
      1 year ago

      How would one realize CSAM protection? You’d need actual ML to check for it, and I do not think there are trained models available. And now find someone that wants to train such a model, somehow. Also, running an ML model would be quite expensive in energy and hardware.

      • NightAuthor@beehaw.org
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        There are models for detecting adult material, idk how well they’d work on CSAM though. Additionally, there exists a hash identification system for known images, idk if it’s available to the public, but I know apple has it.

        Idk, but we gotta figure out something