A nightmare scenario previously only imagined by AI researchers, where AI image generators accidentally spit out non-consensual pornography of real people, is now reality.

  • bioemerl@kbin.social
    link
    fedilink
    arrow-up
    5
    ·
    1 year ago

    I think the idea is that if the model doesn’t know what Jennifer Lopez is, it couldn’t make imitations of her naked.

    Realistically that ship has sailed and AI is capable enough now that even if the data wasn’t there it could be pretty easily added.

    It will need to become a simple fact of life. If we can imagine something now, we can have pictures of it. There is no putting this back in the bottle.

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      1 year ago

      “Knowing what Jennifer Lopez looks like” is a very distinct thing from “reproducing an exact replica” of training data. OP appears to be arguing that the former is not true because he thinks the latter is true, but it’s actually the opposite. That’s the crux of what I’m arguing here, OP is simply factually wrong about his position.

      Edit: OP has pointed out that he doesn’t actually think there are exact replicas being produced, which just makes this even more confusing.

      • bioemerl@kbin.social
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        1 year ago

        OP has pointed out that he doesn’t actually think there are exact replicas being produced, which just makes this even more confusing.

        Your misread their first comment, I think.

        They were saying that DESPITE the common arguments that AI only learns and doesn’t copy exactly it might still be good to require consent for people’s content to be in training data.