• CptBread@lemmy.world
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    2
    ·
    3 days ago

    There is also the angle of generated CSAM looking real adding difficulty in prosecuting real CSAM producers.

    • PM_Your_Nudes_Please@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      24 hours ago

      This is actually why I’d be in favor of AI generators creating a hash database of their generated images. If legalized, they should be required to maintain records of the images they have produced. So that if those images appear elsewhere, they can be verified as AI generated.

      It would be a monumental effort to actually get the AI companies to agree to it willingly. But that’s why legislation exists.

    • MagicShel@lemmy.zip
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      1
      ·
      2 days ago

      This, above any other reason, is why I’m most troubled with AI CSAM. I don’t care what anyone gets off to if no one is harmed, but the fact that real CSAM could be created and be indistinguishable from AI created, is a real harm.

      And I instinctively ask, who would bother producing it for real when AI is cheap and harmless? But people produce it for reasons other than money and there are places in this world where a child’s life is probably less valuable than the electricity used to create images.

      I fundamentally think AI should be completely uncensored. Because I think censorship limits and harms uses for it that might otherwise be good. I think if 12 year old me could’ve had an AI show me where the clitoris is on a girl or what the fuck a hymen looks like, or answer questions about my own body, I think I would’ve had a lot less confusion and uncertainty in my burgeoning sexuality. Maybe I’d have had less curiosity about what my classmates looked like under their clothes, leading to questionable decisions on my part.

      I can find a million arguments why AI shouldn’t be censored. Like, do you know ChatGPT can be convinced to describe vaginal and oral sex in a romantic fiction is fine, but if it’s anal sex, it has a much higher refusal rate? Is that subtle anti-gay encoding in the training data? It also struggles with polyamory when it’s two men and a woman but less when it’s two women and a man. What’s the long-term impact when these biases are built into everyday tools? These are concerns I consider all the time.

      But at the end of the day, the idea that there are children out there being abused and consumed and no one will even look for them because “it’s probably just AI” isn’t something I can bear no matter how firm my convictions are about uncensored AI. It’s something I struggle to reconcile.

      • Ænima@lemm.ee
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 day ago

        Maybe the weird, extra human finger and appendage issues in AI images are a feature, not bugs. Maybe it’s a naturally occurring, unintended consequences of their learning and feedback process to sabotage the output they generate in order to make it obvious the image is fake.

        /s (sort of)

    • Eggyhead@fedia.io
      link
      fedilink
      arrow-up
      3
      ·
      2 days ago

      I’d say that’s more of an AI industry issue than anything else. All AI art needs to be easily identifiable and sourced as such, but I doubt AI producers will want to hide tags on all their AI generated work though.