• not_that_guy05@lemmy.world
    link
    fedilink
    arrow-up
    33
    arrow-down
    1
    ·
    11 months ago

    Fuck that guy first of all.

    What makes me think is, what about all that cartoon porn showing cartoon kids? What about hentai showing younger kids? What’s the difference if all are fake and being distributed online as well?

    Not defending him.

    • jeffw@lemmy.worldOPM
      link
      fedilink
      arrow-up
      22
      arrow-down
      1
      ·
      11 months ago

      I think there’s certainly an argument here. What if the hentai was more lifelike? What if the AI stuff was less realistic? Where’s the line?

      At least in the US, courts have been pretty shitty at defining things like “obscenity”. This AI stuff might force them to delineate more clearly.

  • prettydarknwild@lemmy.world
    link
    fedilink
    arrow-up
    14
    ·
    edit-2
    11 months ago

    oh man, i love the future, we havent solved world hunger, or reduce carbon emissions to 0, and we are on the brink of a world war, but now we have AI’s that can generate CSAM and fake footage on the fly 💀

    • Dasus@lemmy.world
      link
      fedilink
      arrow-up
      19
      arrow-down
      1
      ·
      11 months ago

      Technically we’ve solved world hunger. We’ve just not fixed it, as the greedy fucks who hoard most of the resources of this world don’t see immediate capital gains from just helping people.

      Pretty much the only real problem is billionaires being in control.

      • ArchRecord@lemm.ee
        link
        fedilink
        arrow-up
        3
        ·
        11 months ago

        True that. We have the means to fix so many problems, we just have a very very very small few that reeeeally don’t like to do anything good with their money, and instead choose to hoard it, at the expense of everyone else.

        • myliltoehurts@lemm.ee
          link
          fedilink
          arrow-up
          2
          ·
          11 months ago

          Oh cmon they don’t hoard the money. They use it to pay each other/politicians to make sure the status quo remains.

    • TheObviousSolution@lemm.ee
      link
      fedilink
      arrow-up
      3
      ·
      11 months ago

      Honestly not as bad as I would have thought it would be by now with fake propaganda videos, but the quality isn’t there yet I suppose.

  • eating3645@lemmy.world
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    1
    ·
    11 months ago

    I find it interesting that the relabeling of CP to CSAM weakens their argument here. “CP generated by AI is still CP” makes sense, but if there’s no abusee, it’s just CSM. Makes me wonder if they would have not rebranded if they knew about the proliferation of AI pornography.

    • Stovetop@lemmy.world
      link
      fedilink
      arrow-up
      16
      arrow-down
      3
      ·
      11 months ago

      The problem is that it abets the distribution of legitimate CSAM more easily. If a government declares “these types of images are okay if they’re fake”, you’ve given probable deniability to real CSAM distributors who can now claim that the material is AI generated, placing the burden on the legal system to prove it to the contrary. The end result will be a lot of real material flying under the radar because of weak evidence, and continued abuse of children.

      Better to just blanket ban the entire concept and save us all the trouble, in my opinion. Back before it was so easy to generate photorealistic images, it was easier to overlook victimless CP because illustrations are easy to tell apart from reality, but times have changed, and so should the laws.

      • kromem@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        11 months ago

        Not necessarily. There’s been a lot of advances in watermarking AI outputs.

        As well, there’s the opposite argument.

        Right now, pedophile rings have very high price points to access CSAM or require users to upload original CSAM content, adding a significant motivator to actually harm children.

        The same way rule 34 artists were very upset with AI being able to create what they were getting commissions to create, AI generated CSAM would be a significant dilution of the market.

        Is the average user really going to risk prison, pay a huge amount of money or harm a child with an even greater prison risk when effectively identical material is available for free?

        Pretty much overnight the CSAM dark markets would lose the vast majority of their market value and the only remaining offerings would be ones that could demonstrate they weren’t artificial to justify the higher price point, which would undermine the notion of plausible deniability.

        Legalization of AI generated CSAM would decimate the existing CSAM markets.

        That said, the real question that needs to be answered from a social responsibility perspective is what the net effect of CSAM access by pedophiles has on their proclivity to offend. If there’s a negative effect then it’s an open and shut case that it should be legalized. If it’s a positive effect than we should probably keep it very much illegal, even if that continues to enable dark markets for the real thing.

        • HereToLurk@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          11 months ago

          Is the average user really going to risk prison, pay a huge amount of money or harm a child with an even greater prison risk when effectively identical material is available for free?

          Average users aren’t pedophiles and it would appear that yes they would considering he did exactly that. He had access to tools that generated the material for free, which he then used to entice boys.

        • solrize@lemmy.world
          link
          fedilink
          arrow-up
          2
          arrow-down
          1
          ·
          edit-2
          11 months ago

          Not necessarily. There’s been a lot of advances in watermarking AI outputs.

          That presumes that the image generation is being done by some corporation or government entity that adds the watermarks to AI outputs and doesn’t add them to non-AI outputs. I’m not thrilled that AI of this sort exists at all, but given that it does, I’d rather not have it controlled by such entities. We’re heading towards a world where we can all run that stuff on our own computers and control the watermarks ourselves. Is that good or bad? Probably bad, but having it under the exclusive control of megacorps has to be even worse.

      • Thorny_Insight@lemm.ee
        link
        fedilink
        arrow-up
        4
        arrow-down
        2
        ·
        11 months ago

        placing the burden on the legal system to prove it to the contrary.

        That’s how it should be. Everyone is innocent until proven otherwise.

        • Stovetop@lemmy.world
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          11 months ago

          Right, but what I am suggesting is that laws should be worded to criminalize any sexualized depiction of children, not just ones with a real victim. It is no longer as simple to prove a photograph or video is actual CSAM with a real victim, making it easier for real abuse to avoid detection.

          • Thorny_Insight@lemm.ee
            link
            fedilink
            arrow-up
            3
            arrow-down
            1
            ·
            11 months ago

            This same “think about the children” -argument is used when advocating for stuff such as banning encryption aswell which in it’s current form enables the easy spreading of such content AI generated or not. I do not agree with that. It’s a slippery slope despite the good intentions. We’re not criminalizing fictional depictions of violence either. I don’t see how this is any different. I don’t care what people are jerking off to as long as they’re not hurting anyone and I don’t think you should either. Banning it haven’t gotten rid of actual CSAM content and it sure wont work for AI generated stuff either. No one benefits from the police running after people creating/sharing fictional content.

            • Stovetop@lemmy.world
              link
              fedilink
              arrow-up
              1
              arrow-down
              1
              ·
              11 months ago

              I think you’re painting a false equivalency. This isn’t about surveillance or incitement or any other pre-crime hypotheticals, but simply adjusting what material is considered infringing in light of new developments which can prevent justice from being carried out on actual cases of abuse.

              How do you prove what is fictional versus what is real? Unless there is some way to determine with near 100% certainty that a given image or video is AI generated and not real, or even that an AI generated image wasn’t trained on real images of abuse, you invite scenarios where real images of abuse get passed off as “fictional content” and make it easier for predators to victimize more children.

    • CaptPretentious@lemmy.world
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      11 months ago

      Have to agree. Because I have no clue what CSAM is. My first glance at the title made me think it was CSPAN (the TV channel)… So CP is better identifier, as of at least recognize the initialism.

      If we could stop turning everything, and especially important things, into acronyms and initialisms that’d be great.

  • ocassionallyaduck@lemmy.world
    link
    fedilink
    arrow-up
    5
    arrow-down
    1
    ·
    11 months ago

    The cats out of the bag on this. It’s enforceable for now to try and ban it, maybe. Because the models are mostly online and intensive.

    In 2028 though, when you can train your own model and generate your own local images without burning a server farm? This has to happen for ML to keep growing and catch on.

    welp. Then there is infinite fake child porn. Because you cannot police every device and model.

    Because of how tech companies have handled this technology, this is not an if scenario. This is guaranteed now.

    • TheObviousSolution@lemm.ee
      link
      fedilink
      arrow-up
      1
      arrow-down
      2
      ·
      11 months ago

      I remember when they tried to do the same with CRISPR. Glad that didn’t take off and remained largely limited to the industry and academia. But then again, Wuhan …

    • wetsoggybread@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      11 months ago

      I read that its more accurate to say “child sexual abuse material” than child porn because it carries the message of just how bad the stuff is better than just calling it porn and it sounds more professional

      • stoly@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        11 months ago

        And I suppose it’s also saying that the form it’s in doesn’t matter. Any type of material is the same.

  • over_clox@lemmy.world
    link
    fedilink
    arrow-up
    6
    arrow-down
    28
    ·
    11 months ago

    Then we should be able to charge AI (the developers moreso) for the same disgusting crime, and shut AI down.

    • jeffw@lemmy.worldOPM
      link
      fedilink
      arrow-up
      13
      arrow-down
      1
      ·
      11 months ago

      I think that’s a bit of a stretch. If it was being marketed as “make your fantasy, no matter how illegal it is,” then yeah. But just because I use a tool someone else made doesn’t mean they should be held liable.

      • over_clox@lemmy.world
        link
        fedilink
        arrow-up
        1
        arrow-down
        6
        ·
        11 months ago

        Check my other comments. My thought was compared to a hammer.

        Hammers aren’t trained to act or respond on their own from millions of user inputs.

              • over_clox@lemmy.world
                link
                fedilink
                arrow-up
                1
                arrow-down
                2
                ·
                11 months ago

                AI hasn’t exactly kicked out a Picasso with a naked young girl missing an ear yet has it?

                I sure hope not!

                But if it can, then that seriously indicates it must have some bad training data in the system…

                I won’t be testing these hypotheses.

      • over_clox@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        11 months ago

        Totally dismissing inappropriate usage, AI can be funny and entertaining, but on the flip side it’s also taking people’s jobs.

        It shouldn’t take a book, let alone 3 seconds of common sense thought, to realize that.