• bread@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    42 minutes ago

    What a reprehensible, disingenuous representation of what he actually said. I’m not a fan of the guy, but PC Gamer is trash as well. Scary to see how people here are reacting just because it’s about X and AI.

    • popcar2@piefed.ca
      link
      fedilink
      English
      arrow-up
      2
      ·
      17 minutes ago

      Yeah nobody in this thread went past the title, but that’s literally not what he said.

      He actually said that demanding X remove AI features is gatekeeping since competitors get to keep them, which is still a dumb take but very very far from “Tim Sweeny loves child porn”…

  • SoloCritical@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    41 minutes ago

    Man that title gives me a stroke trying to decipher it… it almost reads like Tim Sweeney wants Twitter banned but clearly that’s not the case…

  • Jankatarch@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    44 minutes ago

    Did he take an oath against common sense? Is he bound by a curse to have bad takes for his entire life? Does he ragebait as a living? What the actual fuck is up with this man?

    • D_C@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      1
      ·
      54 minutes ago

      I think it’s more a:
      “Guy who enjoys making child porn on xhitter gets angry when decent people want to ban it.”
      type situation.

      If I see someone arguing for something then that’s because they want it. Or want to use it.

  • WatDabney@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    102
    arrow-down
    1
    ·
    8 hours ago

    If you can be effectively censored by the banning of a site flooded with CSAM, that’s very much your problem and nobody else’s.

    • mindbleach@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      38
      ·
      8 hours ago

      Nothing made-up is CSAM. That is the entire point of the term “CSAM.”

      It’s like calling a horror movie murder.

      • ruuster13@lemmy.zip
        link
        fedilink
        English
        arrow-up
        10
        arrow-down
        2
        ·
        5 hours ago

        The Rape, Abuse, & Incest National Network (RAINN) defines child sexual abuse material (CSAM) as “evidence of child sexual abuse” that "includes both real and synthetic content

        Were you too busy fapping to read the article?

      • ryper@lemmy.ca
        link
        fedilink
        English
        arrow-up
        18
        arrow-down
        1
        ·
        edit-2
        7 hours ago

        It’s too hard to tell real CSAM from AI-generated CSAM. Safest to treat it all as CSAM.

        • greenskye@lemmy.zip
          link
          fedilink
          English
          arrow-up
          7
          ·
          5 hours ago

          I get this and I don’t disagree, but I also hate that AI fully brought back thought crimes as a thing.

          I don’t have a better approach or idea, but I really don’t like that simply drawing a certain arrangement of lines and colors is now a crime. I’ve also seen a lot of positive sentiment at applying this to other forms of porn as well, ones less universally hated.

          Not supporting this use case at all and on balance I think this is the best option we have, but I do think thought crimes as a concept are just as concerning, especially given the current political climate.

          • shani66@ani.social
            link
            fedilink
            English
            arrow-up
            3
            ·
            4 hours ago

            Sure, i think it’s weird to really care about loli or furry or any other niche, but ai generating material of actual children (and unwilling people besides) is actually harmful. If they can’t have effective safeguards against that harm it makes sense to restrict it legally.

        • mindbleach@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          15
          ·
          7 hours ago

          You can insist every frame of Bart Simspon’s dick in The Simpsons Movie should be as illegal as photographic evidence of child rape, but that does not make them the same thing. The entire point of the term CSAM is that it’s the actual real evidence of child rape. It is nonsensical to use the term for any other purpose.

          • deranger@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            10
            arrow-down
            2
            ·
            edit-2
            6 hours ago

            The *entire point* of the term CSAM is that it’s the actual real evidence of child rape.

            You are completely wrong.

            https://rainn.org/get-the-facts-about-csam-child-sexual-abuse-material/what-is-csam/

            “CSAM (“see-sam”) refers to any visual content—photos, videos, livestreams, or AI-generated images—that shows a child being sexually abused or exploited.”

            “Any content that sexualizes or exploits a child for the viewer’s benefit” <- AI goes here.

            • mindbleach@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              9
              ·
              edit-2
              6 hours ago

              RAINN has completely lost the plot by conflating the explicit term for Literal Photographic Evidence Of An Event Where A Child Was Raped with made-up bullshit.

              We will inevitably develop some other term like LPEOAEWACWR, and confused idiots will inevitably misuse that to refer to drawings, and it will be the exact same shit I’m complaining about right now.

              • deranger@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                3
                ·
                6 hours ago

                Dude, you’re the only one who uses that strict definition. Go nuts with your course of prescriptivism but I’m pretty sure it’s a lost cause.

          • VeganBtw@piefed.social
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            2
            ·
            edit-2
            6 hours ago

            Child pornography (CP), also known as child sexual abuse material (CSAM) and by more informal terms such as kiddie porn, is erotic material that involves or depicts persons under the designated age of majority.
            […]
            Laws regarding child pornography generally include sexual images involving prepubescents, pubescent, or post-pubescent minors and computer-generated images that appear to involve them.
            (Emphasis mine)

            https://en.wikipedia.org/wiki/Child_pornography

            • mindbleach@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              3
              ·
              6 hours ago

              ‘These several things are illegal, including the real thing and several made-up things.’

              Please stop misusing the term that explicitly refers to the the real thing.

              ‘No.’

      • baguettefish@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        4
        ·
        edit-2
        5 hours ago

        AI CSAM was generated from real CSAM

        AI being able to accurately undress kids is a real issue in multiple ways

          • rainwall@piefed.social
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            2
            ·
            edit-2
            4 hours ago

            It used real images of shrek and the moon to do that. It didnt “invent” or “imagine” either.

            The child porn it’s generating is based on literal child porn, if not itself just actual child porn.

            • mindbleach@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              4
              ·
              edit-2
              4 hours ago

              You think these billion-dollar companies keep hyper-illegal images around, just to train their hideously expensive models to do the things they do not want those models to do?

              Like combining unrelated concepts isn’t the whole fucking point?

              • mcv@lemmy.zip
                link
                fedilink
                English
                arrow-up
                2
                ·
                2 hours ago

                No, I think these billion dollar companies are incredibly sloppy about curating the content they steal to train their systems on.

              • stray@pawb.social
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                1
                ·
                3 hours ago

                It literally can’t combine unrelated concepts though. Not too long ago there was the issue where one (Dall-E?) couldn’t make a picture of a full glass of wine because every glass of wine it had been trained on was half full, because that’s generally how we prefer to photograph wine. It has no concept of “full” the way actual intelligences do, so it couldn’t connect the dots. It had to be trained on actual full glasses of wine to gain the ability to produce them itself.

    • andybytes@programming.dev
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      5 hours ago

      Maybe we are the only people that don’t f kids. Maybe this is "H’ “E” Double Hockey Sticks.

    • fonix232@fedia.io
      link
      fedilink
      arrow-up
      8
      arrow-down
      1
      ·
      8 hours ago

      It is when one side of the political palette is “against” it but keeps supporting people who think CSAM is a-okay, while the other side finds it abhorrent regardless who’s pushing it.

    • andybytes@programming.dev
      link
      fedilink
      English
      arrow-up
      1
      ·
      5 hours ago

      I mean the capitalist are the ones calling the shots since the imperial core is no democracy. This is their battle we are their dildos.

  • bearboiblake@pawb.social
    link
    fedilink
    English
    arrow-up
    30
    ·
    8 hours ago

    inb4 “In a stunning 5-4 decision, the Supreme Court has ruled that AI-generated CSAM is constitutionally protected speech”

    • mindbleach@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      18
      ·
      8 hours ago

      There is no such thing as generated CSAM, because the term exists specifically to distinguish anything made-up from photographic evidence of child rape. This term was already developed to stop people from lumping together Simpsons rule 34 with the kind of images you report to the FBI. Please do not make us choose yet another label, which you would also dilute.

        • mindbleach@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          17
          ·
          7 hours ago

          ‘If you care about child abuse please stop conflating it with cartoons.’

          ‘Pedo.’

          Fuck off.

      • deranger@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        11
        arrow-down
        3
        ·
        edit-2
        7 hours ago

        Generating images of a minor can certainly fulfill the definition of CSAM. It’s a child, It’s sexual, It’s abusive, It’s material. It’s CSAM dude.

        These are the images you report to the FBI. Your narrow definition is not the definition. We don’t need to make a separate term because it still impacts the minor even if it’s fake. I say this as a somewhat annoying prescriptivist pedant.

        • mindbleach@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          14
          ·
          7 hours ago

          There cannot be material from the sexual abuse of a child if that sexual abuse did not fucking happen. The term does not mean ‘shit what looks like it could be from the abuse of some child I guess.’ It means, state’s evidence of actual crimes.

          • Sas@piefed.blahaj.zone
            link
            fedilink
            English
            arrow-up
            8
            ·
            6 hours ago

            It is sexual abuse even by your definition if photos of real children get sexualised by AI and land on xitter. And afaik know that is what’s happened. These kids did not consent to have their likeness sexualised.

            • mindbleach@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              9
              ·
              edit-2
              6 hours ago

              Nothing done to your likeness is a thing that happened to you.

              Do you people not understand reality is different from fiction?

          • deranger@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            2
            ·
            6 hours ago

            CSAM is abusive material of a sexual nature of a child. Generated or real, both fit this definition.

              • deranger@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                6
                arrow-down
                1
                ·
                edit-2
                6 hours ago

                You’re the only one using that definition. There is no stipulation that it’s from something that happened.

                Where is your definition coming from?

                • mindbleach@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  5
                  ·
                  5 hours ago

                  My definition is from what words mean.

                  We need a term to specifically refer to actual photographs of actual child abuse. What the fuck are we supposed to call that, such that schmucks won’t use the same label to refer to drawings?

              • queermunist she/her@lemmy.ml
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                1
                ·
                6 hours ago

                How do you think a child would feel after having a pornographic image generated of them and then published on the internet?

                Looks like sexual abuse to me.

      • ruuster13@lemmy.zip
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        5 hours ago

        The Rape, Abuse, & Incest National Network (RAINN) defines child sexual abuse material (CSAM) as “evidence of child sexual abuse” that "includes both real and synthetic content