Each of these reads like an extremely horny and angry man yelling their basest desires at Pornhub’s search function.

  • diffuselight@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    5
    ·
    1 year ago

    I just retained an LLM on your comment you put on the public internet. You feel violated enough to equate it to physical violation?

    • GBU_28@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Why would I? Folks who have had real nudes of them posted on the Internet haven’t felt “physical violation” but they’ve certainly been violated.

      If you had photos of me and trained a porn generating LLM on my photos and shared porn of me, in an identifiable way, I would consider that violation.

      But simply taking my words in that simple sentence isn’t identifiable, unique, or revealing. So no.

      Further, the original point was about the ethics of AI porn. You can’t get something from nothing.

    • urist@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      7
      ·
      1 year ago

      How would you respond to photo realistic porn that looks like your mother, daughter, [insert person you care about here] especially if they found it distressing?

      How would you feel if it was posted on facebook? How would you feel if they had to deal with it at work? From coworkers? From clients?

      We are entering uncharted waters. You know why this is different than training a model on text, and your reply to @GBU_28@lemm.ee is hostile and doesn’t acknowledge why people would be upset about AI porn featuring their likeness.

      • diffuselight@lemmy.world
        link
        fedilink
        English
        arrow-up
        11
        ·
        edit-2
        1 year ago

        you are answering a question with a different question. LLMs don’t make pictures of your mom. And this particular question?. One that has roughly existed since Photoshop existed.

        It just gets easier every year. It was already easy. You could already pay someone 15 bucks on Fiver to do all of that, for years now.

        Nothing really new here.

        The technology is also easy. Matrix math. About as easy to ban as mp3 downloads. Never stopped anyone. It’s progress. You are a medieval knight asking to put gunpowder back into the box, but it’s clear it cannot be put back - it is already illegal to make non consensual imagery just as it is illegal to copy books. And yet printers exist and photocopiers exist.

        Let me be very clear - accepting the reality that the technology is out there, it’s basic, easy to replicate and on a million computers now is not disrespectful to victims of no consensual imagery.

        You may not want to hear it, but just like with encryption, the only other choice society has is full surveillance of every computer to prevent people from doing “bad things”. everything you complain about is already illegal and has already been possible - it just gets cheaper every year. What you want to have protection from is technological progress because society sucks at dealing with the consequences of it.

        To be perfectly blunt, you don’t need to train any generative AI model for powerful deepfakes. You can use technology like Roop and Controlnet to synthesize any face on any image from a singe photograph. Training not necessary.

        When you look at it that way, what point is there to try to legislate training with these arguments? None.

        • urist@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          5
          ·
          1 year ago

          I’m not making an argument to ban it. I’m just pointing out you’re pretending a model from text someone wrote is similar to a model that makes nonconcentual porn.

          I don’t think it can be banned, it’s just something they will need to encorperate into revenge porn laws, if it isn’t already covered.

          I’m just pointing out your comment sucked.

          • diffuselight@lemmy.world
            link
            fedilink
            English
            arrow-up
            7
            ·
            edit-2
            1 year ago

            It’s already covered under those laws. So what are you doing that’s different from ChatGPT hallucinating here ?

            Those laws don’t spell out the tools (photoshop); they hinge on reproducing likeness.

            • urist@lemmy.blahaj.zone
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              7
              ·
              1 year ago

              Oh good, someone who has read every revenge porn law, ever. I’m glad they work exactly the same, in every nation and state.

              Anyway, I must be hallucinating, true, because it seems you keep attacking what I’m saying, instead of defending the comment you made earlier that I took issue with, the one that points out you’re being needlessly hostile.

      • stevedidWHAT@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        1
        ·
        edit-2
        1 year ago

        I can do this right now with photoshop dude what are you talking about. This just points to the need for more revenge porn laws.

        We don’t have to sit in the fire when we can crawl out. Are we still on fire? Yeah. Can we do something about that? Yeah!

        It seems like so many people these days want perfect solutions but the reality is that sometimes we have to make incremental solutions to erase the problem as much as we can.

        • polymer@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          edit-2
          1 year ago

          And incidentally, this need for revenge porn laws is also a symptomatic issue with a separate cause. Technology always moved forward and with no relation to social advancement, where there is also no realistic “Genie being forced back in the bottle” scenario either.

          That being said, easier access to more powerful technology with lackluster recognition of personal responsibility doesn’t exactly bring happy prospects. lol…

      • Donjuanme@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        Revenge porn/blackmail/exploitation will hopefully become much less obscene, not to the “let’s not prosecute this” levels, but maybe people can stop living in fear of their lives being ruined by malicious actors (unless that’s your kink, you do you).

        It will take/drive/demand a massive cultural shift, but if you asked me which world I would rather live in, and the options are one where people are abused and exploited, or one where people can visualize their perversions more easily (but content creators have a harder time making a living) I’ll take the second. Though I may have straw-manned a bit, it’s not something I’ve thought of outside of this forum thread.

      • afraid_of_zombies@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        I wouldn’t be happy about it but me not being happy about something doesn’t mean I just get an override.

        I think the boat has sailed a bit on this one. You can’t really copyright your own image and even if you were some famous person who is willing to do this and fight the legal battles you still have to go up against the fact that no one is making money off of it. You might be able to get a news source to take down that picture of you but it is another thing to make it so the camera company can’t even record you.

        But hey I was saying for years that we need to change the laws forbidding photography of people and property without consent and everyone yelled at me that they have the right to use a telescoping lense to view whomever they wanted blocks away.

        The creeps have inherited the earth.