Nonconsensual AI porn is hated on the left and right. Can Congress act on it?

jeffw@lemmy.world to politics @lemmy.world – 88 points –
politico.com
44

You are viewing a single comment

Can AIs really consent, though?

The issue is not with all forms of pornographic AI, but more about deepfakes and nudifying apps that create nonconsensual pornography of real people. It is those people's consent that is being violated.

No-one cares if you consent to being drawn. The problem here isn't the consent of the depicted person, it's that the viewer is being misled. That's why the moral quandary goes away entirely if the AI porn is clearly labeled as such.

I dont think that this is really true, I strongly suspect that most people I know would consider someone drawing porn of them without consent a majorly icky thing to do, and would probably consider someone doing that to someone else to be a creep for doing so. The reason such drawings are less an issue is at least partly that the barrier to entry is lower with AI, since it takes a certain amount of skill and time investment to draw something like that such as to be clearly recognizable as any specific real person.

I still don't understand why this is now an issue but decades of photo editing did not bother anyone at all.

I mean, it did bother people, it just took more skill and time at using photo manipulation software to make it look convincing such that it was rare for someone to both have the expertise and be willing to put in the time, so it didnt come up often enough to be a point of discussion. The AI just makes it quick and easy enough to become more common.

Regular editing is much easier and quicker than installing, configuring and using stable diffusion. People acting like "AI" is a 1 click solution that gets you convincing looking images have probably never used it.

It literally is a one-click solution. People are running nudifying sites that use CLIP, GroundingDINO, SegmentAnything, and Stable Diffusion to autonomously nudify people's pictures.

These sites (which I won't even mention the names of), just ask for a decent quality photo of a woman wearing a crop top or bikini for best results.

The people who have the know-how to set up Stable Diffusion and all these other AI photomanipulation tools are using those skills to monetize sexual exploitation services. They're making it so you don't need to know what you're doing to participate.

And sites like Instagram, which are filled with millions of exploitable images of women and girls, has allowed these perverted services to advertise their warez to their users.

It is now many orders of magnitude easier than it ever has been in history to sexually exploit people's photographs. That's a big deal.

If you wanna pay for that then you do you. lol But at that point you could've also paid a shady artist to do the work for you too.

Also, maybe don't pose half naked on the internet already if you don't want people to see you in a sexual way. That's just weird, just like this whole IG attention whoring of people nowadays. And no, this isn't even just a women thing. Just look how thirsty women get under the images of good looking dudes that pose topless, or just your ordinary celeb doing ordinary things (Pedro Pascal = daddy, and yes, that includes more explicit comments too).

This hypocritical fake outrage is just embarrassing.

2 more...
2 more...
2 more...
2 more...
2 more...
2 more...
2 more...