Stanford researchers find Mastodon has a massive child abuse material problem
theverge.com
Not a good look for Mastodon - what can be done to automate the removal of CSAM?
You are viewing a single comment
Not a good look for Mastodon - what can be done to automate the removal of CSAM?
No possible definition of child sexual abuse can include drawings, no matter who writes it. That's not what those words mean.
Drawings... aren't children. It is literally that simple.
And if you think any of this is identically illegal to actual photos of child abuse - one, there's a whole network of shamelessly public US-hosted sites for you to turn in and be a national hero, and two, you might be wholly incapable of remembering what you're arguing. Whether you think these things are equivalent oscillates between letters.
Step up the reading comprehension please :)
It's pretty funny having you state, re-state and re-re-state the exact same obvious things that everyone understands while not seeing that everyone gets that, that you're missing the point, and that you're yelling in a hole
Why do you keep going?
Are you that painfully unconvincing in real life like cmon
Step it up
You forgot to say anything.