this post was submitted on 20 Jul 2023
1832 points (98.3% liked)

Technology

59577 readers
3239 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] HelloHotel@lemmy.world 7 points 1 year ago (1 children)

CSAM, as defined by apple, SPOILER that could be anything, including, and I could rattle off names, anything that threatens the government or those who got their tendrils into it, if we, For example have authoritarians change us to be facist, or re-introduce slavery or segrogation. A mere picture of your bedroom or face could have a somthing in it that allows you to be put into a cohort for later use (legal or not)

[–] Whirlybird@aussie.zone 4 points 1 year ago* (last edited 1 year ago) (1 children)

No, that's not at all what it was defined as or what it could be. CSAM is Child Sex Abuse Material. It wasn't going to be memes of winny the pooh like people argued.

That's also not how CSAM matching works. It simply compares hashes of images. If you take a photo of you in your bedroom with a sign saying "fuck the government" it will not match any CSAM database hashes no matter how authoritarian or fascist the government is, because they don't have that same photo in their CSAM databases.

You're doing what the outraged did back then and thinking CSAM scanning is some sort of AI powered image recognition that scans images for specific things. It's not that at all. It is a database of known CSAM images that have been hashed and that have been confirmed by multiple different governments (multiple different ones so one government can't just put an image of their president that they don't like in theirs and then find out who has uploaded that photo. If it only appears in one government CSAM database it will not be checked). It takes your photo, hashes it, and then checks to see if that hash is in the CSAM database. It won't be, ever.

You know what will be in there and matched? If you download child porn that is already out there on the web.

[–] Archer@lemmy.world 6 points 1 year ago

You're naive if you think that is all it will ever be, and that there will never be scope creep, especially malicious scope creep that turns into overreach