Meta tool to block nude images in teens’ private messages | …note, since this does not report outwards this is not client-side scanning, no more than a spellchecker would be

The test of Client-Side Scanning (CSS) is:

  • might it leak 1 or more bits of message or attachment content to a non-participant, and…
  • can it be disabled, and…
  • whom does it serve or offer intended benefit?

…the answers (for CSS) to which are: “yes”, “no”, and “any non-participant”, respectively.

Something mandatory which is scanning for & reporting CSAM on behalf of “society” is CSS.

Whereas Grammarly is “maybe/yes/the participant” is not CSS.

Something which nudges you with “this selfie you just took has a lot of fleshy bits in it, are you sure you trust the other person to keep it safely?” — is not CSS.

Meta has said it will launch a new safety tool to block children from receiving and discourage them from sending nude images, including in encrypted chats later this year. The tool is likely to be optional and available to adults too on Instagram and Facebook.

https://www.bbc.co.uk/news/technology-68093343

Comments

One response to “Meta tool to block nude images in teens’ private messages | …note, since this does not report outwards this is not client-side scanning, no more than a spellchecker would be”

  1. Mark Truman

    So once again Andy burrows doesn’t know what he’s talking about when he tweeted

    “Kudos to
    @frankhersey
    for cutting through Meta’s spin.

    Announcing client-side measures may tactically help them at the Senate hearing next week, but it also makes it much more likely
    @ofcom
    can consider it proportionate & technically feasible to issue Tech Warning Notices”

Leave a Reply

Your email address will not be published. Required fields are marked *