Comment by mortarion

20 hours ago

CSAM does not have a universal definition. In Sweden for instance, CSAM is any image of an underage subject (real or realistic digital) designed to evoke a sexual response. If you take a picture of a 14 year old girl (age of consent is 15) and use Grok to give her bikini, or make her topless, then you are most definately producing and possessing CSAM.

No abuse of a real minor is needed.

[flagged]

  • He made no judgement in his comment, he just observed the fact that the term csam - in at least the specified jurisdiction - applies to generated pictures of teenagers, wherever real people were subjected to harm or not.

    I suspect none of us are lawyers with enough legal knowledge of the French law to know the specifics of this case

> CSAM does not have a universal definition.

Strange that there was no disagreement before "AI", right? Yet now we have a clutch of new "definitions" all of which dilute and weaken the meaning.

> In Sweden for instance, CSAM is any image of an underage subject (real or realistic digital) designed to evoke a sexual response.

No corroboration found on web. Quite the contrary, in fact:

"Sweden does not have a legislative definition of child sexual abuse material (CSAM)"

https://rm.coe.int/factsheet-sweden-the-protection-of-childr...

> If you take a picture of a 14 year old girl (age of consent is 15) and use Grok to give her bikini, or make her topless, then you are most definately producing and possessing CSAM.

> No abuse of a real minor is needed.

Even the Google "AI" knows better than that. CSAM "is considered a record of a crime, emphasizing that its existence represents the abuse of a child."

Putting a bikini on a photo of a child may be distasteful abuse of a photo, but it is not abuse of a child - in any current law.

  • " Strange that there was no disagreement before "AI", right? Yet now we have a clutch of new "definitions" all of which dilute and weaken the meaning. "

    Are you from Sweden? Why do you think the definition was clear across the world and not changed "before AI"? Or is it some USDefaultism where Americans assume their definition was universal?

    • > Are you from Sweden?

      No. I used this interweb thing to fetch that document from Sweden, saving me a 1000-mile walk.

      > Why do you think the definition was clear across the world and not changed "before AI"?

      I didn't say it was clear. I said there was no disagreement.

      And I said that because I saw only agreement. CSAM == child sexual abuse material == a record of child sexual abuse.

      4 replies →

  • > Even the Google "AI" knows better than that. CSAM "is [...]"

    Please don't use the "knowledge" of LLMs as evidence or support for anything. Generative models generate things that have some likelihood of being consistent with their input material, they don't "know" things.

    Just last night, I did a Google search related to the cell tower recently constructed next to our local fire house. Above the search results, Gemini stated that the new tower is physically located on the Facebook page of the fire department.

    Does this support the idea that "some physical cell towers are located on Facebook pages"? It does not. At best, it supports that the likelihood that the generated text is completely consistent with the model's input is less than 100% and/or that input to the model was factually incorrect.

  • > - in any current law.

    It has been since at least 2012 here in Sweden. That case went to our highest court and they decided a manga drawing was CSAM (maybe you are hung up on this term though, it is obviously not the same in Swedish).

    The holder was not convicted but that is besides the point about the material.

    • > It has been since at least 2012 here in Sweden. That case went to our highest court

      This one?

      "Swedish Supreme Court Exonerates Manga Translator Of Porn Charges"

      https://bleedingcool.com/comics/swedish-supreme-court-exoner...

      It has zero bearing on the "Putting a bikini on a photo of a child ... is not abuse of a child" you're challenging.

      > and they decided a manga drawing was CSAM

      No they did not. They decided "may be considered pornographic". A far lesser offence than CSAM.

  • In Swedish:

    https://www.regeringen.se/contentassets/5f881006d4d346b199ca...

    > Även en bild där ett barn t.ex. genom speciella kameraarrangemang framställs på ett sätt som är ägnat att vädja till sexualdriften, utan att det avbildade barnet kan sägas ha deltagit i ett sexuellt beteende vid avbildningen, kan omfattas av bestämmelsen.

    Which translated means that the children does not have to be apart of sexual acts and indeed undressing a child using AI could be CSAM.

    I say "could" because all laws are open to interpretation in Sweden and it depends on the specific image. But it's safe to say that many images produces by Grok are CSAM by Swedish standards.

  • The lady doth protest too much, methinks.

    • That's the problem with CSAM arguments, though. If you disagree with the current law and think it should be loosened, you're a disgusting pedophile. But if you think it should be tightened, you're a saint looking out for the children's wellbeing. And so laws only go one way...

  • "Sweden does not have a legislative definition of child sexual abuse material (CSAM)"

    Because that is up to the courts to interpret. You cant use your common law experience to interpret the law in other countries.

    • > You cant use your common law experience to interpret the law in other countries.

      That interpretation wasn't mine. It came from the Court of Europe doc I linked to. Feel free to let them know its wrong.

      1 reply →