cross-posted from: https://lemmy.zip/post/15863526
Steven Anderegg allegedly used the Stable Diffusion AI model to generate photos; if convicted, he could face up to 70 years in prison
cross-posted from: https://lemmy.zip/post/15863526
Steven Anderegg allegedly used the Stable Diffusion AI model to generate photos; if convicted, he could face up to 70 years in prison
Sure.
I mostly referred to the second paragraph. Probably, this person meant that it’s better no child has been harmed in this 13k images’ production, but the wording irked me, especially the ‘win’. It got me a bit salty and I didn’t elaborate, so I don’t know what exactly people thought I’ve meant.
So I don’t consider this a ‘win’ because it doesn’t help their urges or make them less dangerous unlike therapy, like vaping was sometimes marketed as a healthier alternative to cigs or a way to give up smoking. I don’t want to dive into ethics of these two kinds of CSAM, but I find that leaving out the aspect of production (victimless?), it’s still harmful to the society as a whole to (generate,) collect and share it. Why brackets? Usually in court there are different levels or different articles that may be involved, and if production itself may be treated as harmless, merely having a collection and participating in trade\share of such materials are criminal offences themselves. And there we are to pick if we treat them as real or not. Returning to vapes: due to not being regular cigs, when it was a novelty many initially thought it’s okay to smoke them at work or in a classroom, but later they were banned as well. That’s not the only case where the nature of what AI produces and responsibility for that causes arguments, and our codified laws aren’t all bleeding edge to cover this, so I guess we are in the time we decide the framework to evaluate, work with them. And as silly as it is, vape pandemic was the first thing I’ve been reminded of, and it’s not great because both this and AI CSAM I’ve heard of because of it’s usage in schools - the second one is an article from months ago about deepfake nudes boys made of peers. Seemingly gated garden keeps being the most vulnerable.