I disagree. They used the word “involved”. And that term means sth has a part somewhere in the process. Which applies if you feed it a real picture of a face or even use it as a template. So a real image was “involved”, disregarding if it looks or is real. However it gets processed or becomes part of the final thing, it was involved nonetheless.
The issue with that is that it’s a very annoying form of bullying. And kids are afraid to tell someone because of the nature of it, they’re being blackmailed or whatever and it’s difficult to cope. I think I even read that lead to teens trying to commit suicide. Which isn’t nice, if true.
But I certainly agree that there is too much hysteria, mixing everything together, spreading FUD, and exploiting child abuse to push for mass surveillance or other political agendas.
Ultimately, I think a phrase with “involved” is a good choice. It makes it clear that it’s not okay if a real child is made part of the process. We can argue if it’s alright to do fictional drawings which seem to be okay in California and for example in Japan. But (ab)using real people no matter where in the process would definitely cross the line for me.
I was just talking with a friend who is a software dev (I’m a Linux Engineer so I do software as part of my job, just not my main focus) and we were just commiserating on how 75-80% of the world doesn’t understand that “AI” is just regurgitating information it has collected and it’s not like Jarvis or Skynet and thinks for itself.
I agree that the term “sexual abuse” is definitely misleading, I think “sexual exploitation” is better. I agree with you it’s no different than face swapping, but the difference is that it’s a lot easier for the general public to do it now than it was 5 or 10 years ago. It’s also pretty fucked that a fake image of you could potentially put you in “hot water” years down the road and you have zero control over it.
While I definitely hate the “AI bubble” that has grown tremendously over the past 2-3 years, we definitely need to figure out how to place limits on it before shit really gets out of hand in another year or two. The problem is that anyone that knows anything about this stuff doesn’t work in or for the government. The woman in the article that said that this needs to be regulated at every point of course doesn’t work in tech, she works for some rights organization 🤦♂️
Removed by mod
I disagree. They used the word “involved”. And that term means sth has a part somewhere in the process. Which applies if you feed it a real picture of a face or even use it as a template. So a real image was “involved”, disregarding if it looks or is real. However it gets processed or becomes part of the final thing, it was involved nonetheless.
The issue with that is that it’s a very annoying form of bullying. And kids are afraid to tell someone because of the nature of it, they’re being blackmailed or whatever and it’s difficult to cope. I think I even read that lead to teens trying to commit suicide. Which isn’t nice, if true.
But I certainly agree that there is too much hysteria, mixing everything together, spreading FUD, and exploiting child abuse to push for mass surveillance or other political agendas.
Ultimately, I think a phrase with “involved” is a good choice. It makes it clear that it’s not okay if a real child is made part of the process. We can argue if it’s alright to do fictional drawings which seem to be okay in California and for example in Japan. But (ab)using real people no matter where in the process would definitely cross the line for me.
I was just talking with a friend who is a software dev (I’m a Linux Engineer so I do software as part of my job, just not my main focus) and we were just commiserating on how 75-80% of the world doesn’t understand that “AI” is just regurgitating information it has collected and it’s not like Jarvis or Skynet and thinks for itself.
I agree that the term “sexual abuse” is definitely misleading, I think “sexual exploitation” is better. I agree with you it’s no different than face swapping, but the difference is that it’s a lot easier for the general public to do it now than it was 5 or 10 years ago. It’s also pretty fucked that a fake image of you could potentially put you in “hot water” years down the road and you have zero control over it.
While I definitely hate the “AI bubble” that has grown tremendously over the past 2-3 years, we definitely need to figure out how to place limits on it before shit really gets out of hand in another year or two. The problem is that anyone that knows anything about this stuff doesn’t work in or for the government. The woman in the article that said that this needs to be regulated at every point of course doesn’t work in tech, she works for some rights organization 🤦♂️