Apple has had quite the rollercoaster ride over plans to scan devices for the presence of child sexual abuse materials (CSAM). After announcing and then withdrawing its own plans for CSAM scanning, it ...
AI-generated child sexual abuse material (CSAM) has been flooding the internet, according to a report by The New York Times. Researchers at organizations like the Internet Watch Foundation and the ...
Throughout last year, Amazon detected the material in its AI training data and reported it to the National Center for Missing ...
For weeks, xAI has faced backlash over undressing and sexualizing images of women and children generated by Grok. One researcher conducted a 24-hour analysis of the Grok account on X and estimated ...
It seems that instead of updating Grok to prevent outputs of sexualized images of minors, X is planning to purge users generating content that the platform deems illegal, including Grok-generated ...