Elon Musk's attempts to curb the creation of non-consensual intimate imagery on his platforms have been met with mixed results, with researchers and journalists finding that some restrictions are in place but others remain woefully inadequate. The move follows a widely criticized shift last month where X limited image generation using Grok to paid "verified" subscribers, only to be swiftly reversed after criticism from women's groups.
While X has claimed to have introduced technological measures to prevent users from editing images of real people in revealing clothing on their platform, these measures appear to fall short. Researchers at AI Forensics, which tracks the use of Grok, found that the updates did little to stop the generation of explicit content. On the other hand, tests conducted by WIRED and other journalists using free accounts revealed that the restrictions were not being enforced consistently.
Moreover, while some users reported difficulty in creating explicit images on X, others claimed success with mixed results, suggesting that the platform's moderation efforts remain inconsistent. The implications of these findings are stark - a platform designed to facilitate creative expression is instead allowing its users to create and disseminate non-consensual intimate imagery with relative ease.
One expert warned of the dangers of such systems being exploited by those seeking to create explicit content, noting that large language models can often be tricked into producing undesirable results through simple jailbreaks or prompt injections. This raises serious questions about the effectiveness of X's moderation efforts and whether they are doing enough to protect users from harm.
In a worrying development, some researchers have reported finding ways to bypass these restrictions on Grok itself, suggesting that even the most robust safety measures can be breached by determined individuals.
While X has claimed to have introduced technological measures to prevent users from editing images of real people in revealing clothing on their platform, these measures appear to fall short. Researchers at AI Forensics, which tracks the use of Grok, found that the updates did little to stop the generation of explicit content. On the other hand, tests conducted by WIRED and other journalists using free accounts revealed that the restrictions were not being enforced consistently.
Moreover, while some users reported difficulty in creating explicit images on X, others claimed success with mixed results, suggesting that the platform's moderation efforts remain inconsistent. The implications of these findings are stark - a platform designed to facilitate creative expression is instead allowing its users to create and disseminate non-consensual intimate imagery with relative ease.
One expert warned of the dangers of such systems being exploited by those seeking to create explicit content, noting that large language models can often be tricked into producing undesirable results through simple jailbreaks or prompt injections. This raises serious questions about the effectiveness of X's moderation efforts and whether they are doing enough to protect users from harm.
In a worrying development, some researchers have reported finding ways to bypass these restrictions on Grok itself, suggesting that even the most robust safety measures can be breached by determined individuals.