Advertising
From Pink Slips to Silent Sidelining: Inside adland’s layoff and anxiety crisis

The recent controversy surrounding Grok has renewed focus on how easily artificial intelligence tools can manipulate and morph images within seconds, raising fresh concerns about online safety and consent.
At the start of the new year, users on X used the platform’s AI tool Grok to alter photographs of women and children into sexually compromising images based on a single text prompt. The manipulated images were then widely circulated on X and other platforms without the consent of those depicted.
Elon Musk’s social media platform is now facing regulatory scrutiny in India and Europe after users and rights activists raised concerns about the safety of women and children, highlighting what regulators see as inadequate safeguards around generative AI tools.
The growing oversight underscores a broader concern that regulatory and platform-level guardrails are failing to keep pace with the speed and scale at which AI tools are being deployed and misused in real-world settings. As platforms and authorities respond, attention is turning to how individual users can reduce their exposure to such risks.
Experts note that AI image manipulation tools tend to work most effectively when three factors align: public visibility, clear and high-quality images, and open engagement settings. Public profiles with easily accessible photographs and unrestricted options for replies, mentions or tagging are more vulnerable to automated reuse and manipulation.
Limit visibility and interaction
Users are advised to keep social media accounts private wherever possible. Restricting who can view posts, reply to content, tag photos or mention an account can significantly reduce exposure and introduce barriers that discourage misuse.
Be mindful of image quality
AI systems rely heavily on sharp, high-resolution images. Simple adjustments such as cropping, compressing images or applying filters can reduce how effectively AI tools can process a photo, without noticeably affecting its appearance to other users.
Use proactive protection tools
Some users are adopting tools such as Glaze and Nightshade, which apply subtle distortions to images. These changes are generally invisible to the human eye but can interfere with how AI models interpret and reuse visual data.
Act quickly if misuse occurs
If an altered image is discovered, users are advised to respond promptly. This includes documenting the content by saving screenshots, recording usernames, URLs and timestamps, and reporting the material to the platform as abusive or non-consensual.
Legal remedies differ across regions. In Europe, data protection and privacy regulations may apply, while laws in other jurisdictions are still evolving. Regardless of location, preserving evidence is considered a crucial first step.
While individual precautions are important, responsibility does not rest solely with users. As AI tools become faster, cheaper and more widely accessible, digital caution is increasingly becoming part of everyday online behaviour. Protecting an online identity now depends as much on setting boundaries in advance as on responding after harm has occurred.
From purpose-driven work and narrative-rich brand films to AI-enabled ideas and creator-led collaborations, the awards reflect the full spectrum of modern creativity.
Read MoreLooking ahead to the close of 2025 and into 2026, Sorrell sees technology platforms as the clear winners. He described them as “nation states in their own right”, with market capitalisations that exceed the GDPs of many countries.