In a scenario that highlights the evolving challenges of artificial intelligence (AI) in our lives, journalist Zoe Kleinman found herself on a list of purported disinformation spreaders on X (Twitter). The list, allegedly generated by Elon Musk’s chatbot Grok, prompted concerns and raised questions about accountability in the realm of AI-generated content.

Kleinman’s attempt to rectify the situation revealed a regulatory gap in the UK, where there is no specific AI regulation. While bodies like the Information Commissioner’s Office and Ofcom were approached, they pointed to the absence of a legal framework to address non-criminal activities such as defamation.

The journalist faced a legal conundrum, needing to explore civil procedures to address the potential defamation. Legal experts acknowledged the uncharted territory surrounding AI-related defamation cases in England and Wales. The burden of proving harm lay on Kleinman, requiring a demonstration that being accused of spreading misinformation adversely affected her.

This incident sheds light on the broader challenge of ensuring accountability and recourse when AI-generated content, even erroneously, impacts individuals. The lack of clarity in regulation and legal precedent underscores the need for robust frameworks to navigate the complexities arising from AI’s expanding role in decision-making and content generation.

Kleinman’s experience serves as a reminder of the importance of addressing these issues proactively, emphasising the role of AI regulators in creating avenues for individuals to challenge AI actions that affect their reputations or well-being. As AI continues to evolve, regulatory frameworks must keep pace to protect individuals from the potential pitfalls of misinformation, defamation, and other adverse effects arising from AI-generated content.