By: Leela Xie
In the realm of content creation, advancements in artificial intelligence (AI) are reshaping the boundaries of creative expression. However, these technological strides also raise ethical concerns, mainly when using AI to recreate the likeness and voices of deceased or missing children. While some creators argue that this approach helps raise awareness, experts warn that it can potentially spread misinformation and cause distress to the victims’ families.
In recent instances, content creators have employed AI to give voice to the silence, portraying these children recounting the tragic events that occurred to them. One disturbing example involves the case of James Bulger, a 2-year-old British child abducted in 1993. In a TikTok video, an AI-generated image of James narrates the heartbreaking details of his abduction and subsequent tragic fate (The Washington Post). Similarly, another video features an AI-generated likeness of Madeleine McCann, a British 3-year-old who disappeared in 2007 in Portugal (The Washington Post).
The use of AI to recreate these children’s voices has drawn significant criticism. While some argue it is a novel way to bring attention to these cases, many experts, including the victims’ families, find it offensive and exploitative (The Washington Post). The videos often incorporate emotional elements like dramatic music and visuals of children with scars, which can potentially re-traumatize those who knew the victims.
As one platform where these AI-generated videos have appeared, TikTok has taken steps to address this issue. The platform’s guidelines explicitly state that synthetic media depicting young individuals is prohibited, and such content is actively removed. TikTok acknowledges the potential of AI to blur the line between fact and fiction, and it requires users to provide disclaimers when sharing manipulated media (The Washington Post).
Felix M. Simon, a communication researcher at the Oxford Internet Institute, notes that while AI tools have made it easier to create such videos, they often exhibit distinct hallmarks of AI-generated content, such as a stylized aesthetic. The availability of affordable AI applications has contributed to the proliferation of these videos on various platforms (The Washington Post).
AI has reached an advanced stage of convincingly mimicking human voices, but the childlike voices in these videos are likely computer-generated. This raises concerns about infringing upon the dignity of the deceased and their families. Social media engagement patterns, where emotionally charged content often garners more attention, encourage sharing these videos, although viewers’ reactions vary widely.
Creators who employ AI in content creation acknowledge the potential for shock value and increased viewership, especially on platforms with younger audiences. Despite the potential for AI to enhance historical storytelling, there remains a fine line between responsible use and exploitative manipulation.
The ethical debate surrounding using AI to recreate the voices of deceased children prompts reflection on societal norms and respect for the dead. While technology’s capabilities continue to evolve, content creators, platforms, and users must navigate the ethical complexities arising when AI intersects with sensitive subjects. As AI’s capabilities progress, the consequences and responsibilities associated with its application will become increasingly pronounced, shaping the future landscape of content creation and consumption.
Source:
AI social media videos depict missing, dead children narrating their stories – The Washington Post