To make sure you never miss out on your favourite NEW stories, we're happy to send you some reminders

Click 'OK' then 'Allow' to enable notifications

Disturbing AI ‘ghostbots’ could stalk you from beyond the grave, experts warn

Disturbing AI ‘ghostbots’ could stalk you from beyond the grave, experts warn

This is literally the plot of a Black Mirror episode

Experts from the University of Cambridge have warned that the world of AI needs protections against an 'emerging digital afterlife industry'.

They've said that psychological harm could be caused by 'deadbots' and 'griefbots' - AI chatbots which mimic the language patterns and personalities of those who have passed away.

There are companies which have already set up these chatbots, which allow people to create avatars of deceased loved ones and have an AI pretending to be them speak as though they were still alive.

If you're thinking this sounds familiar, then that's because this is quite literally the plot of a Black Mirror episode, namely the season two opening episode Be Right Back.

When Martha's (Hayley Atwell) boyfriend Ash (Domhnall Gleeson) dies suddenly, she starts speaking to one of these AIs which scoop up Ash's social media posts to answer as he would.

They're not real, but unless they're regulated it could lead to grief and suffering. (Getty Stock Photo)
They're not real, but unless they're regulated it could lead to grief and suffering. (Getty Stock Photo)

Martha dives into talking with the AI replication of her dead boyfriend, to the detriment of her real life relationships, but things get more difficult as Martha has to confront the fact that it isn't really Ash.

Now, a team of AI ethicists from the University of Cambridge have said there's a 'high risk' to people due to these 'ghostbots' which can pretend to be talking to you from beyond the grave.

They warned that these ghoulish chatbots could be used by companies to surreptitiously advertise products to people who were grieving, all while posing as a simulation of their dearly departed.

Companies which do this already exist in the form of apps and platforms such as HereAfter and Project December.

Depending on how this technology develops, people could end up 'stalked by the dead' as the chatbots send them notifications and try to communicate with them unprompted, which is a very real concern if safeguards aren't put in place to restrict this technology.

Would you want to speak with an AI recreation of a dead loved one? Would you be comfortable ending the conversation? (Getty Stock Photo)
Would you want to speak with an AI recreation of a dead loved one? Would you be comfortable ending the conversation? (Getty Stock Photo)

The experts have also urged caution as even those who are at first comforted by a 'ghostbot' might come to find the interactions emotionally draining.

"Rapid advancements in generative AI mean that nearly anyone with Internet access and some basic know-how can revive a deceased loved one," said study co-author Dr. Katarzyna Nowaczyk-Basińska.

"This area of AI is an ethical minefield. It’s important to prioritise the dignity of the deceased, and ensure that this isn’t encroached on by financial motives of digital afterlife services, for example.

“At the same time, a person may leave an AI simulation as a farewell gift for loved ones who are not prepared to process their grief in this manner.

"The rights of both data donors and those who interact with AI afterlife services should be equally safeguarded."

Dr Tomasz Hollanek, who co-authored the study, added: "It is vital that digital afterlife services consider the rights and consent not just of those they recreate, but those who will have to interact with the simulations.

"These services run the risk of causing huge distress to people if they are subjected to unwanted digital hauntings from alarmingly accurate AI recreations of those they have lost."

He recommended a 'digital funeral' to 'retire' AI chatbots.

Featured Image Credit: Getty Stock Images

Topics: AI, Technology, Black Mirror, Mental Health