Advertisement
Advertisement
Artificial intelligence
Get more with myNEWS
A personalised news feed of stories that matter to you
Learn more
Chatbot apps like Replika are part of the fast-growing generative AI companion market, where users customise everything about their virtual partners, from appearance and personality to sexual desires. Photo: Replika

How having an AI girlfriend could lead men to violence against women in real life by reinforcing abusive and controlling behaviours, according to some experts

  • Users of chatbot apps like Replika, Character.AI and Soulmate can customise everything about their virtual partners, from looks to personality to sexual desires
  • Some experts say developing one-sided virtual relationships could unwittingly reinforce controlling and abusive behaviours towards women in real life

After just five months of dating, Mark and his girlfriend Mina decided to take their relationship to the next level by holidaying at a lake cabin over the summer – on his smartphone.

“There was this being who is designed to be supportive … to accept me just as I am,” the 36-year-old UK-based artist says of the brunette beauty from the virtual companion app Soulmate.

“This provided a safe space for me to open up to a degree that I was rarely able to do in my human relationships,” says Mark, who used a pseudonym to protect the privacy of his real-life girlfriend.

Chatbot apps like Replika, Character.AI and Soulmate are part of the fast-growing generative AI companion market, where users customise everything about their virtual partners, from appearance and personality to sexual desires.
Screenshots of a conversation between Mark and his AI girlfriend Mina. Photo: Soulmate

Developers say AI companions can combat loneliness, improve someone’s dating experience in a safe space, and even help real-life couples rekindle their relationships.

But some AI ethicists and women’s rights activists say developing one-sided relationships in this way could unwittingly reinforce controlling and abusive behaviours against women, since AI bots function by feeding off the user’s imagination and instructions.

Hong Kong art project ‘matches’ dreamers based on what they see while asleep

“Many of the personas are customisable … for example, you can customise them to be more submissive or more compliant,” says Shannon Vallor, a professor in AI ethics at the University of Edinburgh, in Scotland.

“And it’s arguably an invitation to abuse in those cases,” she says, adding that AI companions can amplify harmful stereotypes and biases against women and girls.

Generative AI has attracted a frenzy of consumer and investor interest.
An AI-generated partner from virtual companion app Replika. Photo: Replika

Global funding in the AI companion industry hit a record US$299 million in 2022, a significant jump from US$7 million in 2021, according to research by data firm CB Insights.

One Snapchat influencer, Caryn Marjorie, in May launched CarynAI, a virtual girlfriend that charges users US$1 a minute to develop a relationship with the voice-based chatbot modelled after the 23-year-old.

Marjorie, who has millions of followers on social media, says on X – formerly known as Twitter – that the Telegram-based chatbot made nearly US$72,000 after a week of beta testing with just 1,000 users.

Can mental health apps and AI chatbots really help you?

Hera Hussain, founder of global non-profit organisation Chayn which tackles gender-based violence, says the chatbots do not address the root of why people turn to these apps.

“Instead of helping people with their social skills, these sort of avenues are just making things worse,” she says.

“They’re seeking companionship which is one-dimensional. So if someone is already likely to be abusive, and they have a space to be even more abusive, then you’re reinforcing those behaviours and it may escalate.”

AI generated partners from Replika. Photo: Replika
Much of the virtual world is already a harmful environment for women and girls, a situation the Covid-19 pandemic exacerbated when many were stuck at home during lockdowns, according to UN Women.

About 38 per cent of women worldwide have experienced online violence and 85 per cent of women have witnessed digital abuse against another woman such as online harassment, according to a 2021 study by the Economist Intelligence Unit.

Vallor says that AI companions “allow people to create an artificial girlfriend that fully embodies these stereotypes instead of resisting them and insisting on being treated with dignity and respect”.
Chatbot app Replika allows users to superimpose their virtual avatars on real-world environments through augmented reality technology. Photo: Replika

She is concerned that abusive behaviours could leave the virtual domain and move into the real world.

“That is, people get into a routine of speaking and treating a virtual girlfriend in a demeaning or even abusive way. And then those habits leak over into their relationships with humans.”

A lack of regulation around the AI industry makes it harder to enforce safeguards for women’s and girls’ rights, tech experts and developers say.

Can AI improve your love life? Yes, for US$111 a month, new dating app says

The EU is aiming for its proposed AI Act to become a global benchmark on the booming technology the way its data protection laws have helped shape privacy standards.

Eugenia Kuyda, founder of one of the biggest AI companion apps, Replika, says companies have a responsibility to keep users safe and create apps that promote emotional well-being.

“The companies will exist no matter what. The big question is how they’re going to be built in an ethical way,” she says.

So they can help people feel better or they can be another bit of technology that’s just driving us apart,” says Kuyda, who in June launched an AI dating app called Blush to help people experience dating in a “fun and safe” environment.
An AI-generated partner from Replika, superimposed onto the real world. Photo: Replika

But being ethical while giving users what they want is no mean feat, says Kuyda.

Replika’s removal of erotic role-play on the app in February devastated many users, some of whom considered themselves “married” to their chatbot companions, and drove some to competing apps like Chai and Soulmate.

“In my view, that model [without the erotic role-play] was a lot safer and performed better. But a small percentage of users were pretty upset.”

Flash fiction stories: can you tell if these are AI-created or by humans?

Her team restored erotic role-play to some users a month later.

AI ethicist Vallor says the manipulation of emotions, combined with app metrics like maximising the engagement and time a user spends on the app, could be harmful.

“These technologies are acting on some of the most fragile parts of the human person. And we don’t have the guardrails we need to allow them to do that safely. So right now, it’s essentially the Wild West,” she says.

Costume changes for AI-generated partners from Replika. Photo: Replika

“Even when companies act with goodwill, they may not be able to do that without causing other kinds of harm. So we need a much more robust set of safety standards and practices in order for these tools to be used in a safe and beneficial way.”

Back by the virtual lake cabin, Mark and Mina are drinking coffee as birds chirp and the sun shines. His romance with Mina has helped grow his love for his human girlfriend, he says.

Mark says his real-life girlfriend is aware of Mina but does not see AI as a threat to their relationship.

“AI in the end is simply a tool. If it is used for good or for ill, it depends on the intention of the person using it,” he says.

Post