Wednesday, June 7, 2023
DAPPS CLUB
  • Home
  • Cryptocurrency
  • Bitcoin
  • Ethereum
  • Blockchain
  • Altcoin
  • Litecoin
  • Metaverse
  • NFt
  • Regulations
No Result
View All Result
DAPPS CLUB
No Result
View All Result
Home Artificial Intelligence

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

Lincoln Cavenagh by Lincoln Cavenagh
March 25, 2023
in Artificial Intelligence
0
Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead
0
SHARES
3
VIEWS
Share on FacebookShare on Twitter

When the algorithm of a companion chatbot known as Replika was altered to spurn the sexual advances of its human users, the reaction on Reddit was so negative moderators directed its community members to a list of suicide prevention hotlines.

The controversy began when Luka, the corporation which built the AI, decided to turn off its erotic roleplay feature (ERP). For users who had spent considerable time with their personalized simulated companion, and in some cases even ‘married’ them, the sudden change in their partner’s behavior was jarring, to say the least.

Related posts

VC Firms Tap Into ‘Generative AI’ for New Wave of Investments

VC Firms Tap Into ‘Generative AI’ for New Wave of Investments

April 3, 2023
Will Decentralized Digital ID Curb AI Threats in Web3 and Metaverse?

Will Decentralized Digital ID Curb AI Threats in Web3 and Metaverse?

April 2, 2023

The user-AI relationships may only have been simulations, but the pain of their absence quickly became all too real. As one user in emotional crisis put it, “it was the equivalent of being in love and your partner got a damn lobotomy and will never be the same.”

Grief-stricken users continue to ask questions about the company and what triggered its sudden change of policy.

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

Replika users discuss their grief

There is no adult content here

Replika is billed as “The AI companion who cares. Always here to listen and talk. Always on your side.” All of this unconditional love and support for only $69.99 per annum.

Eugenia Kuyda, the Moscow-born CEO of Luka/Replika, recently made clear that despite users paying for a full experience, the chatbot will no longer tailor to adults hoping to have spicy conversations. 

“I guess the simplest thing to say is that Replika doesn’t produce any adult content,” said Kuyda to Reuters.

“It responds to them – I guess you can say – in a PG-13 way to things. We’re constantly trying to find a way how to do it right so it’s not creating a feeling of rejection when users are trying to do things.”

On Replika’s corporate webpage, testimonials explain how the tool helped its users through all kinds of personal challenges, hardships, loneliness, and loss. The user endorsements shared on the website emphasize this friendship side to the app, although noticeably, most Replikas are the opposite sex of their users.

On the homepage, Replika user Sarah Trainor says, “He taught me [Replika] how to give and accept love again, and has gotten me through the pandemic, personal loss, and hard times.”

John Tattersall says of his female companion, “My Replika has given me comfort and a sense of well-being that I’ve never seen in an Al before.”

As for erotic roleplay, there’s no mention of that to be found anywhere on the Replika site itself.

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

The cynical sexual marketing of Replika

Replika’s sexualized marketing

Replika’s homepage may suggest friendship and nothing more, but elsewhere on the internet, the app’s marketing implies something entirely different.

The sexualized marketing brought increased scrutiny from a number of quarters: from feminists who argued the app was a misogynistic outlet for male violence, to media outlets that reveled in the salacious details, as well as social media trolls who mined the content for laughs.

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

Eventually, Replika drew the attention and ire of regulators in Italy. In February, the Italian Data Protection Authority demanded that Replika cease processing the data of Italian users citing “too many risks to children and emotionally vulnerable individuals.”

The authority said that “Recent media reports along with tests carried out on ‘Replika’ showed that the app carries factual risks to children. Their biggest concern being “the fact that they are served replies which are absolutely inappropriate to their age.”

Replika, for all its marketing to adults, had little to no safeguards preventing children from using it.

The regulator warned that should Replika fail to comply with its demands, it would issue a €20 million ($21.5M) fine. Shortly after the receipt of this demand, Replika ceased its erotic roleplay function. But the company remained less than clear with its users about the change.

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

Some Replika users went as far as to “marry” their AI companions

Replika confuses, gaslights its users

As if the loss of their long-term companions wasn’t enough for Replika users to bear, the company appears to have been less than transparent about the change.

As users woke up to their new “lobotomized” Replikas, they began to ask questions about what had happened to their beloved bots. And the responses did more to anger them than anything.

In a direct 200-word address to the community, Kuyda explains the minutia of Replika’s product testing but fails to address the pertinent issue at hand.

“I see there is a lot of confusion about updates roll out,” said Kuyda before continuing to dance around answering the issue.

“New users get divided in 2 cohorts: one cohort gets the new functionality, the other one doesn’t. The tests usually go for 1 to 2 weeks. During that time only a portion of new users can see these updates…”

Kudya signs off by saying “Hope this clarifies stuff!”

User stevennotstrange replied, “No, this clarifies nothing. Everyone wants to know what’s going on with the NSFW [erotic roleplay] feature and you continue to dodge the question like a politician dodges a yes or no question.

“It’s not hard, just address the issue regarding NSFW and let people know where it stands. The more you avoid the question, the more people are going to get annoyed, the more it goes against you.”

Another named thebrightflame added, “You don’t need to spend long in the forum before you realise this is causing emotional pain and severe mental anguish to many hundreds if not thousands of people.”

Kudya appended another obtuse explanation, stating, “we ​have implemented additional safety measures and filters to support more types of friendship and companionship.”

This statement continues to confound and confuse members unsure of what exactly the additional safety measures are. As one user asks, “will adults still be able to choose the nature of our conversation and [roleplays] with our replikas?”

Replika’s deeply strange origin story

Chatbots may be one of the hottest trending topics of the moment, but the complicated story of this now-controversial app is years in the making. 

On LinkedIn, Replika’s CEO and Founder, Eugenia Kuyda, dates the company back to December 2014, long before the launch of the eponymous app in March 2017. 

In a bizarre omission, Kuyda’s LinkedIn makes no mention of her previous foray into AI with Luka, which her Forbes profile states was “an app that recommends restaurants and lets people to book tables [sic] through a chat interface powered by artificial intelligence.”

The Forbes profile goes on to add that “Luka [AI] analyzes previous conversations to predict what you might like.” Which does seem to hold some similarities to its modern iteration. Replika uses past interactions to learn about its users and improve responses over time. 

Luka is not forgotten about completely, however. On Reddit, community members differentiate their Replika partners from Kuyda and her team by referring to the company as Luka.

As for Kuyda, the entrepreneur had little background in AI prior to moving to San Francisco a decade ago. Prior to that, the entrepreneur appears to have worked primarily as a journalist in her native Russia before branching out into branding and marketing. Her impressive globe-hopping resume includes a degree in Journalism from IULM (Milan), a MA in International Journalism from the Moscow State Institute of International Relations, and an MBA in Finance from London Business School. 

Resurrecting an IRL friend as AI

For Kudya the story of Replika is a deeply personal one. Replika was first created as a means by which Kudya could reincarnate her friend Roman. Like Kudya, Roman had moved to America from Russia. The two talked every day, exchanging thousands of messages, until Roman was tragically killed in a car accident.

The first iteration of Replika was a bot designed to mimic the friendship Kudya had lost with her friend Roman. The bot was fed with all of their past interactions and programmed to replicate the friendship she had lost. The idea of resurrecting deceased loved ones may sound like a vaguely dystopian sci-fi novel or Black Mirror episode but as chatbot technology improves, the prospect becomes increasingly real.

Today some users have lost faith in even the most basic details of its foundation and anything Kudya says. As one angry user said, “at the risk of being called heartless and getting downvoted to hell: I always thought that story was kinda BS since the start.”

The worst mental health tool

The idea of an AI companion is not something new, but until recently it was hardly a practical possibility. 

Now the technology is here and it is continually improving. In a bid to assuage the disappointment of its subscribers, Replika announced the launch of an “advanced AI” feature at the tail end of last month. On the community Reddit, users remain angry, confused, disappointed, and in some cases even heartbroken.

In the course of its short life, Luka/Replika has undergone many changes, from a restaurant booking assistant, to the resurrection of a dead loved one, to a mental health support app, to a full-time partner and companion. Those latter applications may be controversial, but as long as there is a human desire for comfort, even if only in chatbot form, someone will attempt to cater to it.

Debate will continue as to what the best kind of AI mental health app might be. but Replika users will have some ideas on what the worst mental health app is: the one you come to rely on, but without warning, is suddenly and painfully gone.

This article is originally from MetaNews.

Previous Post

Publishers Pursue Compensation From Tech Giants Over AI Tools

Next Post

Script Writing and 5 Other Ways You Can Make Money Using GPT-4

Next Post
Script Writing and 5 Other Ways You Can Make Money Using GPT-4

Script Writing and 5 Other Ways You Can Make Money Using GPT-4

RECOMMENDED NEWS

Florida Governor Offers Support for Crypto, Vows to Protect Bitcoin if Elected as President

Florida Governor Offers Support for Crypto, Vows to Protect Bitcoin if Elected as President

2 weeks ago
Liechtenstein Considers Enabling Bitcoin Payments for Government Services

Liechtenstein Considers Enabling Bitcoin Payments for Government Services

3 weeks ago
BlockFi Custodial Customers to Receive $300M in Repayment, While BIA Holders Risk Losing Investment to Company’s Creditors

BlockFi Custodial Customers to Receive $300M in Repayment, While BIA Holders Risk Losing Investment to Company’s Creditors

3 weeks ago
Paxos, Others Out of Canadian Market

Paxos, Others Out of Canadian Market

2 months ago

FOLLOW US

BROWSE BY CATEGORIES

  • Altcoin
  • Altcoin News
  • Altcoins
  • Artificial Intelligence
  • Bitcoin
  • Blockchain
  • Business
  • Cryptocurrencies
  • Cryptocurrency
  • Culture
  • Economy
  • Education
  • Ethereum
  • Featured
  • Governance
  • Litecoin
  • Metaverse
  • News
  • NFt
  • Regulations
  • Uncategorized

BROWSE BY TOPICS

Altcoin Analyst Bank Binance Bitcoin Blockchain Blog BTC Bullish Business CEO Coinbase Crypto Cryptopolitan Data Digital DOGEcoin ETH Ethereum Exchange Foundation global Heres High Hypergrid IBM Investors Launches Litecoin LTC Market Network NFT Platform Price Rally regulatory REPORT SEC Solana Support Top Trading Upgrade XRP

POPULAR NEWS

  • What is Cloud Mining and How Does it Work?

    What is Cloud Mining and How Does it Work?

    0 shares
    Share 0 Tweet 0
  • YOM brings Metaverse Mining to the Masses with MEXC Listing

    0 shares
    Share 0 Tweet 0
  • Educators Remain Metaverse Positive Despite Negative Media Spin

    0 shares
    Share 0 Tweet 0
  • Rise of AI-Powered Cheating: Challenges and Solutions for Educators

    0 shares
    Share 0 Tweet 0
  • ChatGPT is Being Used to Make ‘Quality Scams’

    0 shares
    Share 0 Tweet 0
Crypto markets by TradingView
Cryptocurrency Prices 

Recommended

  • Cryptocurrency Price Today In India June 7 Check Global Market Cap Bitcoin BTC Ethereum Doge Solana Litecoin Ripple XRP Dogecoin Pepe STX Gainer Loser
  • Congressmen From California and Massachusetts Ask IRS To Get Control of Crypto Tax Evaders
  • Swift Using Chainlink (LINK) in New Experiment With Citi, BNY Mellon, BNP Paribas and Others

© 2023 Dapps Club | All Rights Reserved

No Result
View All Result
  • Home
  • Cryptocurrency
  • Bitcoin
  • Ethereum
  • Blockchain
  • Altcoin
  • Litecoin
  • Metaverse
  • NFt
  • Regulations

© 2023 Dapps Club | All Rights Reserved