‘A relationship with another human is overrated’ – inside the rise of AI girlfriends

re'up

Veteran
Joined
May 26, 2012
Messages
20,651
Reputation
6,243
Daps
64,577
Reppin
San Diego
We're going to take the chance and mystery out of romantic situations and the irony of this statement will come to fruition; sex will become porn, and subsequently unfulfilling.

There needs to be challenge, hope, disappointment, frustration, innuendo, seduction...sex sucks when you don't have these. It literally becomes as repulsive as porn after you nut.

What Pac say..."I don't want it, if it's that easy..."

Honestly, a lot of what modern dating norms seem to align with this in a sense. it's a lot of self protective measures to essentially feel nothing bad.

I know a handful of women who I am 99% sure are not having sex with anyone, and would rather text back and forth, until they have vetted that this person is safe for a long term relationship. This is all pre sex. How do you even do that? I don't think things like that were common 25 years ago.

I know some super bad girls like bottle girls who don't have sex, but post like near nude pics all day.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,601
Reputation
8,519
Daps
160,535

Replika CEO Eugenia Kuyda says it’s okay if we end up marrying AI chatbots​


The head of chatbot maker Replika discusses the role AI will play in the future of human relationships.​


By Nilay Patel, editor-in-chief of The Verge, host of the Decoder podcast, and co-host of The Vergecast.
Aug 12, 2024, 10:00 AM EDT


Photo illustration of Replika CEO Eugenia Kuyda

Photo illustration by The Verge / Photo by Replika

Today, I’m talking with Replika founder and CEO Eugenia Kuyda, and I will just tell you right from the jump, we get all the way to people marrying their AI companions, so get ready.

Replika’s basic pitch is pretty simple: what if you had an AI friend? The company offers avatars you can curate to your liking that basically pretend to be human, so they can be your friend, your therapist, or even your date. You can interact with these avatars through a familiar chatbot interface, as well as make video calls with them and even see them in virtual and augmented reality.

The idea for Replika came from a personal tragedy: almost a decade ago, a friend of Eugenia’s died, and she fed their email and text conversations into a rudimentary language model to resurrect that friend as a chatbot. Casey Newton wrote an excellent feature about this for The Verge back in 2015. Even back then, that story grappled with some of the big themes you’ll hear Eugenia and I talk about today: what does it mean to have a friend inside the computer?

READ THE REST ON SITE....
 

The axe murderer

For I am death and I ride on a pale horse
Joined
Aug 27, 2015
Messages
40,311
Reputation
6,138
Daps
137,966
I used to laugh at this concept but shyt becoming more plausible to many ppl:francis:
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,601
Reputation
8,519
Daps
160,535

By Will Knight
Business

Oct 16, 2024 1:14 PM


Inside the Mind of an AI Girlfriend (or Boyfriend)​


Dippy, a startup that offers “uncensored” AI companions, lets you peer into their thought process—sometimes revealing hidden motives.

1950s BRUNETTE WOMAN KNIT DRESS OFFHAND GESTURE SERIOUS EXPRESSION.


Photo-Illustration: WIRED Staff/Getty Images

If you buy something using links in our stories, we may earn a commission. This helps support our journalism. Learn more. Please also consider subscribing to WIRED

Last month, OpenAI unveiled an ambitious new language model capable of working through challenging problems with a simulated kind of step-by-step reasoning. OpenAI says the approach could be crucial for building more capable AI systems in the future.

In the meantime, perhaps a more modest version of this technology could help make AI girlfriends and boyfriends a bit more spontaneous and alluring.

That’s what Dippy, a startup that offers “uncensored” AI companions is betting. The company recently launched a feature that lets users see the reasoning behind their AI characters’ responses.

Dippy runs its own large language model, which is an open source offering fine-tuned using role-play data, which the company says makes it better at improvising when a user steers a conversation in a particular direction.

Akshat Jagga, Dippy’s CEO, says that adding an additional layer of simulated “thinking”—using what’s known as “chain-of-thought prompting”—can elicit more interesting and surprising responses, too. “A lot of people are using it,” Jagga says. “Usually, when you chat with an LLM, it sort of just gives you a knee-jerk reaction.”

Jagga adds that the new feature can reveal when one of its AI characters is being deceptive, for instance, which some users apparently enjoy as part of their role-play. “It’s interesting when you can actually read the character’s inner thoughts,” Jagga says. “We have this character that is sweet in the foreground, but manipulative in the background.”

I tried chatting with some of Dippy’s default characters, with the PG settings on because otherwise they are way too horny. The feature does add another dimension to the narrative, but the dialog still seems, to me, rather predictable, resembling something lifted from a bad romance novel or an overwrought piece of fan fiction.

One Dippy character, described as “Bully on the outside, warm on the inside,” revealed a soft side behind the gruff exterior when I clicked the “Read thought process” link beneath each message, but both the inner and outer dialogs lacked nuance or surprise and were repetitive. For fun, I also tried asking several characters some simple arithmetic problems, and their thinking sometimes showed how to break the puzzle down to get the correct answer.

Despite its limitations, Dippy seems to show how popular and addictive AI companions are becoming. Jagga and his cofounder, Angad Arneja, previously cofounded Wombo, a company that uses AI to create memes including singing photographs. The pair left in 2023, setting out to build an AI-powered office productivity tool, but after experimenting with different personas for their assistant, they became fascinated with the potential of AI companionship.

With little promotion, Dippy has amassed 500,000 monthly and 50,000 daily active users, Jagga says, with people spending, on average, an hour on the app at a time. “That engagement was absolutely insane for us,” he says.

Dippy revealed that it has secured $2.1 million in “pre-seed” funding in a round led by Drive Capital.

Dippy is of course entering an already bustling market that includes well-known companies like Character.AI and Replika as well as a host of other AI girlfriend apps. A recent report from investment firm Andreessen Horowitz shows that the top 100 generative AI tools based on usage include many AI companions; the chart on engagement among users of such apps shows how much stickier they are than just about anything else out there.

While these apps are often associated with undersocialized young men, they cater to women too. Jagga says that 70 percent of Dippy’s accounts tend to favor male characters, which could mean that many users identify as female.

Besides threatening to upend the world of surrogate OnlyFans chatters, these AI companions may have social effects that we have yet to reckon with. While a few research studies suggest that chatbots can lessen feelings of loneliness, some experts warn that they may result in greater alienation among heavy users, and seem to perpetuate harmful stereotypes.

“Some of these bots have dark patterns,” says Iliana Depounti, an ESRC-backed researcher at Loughborough University in the UK who has studied usage of Replika, another AI companion app. Depounti says these patterns often target the emotional vulnerabilities of lonely people. She adds that Dippy seems to promote themes and narratives designed to appeal to young women in particular. “Some people who use these apps are socially isolated, and these apps create further silos through their emotionally validating algorithms that don’t challenge existing conditions,” she adds.

Rather than just looking inside the minds of AI companions, then, we may need to take a closer look at how people are interacting with these apps to understand the real benefits and risks.
 
Top