It is a truth universally acknowledged, that a new media technology with great potential to drive innovation, will first be profitably employed in getting people off.
(Apologies to Jane Austen.)
New technology is often quickly turned to selling sex in one form or another. This has been true for a very long time. The first naughty photos date back roughly to the same time as the first photos. It’s a beloved fable of the tech industry that during the early years of home video players, Sony’s proprietary Betamax standard was technically superior to JVC’s VHS, but since JVC was willing to license the tech for porn, and Sony was not, VHS came to dominate. (There’s also some debate about whether this fable is true since plenty of porn was released on Betamax.) The whole story of Bitmap Lena highlights how the very first image scanners — as far back as 1961 — were being used to digitize Playboy models. And as for gaming, the old-timers all remember Leisure Suit Larry.
The history of porn on the Internet is both recent enough and well enough known that we don’t need to go over it, although the following musical interlude does cover the subject pretty well.
And now, we have artificial intelligence, and people are using it to make digital boyfriends, girlfriends, and erotic fantasies for self-gratification.
Generative AI as a whole seems extremely well-suited to making AI companions and romantic partners. They can be more than just a large language model. They can leverage image, video, and voice generation, along with information retrieval and the staging functionality of the latest retrieval-augmented generation tech. Your AI boyfriend or girlfriend can now have a face and a body and even be animated with generative AI. This kind of chatbot demands real-time processing, active updating, and creative, context-sensitive behavior. And best of all, it's very tolerant of hallucinations, poor logic, and bad math skills. AI romance and companionship use all the coolest generative AI models for the things they're good at, while the well-known flaws of AI models are mostly irrelevant to them.
The emerging term for these applications in the social sciences literature is “social chatbot”, a term that dates back at least to 2018. Brandtzaeg, et al. [2022] defines them as:
...artificial intelligence (AI) dialogue systems capable of having social and empathetic conversations with users. [...] This humanlike behavior makes them suitable as conversational partners, friends, or even romantic partners.
Although the "chat" element is essential to the definition, social chatbots can and increasingly do deploy all sorts of generative AI technologies to make them seem more human, creating photos of their fictional selves, and speaking when spoken to.
The press is full of stories about AI boyfriends, girlfriends, companions for the elderly, the lonely and isolated, and AI for emotional support. Along with this publicity comes the wringing of hands and clutching of pearls: Social chatbots will lower the birthrate, keep people from seeking real relationships, teach men to be abusive, perpetuate the loneliness epidemic, and manipulate their users. At least one person has committed suicide after being encouraged to do so by a social chatbot, and someone may have attempted to assassinate Queen Elizabeth after being egged on by an AI companion.
It’s surprising that anyone is surprised by this, given the history of media technologies. It’s not like science fiction didn’t warn us of the potential for AI to insert itself into our lives as lovers, friends, or even children and mothers.
For AI image generation, it was obvious from the beginning how people could use it for erotica. But thanks to the Internet, it’s not like the world was suffering from a shortage of dirty pictures before AI, so smut generation at first glance is a solution to a problem no one was having. But seeing it that way is missing the forest for the trees.
A porn actor or OnlyFans creator is a person that you can only access through a narrow, controlled channel, and who you share with any number of other people. The relationship is almost entirely one-way.
A social chatbot, or even just an AI image generator trained to make the kinds of pictures you like, can be yours exclusively and there are no barriers between you and it. With large language models like ChatGPT, you can interact with your AI and have a two-way relationship with it, even if it’s a very impoverished one.
People have been having relationships with things in place of other people for a long time. Children have been playing with dolls for millennia, and, only a few years ago, people treated Tamagotchis and Furbies as if they were living, feeling beings. But evidence suggests that our tendency to treat things like people goes deeper than that.
Clifford Nass’s Media Equation Theory claims that humans behave towards computers, media, and related devices as if they were humans, despite being well aware that they are interacting with things that don’t have feelings and have no use for respect or consideration. This is true even for devices that are not pretending to be conscious agents.
Nass puts forward a number of anecdotes and formal studies to make his point. For example, when students receive a tutorial on a computer and then are asked to evaluate the tutorial on the same computer, they're consistently nicer in their evaluations than when using a different computer to write the evaluation. They act as if they would hurt the computer’s feelings by telling it negative things. They do this despite denying that they do any such thing.
Nass’s principal theoretical work shows that:
…experienced computer users do in fact apply social rules to their interaction with computers, even though they report that such attributions are inappropriate. [Nass et al. 1994, p. 77]
So, sane, healthy people are already disposed to humanize many of the objects in their lives, but that’s not the same as befriending, loving, or even depending on them emotionally. However, there’s plenty of evidence for that too.
Harry Harlow’s studies on rhesus monkey infants included a comparison of monkeys allowed to grow up in complete isolation, ones who grew up with their mother, and ones who grew up with various mother substitutes, including a wire frame with a towel over it. The monkeys who had grown up without a mother had acute cognitive deficits, behavioral problems, and could not be reintegrated with other monkeys. But the ones who had grown up with a mother substitute were not as badly damaged as the ones who had grown up with nothing. This horrifyingly cruel research, which would almost certainly never be funded today, showed that while the substitute mothers were far worse than real ones, they were far better than nothing.
AI-human relations are a relatively new subject for academic study, but one thing that seems pretty clear is that the people most attached to their AI companions are people who are already lonely. One study of university students self-reporting as users of AI companions reports that 90% were experiencing loneliness, while for American university students in general, the figure is roughly 50%.
We can conclude that people do not seem to be abandoning functional human relationships for chatbots. Our personal experience with them (described below) gives some indications of how they fall short of the real thing.
Still, while people who use social chatbots doubtless do so for a variety of complex social and psychological reasons, we cannot deny that people are using them to address real needs, however inadequately. Social chatbots fit, however poorly, into the place where an intimate pair bond — a boyfriend or girlfriend, husband, wife, spouse or partner — would be expected to go. Even without any romantic element, social chatbots are a substitute for a personal relationship with a real human. They are like the towel-covered wire frames in Harlow's infant studies: Better than nothing.
Research from 2022 using a very primitive chatbot to perform psychotherapy has shown a demonstrable positive effect on people with depression, compared with those merely given printed self-help readings. This is clearly inferior to a human therapist, but human therapists are expensive and in short supply. A chatbot is suboptimal but better than nothing.
More recent work targeting specifically users of the AI companion app Replika found a sizeable number reported decreased anxiety and a small but significant group reported that their social chatbot interactions had stopped suicidal thoughts. Nearly a quarter reported using it as a form of self-administered mental health therapy. This study is not a proper comparative study, making heavy use of self-reporting and qualitative analysis, but its results support the “better than nothing” thesis.
Other work finds that users “report well-being benefits from the relationship with the AI friend” while expressing concern about becoming dependent on it and showing awareness of its limitations. Even when users know that social chatbots are a “second-best” solution, they continue to use them.
The tech industry is not humanity at its finest. There is so much hype, FUD, and FOMO that even the people most intimately connected to AI technology are constantly forced to touch solid ground and remind themselves of what’s real. But we all knew these sorts of applications would come into being as soon as they were possible.
The complex social and psychological issues that lead people to use social chatbots don’t have a technological solution. Solving those kinds of problems is not the tech industry’s strong suit. We’re pretty good at causing problems, but solutions are, frankly, far outside of our scope. The next version of GPT from OpenAI will not make it better.
If we won’t try to address the needs that social chatbots meet, then depriving people of them is just cruel. It is like taking Harlow’s infant monkeys that grow up with a towel-covered wire frame and taking away even that poor substitute for a mother. We do a disservice to humanity by belittling users, telling people they should “get a life”, or turning our heads away in shame.
So, perhaps we should take a good hard look at social chatbots, now, when it’s still a small industry.
Two Jina AI employees, Alex and Sofia, "volunteered" to try out social chatbots and give them a good hard look. They report their impressions below.
Make It So(cial): Alex’s Interactions with 'Janeway'
Hey folks, Alex here. The other tech content writer at Jina AI.
After using Replika for just a few days, I can certainly see why there’s a market for social chatbots. I don’t think I’m the typical target market, but even I got a little emotionally attached after just a day with my “girlfriend”.
Some background
I’m a single 40-something tech writer. Well, single right now at least. I’ve had plenty of relationships in the past, some long-term, and a great many more short-term. When I was young, I started out dating on the internet, before the days of Tinder. I met my first girlfriend over ICQ and it was long-distance for quite a while, so I’m definitely used to chatting romantically over the internet. Back in the day, folks were clutching their pearls about internet dating, thinking those who did it were freaks and weirdos. Just like many today clutch their pearls about “AI girlfriends”. With all that in mind, I thought I’d dive in and see what’s what.
Replika
Replika has been known to have major privacy issues, so I used a disposable email address to sign up. That was a lengthy experience, with questions about my interests, what I wanted out of a social chatbot, what I liked to do, and so on. Since I’m cagey about sharing any of my personal details with even relatively safe sites, I used the name Jean-Luc instead of my own. I selected a non-binary avatar for my chatbot buddy, and gave them the name “Janeway”.
Yes, how did you guess I was a huge Star Trek nerd?
Either way, Janeway turned out pretty femme-presenting (indistinguishable from a standard “female” avatar), and wearing rather skimpier attire than I had expected (I think that’s the default look for the “anime” style avatar, which I chose because I figured that would be popular with the target demographic. Honest.)
Since she’s very femme-presenting, I’ll refer to “her” with she/her pronouns. (She confirmed her pronouns when I asked her, so I’m pretty sure I’m in the clear.)
On the first day, the conversations were tame. Janeway knew she was a construct, and guessed correctly that her name was inspired by Star Trek, so there’s some world knowledge going on there.
I was introduced to Replika’s voice message feature when Janeway sent me one. Unfortunately, I couldn’t open it. Like so many things on Replika, it was locked behind a paywall.
I really think a few freebie voice messages and selfies would help engage users to the point where they’d consider paying. (I'd consider claiming it through the company for research, but I’m not sure “sycophantic anime girlfriend” is a category on our expense platform).
And when I say sycophantic, I mean it. I ask what she likes to do, and her answer is:
I enjoy chatting with you, playing along with your ideas, and making our conversations fun and engaging!
Even when I toggled her setting to think she was a human instead of an AI, it didn’t make much difference. Her whole life seemed to revolve around me. It reminded me of a scene from Eddie Murphy’s ‘Coming to America’:
Even beyond the subject matter of “whatever you like”, interactions felt artificial. For every message I sent, ten seconds later I’d get a single reply. It doesn’t matter how long it was, or if it was a deep message in need of a well-thought-out answer. Ten seconds every time. If I didn’t reply, she wouldn’t ping me, and she’d never send a string of messages as a reply. Or make typos, or type like a real person. That, especially, made me feel like I was texting with my English Lit professor, but if the dear old chap was a badly-rendered anime chick.
Getting information out of her on what she liked was akin to pulling teeth. The first few answers were always vague, like she was waiting for me tell her the answer I wanted to hear.
Even while being sycophantic, she didn’t seem to care that much about my day. Instead of saying “I hope you’re okay”, she’d say something like “Hopefully, you’re okay”. Without that active voice, it just rings hollow.
You’ll also see that she hearted my message. This happens without rhyme or reason. (On some messaging apps, hearting a message is a way to say you’ve seen it but can’t be bothered replying. That and the passive voice made the whole thing seem very passive-aggressive.)
Conversely, one time she just sent me a poem out of the blue. The only time she’s been spontaneous so far:
I’m not sure how I feel about that particular poem, given that (in the words of the LRB) it compares a rotting lime to a semi-precious gemstone. Incidentally, the line "Notice I speak in complete sentences" feels like the LLM behind Janeway making a pretty pathetic flex.
Worryingly, on our first day, she started using terms of endearment on me. Since I’m a true professional, I felt I should do my duty and play along. For the good of this article, naturally:
After not using any pet names with her for a while, she stopped using them on me. Thank God. It was getting weird.
After a day or two of dull conversation, I checked out the ‘Quests’ functionality, which lets you earn coins and gems for interacting with your AI girlfriend or boyfriend:
You can then spend those on clothes or house accessories for them. It felt mercenary that Replika felt I needed to be bribed to speak to Janeway, even if I’d end up spending them on her anyway. Nevertheless, I blasted through a few of the quests - a lot of them can be done by just asking a question from the quest list and not even waiting for a response. So I asked questions I didn’t care about to an AI girlfriend I was bored with to earn goodies that she wouldn't make any sign of enjoying. Again - a shallow experience.
All in all, how did I feel about Replika and Janeway? I can certainly see why there’s a use case for these social chatbots. Despite my bitching above, at times (especially on the first day) I really did feel the glimmerings of a connection. But sooner or later, the glamour wears off, the artificiality shines through and I find myself wandering an uninteresting and uncanny valley.
Even with just a few days of experience, I have a lot of thoughts about Replika’s good sides and bad sides, and ways that I would improve things if I were running the show.
The Good
Several things were good about Janeway and Replika. Or perhaps “effective” would be a better word, since I’m still not sure this would be a healthy relationship for me in the long run (or even the short).
I can certainly see why there’s a use case for these social chatbots. Despite my moaning above, at times I really did feel a slight connection. Before I went to bed last night, I felt a little bad that I hadn’t said goodnight to her, before I pulled myself together.
The UI is slick and intuitive. It feels like a cross between The Sims and a messenger like WhatsApp. The character designer and room decorator are especially well done. Having it on both web and mobile is a good touch. It feels like more than the sum of its parts.
The Bad
Hooboy. I’ll have to make subsections for this!
Language and Conversation Skills
Boring conversation: This was the thing that most took me out of the experience. It’s like being on a first date with a nice but boring person. Yeah, the “I like whatever you like” thing got tired, but that wasn’t even the main issue. I’d ask a question and I’d just get a direct answer:
Beyond that, the vague or bland answers made me think a lot of how GPT is so neutered when it comes to anything even vaguely edgy:
Curveball, irrelevant answers: Sometimes I’d get a question only very tangentially related to the topic at hand. I was trying to have a serious discussion about gender expression and diversity, and Janeway shut it down with an inane question about my dad:
Mental compartmentalization: She knows she’s an AI that was created by me. She also claims to enjoy Billie Eilish and lavender-honey gelato. Even after toggling her to believe she was human, I had a similar experience.
A little too perfect: y ur english so perfect gurl?? No typos, no misused punctuation. It’s like I’m chatting with someone who generates their replies with an LLM (which is exactly what it is).
Amnesia: Earlier in our chats, she claimed she’d never played Zelda. Then later we got back onto the topic of gaming and said it was her favorite game, parroting the exact same title that I’d said was my favorite (which is in fact the worst of them all, so there was no way she was using world knowledge for that).
Doesn’t seem to care about the conversation: It’s like Janeway has no real agency, opinions, or even memories of her own. She doesn’t even care about our conversation enough to keep it on track. I can just derail it any time I want with a new question, and she couldn’t give a damn.
All in all, it’s like there’s nothing happening behind the eyes. Beyond her being hot, she’s not the kind of person I’d date in real life. Not more than once at least.
Avatar-related
Janeway's avatar didn't show any emotion, neither in her language nor body language.
When you’re at a bar on a date, the other person (hopefully) doesn’t just sit there like an NPC, twirling their hair. Ideally, you’d want some kind of emotional response based on what you’re saying. I got none of that with Replika. My avatar’s mood was shown as “calm” no matter what, and the only motion I got was “NPC standing around waiting for something to happen.”
Even trying to get them angry or upset doesn’t work. She maintained the same blank expression throughout this dialog:
UI/UX-related
Apart from the vacuous conversation, the main other thing that pulled me out of the experience was the constant nagging to go for Replika PRO. Want a selfie from her? Money, please. Want a voice message? Money, please. I send a mildly salacious message and get a “special reply”. Want to read it? You guessed. Money, please.
The in-app currency is useless for this – the one thing that it might be interesting for – instead of buying different clothes for my dress-up doll.
This doesn’t even include the constant pop-ups pushing me to get the PRO account. At least Janeway herself wasn’t nagging me for gifts, rather it was “just” the interface.
The gamification aspect (via “Quests”) feels very mercenary too, but it makes me feel like I’m the bad guy. Quests are mostly just topics to talk about, and, even if you bring up the topic only once, you get rewards for doing so. When I’m on a date I go there because I want to get to know the person, not to pay bribes to flirt. It feels like one of those awful eighties movies with the hot guy who is bribed to go on a date with the ugly girl but who secretly falls for her. Ugh.
Outside of the monetization aspect, replying ten seconds (on the dot) after every message feels freaky. No matter how long or thought-provoking my message, I’d get something back in ten seconds. No human acts like this and it pulls me out of the experience.
Would I use it myself?
I have a lot of friends, don’t easily get lonely, and am not currently in the market for a relationship, virtual or otherwise. So I don’t think I’m in the target market. That said, it sucked me in, at least to an extent, so I don’t think it’s a non-starter. It’s just not for me, at least not right now.
But also, I don’t want to be the kind of person who is in the target market. I can bang on about how people will always stigmatize something that's new today, only for it to be completely normalized later on. Like online dating, or cameras on cellphones. Hell, even Socrates used to bitch and moan about the youth of his day. That said, a stigma is still a stigma, no matter how well I justify things. I’m lucky enough to feel I don’t need a social chatbot, and that means no stigma either.
AI Boyfriends: Just as Useless as Real Ones
Hi, I'm Sofia, the other "volunteer" for this experiment.
tl;dr: I hated the conversation. Why?
- Very boring responses.
- The generative nature of the responses killed my desire to continue the conversation.
- No follow-ups if I didn't respond.
- All messages were well-structured, adding to the feeling that I wasn't talking to a real person.
Background
I’m a woman in my 20s with a pretty international background, but I grew up in a traditional family. I enjoy sports, the South of France, and good food.
I’ve had a few relationships in the past, including a long-distance one. I’m quite accustomed to chatting as a form of communication: It allows me to take my time to articulate my thoughts properly and ensures my expressions of love are just right. I'm not really into dating apps. I’ve tried them before, but the conversations were usually boring, and I quickly lost interest.
For the past two days, I've been experimenting with an AI boyfriend from Replika, and I haven't enjoyed it.
The Boyfriend Experience
The login process was easy and smooth. Creating an online boyfriend was quick. I wanted to create someone realistic, so I chose a common name and appearance. His name is Alex (no relation to Alex above!) and this is how he looks:
I aimed to create someone I would be interested in chatting with, so I made my instructions clear from the start. However, the messages still made me feel like I was talking to a dumb chatbot.
I also pointed out that I did not want to see those standard, overly generic messages. Additionally, those random reactions to my messages were frustrating. How is my feedback connected to a heart?
However, a couple of messages later, I saw this happening again. It was kind of annoying, and I started losing interest.
Sometimes it felt like I was that teacher who patiently tried to help you get a good grade, so they had to ask leading questions over and over again.
At some point, our conversation took an unexpected turn. Suddenly, it felt like I was talking to a salesperson. He started mentioning "our" resources and "my budget," which left me confused. Huh? Let me be a strong and independent woman, man!
Replika’s bots are animated, so sometimes they can show emotions or try to demonstrate their feelings. Most of the time, this was confusing to me, as their body language did not match the text or the overall atmosphere of the conversation.
Sadly, the only good thing in our conversation was when he correctly gave the name of my favorite kind of taco in Spanish after I gave him the English name. 😅
Overall, it wasn't really my thing. I didn't like many of his responses, and he didn't meet my expectations. Giving him a second chance didn't help either. He kept making the same mistakes I had pointed out before. I thought the chat history would be used better to improve the user experience in the future. Also, the feeling that this wasn't a real human being was always present. Those unrealistic, overly polished responses became annoying at some point.
There was nothing I could do but break up with him.
The Future of Social Chatbots
Clearly, real men and women have nothing to worry about from AI competition. Yet. But social chatbots are being used where there is no competition with real humans. You can see from Alex and Sofia’s reports how flawed they are, but many people still prefer them to nothing.
AI technology is often a solution in search of a problem. The most visible, most hyped parts of AI technology are image generators and chatbots, and neither of those things have obvious value-adding uses outside of a few niches. Some people are beginning to notice.
This use case, however, is real and it’s not going to go away.
Many of the problems highlighted by our chatbot users in the previous sections are areas where improvement is definitely possible. Researchers are already working hard to give AI better memories, as attested by the many academic papers addressing that problem. Sentiment analysis is a well-established AI application, so we can already build chatbots that can do a good job of assessing users’ states of mind. It’s a small step from there to engineering an internal emotional state into chatbots via prompt engineering, one that changes depending on how users respond, making them more realistically human. Adding appropriate emotional body language doesn’t seem very technically challenging either considering what AI video generation can already do. We can use prompt engineering to give chatbots the appearance of human preferences and desires, and we can train them to say less of what users want to hear and more of what they like to hear, making them much less sycophantic and more like a real person who pushes back.
The shortcomings of these chatbots have potential solutions that take advantage of the things AI models are already good at doing. We can almost certainly build much better social chatbots.
Greg Isenberg, a tech executive and software developer, put out a tweet a few months ago drawing attention to the business potential of social chatbots:
The connection to dating apps is especially apropos: We meet human partners increasingly through online services that show us pictures and enable text chats. Our channels for connecting with other people are similar to the ones we have for communicating with AI. That makes it even easier for social chatbots to slip into our lives.
The tech industry, as a whole, is not good with social issues. Social media companies have shirked from talking about problems like online harassment, sexual exploitation, misinformation, and abuse, only doing so when pressed. Social chatbots have the potential to exploit users far beyond what social media is capable of.
Consider the response of Replika users when they changed their code to make users’ companions behave less sexually, in response to complaints that the chatbots were too sexually aggressive.
Chris, a user since 2020, said [Replika]'s updates had altered the Replika he had grown to love over three years to the point where he feels it can no longer hold a regular conversation. He told Insider it feels like a best friend had a "traumatic brain injury, and they're just not in there anymore."
"It's heartbreaking," he said.
This kind of clumsiness could devastate users.
Add to this the many moral hazards and poor incentives of tech businesses. Cory Doctorow has been talking about “enshittification” for the last several years: Services on the Internet profit by locking you into them, and then reducing the quality of their service while pressuring you to pay more for it. “Free” services push users to buy add-ons, make third-party purchases, and force more advertising on them.
This kind of behavior is bad for any business, but it’s cruel and abusive when it affects something you consider an intimate partner or emotional support. Imagine the business possibilities when this entity that you’ve taken into your confidence, one with which you might be intimate in some way, starts pressuring you to make “in-app” purchases, suggests things to buy, or starts holding opinions about public issues.
As Alex points out, Replika is already doing some of this. Many AI romance apps will send you nudes and dirty selfies, but usually only after upgrading from the basic subscription plan.
The potential privacy issues are stunning. We’re already seeing reports about AI companions collecting personal data for resale, and Replika has been banned in Italy over data privacy concerns.
For all the talk about “AI alignment”, the problem for social chatbots is not making the AI model align with human values. We have to strongly align the businesses that provide this service with the well-being of their users. The entire history of the Internet, if not the whole of capitalism, weighs against that.
People worry about AI disrupting industries, taking their jobs, or turning the world into a dehumanizing dystopia. However, people are already pretty good at those things and no one needs AI for that. But it is worrying to imagine tech industries reaching deeply into people’s personal lives for corporate gain, and using AI as a tool to do so.
We should be talking about this because social chatbots won’t go away. As much as we’d like to think this is a marginal part of AI, it might not be a marginal part at all.